Skip to content

Latest commit

 

History

History
45 lines (38 loc) · 2.01 KB

File metadata and controls

45 lines (38 loc) · 2.01 KB

annInferenceApp - Client Application

This Sample Inference Application supports:

  • connect to inference_server_app

    • queries number of GPUs on the server
    • queries pre-loaded neural network models on the server
    • upload pre-trained CAFFE models (optional)
    • specify input width and height dimensions
    • browse and pick deploy.prototxt and weights.caffemodel files
    • specify input preprocesor values for normalization (if needed)
    • specify input channel order (RGB or BGR)
    • optionally, you can publish your neural network for use by others
      • select "PublishAs"
      • give a name to your model
      • type password (default passowrd of inference_server_app is "radeon")
  • pick a CNN model, if upload option is not used

  • run inference on images from a folder

    • browse and pick labels (i.e., synset words for each output label)
    • browse and pick input images to be classified
    • optionally, pick list of input images from a .txt or .csv file
    • optionally, specify max number of images to be processed from folder or image list

If you want to run a simple test, use annInferenceApp.py (a python script) to simply pick a network model to classify local images.

% python annInferenceApp.py [-v] -host:<hostname> -port:<port> -model:<modelName>
                            [-synset:<synset.txt>] [-output:<output.csv>]
                            <folder>|<file(s)>
  • Using QMake

    If you want to avoid installing QtCreator application (Qt creator app is only needed for debugging and designing purposes), you can also build the app using QMake

      sudo apt-get install qt5-qmake
      cd MIVisionX/apps/cloud_inference/client_app/
      qmake -o Makefile annInferenceApp.pro
    

    The above qmake command will generate a Makefile for the application, which can be used to build the application. When your application is successfully built, run the app as you are running ordinary c application.

      make
      ./annInferenceApp