This README describes the Noise Suppression demo application.
On startup the demo application reads command line parameters and loads a network to Inference engine. It also read user-provided sound file with mix of speech and some noise to feed it into the network by small sequential patches. The output of network is also sequence of audio patches with clean speech. The patches collected together and save into output audio file.
The list of models supported by the demo is in <omz_dir>/demos/noise_suppression_demo/python/models.lst
file.
This file can be used as a parameter for Model Downloader and Converter to download and, if necessary, convert models to OpenVINO Inference Engine format (*.xml + *.bin).
An example of using the Model Downloader:
omz_downloader --list models.lst
An example of using the Model Converter:
omz_converter --list models.lst
- noise-suppression-denseunet-ll-0001
- noise-suppression-poconetlike-0001
NOTE: Refer to the tables Intel's Pre-Trained Models Device Support and Public Pre-Trained Models Device Support for the details on models inference support at different devices.
You can use the following command to try the demo (assuming the model from the Open Model Zoo, downloaded with the Model Downloader executed with "--name noise-suppression*"):
python3 noise_suppression_demo.py \
--model=<path_to_model>/noise-suppression-poconetlike-0001.xml \
--input=noisy.wav \
--output=cleaned.wav
The application reads audio wave from the input file with given name. The input file has to have 16kHZ discretization frequency The model is also required demo arguments.
The application outputs cleaned wave to output file. The demo reports
- Latency: total processing time required to process input data (from reading the data to displaying the results).