Skip to content

Latest commit

 

History

History
42 lines (30 loc) · 1.77 KB

README.md

File metadata and controls

42 lines (30 loc) · 1.77 KB

whisper.wasm

Inference of OpenAI's Whisper ASR model inside the browser

This example uses a WebAssembly (WASM) port of the whisper.cpp implementation of the transformer to run the inference inside a web page. The audio data does not leave your computer - it is processed locally on your machine. The performance is not great but you should be able to achieve x2 or x3 real-time for the tiny and base models on a modern CPU and browser (i.e. transcribe a 60 seconds audio in about ~20-30 seconds).

This WASM port utilizes WASM SIMD 128-bit intrinsics so you have to make sure that your browser supports them.

The example is capable of running all models up to size small inclusive. Beyond that, the memory requirements and performance are unsatisfactory. The implementation currently support only the Greedy sampling strategy. Both transcription and translation are supported.

Since the model data is quite big (74MB for the tiny model) you need to manually load the model into the web-page.

The example supports both loading audio from a file and recording audio from the microphone. The maximum length of the audio is limited to 120 seconds.

Live demo

Link: https://whisper.ggerganov.com

image

Build instructions

# build using Emscripten
git clone https://github.com/ggerganov/whisper.cpp
cd whisper.cpp
mkdir build-em && cd build-em
emcmake cmake ..
make -j

# copy the produced page to your HTTP path
cp bin/whisper.wasm/*       /path/to/html/
cp bin/libwhisper.worker.js /path/to/html/