Committed by
GitHub
Add WebAssembly example for VAD + Moonshine models. (#1535)
正在显示
3 个修改的文件
包含
30 行增加
和
1 行删除
| @@ -109,6 +109,7 @@ We also have spaces built using WebAssembly. They are listed below: | @@ -109,6 +109,7 @@ We also have spaces built using WebAssembly. They are listed below: | ||
| 109 | |Real-time speech recognition (English) |[Click me][wasm-hf-streaming-asr-en-zipformer] |[地址][wasm-ms-streaming-asr-en-zipformer]| | 109 | |Real-time speech recognition (English) |[Click me][wasm-hf-streaming-asr-en-zipformer] |[地址][wasm-ms-streaming-asr-en-zipformer]| |
| 110 | |VAD + speech recognition (Chinese + English + Korean + Japanese + Cantonese) with [SenseVoice][SenseVoice]|[Click me][wasm-hf-vad-asr-zh-en-ko-ja-yue-sense-voice]| [地址][wasm-ms-vad-asr-zh-en-ko-ja-yue-sense-voice]| | 110 | |VAD + speech recognition (Chinese + English + Korean + Japanese + Cantonese) with [SenseVoice][SenseVoice]|[Click me][wasm-hf-vad-asr-zh-en-ko-ja-yue-sense-voice]| [地址][wasm-ms-vad-asr-zh-en-ko-ja-yue-sense-voice]| |
| 111 | |VAD + speech recognition (English) with [Whisper][Whisper] tiny.en|[Click me][wasm-hf-vad-asr-en-whisper-tiny-en]| [地址][wasm-ms-vad-asr-en-whisper-tiny-en]| | 111 | |VAD + speech recognition (English) with [Whisper][Whisper] tiny.en|[Click me][wasm-hf-vad-asr-en-whisper-tiny-en]| [地址][wasm-ms-vad-asr-en-whisper-tiny-en]| |
| 112 | +|VAD + speech recognition (English) with [Moonshine tiny][Moonshine tiny]|[Click me][wasm-hf-vad-asr-en-moonshine-tiny-en]| [地址][wasm-ms-vad-asr-en-moonshine-tiny-en]| | ||
| 112 | |VAD + speech recognition (English) with Zipformer trained with [GigaSpeech][GigaSpeech] |[Click me][wasm-hf-vad-asr-en-zipformer-gigaspeech]| [地址][wasm-ms-vad-asr-en-zipformer-gigaspeech]| | 113 | |VAD + speech recognition (English) with Zipformer trained with [GigaSpeech][GigaSpeech] |[Click me][wasm-hf-vad-asr-en-zipformer-gigaspeech]| [地址][wasm-ms-vad-asr-en-zipformer-gigaspeech]| |
| 113 | |VAD + speech recognition (Chinese) with Zipformer trained with [WenetSpeech][WenetSpeech] |[Click me][wasm-hf-vad-asr-zh-zipformer-wenetspeech]| [地址][wasm-ms-vad-asr-zh-zipformer-wenetspeech]| | 114 | |VAD + speech recognition (Chinese) with Zipformer trained with [WenetSpeech][WenetSpeech] |[Click me][wasm-hf-vad-asr-zh-zipformer-wenetspeech]| [地址][wasm-ms-vad-asr-zh-zipformer-wenetspeech]| |
| 114 | |VAD + speech recognition (Japanese) with Zipformer trained with [ReazonSpeech][ReazonSpeech]|[Click me][wasm-hf-vad-asr-ja-zipformer-reazonspeech]| [地址][wasm-ms-vad-asr-ja-zipformer-reazonspeech]| | 115 | |VAD + speech recognition (Japanese) with Zipformer trained with [ReazonSpeech][ReazonSpeech]|[Click me][wasm-hf-vad-asr-ja-zipformer-reazonspeech]| [地址][wasm-ms-vad-asr-ja-zipformer-reazonspeech]| |
| @@ -240,7 +241,7 @@ for more models. The following table lists only **SOME** of them. | @@ -240,7 +241,7 @@ for more models. The following table lists only **SOME** of them. | ||
| 240 | |Name | Supported Languages| Description| | 241 | |Name | Supported Languages| Description| |
| 241 | |-----|-----|----| | 242 | |-----|-----|----| |
| 242 | |[Whisper tiny.en](https://github.com/k2-fsa/sherpa-onnx/releases/download/asr-models/sherpa-onnx-whisper-tiny.en.tar.bz2)|English| See [also](https://k2-fsa.github.io/sherpa/onnx/pretrained_models/whisper/tiny.en.html)| | 243 | |[Whisper tiny.en](https://github.com/k2-fsa/sherpa-onnx/releases/download/asr-models/sherpa-onnx-whisper-tiny.en.tar.bz2)|English| See [also](https://k2-fsa.github.io/sherpa/onnx/pretrained_models/whisper/tiny.en.html)| |
| 243 | -|[Moonshine tiny](https://github.com/k2-fsa/sherpa-onnx/releases/download/asr-models/sherpa-onnx-moonshine-tiny-en-int8.tar.bz2)|English|See [also](https://github.com/usefulsensors/moonshine)| | 244 | +|[Moonshine tiny][Moonshine tiny]|English|See [also](https://github.com/usefulsensors/moonshine)| |
| 244 | |[sherpa-onnx-sense-voice-zh-en-ja-ko-yue-2024-07-17][sherpa-onnx-sense-voice-zh-en-ja-ko-yue-2024-07-17]|Chinese, Cantonese, English, Korean, Japanese| 支持多种中文方言. See [also](https://k2-fsa.github.io/sherpa/onnx/sense-voice/index.html)| | 245 | |[sherpa-onnx-sense-voice-zh-en-ja-ko-yue-2024-07-17][sherpa-onnx-sense-voice-zh-en-ja-ko-yue-2024-07-17]|Chinese, Cantonese, English, Korean, Japanese| 支持多种中文方言. See [also](https://k2-fsa.github.io/sherpa/onnx/sense-voice/index.html)| |
| 245 | |[sherpa-onnx-paraformer-zh-2024-03-09][sherpa-onnx-paraformer-zh-2024-03-09]|Chinese, English| 也支持多种中文方言. See [also](https://k2-fsa.github.io/sherpa/onnx/pretrained_models/offline-paraformer/paraformer-models.html#csukuangfj-sherpa-onnx-paraformer-zh-2024-03-09-chinese-english)| | 246 | |[sherpa-onnx-paraformer-zh-2024-03-09][sherpa-onnx-paraformer-zh-2024-03-09]|Chinese, English| 也支持多种中文方言. See [also](https://k2-fsa.github.io/sherpa/onnx/pretrained_models/offline-paraformer/paraformer-models.html#csukuangfj-sherpa-onnx-paraformer-zh-2024-03-09-chinese-english)| |
| 246 | |[sherpa-onnx-zipformer-ja-reazonspeech-2024-08-01][sherpa-onnx-zipformer-ja-reazonspeech-2024-08-01]|Japanese|See [also](https://k2-fsa.github.io/sherpa/onnx/pretrained_models/offline-transducer/zipformer-transducer-models.html#sherpa-onnx-zipformer-ja-reazonspeech-2024-08-01-japanese)| | 247 | |[sherpa-onnx-zipformer-ja-reazonspeech-2024-08-01][sherpa-onnx-zipformer-ja-reazonspeech-2024-08-01]|Japanese|See [also](https://k2-fsa.github.io/sherpa/onnx/pretrained_models/offline-transducer/zipformer-transducer-models.html#sherpa-onnx-zipformer-ja-reazonspeech-2024-08-01-japanese)| |
| @@ -320,6 +321,8 @@ Video demo in Chinese: [辷コシ∫き逾樊蕗菴蠑謇灘ュ玲撃∫悄豁」蠖ア蜩崎邇 | @@ -320,6 +321,8 @@ Video demo in Chinese: [辷コシ∫き逾樊蕗菴蠑謇灘ュ玲撃∫悄豁」蠖ア蜩崎邇 | ||
| 320 | [wasm-ms-vad-asr-zh-en-ko-ja-yue-sense-voice]: https://www.modelscope.cn/studios/csukuangfj/web-assembly-vad-asr-sherpa-onnx-zh-en-jp-ko-cantonese-sense-voice | 321 | [wasm-ms-vad-asr-zh-en-ko-ja-yue-sense-voice]: https://www.modelscope.cn/studios/csukuangfj/web-assembly-vad-asr-sherpa-onnx-zh-en-jp-ko-cantonese-sense-voice |
| 321 | [wasm-hf-vad-asr-en-whisper-tiny-en]: https://huggingface.co/spaces/k2-fsa/web-assembly-vad-asr-sherpa-onnx-en-whisper-tiny | 322 | [wasm-hf-vad-asr-en-whisper-tiny-en]: https://huggingface.co/spaces/k2-fsa/web-assembly-vad-asr-sherpa-onnx-en-whisper-tiny |
| 322 | [wasm-ms-vad-asr-en-whisper-tiny-en]: https://www.modelscope.cn/studios/csukuangfj/web-assembly-vad-asr-sherpa-onnx-en-whisper-tiny | 323 | [wasm-ms-vad-asr-en-whisper-tiny-en]: https://www.modelscope.cn/studios/csukuangfj/web-assembly-vad-asr-sherpa-onnx-en-whisper-tiny |
| 324 | +[wasm-hf-vad-asr-en-moonshine-tiny-en]: https://huggingface.co/spaces/k2-fsa/web-assembly-vad-asr-sherpa-onnx-en-moonshine-tiny | ||
| 325 | +[wasm-ms-vad-asr-en-moonshine-tiny-en]: https://www.modelscope.cn/studios/csukuangfj/web-assembly-vad-asr-sherpa-onnx-en-moonshine-tiny | ||
| 323 | [wasm-hf-vad-asr-en-zipformer-gigaspeech]: https://huggingface.co/spaces/k2-fsa/web-assembly-vad-asr-sherpa-onnx-en-zipformer-gigaspeech | 326 | [wasm-hf-vad-asr-en-zipformer-gigaspeech]: https://huggingface.co/spaces/k2-fsa/web-assembly-vad-asr-sherpa-onnx-en-zipformer-gigaspeech |
| 324 | [wasm-ms-vad-asr-en-zipformer-gigaspeech]: https://www.modelscope.cn/studios/k2-fsa/web-assembly-vad-asr-sherpa-onnx-en-zipformer-gigaspeech | 327 | [wasm-ms-vad-asr-en-zipformer-gigaspeech]: https://www.modelscope.cn/studios/k2-fsa/web-assembly-vad-asr-sherpa-onnx-en-zipformer-gigaspeech |
| 325 | [wasm-hf-vad-asr-zh-zipformer-wenetspeech]: https://huggingface.co/spaces/k2-fsa/web-assembly-vad-asr-sherpa-onnx-zh-zipformer-wenetspeech | 328 | [wasm-hf-vad-asr-zh-zipformer-wenetspeech]: https://huggingface.co/spaces/k2-fsa/web-assembly-vad-asr-sherpa-onnx-zh-zipformer-wenetspeech |
| @@ -405,3 +408,4 @@ Video demo in Chinese: [辷コシ∫き逾樊蕗菴蠑謇灘ュ玲撃∫悄豁」蠖ア蜩崎邇 | @@ -405,3 +408,4 @@ Video demo in Chinese: [辷コシ∫き逾樊蕗菴蠑謇灘ュ玲撃∫悄豁」蠖ア蜩崎邇 | ||
| 405 | [sherpa-onnx-telespeech-ctc-int8-zh-2024-06-04]: https://github.com/k2-fsa/sherpa-onnx/releases/download/asr-models/sherpa-onnx-telespeech-ctc-int8-zh-2024-06-04.tar.bz2 | 408 | [sherpa-onnx-telespeech-ctc-int8-zh-2024-06-04]: https://github.com/k2-fsa/sherpa-onnx/releases/download/asr-models/sherpa-onnx-telespeech-ctc-int8-zh-2024-06-04.tar.bz2 |
| 406 | [sherpa-onnx-sense-voice-zh-en-ja-ko-yue-2024-07-17]: https://github.com/k2-fsa/sherpa-onnx/releases/download/asr-models/sherpa-onnx-sense-voice-zh-en-ja-ko-yue-2024-07-17.tar.bz2 | 409 | [sherpa-onnx-sense-voice-zh-en-ja-ko-yue-2024-07-17]: https://github.com/k2-fsa/sherpa-onnx/releases/download/asr-models/sherpa-onnx-sense-voice-zh-en-ja-ko-yue-2024-07-17.tar.bz2 |
| 407 | [sherpa-onnx-streaming-zipformer-fr-2023-04-14]: https://github.com/k2-fsa/sherpa-onnx/releases/download/asr-models/sherpa-onnx-streaming-zipformer-fr-2023-04-14.tar.bz2 | 410 | [sherpa-onnx-streaming-zipformer-fr-2023-04-14]: https://github.com/k2-fsa/sherpa-onnx/releases/download/asr-models/sherpa-onnx-streaming-zipformer-fr-2023-04-14.tar.bz2 |
| 411 | +[Moonshine tiny]: https://github.com/k2-fsa/sherpa-onnx/releases/download/asr-models/sherpa-onnx-moonshine-tiny-en-int8.tar.bz2 |
| @@ -52,6 +52,24 @@ def get_models(): | @@ -52,6 +52,24 @@ def get_models(): | ||
| 52 | """, | 52 | """, |
| 53 | ), | 53 | ), |
| 54 | Model( | 54 | Model( |
| 55 | + model_name="sherpa-onnx-moonshine-tiny-en-int8", | ||
| 56 | + hf="k2-fsa/web-assembly-vad-asr-sherpa-onnx-en-moonshine-tiny", | ||
| 57 | + ms="csukuangfj/web-assembly-vad-asr-sherpa-onnx-en-moonshine-tiny", | ||
| 58 | + short_name="vad-asr-en-moonshine_tiny", | ||
| 59 | + cmd=""" | ||
| 60 | + pushd $model_name | ||
| 61 | + mv -v preprocess.onnx ../moonshine-preprocessor.onnx | ||
| 62 | + mv -v encode.int8.onnx ../moonshine-encoder.onnx | ||
| 63 | + mv -v uncached_decode.int8.onnx ../moonshine-uncached-decoder.onnx | ||
| 64 | + mv -v cached_decode.int8.onnx ../moonshine-cached-decoder.onnx | ||
| 65 | + mv -v tokens.txt ../ | ||
| 66 | + popd | ||
| 67 | + rm -rf $model_name | ||
| 68 | + sed -i.bak 's/Zipformer/Moonshine tiny supporting English 英文/g' ../index.html | ||
| 69 | + git diff | ||
| 70 | + """, | ||
| 71 | + ), | ||
| 72 | + Model( | ||
| 55 | model_name="sherpa-onnx-sense-voice-zh-en-ja-ko-yue-2024-07-17", | 73 | model_name="sherpa-onnx-sense-voice-zh-en-ja-ko-yue-2024-07-17", |
| 56 | hf="k2-fsa/web-assembly-vad-asr-sherpa-onnx-zh-en-ja-ko-cantonese-sense-voice", | 74 | hf="k2-fsa/web-assembly-vad-asr-sherpa-onnx-zh-en-ja-ko-cantonese-sense-voice", |
| 57 | ms="csukuangfj/web-assembly-vad-asr-sherpa-onnx-zh-en-jp-ko-cantonese-sense-voice", | 75 | ms="csukuangfj/web-assembly-vad-asr-sherpa-onnx-zh-en-jp-ko-cantonese-sense-voice", |
| @@ -111,6 +111,13 @@ function initOfflineRecognizer() { | @@ -111,6 +111,13 @@ function initOfflineRecognizer() { | ||
| 111 | }; | 111 | }; |
| 112 | } else if (fileExists('telespeech.onnx')) { | 112 | } else if (fileExists('telespeech.onnx')) { |
| 113 | config.modelConfig.telespeechCtc = './telespeech.onnx'; | 113 | config.modelConfig.telespeechCtc = './telespeech.onnx'; |
| 114 | + } else if (fileExists('moonshine-preprocessor.onnx')) { | ||
| 115 | + config.modelConfig.moonshine = { | ||
| 116 | + preprocessor: './moonshine-preprocessor.onnx', | ||
| 117 | + encoder: './moonshine-encoder.onnx', | ||
| 118 | + uncachedDecoder: './moonshine-uncached-decoder.onnx', | ||
| 119 | + cachedDecoder: './moonshine-cached-decoder.onnx' | ||
| 120 | + }; | ||
| 114 | } else { | 121 | } else { |
| 115 | console.log('Please specify a model.'); | 122 | console.log('Please specify a model.'); |
| 116 | alert('Please specify a model.'); | 123 | alert('Please specify a model.'); |
-
请 注册 或 登录 后发表评论