Fangjun Kuang
Committed by GitHub

Add speaker identification and verification exmaple for Dart API (#1194)

... ... @@ -4,6 +4,11 @@ set -ex
cd dart-api-examples
pushd speaker-identification
echo '----------3d speaker----------'
./run-3d-speaker.sh
popd
pushd add-punctuations
echo '----------CT Transformer----------'
./run-ct-transformer.sh
... ...
... ... @@ -112,6 +112,7 @@ jobs:
cp scripts/dart/vad-non-streaming-asr-pubspec.yaml dart-api-examples/vad-with-non-streaming-asr/pubspec.yaml
cp scripts/dart/audio-tagging-pubspec.yaml dart-api-examples/audio-tagging/pubspec.yaml
cp scripts/dart/add-punctuations-pubspec.yaml dart-api-examples/add-punctuations/pubspec.yaml
cp scripts/dart/speaker-id-pubspec.yaml dart-api-examples/speaker-identification/pubspec.yaml
cp scripts/dart/sherpa-onnx-pubspec.yaml flutter/sherpa_onnx/pubspec.yaml
... ...
... ... @@ -9,14 +9,15 @@ https://pub.dev/packages/sherpa_onnx
| Directory | Description |
|-----------|-------------|
| [./add-punctuations](./add-punctuations)| Example for adding punctuations to text.|
| [./audio-tagging](./audio-tagging)| Example for audio tagging.|
| [./keyword-spotter](./keyword-spotter)| Example for keyword spotting|
| [./non-streaming-asr](./non-streaming-asr)| Example for non-streaming speech recognition|
| [./speaker-identification](./speaker-identification)| Example for speaker identification and verification.|
| [./streaming-asr](./streaming-asr)| Example for streaming speech recognition|
| [./tts](./tts)| Example for text to speech|
| [./vad](./vad)| Example for voice activity detection|
| [./vad-with-non-streaming-asr](./vad-with-non-streaming-asr)| Example for voice activity detection with non-streaming speech recognition. You can use it to generate subtitles.|
| [./audio-tagging](./audio-tagging)| Example for audio tagging.|
| [./add-punctuations](./add-punctuations)| Example for adding punctuations to text.|
| [./vad](./vad)| Example for voice activity detection|
## How to create an example in this folder
... ...
# https://dart.dev/guides/libraries/private-files
# Created by `dart pub`
.dart_tool/
... ...
# Introduction
This example shows how to use the Dart API from sherpa-onnx for speaker identification.
| File | Description|
|------|------------|
|[./bin/speaker_id.dart](./bin/speaker_id.dart)| Use a speaker embedding extractor model for speaker identification and verification. See also [./run-3d-speaker.sh](./run-3d-speaker.sh)|
... ...
# This file configures the static analysis results for your project (errors,
# warnings, and lints).
#
# This enables the 'recommended' set of lints from `package:lints`.
# This set helps identify many issues that may lead to problems when running
# or consuming Dart code, and enforces writing Dart using a single, idiomatic
# style and format.
#
# If you want a smaller set of lints you can change this to specify
# 'package:lints/core.yaml'. These are just the most critical lints
# (the recommended set includes the core lints).
# The core lints are also what is used by pub.dev for scoring packages.
include: package:lints/recommended.yaml
# Uncomment the following section to specify additional rules.
# linter:
# rules:
# - camel_case_types
# analyzer:
# exclude:
# - path/to/excluded/files/**
# For more information about the core and recommended set of lints, see
# https://dart.dev/go/core-lints
# For additional information about configuring this file, see
# https://dart.dev/guides/language/analysis-options
... ...
../../vad/bin/init.dart
\ No newline at end of file
... ...
// Copyright (c) 2024 Xiaomi Corporation
import 'dart:io';
import 'dart:typed_data';
import 'package:args/args.dart';
import 'package:sherpa_onnx/sherpa_onnx.dart' as sherpa_onnx;
import './init.dart';
Float32List computeEmbedding(
{required sherpa_onnx.SpeakerEmbeddingExtractor extractor,
required String filename}) {
final waveData = sherpa_onnx.readWave(filename);
final stream = extractor.createStream();
stream.acceptWaveform(
samples: waveData.samples,
sampleRate: waveData.sampleRate,
);
stream.inputFinished();
final embedding = extractor.compute(stream);
stream.free();
return embedding;
}
void main(List<String> arguments) async {
await initSherpaOnnx();
final parser = ArgParser()..addOption('model', help: 'Path to model.onnx');
final res = parser.parse(arguments);
if (res['model'] == null) {
print(parser.usage);
exit(1);
}
final model = res['model'] as String;
/*
Please download test data by yourself
curl -SL -o sr-data.tar.gz https://github.com/csukuangfj/sr-data/archive/refs/tags/v1.0.0.tar.gz
tar xvf sr-data.tar.gz
mv sr-data-1.0.0 sr-data
*/
final config = sherpa_onnx.SpeakerEmbeddingExtractorConfig(
model: model,
numThreads: 1,
debug: true,
provider: 'cpu',
);
final extractor = sherpa_onnx.SpeakerEmbeddingExtractor(config: config);
final manager = sherpa_onnx.SpeakerEmbeddingManager(extractor.dim);
final spk1Files = [
"./sr-data/enroll/fangjun-sr-1.wav",
"./sr-data/enroll/fangjun-sr-2.wav",
"./sr-data/enroll/fangjun-sr-3.wav",
];
final spk1Vec = <Float32List>[];
for (final f in spk1Files) {
final embedding = computeEmbedding(extractor: extractor, filename: f);
spk1Vec.add(embedding);
}
final spk2Files = [
"./sr-data/enroll/leijun-sr-1.wav",
"./sr-data/enroll/leijun-sr-2.wav",
];
final spk2Vec = <Float32List>[];
for (final f in spk2Files) {
final embedding = computeEmbedding(extractor: extractor, filename: f);
spk2Vec.add(embedding);
}
if (!manager.addMulti(name: "fangjun", embeddingList: spk1Vec)) {
// Note you should free extractor and manager in your app to avoid memory leak
print("Failed to register fangjun");
return;
}
if (!manager.addMulti(name: "leijun", embeddingList: spk2Vec)) {
print("Failed to register leijun");
return;
}
if (manager.numSpeakers != 2) {
print("There should be two speakers");
return;
}
if (!manager.contains("fangjun")) {
print("It should contain the speaker fangjun");
return;
}
if (!manager.contains("leijun")) {
print("It should contain the speaker leijun");
return;
}
print("---All speakers---");
final allSpeakers = manager.allSpeakerNames;
for (final s in allSpeakers) {
print(s);
}
print("------------");
final testFiles = [
"./sr-data/test/fangjun-test-sr-1.wav",
"./sr-data/test/leijun-test-sr-1.wav",
"./sr-data/test/liudehua-test-sr-1.wav",
];
final threshold = 0.6;
for (final file in testFiles) {
final embedding = computeEmbedding(extractor: extractor, filename: file);
var name = manager.search(embedding: embedding, threshold: threshold);
if (name == '') {
name = "<Unknown>";
}
print("$file: $name");
}
if (!manager.verify(
name: "fangjun",
embedding: computeEmbedding(extractor: extractor, filename: testFiles[0]),
threshold: threshold)) {
print("{$testFiles[0]} should match fangjun!");
return;
}
if (!manager.remove("fangjun")) {
print("Failed to remove fangjun");
return;
}
if (manager.verify(
name: "fangjun",
embedding: computeEmbedding(extractor: extractor, filename: testFiles[0]),
threshold: threshold)) {
print("${testFiles[0]} should match no one!");
return;
}
if (manager.numSpeakers != 1) {
print("There should only 1 speaker left.");
return;
}
extractor.free();
manager.free();
}
... ...
name: speaker_identification
description: >
This example demonstrates how to use the Dart API for speaker identification.
version: 1.0.0
environment:
sdk: ^3.4.0
dependencies:
sherpa_onnx: ^1.10.20
path: ^1.9.0
args: ^2.5.0
dev_dependencies:
lints: ^3.0.0
... ...
#!/usr/bin/env bash
set -ex
dart pub get
if [ ! -f ./3dspeaker_speech_eres2net_base_sv_zh-cn_3dspeaker_16k.onnx ]; then
curl -SL -O https://github.com/k2-fsa/sherpa-onnx/releases/download/speaker-recongition-models/3dspeaker_speech_eres2net_base_sv_zh-cn_3dspeaker_16k.onnx
fi
if [ ! -f ./sr-data/enroll/leijun-sr-1.wav ]; then
curl -SL -o sr-data.tar.gz https://github.com/csukuangfj/sr-data/archive/refs/tags/v1.0.0.tar.gz
tar xvf sr-data.tar.gz
mv sr-data-1.0.0 sr-data
fi
dart run \
./bin/speaker_id.dart \
--model ./3dspeaker_speech_eres2net_base_sv_zh-cn_3dspeaker_16k.onnx
... ...
... ... @@ -15,3 +15,7 @@
|Non-Streaming speech recognition| [Address](https://github.com/k2-fsa/sherpa-onnx/tree/master/dart-api-examples/non-streaming-asr)| macOS, Windows, Linux|
|Text to speech| [Address](https://github.com/k2-fsa/sherpa-onnx/tree/master/dart-api-examples/tts)| macOS, Windows, Linux|
|Voice activity detection (VAD)| [Address](https://github.com/k2-fsa/sherpa-onnx/tree/master/dart-api-examples/vad)| macOS, Windows, Linux|
|Voice activity detection (VAD) with non-streaming speech recognition| [Address](https://github.com/k2-fsa/sherpa-onnx/tree/master/dart-api-examples/vad-with-non-streaming-asr)| macOS, Windows, Linux|
|Speaker identification and verification| [Address](https://github.com/k2-fsa/sherpa-onnx/tree/master/dart-api-examples/speaker-identification)| macOS, Windows, Linux|
|Audio tagging| [Address](https://github.com/k2-fsa/sherpa-onnx/tree/master/dart-api-examples/audio-tagging)| macOS, Windows, Linux|
|Keyword spotter| [Address](https://github.com/k2-fsa/sherpa-onnx/tree/master/dart-api-examples/keyword-spotter)| macOS, Windows, Linux|
... ...
... ... @@ -107,7 +107,7 @@ public class SpeakerIdentification {
// test verify
if (!manager.verify("fangjun", computeEmbedding(extractor, testFiles[0]), threshold)) {
System.out.printf("testFiles[0] should match fangjun!");
System.out.printf("%s should match fangjun!\n", testFiles[0]);
return;
}
... ...
name: speaker_identification
description: >
This example demonstrates how to use the Dart API for speaker identification.
version: 1.0.0
environment:
sdk: ^3.4.0
dependencies:
sherpa_onnx:
path: ../../flutter/sherpa_onnx
path: ^1.9.0
args: ^2.5.0
dev_dependencies:
lints: ^3.0.0
... ...