You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardexpand all lines: README.md
+63-34
Original file line number
Diff line number
Diff line change
@@ -1,36 +1,46 @@
1
1
Using `pyannote.audio` open-source toolkit in production?
2
2
Consider switching to [pyannoteAI](https://www.pyannote.ai) for better and faster options.
3
3
4
-
# `pyannote.audio` speaker diarization toolkit
4
+
# `pyannote` speaker diarization toolkit
5
5
6
6
`pyannote.audio` is an open-source toolkit written in Python for speaker diarization. Based on [PyTorch](https://pytorch.org) machine learning framework, it comes with state-of-the-art [pretrained models and pipelines](https://hf.co/pyannote), that can be further finetuned to your own data for even better performance.
-:exploding_head: state-of-the-art performance (see [Benchmark](#benchmark))
16
+
-:hugs: pretrained [pipelines](https://hf.co/models?other=pyannote-audio-pipeline) (and [models](https://hf.co/models?other=pyannote-audio-model)) on [:hugs: model hub](https://huggingface.co/pyannote)
17
+
-:rocket: built-in support for [pyannoteAI](https://pyannote.ai) premium speaker diarization
18
+
-:snake: Python-first API
19
+
-:zap: multi-GPU training with [pytorch-lightning](https://pytorchlightning.ai/)
20
+
21
+
## Open-source speaker diarization pipeline
13
22
14
23
1. Install [`pyannote.audio`](https://github.com/pyannote/pyannote-audio) with `pip install pyannote.audio`
15
24
2. Accept [`pyannote/segmentation-3.0`](https://hf.co/pyannote/segmentation-3.0) user conditions
16
25
3. Accept [`pyannote/speaker-diarization-3.1`](https://hf.co/pyannote/speaker-diarization-3.1) user conditions
17
-
4. Create access token at [`hf.co/settings/tokens`](https://hf.co/settings/tokens).
26
+
4. Create Huggingface access token at [`hf.co/settings/tokens`](https://hf.co/settings/tokens).
18
27
19
28
```python
29
+
import torch
20
30
from pyannote.audio import Pipeline
21
31
from pyannote.audio.pipelines.utils.hook import ProgressHook
Visit [`docs.pyannote.ai`](https://docs.pyannote.ai) to learn about other pyannoteAI features (voiceprinting, confidence scores, ...)
79
+
80
+
## Benchmark
81
+
82
+
Out of the box, `pyannote.audio` speaker diarization [pipeline v3.1](https://hf.co/pyannote/speaker-diarization-3.1) is expected to be much better (and faster) than v2.x. [`pyannoteAI`](https://www.pyannote.ai) premium model goes one step further. Those numbers are diarization error rates (in %) - the lower the better.
[Diarization error rate](http://pyannote.github.io/pyannote-metrics/reference.html#diarization) (in %)
45
101
46
-
-:hugs: pretrained [pipelines](https://hf.co/models?other=pyannote-audio-pipeline) (and [models](https://hf.co/models?other=pyannote-audio-model)) on [:hugs: model hub](https://huggingface.co/pyannote)
47
-
-:exploding_head: state-of-the-art performance (see [Benchmark](#benchmark))
48
-
-:snake: Python-first API
49
-
-:zap: multi-GPU training with [pytorch-lightning](https://pytorchlightning.ai/)
50
102
51
103
## Documentation
52
104
@@ -78,29 +130,6 @@ for turn, _, speaker in diarization.itertracks(yield_label=True):
78
130
- 2024-04-05 > [Offline speaker diarization (speaker-diarization-3.1)](tutorials/community/offline_usage_speaker_diarization.ipynb) by [Simon Ottenhaus](https://github.com/simonottenhauskenbun)
Out of the box, `pyannote.audio` speaker diarization [pipeline](https://hf.co/pyannote/speaker-diarization-3.1) v3.1 is expected to be much better (and faster) than v2.x.
0 commit comments