Jyhan003 nielsr HF Staff commited on
Commit
a239db6
·
verified ·
1 Parent(s): 5276400

Add pipeline tag, library_name and link to paper (#1)

Browse files

- Add pipeline tag, library_name and link to paper (2fa824d82607cfe396880595d287fd9297e6b72f)


Co-authored-by: Niels Rogge <[email protected]>

Files changed (1) hide show
  1. README.md +5 -3
README.md CHANGED
@@ -1,7 +1,9 @@
1
  ---
2
  license: mit
 
 
3
  tags:
4
- - speaker
5
  - speaker-diarization
6
  - meeting
7
  - wavlm
@@ -12,7 +14,7 @@ tags:
12
  ---
13
 
14
  ## Overview
15
- This hub features the pre-trained model by [DiariZen](https://github.com/BUTSpeechFIT/DiariZen). The EEND component is built upon WavLM-Large and Conformer layers. The model was pre-trained on far-field, single-channel audio from a diverse set of public datasets, including AMI, AISHELL-4, AliMeeting, NOTSOFAR-1, MSDWild, DIHARD3, RAMC, and VoxConverse. Then structured pruning at 80% sparsity is applied. Finally, the pruned model is fine-tuned with [MLC-SLM](https://www.nexdata.ai/competition/mlc-slm) data.
16
 
17
 
18
  ## Usage
@@ -57,4 +59,4 @@ DER evaluation of [Pyannote baseline](https://github.com/mubingshen/MLC-SLM-Base
57
  | Spanish | 12.92 | 10.82 |
58
  | Thai | 10.90 | 10.62 |
59
  | Vietnamese | 14.64 | 12.69 |
60
- | **Average** | **16.44**| **12.71**|
 
1
  ---
2
  license: mit
3
+ library_name: transformers
4
+ pipeline_tag: voice-activity-detection
5
  tags:
6
+ - speaker
7
  - speaker-diarization
8
  - meeting
9
  - wavlm
 
14
  ---
15
 
16
  ## Overview
17
+ This hub features the pre-trained model by [DiariZen](https://github.com/BUTSpeechFIT/DiariZen) as described in [BUT System for the MLC-SLM Challenge](https://huggingface.co/papers/2506.13414). The EEND component is built upon WavLM-Large and Conformer layers. The model was pre-trained on far-field, single-channel audio from a diverse set of public datasets, including AMI, AISHELL-4, AliMeeting, NOTSOFAR-1, MSDWild, DIHARD3, RAMC, and VoxConverse. Then structured pruning at 80% sparsity is applied. Finally, the pruned model is fine-tuned with [MLC-SLM](https://www.nexdata.ai/competition/mlc-slm) data.
18
 
19
 
20
  ## Usage
 
59
  | Spanish | 12.92 | 10.82 |
60
  | Thai | 10.90 | 10.62 |
61
  | Vietnamese | 14.64 | 12.69 |
62
+ | **Average** | **16.44**| **12.71**|