nithinraok
commited on
Commit
•
2326ee7
1
Parent(s):
f5d70c1
Update README.md
Browse files
README.md
CHANGED
@@ -241,7 +241,7 @@ The tokenizers for these models were built using the text transcripts of the tra
|
|
241 |
|
242 |
The model was trained on 64K hours of English speech collected and prepared by NVIDIA NeMo and Suno teams.
|
243 |
|
244 |
-
The training dataset consists of private subset with 40K hours of English speech plus
|
245 |
|
246 |
- Librispeech 960 hours of English speech
|
247 |
- Fisher Corpus
|
|
|
241 |
|
242 |
The model was trained on 64K hours of English speech collected and prepared by NVIDIA NeMo and Suno teams.
|
243 |
|
244 |
+
The training dataset consists of private subset with 40K hours of English speech plus 24K hours from the following public datasets:
|
245 |
|
246 |
- Librispeech 960 hours of English speech
|
247 |
- Fisher Corpus
|