Commit
·
097ffc5
1
Parent(s):
de27b68
Update README.md
Browse files
README.md
CHANGED
|
@@ -240,7 +240,7 @@ The tokenizers for these models were built using the text transcripts of the tra
|
|
| 240 |
|
| 241 |
The model was trained on 64K hours of English speech collected and prepared by NVIDIA NeMo and Suno teams.
|
| 242 |
|
| 243 |
-
The training dataset consists of private subset with 40K hours of English speech plus
|
| 244 |
|
| 245 |
- Librispeech 960 hours of English speech
|
| 246 |
- Fisher Corpus
|
|
|
|
| 240 |
|
| 241 |
The model was trained on 64K hours of English speech collected and prepared by NVIDIA NeMo and Suno teams.
|
| 242 |
|
| 243 |
+
The training dataset consists of private subset with 40K hours of English speech plus 24K hours from the following public datasets:
|
| 244 |
|
| 245 |
- Librispeech 960 hours of English speech
|
| 246 |
- Fisher Corpus
|