Update README.md
Browse files
README.md
CHANGED
@@ -23,7 +23,7 @@ pipeline_tag: text-to-speech
|
|
23 |
|
24 |
**Main Idea:**
|
25 |
This model enhances previous Llasa TTS by incorporating multilingual data. The approach leverages the LLAMA-initialized text BPE tokenizer,
|
26 |
-
which
|
27 |
Although the multilingual training data is limited—using only the MLS (En/Fr/De/Nl/Es/It/Pt/Pl) and Emilia (En/Zh/De/Fr/Ja/Ko) datasets—resulting in potentially less optimal performance for some languages due to data scarcity,
|
28 |
our model can serve as a base TTS model. It is particularly suitable for fine-tuning for a specific language, as texts in various languages can be uniformly processed using the BPE tokenizer from Llama.
|
29 |
|
|
|
23 |
|
24 |
**Main Idea:**
|
25 |
This model enhances previous Llasa TTS by incorporating multilingual data. The approach leverages the LLAMA-initialized text BPE tokenizer,
|
26 |
+
which can handle multilingual text without the need to design language-specific G2P (grapheme-to-phoneme) systems.
|
27 |
Although the multilingual training data is limited—using only the MLS (En/Fr/De/Nl/Es/It/Pt/Pl) and Emilia (En/Zh/De/Fr/Ja/Ko) datasets—resulting in potentially less optimal performance for some languages due to data scarcity,
|
28 |
our model can serve as a base TTS model. It is particularly suitable for fine-tuning for a specific language, as texts in various languages can be uniformly processed using the BPE tokenizer from Llama.
|
29 |
|