Fix typo(s) in README.md
Browse files
README.md
CHANGED
|
@@ -36,7 +36,7 @@ LFM2-Audio is an end-to-end multimodal speech and text language model, and as su
|
|
| 36 |
Our model consists of a pretrained LFM2 model as its multimodal backbone, along with a FastConformer based audio encoder to handle continuous audio inputs, and a RQ-transformer generating discrete Mimi tokens as audio output.
|
| 37 |
|
| 38 |
LFM2-Audio supports two distinct generation routines, each suitable for a set of tasks.
|
| 39 |
-
Interleaved
|
| 40 |
Sequential generation is suited for non-conversational tasks such as ASR or TTS, and allows the model to switch generated modality on the fly.
|
| 41 |
|
| 42 |
## 📄 Model details
|
|
|
|
| 36 |
Our model consists of a pretrained LFM2 model as its multimodal backbone, along with a FastConformer based audio encoder to handle continuous audio inputs, and a RQ-transformer generating discrete Mimi tokens as audio output.
|
| 37 |
|
| 38 |
LFM2-Audio supports two distinct generation routines, each suitable for a set of tasks.
|
| 39 |
+
Interleaved generation enables real-time speech-to-speech conversational chatbot capabilities, where audio generation latency is key.
|
| 40 |
Sequential generation is suited for non-conversational tasks such as ASR or TTS, and allows the model to switch generated modality on the fly.
|
| 41 |
|
| 42 |
## 📄 Model details
|