|
16 | 16 | inference, |
17 | 17 | ) |
18 | 18 | from livekit.plugins import noise_cancellation, silero |
19 | | -from livekit.plugins.turn_detector.english import EnglishModel |
| 19 | +from livekit.plugins.turn_detector.multilingual import MultilingualModel |
20 | 20 |
|
21 | 21 | logger = logging.getLogger("agent") |
22 | 22 |
|
@@ -68,13 +68,13 @@ async def entrypoint(ctx: JobContext): |
68 | 68 | llm="azure/gpt-4o-mini", |
69 | 69 | # Speech-to-text (STT) is your agent's ears, turning the user's speech into text that the LLM can understand |
70 | 70 | # See all available models at https://docs.livekit.io/agents/models/stt/ |
71 | | - stt=inference.STT(language="en"), |
| 71 | + stt=inference.STT(language="multi"), |
72 | 72 | # Text-to-speech (TTS) is your agent's voice, turning the LLM's text into speech that the user can hear |
73 | 73 | # See all available models as well as voice selections at https://docs.livekit.io/agents/models/tts/ |
74 | | - tts="cartesia/sonic-2:6f84f4b8-58a2-430c-8c79-688dad597532", |
| 74 | + tts="cartesia/sonic-2:f786b574-daa5-4673-aa0c-cbe3e8534c02", |
75 | 75 | # VAD and turn detection are used to determine when the user is speaking and when the agent should respond |
76 | 76 | # See more at https://docs.livekit.io/agents/build/turns |
77 | | - turn_detection=EnglishModel(), |
| 77 | + turn_detection=MultilingualModel(), |
78 | 78 | vad=ctx.proc.userdata["vad"], |
79 | 79 | # allow the LLM to generate a response while waiting for the end of turn |
80 | 80 | # See more at https://docs.livekit.io/agents/build/audio/#preemptive-generation |
|
0 commit comments