Skip to content

Commit da8ee5f

Browse files
authored
Merge pull request #105 from second-state/alabulei1-patch-8
Add documentation for using Groq with EchoKit
2 parents b47b4e9 + d3e26a0 commit da8ee5f

File tree

1 file changed

+91
-0
lines changed

1 file changed

+91
-0
lines changed
Lines changed: 91 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,91 @@
1+
---
2+
slug: echokit-30-days-day-11-groq
3+
title: "Day 11: Switching EchoKit’s LLM to Groq — And Experiencing Real Speed | The First 30 Days with EchoKit"
4+
tags: [echokit30days]
5+
---
6+
7+
8+
Over the past few days, we’ve been exploring how flexible EchoKit really is — from changing the welcome voice and boot screen to swapping between different ASR providers like Groq Whisper, OpenAI Whisper, and local models.
9+
10+
This week, we shifted our focus to the LLM part of the pipeline. After trying OpenAI and OpenRouter, today we’re moving on to something exciting — Groq, known for its incredibly fast inference.
11+
12+
13+
## Why Groq? Speed. Real, noticeable speed.
14+
15+
Groq runs Llama and other open source models on its LPU™ hardware, which is built specifically for fast inference.
16+
When you pair Groq with EchoKit:
17+
18+
* Responses feel **snappier**,
19+
* Interactions become smoother
20+
21+
If you want your EchoKit to feel **ultra responsive**, Groq is one of the best providers to try.
22+
23+
24+
## **How to Use Groq as Your EchoKit LLM Provider**
25+
26+
Just like yesterday’s setup, all changes happen in your `config.toml` of your EchoKit server.
27+
28+
### **Step 1 — Update your LLM section**
29+
30+
Locate the llm section and replace the existing LLM provider with something like:
31+
32+
```toml
33+
[llm]
34+
chat_url = "https://api.groq.com/openai/v1/chat/completions"
35+
api_key = "YOUR_GROQ_API_KEY"
36+
model = "openai/gpt-oss-120b"
37+
history = 5
38+
```
39+
40+
Replace the LLM endpoint URL, API key and model name. [The production models from Groq](https://console.groq.com/docs/models) are `llama-3.1-8b-instant`, `llama-3.3-70b-versatile`, `meta-llama/llama-guard-4-12b`, `openai/gpt-oss-120b`, and `openai/gpt-oss-20b`.
41+
42+
### **Step 2 — Restart your EchoKit server**
43+
44+
After editing the config.toml, you will need to restart your EchoKit server.
45+
46+
Docker users:
47+
48+
```
49+
docker run --rm \
50+
-p 8080:8080 \
51+
-v $(pwd)/config.toml:/app/config.toml \
52+
secondstate/echokit:latest-server-vad &
53+
```
54+
55+
Or restart the Rust binary if you’re running it locally.
56+
57+
```
58+
# Enable debug logging
59+
export RUST_LOG=debug
60+
61+
# Run the EchoKit server in the background
62+
nohup target/release/echokit_server &
63+
```
64+
65+
66+
Then return to the setup page, pair the device if needed. You should immediately feel the speed difference — especially on follow-up questions.
67+
68+
---
69+
70+
## **A Few Tips for Groq Users**
71+
72+
* Groq works best with **Llama models**
73+
* You can experiment with smaller or larger models depending on your device’s use case
74+
* For learning or exploring, the default Groq Llama models are a great starting point
75+
76+
Groq is known for ultra-fast inference, and pairing it with EchoKit makes conversations feel almost instant.
77+
78+
If you’re building a responsive voice AI agent, Groq is definitely worth trying.
79+
80+
---
81+
82+
If you want to share your experience or see what others are building with EchoKit + Groq:
83+
84+
* Join the **[EchoKit Discord](https://discord.gg/Fwe3zsT5g3)**
85+
* Or share your latency tests, setups, and experiments — we love seeing them
86+
87+
Want to get your own EchoKit device?
88+
89+
* [EchoKit Box](https://echokit.dev/echokit_box.html)
90+
* [EchoKit DIY Kit](https://echokit.dev/echokit_diy.html)
91+

0 commit comments

Comments
 (0)