Skip to content

Commit be0366c

Browse files
authored
Add documentation for using Groq with EchoKit
This document outlines the transition to using Groq as the LLM provider for EchoKit, highlighting its speed advantages and providing setup instructions.
1 parent b47b4e9 commit be0366c

File tree

1 file changed

+91
-0
lines changed

1 file changed

+91
-0
lines changed
Lines changed: 91 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,91 @@
1+
---
2+
slug: echokit-30-days-day-10-openrouter
3+
title: "Day 11: Switching EchoKit’s LLM to Groq — And Experiencing Real Speed | The First 30 Days with EchoKit"
4+
tags: [echokit30days]
5+
---
6+
7+
8+
Over the past few days, we’ve been exploring how flexible EchoKit really is — from changing the welcome voice and boot screen to swapping between different ASR providers like Groq Whisper, OpenAI Whisper, and local models.
9+
10+
This week, we shifted our focus to the LLM part of the pipeline. After trying OpenAI and OpenRouter, today we’re moving on to something exciting — Groq, known for its incredibly fast inference.
11+
12+
13+
## Why Groq? Speed. Real, noticeable speed.
14+
15+
Groq runs Llama and other open source models on its LPU™ hardware, which is built specifically for fast inference.
16+
When you pair Groq with EchoKit:
17+
18+
* Responses feel **snappier**,
19+
* Interactions become smoother
20+
21+
If you want your EchoKit to feel **ultra responsive**, Groq is one of the best providers to try.
22+
23+
24+
## **How to Use Groq as Your EchoKit LLM Provider**
25+
26+
Just like yesterday’s setup, all changes happen in your `config.toml` of your EchoKit server.
27+
28+
### **Step 1 — Update your LLM section**
29+
30+
Locate the llm section and replace the existing LLM provider with something like:
31+
32+
```toml
33+
[llm]
34+
chat_url = "https://api.groq.com/openai/v1/chat/completions"
35+
api_key = "YOUR_GROQ_API_KEY"
36+
model = "openai/gpt-oss-120b"
37+
history = 5
38+
```
39+
40+
Replace the LLM endpoint URL, API key and model name. [The production models from Groq](https://console.groq.com/docs/models) are `llama-3.1-8b-instant`, `llama-3.3-70b-versatile`, `meta-llama/llama-guard-4-12b`, `openai/gpt-oss-120b`, and `openai/gpt-oss-20b`.
41+
42+
### **Step 2 — Restart your EchoKit server**
43+
44+
After edit the config.toml, you will need to restart your EchoKit server.
45+
46+
Docker users:
47+
48+
```
49+
docker run --rm \
50+
-p 8080:8080 \
51+
-v $(pwd)/config.toml:/app/config.toml \
52+
secondstate/echokit:latest-server-vad &
53+
```
54+
55+
Or restart the Rust binary if you’re running it locally.
56+
57+
```
58+
# Enable debug logging
59+
export RUST_LOG=debug
60+
61+
# Run the EchoKit server in the background
62+
nohup target/release/echokit_server &
63+
```
64+
65+
66+
Then return to the setup page, pair the device if needed.You should immediately feel the speed difference — especially on follow-up questions.
67+
68+
---
69+
70+
## **A Few Tips for Groq Users**
71+
72+
* Groq works best with **Llama models**
73+
* You can experiment with smaller or larger models depending on your device’s use case
74+
* For learning or exploring, the default Groq Llama models are a great starting point
75+
76+
Groq is known for ultra-fast inference, and pairing it with EchoKit makes conversations feel almost instant.
77+
78+
If you’re building a responsive voice AI agent, Groq is definitely worth trying.
79+
80+
---
81+
82+
If you want to share your experience or see what others are building with EchoKit + Groq:
83+
84+
* Join the **[EchoKit Discord](https://discord.gg/Fwe3zsT5g3)**
85+
* Or share your latency tests, setups, and experiments — we love seeing them
86+
87+
Want to get your own EchoKit device?
88+
89+
* [EchoKit Box](https://echokit.dev/echokit_box.html)
90+
* [EchoKit DIY Kit](https://echokit.dev/echokit_diy.html)
91+

0 commit comments

Comments
 (0)