You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
feat: bump V2 to 0.4.10 with streaming ONNX and reInferWithEmotion()
- V2 CDN 0.4.8 → 0.4.10 across all examples and configs
- Guide emotion sliders now use reInferWithEmotion() for real-time
emotion changes during file playback (debounced 300ms)
- Mic streaming uses setEmotion() only (no reInfer to avoid LSTM reset)
- API Reference: add reInferWithEmotion() with usage notes and caveats
- V2 pipeline docs: streaming ONNX (UniLSTM + CausalTransformer + FiLM),
5-frame chunks (~167ms), LSTM state carry
Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
Copy file name to clipboardExpand all lines: .well-known/agent-card.json
+1-1Lines changed: 1 addition & 1 deletion
Original file line number
Diff line number
Diff line change
@@ -3,7 +3,7 @@
3
3
"name": "AnimaSync",
4
4
"description": "Voice-driven 3D avatar animation engine for the browser. Extracts emotion from speech and generates lip sync, facial expressions, and body motion in real time — entirely client-side via Rust/WASM and ONNX inference.",
Re-run ONNX inference on cached audio features with a new emotion vector, without re-uploading or re-decoding audio. Requires a prior `processFile()`/`processAudio()` call (uses internally cached features). Do NOT call during mic streaming — it resets LSTM state.
144
+
145
+
```javascript
146
+
const result = await lipsync.processFile(audioFile);
| Emotion control | Not included | 5-dim FiLM conditioning (neutral, joy, anger, sadness, surprise) via setEmotion()/getEmotion() |
185
+
| Emotion control | Not included | 5-dim FiLM conditioning (neutral, joy, anger, sadness, surprise) via setEmotion()/getEmotion()/reInferWithEmotion() |
175
186
| Body motion | VRMA idle/speaking + VAD auto-switch (LoopPingPong, asymmetric crossfade) | VRMA idle/speaking (LoopPingPong, asymmetric crossfade 0.8s/1.0s) |
176
187
| Best for | Full expression control, custom avatars | Emotion-aware lip sync, quick integration |
0 commit comments