OPEN_SOURCE ↗
REDDIT · REDDIT// 25d agoOPENSOURCE RELEASE
Realtime-AI-Translator hits latency, audio wall
A LocalLLaMA builder shared an open-source macOS offline simultaneous translator that chains `faster-whisper` and Ollama `qwen3.5:9b` in a three-thread pipeline with a PyQt floating subtitle UI and 29-language support. The project is functional but currently stuck at roughly 3-5s end-to-end delay and unreliable Aggregate Device capture (BlackHole + mic), and is asking the community for optimization guidance.
// ANALYSIS
This is a strong local-first prototype that proves the stack works, but it also shows where real-time translation pipelines still break in practice on consumer hardware.
- –The repo documents a clear async queue architecture and multiple stabilization passes, which makes it useful for other builders even before performance is “production-grade.”
- –Latency appears to be an architecture tradeoff: shorter chunks improve responsiveness but hurt semantic coherence, while clause-level buffering preserves quality but adds delay.
- –Community feedback points toward specialized translation models (for example NLLB-class models) over general LLMs for lower-latency translation quality on-device.
- –macOS audio routing remains a frequent failure point in local voice apps; channel mapping and multi-output device setup are often as hard as model tuning.
- –Early traction is small (new repo, minimal stars/comments), but the implementation notes make it a practical reference for offline speech+LLM workflows.
// TAGS
realtime-ai-translatorspeechllmopen-sourceself-hostedinferenceqwen3.5whisper
DISCOVERED
25d ago
2026-03-17
PUBLISHED
26d ago
2026-03-17
RELEVANCE
7/ 10
AUTHOR
Levine_C