Top AI Product

Every day, hundreds of new AI tools launch across Product Hunt, Hacker News, and GitHub. We dig through the noise so you don't have to — surfacing only the ones worth your attention with honest, no-fluff reviews. Explore our latest picks, deep dives, and curated collections to find your next favorite AI tool.


Sakana AI KAME hits 6.43 on MT-Bench by giving voice models two brains

Sakana AI just open-sourced KAME, a tandem speech-to-speech architecture that splits voice AI in two: a fast S2S model handles the mouth, a slow LLM handles the brain. The S2S responds instantly while the LLM reasons in the background and injects “oracle” signals as they arrive. The model talks while it’s still thinking.

Why the numbers matter

Voice models have lived with one trade-off for years: fast and dumb, or smart and laggy. KAME breaks it. MT-Bench jumped from 2.05 to 6.43 — roughly 3x — with latency still near zero. The front-end keeps Moshi’s 80ms audio token cycle, so responses start before you finish your sentence. The training trick is Simulated Oracle Augmentation: a simulator LLM generates 56,582 synthetic dialogues across six hint levels, teaching the front-end how to integrate partial reasoning at any moment.

The hot-swap part is the real headline

The backend LLM is fully replaceable. Train once with gpt-4.1-nano, deploy with Claude Opus 4.1 or Gemini 2.5 Flash without retraining a single parameter. KAME is open source on Hugging Face as SakanaAI/kame — self-host the front-end, point it at any LLM API. Customer service bots, voice tutors, real-time translators all become viable now that “smart but slow” stopped being the only option for voice agents.


You Might Also Like


Discover more from Top AI Product

Subscribe to get the latest posts sent to your email.



Leave a comment