Beyond the CB Radio Effect: How New AI Models Are Revolutionizing Real-Time Voice Conversation
The Current State of AI Voice Chat: A Digital CB Radio
Engaging in a voice conversation with today's artificial intelligence often feels like stepping back into the era of citizen's band radio. You speak, then you wait. The AI responds, then it waits. Though modern voice assistants no longer require the explicit "over" and "out" signals of CB radio etiquette, the underlying interaction pattern remains the same: a rigid, turn-by-turn exchange that strips conversation of its natural fluidity.

Behind the scenes, systems like ChatGPT's voice mode or Google's Gemini operate as little more than text-based models with a speech layer grafted on. While you are speaking, the AI cannot process anything else—not the passage of time, not the context of your environment, not even the nuance of what you are saying until you finish. Similarly, when the AI generates its response, it is too consumed with producing words to think about anything beyond the current output. This single-threaded approach forces conversations into an artificial stop-and-go pattern, which is why many users rarely bother with voice features despite their growing availability.
What’s Missing: Continuous Awareness and Real-Time Interaction
The core deficiency in current AI voice chat is the lack of continuous awareness. Human conversation thrives on overlapping cues—interjections, backchannels, simultaneous processing of speech and environment. A person can nod while listening, notice a change in tone, or interrupt to correct a misunderstanding. Today's AI models cannot do any of this. They operate in discrete chunks: listen, process, respond. There is no ability to think while listening or react while speaking. The model is effectively "blind" during both input and output phases, making interaction feel mechanical and disjointed.
This limitation is not just a matter of user experience—it hampers the potential of voice interfaces for complex tasks like real-time tutoring, collaborative problem-solving, or natural back-and-forth in video calls. As a result, voice AI remains a niche feature rather than a primary mode of interaction.
Thinking Machines: A New Approach to Conversational AI
A startup called Thinking Machines, founded by former OpenAI executive Mira Murati, claims to have addressed this fundamental flaw. Their research preview introduces what they call "interaction models"—a new generation of AI designed to follow the natural ebb and flow of conversation, including the ability to interrupt and react in real time while listening.
These interaction models depart radically from the single-threaded architecture used by current chatbots. Instead of processing input and output as separate, sequential blocks, Thinking Machines employs a multi-stream, micro-turn configuration. This architecture allows the AI to continuously process inputs—including sights and sounds—even while it is listening to you. It can then decide to interrupt based on what you are saying, without waiting for you to finish your sentence.
Demos: Real-Time Listening, Interrupting, and Correcting
In a series of demo videos, Thinking Machines showcases its models engaging in video chats with human participants. In one demo, the model identifies products the user holds up to the camera—such as a box of acai—and keeps a running tally of "animal" words like "deer" and "sheep" as the human continues speaking. The AI updates its count in real time, demonstrating that it is not pausing to process but rather listening and thinking simultaneously.
Another striking example shows the model exercising restraint. When a human participant pauses mid-sentence to take a sip of coffee, the AI waits patiently rather than jumping in. This natural pause handling is something current AI struggles with, often either interrupting awkwardly or remaining silent too long. The model also demonstrates the ability to interrupt when instructed. In a segment, it corrects a speaker who mispronounces the word "acai" (properly: ah-sah-EE) and then counters her intentionally false claim that acai bowls originated in Argentina. While the behavior might seem pedantic, it proves the AI can react while it listens, not just after the user finishes speaking.

How It Works: Two Models Working in Tandem
Thinking Machines achieves this real-time interaction through a clever dual-model architecture. The system employs two specialized AI models operating in parallel.
- The Interaction Model: This is the "front-facing" model that remains continuously present with the user. It processes inputs and outputs in rapid-fire 200-millisecond chunks, enabling near-instantaneous responsiveness. It handles the back-and-forth of conversation, manages turn-taking cues, and performs simple real-time tasks like counting words or identifying objects.
- The Background Model: This second model operates in the background, handling more complex cognitive tasks—such as reasoning, planning, or deep knowledge retrieval. Once the background model completes its work, it passes the results to the interaction model, which seamlessly incorporates them into the ongoing conversation without a perceptible delay.
This separation allows the AI to maintain conversational presence while also tackling heavy lifting. The interaction model never goes offline; it is always listening, always processing, and always ready to respond—even while waiting for the background model to finish a deeper analysis. By contrast, today's monolithic models either bog down entirely during complex processing or force the user to wait for a full response cycle.
The Road Ahead: Opportunities and Challenges
Thinking Machines' interaction models are still in a research preview, meaning they are not yet available for public use. The startup faces significant hurdles—including computational cost, latency under real-world network conditions, and the need to handle multiple simultaneous conversation streams gracefully. However, the demos offer a compelling vision of what AI voice chat could become.
If successfully commercialized, this technology could transform industries like customer service, online education, and virtual collaboration. A language tutor that can interrupt to correct pronunciation in real time, a tech support bot that identifies your hardware via camera while you describe the problem, or a meeting assistant that tracks action items as people speak—all become feasible with continuous, multi-stream interaction models.
Ultimately, Thinking Machines is not just adding a better voice layer to existing text AI. They are rethinking the very architecture of conversational interaction. Whether they can deliver on the promise remains to be seen, but the shift from single-threaded to multi-stream micro-turn processing marks a genuine step toward AI that truly listens—not just waits.
Related Articles
- Unmasking the Opt-In Trap: Using Propensity Scores for Causal Inference in AI Feature Experiments
- How to Evaluate AI Chatbot Accuracy: The Strawberry Letter Test and Beyond
- MIT's 'SEAL' Framework Lets AI Rewrite Its Own Code: A Leap Toward Self-Evolving Intelligence
- Smart LLM Routing on a Budget: A NadirClaw Q&A Guide
- Massive Transformers and 90km Cable: Inside the Logistical Nightmare Bringing Marinus Link to Victoria's Coal Country
- 7 Surprising Ways Anthropic's Natural Language Autoencoders Reveal Claude's Hidden Thoughts
- Mastering AWS 2026: A Hands-On Guide to Amazon Quick’s New Desktop App and Amazon Connect’s Agentic AI Solutions
- Ubuntu Set to Integrate On-Device AI Features in 2026, Canonical Emphasizes Principled Approach