Thinking Machines Lab says it’s building full duplex AI, which means an AI system can take in what someone is saying while generating a response. In plain English, it’s closer to a phone call than a walkie-talkie.
The startup, founded last year by former OpenAI CTO Mira Murati, announced interaction models, starting with TML-Interaction-Small. It says the system can respond in 0.40 seconds, a pace that puts it near ordinary human back-and-forth.
There’s a catch for anyone hoping to try it today. This remains a research preview, with limited access planned in the next few months and a broader release expected later this year.
A faster kind of AI exchange
The core idea is easy to understand, and the change is meaningful. Instead of waiting for someone to finish speaking before working on an answer, the model processes incoming speech while preparing its response.
That delay matters because pauses make AI assistants sound artificial. Thinking Machines Lab frames TML-Interaction-Small’s 0.40-second response time as close to natural conversation speed, which would be a noticeable shift for voice tools.
It also claims that pace is faster than comparable models from OpenAI and Google. The benchmark gives the announcement weight, but outside users still need to test whether the experience works as smoothly as the number suggests.
When speed becomes behavior
An assistant that answers while it’s still taking in information changes what users expect from a voice chat. The conversation can move faster, but the system also has to manage timing with much more care.
That tradeoff matters when someone wants quick clarification instead of a long generated reply. Faster responses won’t help much if the assistant jumps in too early, misunderstands the speaker, or breaks the flow it’s supposed to improve.

For now, the architecture is the news. The real product test is whether the interaction model can make better timing feel automatic.
What to watch before launch
The release timeline is the key detail now. Thinking Machines Lab says a limited research preview is coming in the next few months, followed by broader access later this year.
Availability, pricing, supported platforms, and performance outside controlled testing are still unclear. Those missing pieces matter because a faster model only helps if people can use it in everyday voice tools.
For anyone who uses AI voice assistants, the practical move is to watch the preview closely. Full duplex AI has promise, but hands-on testing should show whether faster responses actually make daily AI conversations easier.






