Back to News
Market Impact: 0.25

OpenAI Ex-CTO Mira Murati Is Building AI That Behaves More Like Humans

Artificial IntelligenceTechnology & InnovationPrivate Markets & VentureProduct Launches
OpenAI Ex-CTO Mira Murati Is Building AI That Behaves More Like Humans

Thinking Machines Lab is previewing a real-time AI interaction model that processes conversations in 200-millisecond chunks, enabling interruptions, contextual reactions and backchanneling rather than turn-by-turn replies. The company says the system uses a two-layer architecture, with a fast conversational engine paired with a heavier background reasoning model. The development is promising for voice-first markets like India, but it remains a research preview and long-session reliability is still unresolved.

Analysis

The first-order read is that this is a model-level UX upgrade, but the investable signal is the migration of AI value capture from generative quality to real-time orchestration. If continuous interaction works, the moat shifts toward systems that can fuse speech, vision, latency, and memory into one loop — which favors platform owners with distribution and on-device/cloud plumbing over pure-play model labs. That is structurally bearish for standalone chatbot wrappers, because conversational smoothness becomes table stakes while the premium accrues to whoever owns the default interface. The second-order effect is on the input stack. Real-time, multimodal, always-on systems should increase inference frequency materially versus turn-based usage, which is bullish for GPU demand, networking, and edge silicon over a 12-24 month horizon. It also raises the bar for voice, transcription, and context-management layers; any company that solves low-latency speech-to-speech plus long-context retention can monetize more sessions per user, but the winner is likely the one that can keep compute costs from exploding as engagement lengthens. The main risk is that this is a demo-driven narrative ahead of product proof. Continuous conversations are exactly where latency, hallucination, interruption handling, and context bloat compound, so the failure mode is high user delight in short demos and poor retention in long sessions. If costs per minute of interaction rise faster than willingness to pay, the market will treat this as a feature, not a new category, and the enthusiasm will fade over the next 1-2 quarters. Contrarian view: the market may be underestimating how much of this is an enterprise workflow story, not a consumer chatbot story. The highest ROI use cases are likely call centers, sales assist, telehealth triage, and copilots for field workers where interrupted speech and backchanneling actually matter; those budgets are larger and more defensible than consumer subscriptions. In that framing, the real beneficiaries are incumbents with enterprise distribution and cloud attach, not the newest model startup.

AllMind AI Terminal

AI-powered research, real-time alerts, and portfolio analytics for institutional investors.

Request a Demo

Market Sentiment

Overall Sentiment

mildly positive

Sentiment Score

0.35

Key Decisions for Investors

  • Over 6-12 months, build a basket long NVDA / AVGO / ANET on the thesis that real-time multimodal AI increases inference intensity and low-latency networking demand; use pullbacks of 5-8% to enter, with a 20%+ upside if agentic/voice workloads begin re-rating compute budgets.
  • Short the most exposed AI-wrapper names that monetize chat UX without owning model or distribution leverage over 3-6 months; prefer a basket short versus long infra (e.g., short thin-margin application names against long MSFT or AMZN) to isolate the transition from novelty to commoditized interface.
  • Initiate a pair trade long MSFT, short a pure-play standalone chatbot/consumer AI name for 6-12 months; the thesis is that continuous interaction will be bundled into existing ecosystems, and distribution plus enterprise integration will outweigh model-label narrative.
  • Buy 12-18 month upside in small-cap edge/voice infrastructure where re-rating is underappreciated; focus on names tied to speech recognition, call-center automation, or low-latency AI pipelines, using call spreads to limit decay while waiting for productization.
  • If this theme accelerates into Q2/Q3 product launches, rotate from model enthusiasm to picks-and-shovels by trimming high-multiple application exposure into strength and adding on any selloff in semis tied to fears of 'AI fatigue'—the demand driver here is session minutes, not prompt counts.