Back to News
Market Impact: 0.25

Apple to Use Gemini AI With “A Lot More Freedom” for Siri in iOS 27

AAPLGOOGLGOOG
Artificial IntelligenceTechnology & InnovationProduct LaunchesCybersecurity & Data PrivacyAntitrust & Competition
Apple to Use Gemini AI With “A Lot More Freedom” for Siri in iOS 27

Apple has secured broad access to Google's Gemini models to produce distilled, smaller models that will power a major Siri upgrade in iOS 27, targeted for announcement at WWDC 2026. Model distillation allows on-device inference to reduce cloud dependence, improve latency and better align with Apple's privacy focus. Apple will combine Gemini-derived student models with ongoing internal model development, and plans user features such as memory and proactive suggestions (e.g., leave-early traffic alerts).

Analysis

If a large consumer OEM successfully shifts meaningful LLM work from cloud inference to distilled on-device students, the P&L mechanics are simple but underappreciated: shaving $0.5–2.0 of annual cloud-inference cost per active device scales to $0.5–2.0B in annual opex savings for a billion-device base, producing a 150–400bp improvement in services gross margin over 12–24 months and freeing cash for other growth levers. Faster local inference also compresses latency and failure modes, which can lift engagement and transaction capture (app store/search/assistant routing) by a few percent — a multiplier on the margin gain rather than just a linear services revenue bump. Winners will skew toward advanced-node fabs and NPU/edge-accelerator IP suppliers: sustained on-device model expansion increases demand for >5nm wafers and mobile NPUs, creating a 12–24 month revenue tailwind for foundries and select chip vendors. Conversely, marginal growth for cloud inference could decelerate, pressuring the growth assumptions embedded in cloud/AI infrastructure valuations and altering the cadence of GPU/equipment orders — not a structural deathblow but a measurable re-rating risk for vendors whose TAM depends on per-request cloud inferencing. Key risks and catalysts are execution and regulation. The two largest reversal triggers are (1) engineering friction: distillation at scale frequently yields capability gaps and knowledge lag versus the teacher model, which can slow user adoption over quarters, and (2) regulatory/contract friction from privileged model access or data handling that could impose restrictions within 6–18 months. Monitor developer previews, device OS release notes, supplier booking cadence, and quarterly commentary on cloud inference spend as 3–12 month catalysts that will validate or reverse the thesis. A contrarian angle: the market tends to oscillate between binary views — “on-device wins” or “cloud remains king.” Reality will be hybrid and choppy; therefore pure cloud shorts are hazardous without an execution/UX failure signal. Relative trades that hedge market beta and capture differential margin expansion are preferable to one-sided bets on either ecosystem.