Back to stories
Models

Apple Is Using Google's Gemini to Train Smaller On-Device AI Models

Michael Ouroumis2 min read
Apple Is Using Google's Gemini to Train Smaller On-Device AI Models

Apple's partnership with Google runs deeper than previously understood. According to a new report from The Information, Apple has been granted "complete access" to Google's Gemini model inside its own data centers — and is using that access to train smaller AI models for deployment on its devices.

The arrangement, stemming from the deal announced in January 2026, leverages a technique called distillation: a frontier-scale "teacher" model is used to generate training data and supervision signals for a smaller "student" model, which can then run efficiently on-device.

Why Distillation Matters

Model distillation has become one of the most important techniques in practical AI deployment. Training a model that can match GPT-4 or Gemini Ultra on a broad range of tasks requires enormous compute resources. But distilling a focused version of those capabilities into a smaller model — one optimized for specific Apple use cases — is far more tractable.

The resulting models can run on iPhone, iPad, and Mac hardware without constant cloud inference, preserving privacy and reducing latency. It's the same general approach Apple used to build its Apple Intelligence features, though the use of Gemini as a teacher is new.

The Strategic Calculus

For Apple, the arrangement is pragmatic. Building frontier-scale models internally would require massive investment in data centers and research talent. Using Google's model as a scaffold allows Apple's teams to focus on the distillation pipeline, device optimization, and Apple-specific fine-tuning — areas where they already excel.

For Google, it deepens the commercial relationship with Apple's enormous install base, even as the two companies compete in AI assistants and mobile software. Notably, it also means Gemini's capabilities — however indirectly — end up powering Apple's on-device AI features.

Privacy Implications

Apple has been careful to position its AI features around on-device processing and Private Cloud Compute. If Gemini is being used purely as a training-time teacher model, with no inference happening via Google's servers during normal device use, that's largely consistent with Apple's privacy narrative.

The more sensitive question is what training data flows through this arrangement — something neither Apple nor Google has commented on publicly.

What to Watch

The distillation strategy suggests Apple is serious about closing the capability gap with Google and OpenAI without abandoning its hardware-first, privacy-first positioning. If the approach works, it could become a template for how device manufacturers at scale build competitive AI without the resources of a frontier lab.

Expect more details to emerge as Apple Intelligence features roll out in the next major iOS and macOS releases.

Learn AI for Free — FreeAcademy.ai

Take "AI Essentials: Understanding AI in 2026" — a free course with certificate to master the skills behind this story.

More in Models

OpenAI Ships GPT-Realtime-2 With Live Translation and Streaming Whisper, Pushing Voice Agents Toward GPT-5 Reasoning
Models

OpenAI Ships GPT-Realtime-2 With Live Translation and Streaming Whisper, Pushing Voice Agents Toward GPT-5 Reasoning

OpenAI launched three new audio models on May 7 — GPT-Realtime-2, GPT-Realtime-Translate, and GPT-Realtime-Whisper — adding GPT-5-class reasoning, a 128K context window, and metered live translation across 70 languages to its already-GA Realtime API.

1 day ago2 min read
Zyphra Releases ZAYA1-8B, the First Frontier-Class Reasoning MoE Trained Entirely on AMD
Models

Zyphra Releases ZAYA1-8B, the First Frontier-Class Reasoning MoE Trained Entirely on AMD

Zyphra's ZAYA1-8B is an Apache-licensed reasoning MoE with under 1B active parameters that matches Claude 4.5 Sonnet and Gemini 2.5 Pro on math benchmarks — and it was pretrained, midtrained, and fine-tuned on AMD Instinct MI300X GPUs.

2 days ago2 min read
OpenAI Makes GPT-5.5 Instant the Default ChatGPT Model With 52.5% Fewer Hallucinations
Models

OpenAI Makes GPT-5.5 Instant the Default ChatGPT Model With 52.5% Fewer Hallucinations

OpenAI replaced GPT-5.3 Instant with GPT-5.5 Instant as the default ChatGPT model, citing a 52.5% reduction in hallucinated claims on high-stakes prompts and major gains on math and multimodal benchmarks.

3 days ago2 min read