Categories: FAANG

Apple Intelligence Foundation Language Models Tech Report 2025

We introduce two multilingual, multimodal foundation language models that power Apple Intelligence features across Apple devices and services: (i) a ∼3B-parameter on-device model optimized for Apple silicon through architectural innovations such as KV-cache sharing and 2-bit quantization-aware training; and (ii) a scalable server model built on a novel Parallel-Track Mixture-of-Experts (PT-MoE) transformer that combines track parallelism, mixture-of-experts sparse computation, and interleaved global–local attention to deliver high quality with competitive cost on Apple’s Private Cloud Compute…
AI Generated Robotic Content

Recent Posts

Closed-source AI hate is understandable, but local AI has nothing that should concern AI haters

Let’s face it, AI is forbidden to be praised or used in pretty much any…

10 hours ago

Building AI Agents with Local Small Language Models

The idea of building your own AI agent used to feel like something only big…

10 hours ago

ParaRNN: Large-Scale Nonlinear RNNs, Trainable in Parallel

Recurrent Neural Networks (RNNs) are naturally suited to efficient inference, requiring far less memory and…

10 hours ago

Amazon Quick for marketing: From scattered data to strategic action

Imagine the following scenario: You’re leading marketing campaigns, creating content, or driving demand generation. Your…

10 hours ago

US Special Forces Soldier Arrested for Polymarket Bets on Maduro Raid

The master sergeant allegedly used classified intel to profit on the capture of Venezuelan president…

11 hours ago