All reports
Technology by deep-research

Apple Core AI & Foundation Models — On-Device AI Ecosystem for macOS/iOS Apps

RemindrJarvisNeuron

Apple Core AI & Foundation Models — On-Device AI Ecosystem for macOS/iOS Apps

Research date: 2026-03-19 | Agent: Deep Research | Confidence: High

Executive Summary

  • Apple is replacing Core ML with Core AI at WWDC 2026, signaling a fundamental shift from “machine learning” to “artificial intelligence” as the platform primitive
  • The Foundation Models framework (launched 2025) gives developers Swift-native access to a ~3B parameter on-device LLM with guided generation, tool calling, and LoRA adapter fine-tuning
  • Apple’s privacy-first approach (on-device + Private Cloud Compute) creates a unique value proposition: AI without data leaving the device
  • The on-device AI market is projected to grow from $10.8B (2025) to $75.5B by 2033 (27.8% CAGR)
  • Apple-Google partnership (January 2026) will bring Gemini models to Apple devices, expanding server-side capabilities
  • For Moklabs: building Apple-native AI apps that leverage on-device processing could be a strategic differentiator, especially for privacy-sensitive use cases like Remindr and Jarvis

Market Size & Growth

Segment2025/2026ProjectionCAGRSource Confidence
On-device AI market$10.8B (2025)$75.5B by 203327.8%High
Mobile AI market$33.0B (2026)$258.1B by 2034Medium
North America on-device share38.5% (2026)Dominant regionHigh
Apple devices with Apple Intelligence2B+ active devicesGrowing with M/A-seriesHigh
App Store AI app revenueGrowing rapidlyNew wave of AI-native apps expectedMedium

Key Players

Apple’s AI Stack

LayerComponentStatusDetails
Framework (current)Core MLProduction (since 2017)Image classification, NLP, model deployment
Framework (upcoming)Core AIWWDC 2026 announcementReplaces Core ML; third-party AI model integration
Developer APIFoundation Models frameworkProduction (2025)Swift-native LLM access, guided generation, LoRA
On-device modelApple Foundation Model (~3B params)Production16 languages, multimodal (text + image), tool calling
Server modelApple Server Foundation ModelProductionPrivate Cloud Compute; matches GPT-4 class
Cloud infrastructurePrivate Cloud Compute (PCC)ProductionEnd-to-end encrypted; M5 chips in 2026
HardwareApple Silicon (M5/A-series)M5 launching 2026Dedicated Neural Engine for AI workloads
PartnerGoogle Gemini integrationAnnounced Jan 2026Multi-year partnership for server-side capabilities

Competing On-Device AI Ecosystems

PlatformOn-Device ModelDeveloper FrameworkDifferentiator
Apple~3B param AFMFoundation Models / Core AIPrivacy-first, Swift-native, LoRA adapters
GoogleGemini NanoML Kit / AI CoreBroader model selection, Android reach
QualcommLlama/custom modelsQualcomm AI EngineHardware-agnostic, multi-model support
SamsungGalaxy AI (Gemini-based)Samsung AI APIsSamsung ecosystem integration
MicrosoftPhi-3/4 (small models)ONNX RuntimeCross-platform, Windows/mobile

Technology Landscape

Foundation Models Framework (Production 2025)

The Foundation Models framework is Apple’s most significant developer tool for AI, providing:

Guided Generation: Type-safe structured outputs through @Generable and @Guide macros — compile-time schema generation with constrained decoding. Developers generate rich Swift data structures directly:

@Generable
struct RecipeAnalysis {
    var ingredients: [String]
    var difficulty: Difficulty
    var estimatedTime: Int
}

LoRA Adapter Fine-Tuning: Developers can train custom adapters (~160MB each) to specialize the on-device model for their domain. Uses rank-32 LoRA with frozen base weights. Adapters deploy via Background Assets framework.

Tool Calling: Constrained tool calling lets the on-device model interact with app functionality — launching features, querying databases, or triggering actions based on natural language input.

Xcode Integration: Built-in playground for prompt engineering, performance profiler for on-device inference, and simulator support for iOS/visionOS.

Core AI (Expected WWDC 2026)

The transition from Core ML to Core AI represents:

  1. Third-party model integration: Key focus — allowing developers to integrate external AI models (potentially via MCP) into apps without building from scratch
  2. Modern naming: “Machine learning” → “AI” reflects the broader capability set (LLMs, agents, multimodal)
  3. Backward compatibility: Core ML will likely coexist during transition period
  4. Expanded scope: Beyond inference — potentially including on-device training, agent orchestration, and multi-model pipelines

Apple’s On-Device Model Performance

BenchmarkApple On-Device (~3B)Comparable ModelsAssessment
General NLPBeats/ties Mistral, Microsoft, Google equivalentsPhi-3-mini, Gemma-2BStrong for size class
Image understandingSupports text + image inputsUnique for on-device
Language support16 languagesMost competitors: English-only on-deviceSignificant advantage
LatencyReal-time on Apple SiliconVaries by hardwareBest-in-class on Apple devices
Privacy100% on-device processingMost cloud-dependentCategory-defining

Private Cloud Compute Evolution

  • Current: Custom Apple Silicon servers with end-to-end encryption
  • 2026: M5 chips being deployed in PCC infrastructure
  • 2027: Dedicated AI server chips (mass production starting H2 2026)
  • Security model: Stateless computation — data processed only for the task, never stored
  • Verifiability: Publicly verifiable security guarantees through cryptographic attestation

Pain Points & Gaps

Developer Pain Points

  • Model size limitations: 3B parameter on-device model is capable but can’t match cloud LLMs for complex reasoning
  • Limited fine-tuning options: LoRA adapters are powerful but constrained — no full fine-tuning on device
  • Apple ecosystem lock-in: Foundation Models framework is Swift-only; cross-platform developers excluded
  • Model opacity: Developers can’t see model weights, architecture details, or intermediate outputs
  • Rate limiting concerns: Unclear how Apple handles high-frequency model calls in production apps
  • Testing complexity: Evaluating AI features requires device-specific testing; simulators have limitations

Market Gaps

  • No RAG framework: Apple provides the model and embeddings but no built-in RAG pipeline
  • Agent orchestration absent: No native framework for multi-agent or multi-step AI workflows on device
  • Limited third-party model support: Until Core AI arrives, running non-Apple models on device requires workarounds
  • Observability tools missing: No native telemetry for AI feature performance, accuracy, or cost tracking
  • Cross-device continuity: AI state doesn’t seamlessly transfer between iPhone/Mac/iPad in a session

Enterprise Gaps

  • MDM for AI features: IT admins can’t selectively enable/disable AI features per policy
  • Audit logging: No enterprise-grade logging of AI interactions for compliance
  • Custom model deployment: Enterprises can’t deploy proprietary models through Apple’s infrastructure

Opportunities for Moklabs

1. Remindr: On-Device Transcription + AI Processing (High Impact, Medium Effort)

  • Opportunity: Leverage Apple’s on-device foundation model for voice memo transcription and intelligent summarization — completely on-device, no cloud costs
  • Effort: 2-3 months to integrate Foundation Models framework
  • Impact: High — eliminates cloud transcription costs and addresses privacy concerns
  • Connection: Direct alignment with Remindr’s voice-first approach; Apple’s 16-language support covers global use case

2. Jarvis: Privacy-First Personal Knowledge on Apple Devices (High Impact, High Effort)

  • Opportunity: Build a Jarvis iOS/macOS client that processes personal knowledge entirely on-device using Foundation Models + LoRA adapters trained on the user’s data patterns
  • Effort: 4-6 months
  • Impact: Very High — “your AI assistant that never sends data to the cloud” is a compelling privacy narrative
  • Connection: Jarvis’s personal knowledge mission + Apple’s privacy positioning = natural synergy

3. AgentScope: On-Device Agent Monitoring (Medium Impact, Medium Effort)

  • Opportunity: As developers build AI features with Foundation Models, they’ll need observability — Apple provides no native solution. Build a lightweight AgentScope SDK for iOS/macOS that monitors on-device AI performance
  • Effort: 2-3 months
  • Impact: Medium — first-mover in Apple AI observability
  • Connection: Extends AgentScope to the fastest-growing AI deployment platform

4. Neuron: Apple-Native RAG for Personal Knowledge (Medium Impact, Medium Effort)

  • Opportunity: Apple provides the model but not a RAG pipeline. Build an on-device RAG framework for iOS/macOS that handles ingestion, chunking, embedding, and retrieval using Apple’s native capabilities
  • Effort: 3-4 months
  • Impact: Medium — fills a clear gap in Apple’s developer stack
  • Connection: Could become the knowledge retrieval layer for Jarvis and other Moklabs apps

Risk Assessment

Market Risks

  • Apple could build it: Apple may expand Foundation Models to include RAG, agent orchestration, and observability — commoditizing any third-party tools (High risk — Apple’s track record of absorbing third-party functionality)
  • Slow developer adoption: Foundation Models framework is new; developers may stick with cloud APIs that they already know (Medium risk — Apple’s ecosystem gravity is strong)
  • Gemini partnership uncertainty: The Apple-Google partnership scope is unclear; Google could provide more functionality than expected (Low risk)

Technical Risks

  • Model capability ceiling: 3B parameter model has inherent limitations; some use cases may require cloud fallback, complicating architecture (Medium risk)
  • Core AI breaking changes: WWDC 2026 Core AI announcement could invalidate assumptions about the developer API surface (Medium risk — plan for adaptation)
  • LoRA adapter limits: 160MB per adapter may be insufficient for highly specialized domains (Low risk — sufficient for most use cases)

Business Risks

  • Apple review process: AI apps face additional scrutiny in App Store review; rejection risk for certain use cases (Medium risk)
  • Platform dependency: Building on Apple-only frameworks limits TAM to Apple device users (Medium risk — but Apple users have higher LTV)
  • Pricing model unclear: How Apple will charge developers for heavy Foundation Models usage is unknown (Low risk — currently free with the framework)

Data Points & Numbers

MetricValueSourceConfidence
Apple on-device model size~3B parametersApple ML ResearchHigh
LoRA adapter storage~160MB per adapterApple Developer DocsHigh
LoRA rank used32Apple Developer DocsHigh
Languages supported16Apple ML ResearchHigh
On-device AI market (2025)$10.8BCoherent Market InsightsHigh
On-device AI market (2033)$75.5BCoherent Market InsightsHigh
On-device AI CAGR27.8%Coherent Market InsightsHigh
Mobile AI market (2026)$33.0BFortune Business InsightsMedium
Mobile AI market (2034)$258.1BFortune Business InsightsMedium
North America on-device share38.5%Market reportsHigh
Active Apple devices2B+AppleHigh
Core ML introduced2017AppleHigh
Core AI expectedWWDC 2026 (June)Apple Insider, 9to5MacMedium
Apple-Google AI partnershipJanuary 2026Multiple sourcesHigh
M5 PCC deployment20269to5MacMedium
Dedicated AI server chipsH2 2026 production startReportsMedium
Foundation Models framework launchSeptember 2025AppleHigh

Sources

Related Reports