
According to a new report from Bloomberg’s Mark Gurman, Apple is poised to execute one of its most significant software transitions in a decade at the upcoming Worldwide Developers Conference (WWDC) 2026. The tech giant is reportedly preparing to deprecate its long-standing Core ML framework in favor of a modernized, successor architecture dubbed "Core AI." This shift, targeted for iOS 27, iPadOS 27, and macOS 27, represents a fundamental re-engineering of how Apple devices handle machine learning and artificial intelligence tasks.
For nearly ten years, Core ML has been the bedrock of Apple’s on-device intelligence, powering features ranging from Face ID to photo analysis. However, as the industry has surged toward Large Language Models (LLMs) and complex generative AI agents, the legacy infrastructure of Core ML has reportedly struggled to keep pace with the computational demands of modern models. Gurman’s report suggests that Core AI is not merely a rebranding but a "ground-up rewrite" designed to make integrating state-of-the-art generative models as simple as importing a UI library.
This move signals Apple's intent to aggressively reclaim leadership in the AI sector, moving beyond the predictive capabilities of the past into a new era of generative, context-aware on-device intelligence.
Launched in 2017, Core ML was built for a different era of machine learning. Its primary focus was on classification, regression, and image recognition—tasks that defined the "smart" features of the late 2010s. While Apple updated the framework annually, adding support for new layers and quantization methods, the underlying architecture remained rooted in traditional neural network processing.
The explosive growth of Generative AI has exposed the limitations of this aging framework. Developers have long complained about the friction involved in converting PyTorch or TensorFlow models into the proprietary .mlmodel format, a process that often resulted in performance degradation or unsupported operators.
The transition to Core AI addresses several critical bottlenecks inherent in the current ecosystem:
Core AI is expected to introduce native support for common industry standards, potentially allowing developers to run models closer to their native formats without the cumbersome translation layer that defined the Core ML era.
While specific technical documentation awaits the WWDC keynote, leaks indicate that Core AI focuses on three pillars: Modularity, Generative Native capabilities, and Unified Memory Architecture (UMA) optimization.
Unlike its predecessor, Core AI is built with Transformers and Diffusion models as first-class citizens. The framework reportedly includes pre-optimized "Foundation Blocks"—building blocks that allow developers to assemble AI pipelines (such as RAG, or Retrieval-Augmented Generation) without writing low-level matrix multiplication code. This could democratize local AI development, allowing a solo developer to implement features that previously required a team of ML engineers.
The new framework is rumored to unlock "Direct Path" access to the Apple Neural Engine (ANE). Previously, the OS managed ANE allocation conservatively to preserve battery life. Core AI supposedly introduces "Burst Mode" inference, allowing apps to command peak NPU performance for short durations—ideal for generating images or summarizing long documents on the fly in iOS 27.
For the Apple developer community, the arrival of Core AI marks a watershed moment. The complexity barrier for adding AI features is expected to drop significantly.
Apple’s goal with Core AI is to make import CoreAI as standard as import SwiftUI. The framework is expected to abstract away the complexities of tokenization, samplers, and context window management. Instead of writing hundreds of lines of code to manage an LLM's state, developers might accomplish the same task with declarative APIs similar to how SwiftUI handles views.
To understand the magnitude of this shift, we can look at the comparative capabilities of the outgoing and incoming frameworks:
Table: Feature Comparison Between Core ML and Core AI
| Feature | Core ML (Legacy) | Core AI (New Framework) |
|---|---|---|
| Primary Era | 2017–2025 (Predictive AI) | 2026+ (Generative AI) |
| Model Format | Proprietary .mlmodel (Conversion required) |
Native / Open Standard Compatibility |
| Hardware Focus | Balanced CPU/GPU/ANE distribution | Neural Engine First (Tensor Optimization) |
| GenAI Support | Limited via external libraries | Native LLM & Diffusion Primitives |
| Memory Handling | Static loading | Dynamic paging & Swap optimization |
| Developer API | Imperative, low-level configuration | Declarative, Intent-based APIs |
Note: The table above reflects reported features based on current leaks and may be subject to change upon official release.
The introduction of Core AI is not just a technical update; it is a strategic maneuver to differentiate Apple’s ecosystem through privacy-centric, on-device processing. Competitors like Google and Microsoft have leaned heavily into cloud-based AI processing. By empowering iOS 27 with a framework capable of running powerful models locally, Apple doubles down on its privacy narrative.
With Core AI, Apple aims to process the vast majority of personal context—emails, messages, health data—strictly on the device. The framework reportedly includes a "Gatekeeper" API that intelligently decides whether a request can be handled locally by the Neural Engine or if it requires Apple's Private Cloud Compute. This ensures that sensitive user data never leaves the device unless absolutely necessary, and even then, it does so under strict anonymization protocols.
This software advance coincides with rumored hardware leaps. The A20 chip expected in the iPhone 18 lineup is rumored to feature a Neural Engine specifically tuned for Core AI instructions, offering a multiplier effect on performance. However, Apple is known for its backward compatibility, and Core AI is expected to bring performance improvements even to older devices running iOS 27, likely starting from the iPhone 15 Pro series.
As June approaches, the tech world will be watching the keynote closely. If Mark Gurman’s reporting holds true, the announcement of Core AI will likely be the centerpiece of the event, overshadowing even new hardware reveals.
Developers should prepare for a transition period. Apple typically allows a deprecation window of 1-2 years. While Core ML will likely remain available in iOS 27 to ensure existing apps do not break, new features and optimizations will be exclusive to Core AI.
Key Milestones to Watch:
The shift to Core AI represents the maturation of Apple's silicon investment. After years of building the fastest mobile chips, Apple is finally releasing the software architecture necessary to unleash their full potential in the age of Generative AI. For Creati.ai readers, this underscores the vital importance of staying adaptable—the tools used to build intelligent apps are evolving just as fast as the AI models themselves.