In a world where technological advancements are not only adding new features but also fundamentally changing the way we experience our day-to-day lives, Apple was once again on the leading edge. It’s been a week of big reveals from tech giants, with Microsoft and Google holding their own conferences. Then, the focus shifted to Apple’s Worldwide Developers Conference 2024, and the payload was not disappointing. Apple revealed its progress in generative AI by having new advancements work on-device across its suite of products looking slicker than ever. This is a huge deal for Apple, and it’s only the beginning.
One of the greatest breakthroughs of Apple’s approach is its emphasis on executing much of the AI workload entirely on the devices. In combination with the company’s sophisticated system-on-a-chip processors and advances in machine learning research, Apple has developed world-class, low-latency AI that have the potential to fundamentally reshape the capabilities of smartphones and computers. In this article, we look at how Apple’s on-device AI is setting a new bar for technical excellence.
At the core of Apple’s on-device AI is a 3-billion parameter model. Although Apple has been tight-lipped about the specific model used – it’s likely a specialised adaptation of the OpenELM-3B of the OpenELM family. Opting for OpenELM is an elegant choice: it provides an experience optimised to run on constrained devices without sacrificing the quality of AI performance. This foundation model trained across massive troves of data, including a partnership with Shutterstock, as well as major news organisations, is at the core of Apple’s AI experience.
In addition to scale, the efficiency of Apple’s on-device AI is also remarkable. Grouped query attention and palletisation are two examples of measures that allow for significantly more inference, while keeping the model lightweight. These optimisations mean that AI services available on Apple’s most recent devices, including MacBooks with M1 chips and the iPhone 15 Pro series, will be not only smarter but also much faster, offering results on the order of milliseconds.
This ethos of not building the same thing twice, especially not in AI, is prevalent throughout Apple’s strategy. For instance, Low-Rank Adaptation (LoRA) adapters can be ‘placed on top’ of the foundational model to fine-tune it for specific tasks without cumbersome re-training of numerous distinct model copies, thereby extending the utility of devices for, say, drafting emails as well as extractive text summarisation. At the same time, tight control over processing and storage requirements on devices will still be feasible.
Why is Apple’s on-device AI also the most prosaic of possibilities? Because what the stakeholders trying to make sense of the demo might like to consider is that the functionalities they got a glimpse of didn’t score well merely because they worked, but also because of the promise this approach seems to be holding the experience of mobile technology back, making it more rather than less enjoyable, simpler rather than more complex. In the fall, when Apple releases the functionalities slated in this demo, tech workers and users will get to see if those system engineers have managed to do just that.
This would not have been possible without the company’s underlying commitment to innovation and attention to the user experience. Apple’s devices are famously both handsome and powerful, and its newest device provides a good glimpse into an innovative future. Moving more of the AI back to the device itself is just the latest chapter in the company’s mission to offer not just ‘gadgets’ but experiences, fundamentally changing the way we live. By creating specialised hardware and developing a tight software package, Apple isn’t just emulating the past – it’s predicting the future of mobile technology.
© 2024 UC Technology Inc . All Rights Reserved.