Apple is taking a deep dive into artificial intelligence technology, according to two recently published research papers showcasing the company’s work. The research shows Apple is working to develop on-device AI tech, including a groundbreaking method to create animatable avatars and a novel way to run large language models from an iPhone or iPad.
Aptly named “LLM in a flash,” Apple’s research on efficiently running LLMs on devices with limited memory enables complex AI applications to run smoothly on iPhones or iPads. This could also involve running a generative-AI-powered Siri on-device that simultaneously assists with various tasks, generates text, and features an improved ability to process natural language.
HUGS stands for Human Gaussian Splats, a method to create fully animatable avatars from short video clips captured on an iPhone in as little as 30 minutes. HUGS is a neural rendering framework capable of training with as little as a few seconds of video to create a detailed avatar that users can animate however they’d like.
What this means for the iPhone and Vision Pro
There have been reports about Apple working on its own AI chatbot, used internally and called ‘Apple GPT.’ The new research shows that the company is making strides in running LLMs by leveraging flash memory on smaller, less powerful devices like an iPhone. This could make sophisticated generative AI tools available on-device and could mean a generative AI-powered Siri.
Beyond Siri’s much-needed improvement, having an efficient LLM inference strategy like the one described in LLM in a Flash could lead to more accessible generative AI tools, significant advancements in mobile technology, and improved performance in a wide range of applications on everyday devices.
Arguably the biggest advancement of the two, HUGS is a method that can create malleable digital avatars from just a few seconds of monocular video, or 50-100 frames, to be exact. These human avatars can be animated and placed on different scenes, as the platform uses a disentangled representation of humans and scenes.
According to Apple, HUGS outperforms competitors at animating human avatars with rendering speeds 100 times faster than previous methods and with a significantly shorter training time of only 30 minutes.
Creating an avatar by leveraging the iPhone’s camera and processing power could deliver a new level of personalization and realism for iPhone users in social media, gaming, educational, and augmented reality (AR) applications.
HUGS could seriously reduce the creep factor for the Apple Vision Pro’s Digital Persona, showcased during the company’s last Worldwide Developers’ Conference (WWDC) last June. Vision Pro users could wield the power of HUGS to create a highly realistic avatar that can move fluidly with a 60fps rendering time.
The speed of HUGS would also allow for real-time rendering, which can be crucial for a smooth AR experience and could enhance social, gaming, and professional applications with realistic, user-controlled avatars.
Apple tends to shy away from using buzzwords like ‘AI’ to describe its product features, preferring to focus on machine learning instead. However, these research papers suggest a deeper involvement in new AI tech. Still, Apple hasn’t publicly acknowledged implementing generative AI into its products and has yet tto confirm its work with Apple GPT officially