top of page

Bringing Generative AI to the Edge

12/24/23

Editorial team at Bits with Brains

Major tech firms are increasingly optimizing large language models for on-device applications in a bid to make generative artificial intelligence more ambient and contextually aware.

Major tech firms are increasingly optimizing large language models for on-device applications in a bid to make generative artificial intelligence more ambient and contextually aware.


Apple has taken the lead with a method to efficiently run models that exceed available DRAM on mobile devices using flash storage. By techniques like "windowing" and "row-column bundling" the company has doubled the size of ML models possible on hardware while accelerating CPU inference fivefold.


This advances Apple's plans to incorporate generative capabilities into core iOS functions through Siri and Messages next year. Longer term, custom AI cores in rumored 'Apple Silicon' systems may further embed generative abilities.


Rivals are following suit - Samsung's Gauss is set for Galaxy phones in 2024, while Google's Gemini Nano aims to enhance the Pixel 8's assistant utilities. By hosting ML models locally, these innovations ease privacy-centric, self-supervised training that avoids transmitting private user data.


The shift also opens novel design possibilities as generative AI seamlessly augments interfaces. Yet integrating such systems presents challenges in cultivating user trust through transparency of responses. Careful testing will prove critical for responsible adoption.


Beyond mobile, flash memory optimizations apply to autonomous vehicles and other latency-sensitive industries. As specialized hardware evolves together with models, generative AI stands to become ubiquitous through diverse applications rooted in personalized, on-device learning.


This suggests growing priority for integrative, on-premises applications versus the more traditional centralized, cloud-based solutions. The move to localized generative AI presages an era of more contextual, integrated digital assistance and suggests privacy and personalization as key future priorities for major tech firms.


Sources:

https://analyticsindiamag.com/apple-optimises-llms-for-edge-use-cases/

Sources

bottom of page