The breakthrough could make it easier for Apple to bring a GPT-like feature to the iPhone.
Apple’s AI researchers say they’ve made a huge breakthrough in their quest to deploy large language models (LLMs) on Apple devices like the iPhone without using a ton of memory. Researchers are instead bringing LLMs to the device using a new flash memory technique, MacRumors reports.
In a research paper called « LLM in a Flash: Efficient Large Language Model Inference with Limited Memory, » researchers note that in the world of mobile phones, flash storage is more prevalent than the RAM that is traditionally used to run LLMs.