The breakthrough could make it easier for Apple to bring a GPT-like feature to the iPhone.
Apple’s AI researchers say they’ve made a huge breakthrough in their quest to deploy large language models (LLMs) on Apple devices like the iPhone without using a ton of memory. Researchers are instead bringing LLMs to the device using a new flash memory technique, MacRumors reports.
In a research paper called «LLM in a Flash: Efficient Large Language Model Inference with Limited Memory,» researchers note that in the world of mobile phones, flash storage is more prevalent than the RAM that is traditionally used to run LLMs.
Домой
United States
USA — software Apple Researchers Detail New Way to Run AI on Devices Like the...