latentbrief
Back to news
Research1w ago

New Technique Reduces Memory Usage for AI Models

arXiv CS.LG

In brief

  • Researchers have developed a new method called LARS (Low-memory Activation-Rank Subspace) that addresses the memory issues faced by large language models during fine-tuning.
  • Unlike existing techniques like LoRA and IA3, which reduce parameters but still face memory constraints, LARS focuses on managing the activation subspace used during training.
    • This approach slashes memory usage by up to 51.95% on CPUs and 33.54% on GPUs, making it more efficient for edge devices like Raspberry Pi.
  • By targeting memory consumption directly, LARS enables better performance on resource-limited hardware while maintaining accuracy and speed.
    • This breakthrough could make advanced AI personalization accessible to a wider range of devices, from smartphones to IoT gadgets.
  • As the demand for on-device AI grows, LARS offers a promising solution to overcome current limitations in processing power and memory.

Terms in this brief

LARS
Low-memory Activation-Rank Subspace — a new method designed to reduce memory usage in large language models during fine-tuning. It focuses on managing the activation subspace used during training, significantly cutting down memory consumption on both CPUs and GPUs, making it more efficient for devices with limited resources like Raspberry Pi.

Read full story at arXiv CS.LG

More briefs