• AI Time to Impact
  • Posts
  • . . AI: Locally Trained and Hosted Models Another Step Closer (3.7.24)

. . AI: Locally Trained and Hosted Models Another Step Closer (3.7.24)

AI Industry, Efficient LLMs, Generative AI

Friends, today's edition offers yet another breakthrough in the world of AI and model training. It really seems like almost every week we are sharing a story of a potential paradigm shift in the resources required to train and run models. Today's development offers another glimpse into a future where running local AI models could become significantly more accessible.

As everyone fights over access to compute power, more effort shifting to increasing efficiency seems to be paying dividends. We also highlight a story where the research on complex reasoning within an AI model went into a public-facing model in just six months—a remarkable achievement that highlights how quick the rate of innovation is here! Thank you for reading; we encourage you to share this newsletter with others who might find it of interest!

-Marshall Kirkpatrick, Editor

First impacted: AI developers, Data scientists
Time to impact: Short

A new training strategy known as GaLore reportedly reduces memory usage by up to 65.5% for LLMs and up to 82.5% for optimizer memory, surpassing traditional methods. GaLore's creators highlight its effectiveness in pre-training on LLaMA 1B and 7B architectures with a C4 dataset, and they say it enables a 7B model to be pre-trained on consumer GPUs with 24GB memory, like the NVIDIA RTX 4090, without requiring parallel, checkpointing, or offloading strategies. Easier pre-training directly impacts the feasibility of updating local AI models more frequently and efficiently for individuals or smaller entities and has big implications for our ability to train and run local AI models! [Paper page - GaLore: Memory-Efficient LLM Training by Gradient Low-Rank Projection] Explore more of our coverage of: GaLore Training Strategy, Memory Usage Reduction, LLaMA Pre-Training. Share this story by email

First impacted: AI researchers, AI model developers
Time to impact: Short

Nous Research says it has launched a new model, Genstruct 7B, which is designed to create complex instructions from raw text and form semi-synthetic instruction datasets. Inspired by the Ada-Instruct paper, the model can generate questions for complex scenarios and is available for download from the company's HuggingFace page. Ada-Instruct seems to be capable of generating instructions with more than 100 steps in them, for things like code generation and dealing with other complex situations. This is super interesting as it shows how quickly the research to release window was, in this instance only 6 months. [via @NousResearch] Explore more of our coverage of: AI Research, Genstruct 7B, Instruction Generation. Share this story by email

First impacted: Robotics Engineers, AI Technology Enthusiasts
Time to impact: Long

Hugging Face, known for its open source AI and machine learning software, is launching a new open source robotics project, according to project leader Remi Cadene, who is a former Tesla employee. The Paris-based initiative will aim to develop affordable robotic systems integrating deep learning and embodied AI technologies, and the company is currently seeking engineers to join the team. [Hugging Face is launching an open source robotics project led by former Tesla scientist] Explore more of our coverage of: Open Source Robotics, Deep Learning, Embodied AI. Share this story by email

First impacted: Inflection users
Time to impact: Short

Inflection has launched Inflection-2.5, an updated personal AI that the company says matches the effectiveness of leading language learning models while using 40% less compute power for training. According to the company's blog post, the new model includes real-time web search capabilities and has experienced an increase in user interaction and retention. [Inflection-2.5: meet the world's best personal AI] Explore more of our coverage of: AI Development, Language Learning Models, Compute Efficiency. Share this story by email

That’s it for today! More AI news coming your way next on Monday.