Fine-tuning Llama 3.1 with Long SequencesSeptember 19, 2024 by Saaketh Narayan, Irene Dea, Brian Chu, Shashank Rajput and Vitaliy Chiley in Generative AI We are excited to announce that Mosaic AI Model Training now supports the full context length of 131K tokens when fine-tuning the Meta...
Training Highly Scalable Deep Recommender Systems on Databricks (Part 1)September 4, 2024 by Rithwik Ediga Lakhamsani, Asfandyar Qureshi, Karan Jariwala, Lin Yuan, Lu Wang (Mosaic AI), Saaketh Narayan and Ning Wang in Generative AI Recommender systems (RecSys) have become an integral part of modern digital experiences, powering personalized content suggestions across various platforms. These sophisticated systems and...
Turbocharged Training: Optimizing the Databricks Mosaic AI Stack With FP8March 21, 2024 by Mihir Patel, Cheng Li, Davis Blalock and Saaketh Narayan in Mosaic Research At Databricks, we believe that the best companies in the world, in every sector, will have AI-powered systems that are trained and customized...