LLM Training and Inference with Intel Gaudi 2 AI AcceleratorsJanuary 4, 2024 by Abhi Venigalla and Daya Khudia in Mosaic Research At Databricks, we want to help our customers build and deploy generative AI applications on their own data without sacrificing data privacy or...
Training LLMs at Scale with AMD MI250 GPUsOctober 30, 2023 by Abhi Venigalla in Mosaic Research Introduction Four months ago, we shared how AMD had emerged as a capable platform for generative AI and demonstrated how to easily and...
Training LLMs with AMD MI250 GPUs and MosaicMLJune 30, 2023 by Abhi Venigalla in Mosaic Research With the release of PyTorch 2.0 and ROCm 5.4, we are excited to announce that LLM training works out of the box on...
Mosaic LLMs: GPT-3 quality for <$500kSeptember 29, 2022 by Abhi Venigalla and Linden Li in Mosaic Research Training large language models (LLMs) costs less than you think. Using the MosaicML platform, we show how fast, cheap, and easy it is...
Mosaic LLMs (Part 1): Billion-Parameter GPT Training Made EasyAugust 11, 2022 by Abhi Venigalla and Linden Li in Mosaic Research In Part 1 of this LLM blog post series, we use the MosaicML platform to train vanilla GPT-3 models up to 1.3B params...