Skip to main content
Page 1

LLM Training and Inference with Intel Gaudi 2 AI Accelerators

January 4, 2024 by Abhi Venigalla and Daya Khudia in
At Databricks, we want to help our customers build and deploy generative AI applications on their own data without sacrificing data privacy or...

Training LLMs at Scale with AMD MI250 GPUs

October 30, 2023 by Abhi Venigalla in
Introduction Four months ago, we shared how AMD had emerged as a capable platform for generative AI and demonstrated how to easily and...

Training LLMs with AMD MI250 GPUs and MosaicML

June 30, 2023 by Abhi Venigalla in
With the release of PyTorch 2.0 and ROCm 5.4, we are excited to announce that LLM training works out of the box on...

Mosaic LLMs: GPT-3 quality for <$500k

September 29, 2022 by Abhi Venigalla and Linden Li in
Training large language models (LLMs) costs less than you think. Using the MosaicML platform, we show how fast, cheap, and easy it is...

Mosaic LLMs (Part 1): Billion-Parameter GPT Training Made Easy

August 11, 2022 by Abhi Venigalla and Linden Li in
In Part 1 of this LLM blog post series, we use the MosaicML platform to train vanilla GPT-3 models up to 1.3B params...