Skip to main content
Page 1
Generative AI

LIMIT: Less Is More for Instruction Tuning

February 10, 2024 by Aditi Jha and Jacob Portes in Mosaic Research
How should you finetune a large language model for general-purpose question answering? One intriguing approach is that of supervised finetuning on a small...
Generative AI

MosaicBERT: Pretraining BERT from Scratch for $20

With the MosaicBERT architecture + training recipe, you can now pretrain a competitive BERT-Base model from scratch on the MosaicML platform for $20...
Generative AI

Efficiently Estimating Pareto Frontiers with Cyclic Learning Rate Schedules

April 8, 2022 by Jacob Portes in Mosaic Research
Benchmarking the tradeoff between model accuracy and training time is computationally expensive. Cyclic learning rate schedules can construct a tradeoff curve in a...