Research interests

None defined yet.

Team members 28

Organization Card
About org cards

About Us

MosaicML’s mission is to make efficient training of ML models accessible. We continually productionize state-of-the-art research on efficient model training, and study the combinations of these methods in order to ensure that model training is ✨ as optimized as possible ✨. These findings are baked into our highly efficient model training stack, the MosaicML platform.

If you have questions, please feel free to reach out to us on Twitter, Email, or join our Slack channel!

LLM Foundry

This repo contains code for training, finetuning, evaluating, and deploying LLMs for inference with Composer and the MosaicML platform.

Composer Library

The open source Composer library makes it easy to train models faster at the algorithmic level. It is built on top of PyTorch. Use our collection of speedup methods in your own training loop or—for the best experience—with our Composer trainer.


Fast, accurate streaming of training data from cloud storage. We built StreamingDataset to make training on large datasets from cloud storage as fast, cheap, and scalable as possible.

It’s specially designed for multi-node, distributed training for large models—maximizing correctness guarantees, performance, and ease of use. Now, you can efficiently train anywhere, independent of your training data location. Just stream in the data you need, when you need it. To learn more about why we built StreamingDataset, read our announcement blog.

StreamingDataset is compatible with any data type, including images, text, video, and multimodal data.

With support for major cloud storage providers (AWS, OCI, and GCS are supported today; Azure is coming soon), and designed as a drop-in replacement for your PyTorch IterableDataset class, StreamingDataset seamlessly integrates into your existing training workflows.

MosaicML Examples Repo

This repo contains reference examples for training ML models quickly and to high accuracy. It's designed to be easily forked and modified.

It currently features the following examples:

MosaicML Platform

The proprietary MosaicML Platform enables you to easily train large AI models on your data, in your secure environment.

With the MosaicML Platform, you can train large AI models at scale with a single command. We handle the rest — orchestration, efficiency, node failures, infrastructure.

Our platform is fully interoperable, cloud agnostic, and enterprise proven. It also seamlessly integrate with your existing workflows, experiment trackers, and data pipelines.