Oumi: Open Universal Machine Intelligence
Everything you need to build state-of-the-art foundation models, end-to-end.
Oumi is a fully open-source platform that streamlines the entire lifecycle of foundation models - from data preparation and training to evaluation and deployment. Whether youβre developing on a laptop, launching large scale experiments on a cluster, or deploying models in production, Oumi provides the tools and workflows you need.
With Oumi, you can:
π Train and fine-tune models from 10M to 405B parameters using state-of-the-art techniques (SFT, LoRA, QLoRA, DPO, and more)
π€ Work with both text and multimodal models (Llama, Qwen, Phi, and others)
π Synthesize and curate training data with LLM judges
β‘οΈ Deploy models efficiently with popular inference engines (vLLM, SGLang)
π Evaluate models comprehensively across standard benchmarks
π Run anywhere - from laptops to clusters to clouds (AWS, Azure, GCP, Lambda, and more)
π Integrate with both open models and commercial APIs (OpenAI, Anthropic, Vertex AI, Parasail, β¦)
All with one consistent API, production-grade reliability, and all the flexibility you need for research. Oumi is currently in beta and under active development.
π Getting Started#
Notebook |
Try in Colab |
Goal |
---|---|---|
π― Getting Started: A Tour |
Quick tour of core features: training, evaluation, inference, and job management |
|
π§ Model Finetuning Guide |
End-to-end guide to LoRA tuning with data prep, training, and evaluation |
|
π Model Distillation |
Guide to distilling large models into smaller, efficient ones |
|
π Model Evaluation |
Comprehensive model evaluation using Oumiβs evaluation framework |
|
βοΈ Remote Training |
Launch and monitor training jobs on cloud (AWS, Azure, GCP, Lambda, etc.) platforms |
|
π LLM-as-a-Judge |
Filter and curate training data with built-in judges |
|
π vLLM Inference Engine |
Fast inference at scale with the vLLM engine |
π» Why use Oumi?#
If you need a comprehensive platform for training, evaluating, or deploying models, Oumi is a great choice.
Here are some of the key features that make Oumi stand out:
π§ Zero Boilerplate: Get started in minutes with ready-to-use recipes for popular models and workflows. No need to write training loops or data pipelines.
π’ Enterprise-Grade: Built and validated by teams training models at scale
π― Research Ready: Perfect for ML research with easily reproducible experiments, and flexible interfaces for customizing each component.
π Broad Model Support: Works with most popular model architectures - from tiny models to the largest ones, text-only to multimodal.
π SOTA Performance: Native support for distributed training techniques (FSDP, DDP) and optimized inference engines (vLLM, SGLang).
π€ Community First: 100% open source with an active community. No vendor lock-in, no strings attached.
π Where to go next?#
While you can dive directly into any section that interests you, we recommend following the suggested path below to get the most out of Oumi.
Category |
Description |
Links |
---|---|---|
π Getting Started |
Get up and running quickly with Oumi |
|
π User Guides |
Learn how to use Oumi effectively |
|
π€ Models |
Explore available models and recipes |
|
π§ Development |
Contribute to Oumi |
|
π API Reference |
Documentation of all modules |
π€ Join the Community!#
Oumi is a community-first effort. Whether you are a developer, a researcher, or a non-technical user, all contributions are very welcome!
To contribute to the
oumi
repository, please check theCONTRIBUTING.md
for guidance on how to contribute to send your first Pull Request.Make sure to join our Discord community to get help, share your experiences, and contribute to the project!
If you are interested by joining one of the communityβs open-science efforts, check out our open collaboration page.
β Need Help?#
If you encounter any issues or have questions, please donβt hesitate to:
Check our FAQ section for common questions and answers.
Open an issue on our GitHub Issues page for bug reports or feature requests.
Join our Discord community to chat with the team and other users.