Oumi AI
May 11, 2026

Oumi can now host your Custom AI Models with a single click (and other news)

Updates for Week of May 11, 2026

By Stefan Webb

🚀 It has been just over a month since we launched the Oumi Platform and we’ve been thrilled to see what you all are using it to build - over 2,000 users from diverse industries have signed up and started solving real, pressing business problems (more about that in the case study below).

But we haven’t laid on our laurels, of course. In our first newsletter, Specialized Intelligence, we’d like to tell you about exciting new developments on the Oumi Platform.

Specialized Intelligence

First, what exactly is Specialized Intelligence and why should you care? Specialized Intelligence is the exact opposite of general purpose intelligence. Examples of general purpose intelligence include frontier LLMs like GPT-5.4, Opus 4.7, and DeepSeek V4.

These models are satisfactory at a wide range of tasks without excelling in any particular one. While good for developing prototypes, builders typically hit a road-block. API costs balloon, rate-limits are hit, latencies are long, and the final metrics don’t match those needed for in-production applications. These are problems that cannot be solved by clever context engineering.

“Custom AI that enterprises fully own and control is not just an advantage, it is the only way for them to stay relevant. Enterprises are maturing in their use of GenAI and quickly realizing it. They are acting on it. The wave from rented general purpose intelligence to fully owned specialized intelligence is quickly growing, even when current tools are lacking.” — Manos Koukoumidis, CEO @ Oumi

Specialized Intelligence, on the other hand, typically involves smaller LLMs, like Qwen-3.6, Gemma 4, and Phi 4 Mini. These models excel at specific tasks, exceeding the performance (accuracy, groundedness, etc.) of models like GPT-5.4 that are one-hundred times as large!

Because they are two-orders-of-magnitude smaller, inference costs can be two-orders-of-magnitude smaller. Because they can be run on-edge, on-premise, or self-hosted, rate-limits are circumvented. Likewise, inference latency can be half or more. We demonstrate small specialized models outperforming frontier LLMs on our blog.

Read our Vision Statement on “The Case for Specialized Intelligence”.

What’s new on the Oumi Platform

Managed Inference: Deploy your custom trained models on the Oumi Platform

  • You can host your fine-tuned models directly from the Oumi Platform (in Beta) and start inference immediately

  • Click on the new Deployments tab and then select “Create Deployment”, following the dialogs

  • We host an Open AI-compatible endpoint for you, accessible with an Oumi API key

New open source models available

  • Qwen 3.6/3.5 27B with LoRA fine-tuning

  • Gemma 4 2B, 4B with full-weight fine-tuning

Don’t see the model you need? Let us know at contact@oumi.ai

New data quality tests

  • Oumi now flags conversations with no user message and conversations where a system message appears anywhere other than the start.

  • Make it easier to identify noisy or invalid examples before using a dataset for evaluation, synthesis, or training.

Specialized Model of the Week

Triaging customer support tickets with a sub-1B Small Language Model that Beats GPT-5.4

In a recent blog post, “Building Customer Support with a sub-1B Small Language Model that Beats GPT-5.4”, we show how to build a custom 0.8B-parameter model with the Oumi platform for triaging customer support tickets outperforming frontier models like GPT-5.4 and Opus 4.6. For accuracy, it outperforms Opus 4.6 by 6.3% (see figure below) and GPT 5.4 by 8.4%. What’s more, it runs on a laptop (or even an iPhone), costs one hour and a few dollars to train, and ten minutes of human time, and we did it entirely by prompting the Oumi Agent—“look ma, no code!” Read more via the link.

Customer Story

DMG achieves 6% higher quality and 100x lower costs for invoice validation

In another recent post, we analyzed how DMG built specialized SLMs with Oumi to solve their business needs. DMG is an intermediary between contractors (carpenters, plumbers, electricians, etc.) and project managers. They have to validate invoices from these constructors, both to prevent fraud and also to ensure they are completed correctly.

DMG built a 0.6B-parameter model with Oumi for validating invoices that outperformed frontier models like GPT-5.2 and Kimi K2 (see table below). Read more via the link.

What have we been up to?

The last few weeks have been quite busy for Oumi. We have sponsored the Builder & Brews meetup, AI Dev 26 x SF (DeepLearning.ai), and the AI Agent Conference. Follow us on LinkedIn and Luma to stay updated on our upcoming events.

Until next time

The Oumi Platform continues to grow stronger — more users, more features, more business problems solved in record time! We look forward to your feedback - if you’d like to get in touch, drop us a line via our website.

Future newsletter will cover, in addition to applications, case studies, and platform updates, a discussion of recent news and research related to SLMs - subscribe now to stay in the loop!

Until then, Happy Building and May the Vibes be With You!

Further Reading