Cut Your LLM Cost

Train a custom model to replace your LLM in one day

curl -fsSL https://distillabs.ai/install.sh | sh Copied! Talk to us →

Leverage Automated Fine-Tuning

Start Training in 30min

distil labs automatically creates synthetic data from your traces and fine-tunes a model for your task. Problem in, model out – the process is very simple:

  1. Upload your traces and system prompt
  2. Train your model
  3. Deploy with distil labs and integrate
Get started →

50-90% Lower Inference Costs

Replace expensive LLM API calls with purpose-built small models. The platform handles the heavy lifting, so you can focus on building new features:

  • 30-200x smaller models
  • Accuracy on par with much larger models
  • Latency as low as 50ms for most tasks
Contact us →

30M+ people use distil labs models today

From Problem to Model

$ distil model create my-classifier
ID:         $MODEL_ID
Name:       my-classifier
Created At: 2026-03-23 15:11:08

$ distil model upload-traces $MODEL_ID --data ./traces
Upload successful. Upload ID: $UPLOAD_ID

01

Upload Traces

Create a model and upload 5k - 10k traces of your current agent. Supports various text processing tasks: classification, QA, tool calling, multi-turn tool calling, and more.

$ distil model run-training $MODEL_ID
Kicked off SLM training ID $TRAINING_ID

$ distil model training $MODEL_ID
Training ID: $TRAINING_ID
Status: Distilling

02

Train Model

Start training with a single command. Get feedback on task performance in minutes and model ready in a few hours. Trained SLMs consistently match frontier models 100x larger.

$ distil model deploy remote $MODEL_ID
Training ID:     $TRAINING_ID
Deployment ID:   efd60b29-...-4f56dbc0b13f
Status: Active
URL:             https://your.deployment.url/$TRAINING_ID/v1
Secrets api_key: QtwT1Ah7Jaf6DPIHBPDMYdNTKT6ujrnn1hZkGtsb21U

$ distil model invoke $MODEL_ID
╭────────────────────────────────────────────────────────────────────────────────────────────────────╮
 Run uv run .../$TRAINING_ID/remote_client.py --question "Your question here" to invoke the model
╰────────────────────────────────────────────────────────────────────────────────────────────────────╯

03

Deploy & Invoke

Deploy your trained model to a hosted endpoint with one command, then invoke it immediately. No infrastructure to set up — just deploy and call.

from openai import OpenAI

# Just change the base URL — everything else stays the same
client = OpenAI(
    base_url="https://your.deployment.url/$TRAINING_ID/v1",
    api_key="your-distil-api-key",
)

response = client.chat.completions.create(
    model="model",
    messages=[{"role": "user", "content": "Classify: I want to return my order"}],
)
print(response.choices[0].message.content)
# → {"label": "return_request", "confidence": 0.97}

04

Integrate

Swap one URL in your existing code — that’s it. The distil labs endpoint is OpenAI-compatible, so any SDK or client that talks to OpenAI works out of the box.

What Our Customers Say

We needed a small model that could power our product on an IBM P11, entirely on-premises. distil labs’ fine-tuned models allowed us to ship a self-contained solution where the SLM and our graph platform coexist on the same hardware. For customers in regulated industries, this means AI-powered query generation with complete data privacy – nothing ever leaves their environment.

David J. Haglin

Co-Founder and CTO at Rocketgraph

Using distil labs, we were able to spin up highly accurate custom small models tailored to our workflows in no time. Those models cut our inference costs by roughly 50% without sacrificing quality. The distil labs team was incredibly supportive as we got started and helped us get to production smoothly.

Lucas Hild

Lucas Hild

Co-Founder & CTO at Knowunity

The distil labs platform accelerated the release of our cybersecurity-specialized language model, KINDI, enabling faster iterations with greater confidence. As a result, we ship InovaGuard improvements sooner and continuously boost investigation accuracy with every release.

Samir Bennacer

Samir Bennacer

Co-Founder and CTO at Octodet

The Team

Selim Nowicki photo

Selim Nowicki

Co-Founder & CEO

Jacek Gołębiowski photo

Jacek Gołębiowski

Co-Founder & CTO

Jonas Verschueren photo

Jonas Verschueren

Founding Engineer

Usman Zafar photo

Usman Zafar

Founding ML Engineer

Maciej Gryka photo

Maciej Gryka

Senior ML Engineer

Bartek Kruszczyński photo

Bartek Kruszczyński

Full Stack Engineer

Gabi Kadlecová photo

Gabi Kadlecová

Machine Learning Researcher

Vineeth Reddy Guda photo

Vineeth Reddy Guda

Founders Associate

Backed by

Project A Ventures logo NAP VC logo helloworld VC logo Amazon Web Services logo Google logo NVIDIA logo

Demos & Blog