We benchmarked 12 small language models across 8 tasks to find the best base model for fine-tuning

December 10, 2025
Fine-tuned 12 small models to find which ones are most tunable and perform best after fine-tuning. Surprise finding: Llama-3.2-1B showed the biggest improvement (most tunable), while Qwen3-4B delivered the best final performance - matching a 120B teacher on 7/8 tasks and outperforming by 19 points on the SQuAD 2.0 dataset.

Vibe-Tuning: The Art of Fine-Tuning Small Language Models with a Prompt

December 7, 2025
Fine-tuning is a pain – you need datasets, ML expertise, and a stack of GPUs just to get started. Not anymore. With model vibe-tuning, you go from prompt to production-ready model without these headaches. This blog post shows you exactly how to build one, starting with just a prompt.

Distil Labs Enables Rocketgraph’s Private AI on IBM Power with Small Language Models

October 31, 2025
In this blog, we discuss how we fine-tuned a small language model to generate OpenCypher queries for the Rocketgraph analytics platform, thereby enabling accurate, efficient and privacy-first AI-powered natural language querying capabilities for Rocketgraph's customers.

Small expert agents from 10 examples

October 15, 2025
Distil labs turns a prompt and a few dozen examples into a small accurate expert agent. Our platform automates data generation, curation, fine-tuning, and evaluation—so you can reach LLM-level results with models 50–400× smaller, deployable almost anywhere, in hours.