OPEN_SOURCE ↗
REDDIT · REDDIT// 18d agoMODEL RELEASE
Devstral-Small-2-24B lands Opus reasoning tune
Adam Jenner fine-tuned Devstral-Small-2-24B on roughly 2.3k Claude 4.6 Opus reasoning traces, then shipped Q4_K_M and Q5_K_M GGUFs plus a LoRA adapter. The goal is to make the local coding model reason explicitly before it writes code, with Q5 recommended for best quality.
// ANALYSIS
This is less a SOTA chase than a behavior-packaging exercise: the value is in getting more deliberate coding out of a 24B model people can actually run locally.
- –The ~2.3k-sample dataset is small, so expect a planning/style lift rather than a dramatic jump in raw capability.
- –The VLM-to-text-only extraction is the real feat; it makes a multimodal base trainable on a single RTX 3090.
- –GGUFs plus a LoRA adapter are the right packaging for this audience: easy local testing, easy merges, easy Q4 vs Q5 comparisons.
- –The epoch-2 checkpoint choice, 2,048-token cap, and no-benchmark caveat suggest the model is tuned for practical generalization, not leaderboard theater.
- –Apache 2.0 licensing and open weights make it a useful sandbox for local coding agents and reasoning distills.
// TAGS
devstral-small-2-24b-opus-reasoningreasoningfine-tuningai-codingopen-weightsllm
DISCOVERED
18d ago
2026-03-24
PUBLISHED
18d ago
2026-03-24
RELEVANCE
8/ 10
AUTHOR
admajic