OPEN_SOURCE ↗
REDDIT · REDDIT// 3h agoOPENSOURCE RELEASE
Gemma 3 270M gets thinking LoRA
firstbober released a LoRA adapter for Gemma 3 270M that pushes the tiny model toward a thinking-style output format, with explicit THINKING and RESPONSE tags. The author says the work started as a local function-calling experiment and evolved into a small-scale reasoning adapter built with synthetic data plus Qwen 3.6 35B A3B and GLM 5.1 outputs.
// ANALYSIS
This is less about raw intelligence and more about coercing a very small model into behaving like a reasoning system. The interesting part is the training recipe: format enforcement, synthetic data, and loss shaping matter more here than model size alone.
- –Gemma 3 270M is already a tiny base, so even a modest adapter that preserves structure is useful for edge and hobbyist experiments
- –The author’s emphasis on tag fidelity suggests the main win is controllable output formatting, not broad world knowledge
- –Procedural data plus teacher-generated examples is a practical path when you do not have enough GPU to train on large curated corpora
- –The post is also a reminder that gradient accumulation changes optimization dynamics, but it is not a literal substitute for more VRAM or a larger effective batch
- –If this continues into FunctionGemma-style adapters, the most interesting applications will likely be structured output and tool use, not open-ended chat
// TAGS
gemma-3-270mfine-tuningreasoningllmopen-source
DISCOVERED
3h ago
2026-04-28
PUBLISHED
4h ago
2026-04-28
RELEVANCE
8/ 10
AUTHOR
Firstbober