BACK_TO_FEEDAICRIER_2
Gemma 3 270M gets thinking LoRA
OPEN_SOURCE ↗
REDDIT · REDDIT// 3h agoOPENSOURCE RELEASE

Gemma 3 270M gets thinking LoRA

firstbober released a LoRA adapter for Gemma 3 270M that pushes the tiny model toward a thinking-style output format, with explicit THINKING and RESPONSE tags. The author says the work started as a local function-calling experiment and evolved into a small-scale reasoning adapter built with synthetic data plus Qwen 3.6 35B A3B and GLM 5.1 outputs.

// ANALYSIS

This is less about raw intelligence and more about coercing a very small model into behaving like a reasoning system. The interesting part is the training recipe: format enforcement, synthetic data, and loss shaping matter more here than model size alone.

  • Gemma 3 270M is already a tiny base, so even a modest adapter that preserves structure is useful for edge and hobbyist experiments
  • The author’s emphasis on tag fidelity suggests the main win is controllable output formatting, not broad world knowledge
  • Procedural data plus teacher-generated examples is a practical path when you do not have enough GPU to train on large curated corpora
  • The post is also a reminder that gradient accumulation changes optimization dynamics, but it is not a literal substitute for more VRAM or a larger effective batch
  • If this continues into FunctionGemma-style adapters, the most interesting applications will likely be structured output and tool use, not open-ended chat
// TAGS
gemma-3-270mfine-tuningreasoningllmopen-source

DISCOVERED

3h ago

2026-04-28

PUBLISHED

4h ago

2026-04-28

RELEVANCE

8/ 10

AUTHOR

Firstbober