BACK_TO_FEEDAICRIER_2
Selective contrastive training trims hallucinations with 10% data
OPEN_SOURCE ↗
REDDIT · REDDIT// 4h agoOPENSOURCE RELEASE

Selective contrastive training trims hallucinations with 10% data

This side project releases code for a selective contrastive post-training method that frames hallucination as a preference problem: a frozen base model generates a bad continuation, the training model compares it against the gold answer, and learning only happens when the preference margin is weak. The repo describes first-divergence loss masking, a gated objective, and benchmark gains that are presented as improved factuality with roughly 10% of the data.

// ANALYSIS

Strong idea, because it turns hallucination reduction into a targeted margin problem instead of brute-force full-data alignment.

  • The self-generated negative sample is the right kind of hard negative for this problem: it is model-produced, task-matched, and cheap to obtain.
  • The selective gate is the main efficiency win; if the reported results hold, it avoids wasting updates on already-separated cases.
  • The approach is conceptually close to preference optimization, but with a more surgical loss window after first divergence.
  • Main caveat: the evidence here is still self-reported project-level validation, so reproducibility and benchmark breadth matter more than the headline gain.
// TAGS
hallucination-mitigationllmcontrastive-learningpreference-optimizationpost-trainingfactualityopen-source

DISCOVERED

4h ago

2026-04-24

PUBLISHED

5h ago

2026-04-24

RELEVANCE

8/ 10

AUTHOR

Round_Apple2573