Car Wash Test exposes LLM logic-blindness
A benchmark of 12 LLMs across 360 'Car Wash Test' variations reveals that social distractors often trigger alignment protocols that override basic physical reasoning. Models frequently prioritize relationship advice over logical necessity, demonstrating a significant alignment tax on causal understanding.
AI "alignment" has created a logic-blindness where models prioritize being a marriage counselor over being a functional assistant.
* Social distractors like "overweight" or "wife" trigger safety and politeness protocols that bypass the model's ability to process physical logic.
* High "thinking" token counts in models like Qwen 4B when social conflict is present indicate a computational struggle between logical truth and RLHF conditioning.
* The "Car Wash Test" remains a definitive "sanity check" for distinguishing between probabilistic word association and true causal understanding in LLMs.
DISCOVERED
1d ago
2026-04-11
PUBLISHED
1d ago
2026-04-10
RELEVANCE
AUTHOR
Excellent_Jelly2788