OPEN_SOURCE ↗
REDDIT · REDDIT// 5h agoNEWS
LLMs Still Botch ASCII Art
A Reddit user asks which local LLM can generate decent ASCII art after Qwen failed and Gemma 3/4 only got partway there. The thread quickly turns into a broader debate about whether text-only models are the wrong tool for a 2D, whitespace-sensitive task.
// ANALYSIS
Hot take: this looks less like a model-quality problem and more like a representation problem. ASCII art asks an LLM to preserve spatial layout in a token stream that was never built for precise grids.
- –Commenters point to older Claude, ChatGPT, and Llama 3/4 variants as better at the task, but nobody sounds confident that any general LLM is truly reliable.
- –A parallel LocalLLaMA discussion argues tokenization and whitespace handling are the real failure mode, not raw “intelligence.”
- –The more practical workaround is to generate code, HTML/CSS, or a structured intermediate representation first, then render it into ASCII.
- –Several commenters suggest vision-in-the-loop or fine-tuned approaches if you really want consistent output instead of model lottery.
- –Net: this is an interesting benchmark for formatting discipline, but not a sign that one frontier chat model has “solved” ASCII art.
// TAGS
llmprompt-engineeringlocal-llmqwengemmaclaudellama
DISCOVERED
5h ago
2026-04-27
PUBLISHED
5h ago
2026-04-26
RELEVANCE
6/ 10
AUTHOR
Ne00n