BACK_TO_FEEDAICRIER_2
Qwen3.5-35B-A3B turns photos into rough 3D scenes
OPEN_SOURCE ↗
REDDIT · REDDIT// 32d agoVIDEO

Qwen3.5-35B-A3B turns photos into rough 3D scenes

A Reddit user demoed Qwen3.5-35B-A3B generating walkable HTML 3D scenes from photos using llama.cpp and a Q4 quant, then shared the results on YouTube. It is clearly an experimental community showcase rather than a product launch, but it highlights how far open multimodal models have pushed into lightweight spatial reasoning and scene reconstruction.

// ANALYSIS

This is messy, impractical, and exactly the kind of hack that hints at where open models are headed next.

  • The interesting part is not visual polish but the model’s ability to infer depth, layout, and object placement from a single image
  • Running it through llama.cpp with a quantized 35B-class model makes the demo more notable for local AI builders than a cloud-only proof of concept would be
  • Outputting HTML 3D scenes instead of a proprietary format suggests a low-friction path for browser demos, agent environments, and synthetic scene prototyping
  • It is still far from production-ready, but it points toward multimodal models becoming useful front ends for lightweight 3D authoring workflows
// TAGS
qwen3-5-35b-a3bllmmultimodalopen-weights

DISCOVERED

32d ago

2026-03-10

PUBLISHED

36d ago

2026-03-06

RELEVANCE

6/ 10

AUTHOR

c64z86