BACK_TO_FEEDAICRIER_2
QwenChat iOS brings Qwen offline to iPhone
OPEN_SOURCE ↗
YT · YOUTUBE// 36d agoPRODUCT LAUNCH

QwenChat iOS brings Qwen offline to iPhone

QwenChat iOS is a native SwiftUI app that runs 4-bit Qwen 3.5 0.8B and 2B models locally on iPhone using Apple’s MLX stack. The project shows that multimodal chat with image input, streamed responses, and zero server-side inference is now practical on consumer iPhones.

// ANALYSIS

Small multimodal models are crossing from lab demo to usable mobile software, and QwenChat iOS is a sharp proof point for AI developers building on-device apps.

  • The app runs fully on-device after the initial Hugging Face model download, which makes privacy and offline usage a real product feature instead of a marketing claim.
  • It supports both text and photo input, making it a concrete example of vision-language UX on iPhone rather than just a basic local chat shell.
  • The stack matters: SwiftUI, MVVM, Apple MLX, and `mlx-swift-lm` give developers a credible template for shipping local inference on Apple hardware.
  • Streaming output, stop controls, model switching, and tokens-per-second metrics make it useful as both a demo and a performance reference for edge AI apps.
  • The hard limit is still platform friction: iOS 18+, Xcode 16+, and first-run model downloads mean this is a strong developer example today, not yet a mass-market mobile AI app.
// TAGS
qwenchat-iosllmmultimodalinferenceedge-ai

DISCOVERED

36d ago

2026-03-07

PUBLISHED

36d ago

2026-03-07

RELEVANCE

8/ 10

AUTHOR

Better Stack