OPEN_SOURCE ↗
REDDIT · REDDIT// 12d agoOPENSOURCE RELEASE
Qwen3.5 4B fits uncensored roleplay
A LocalLLaMA help thread asks for a totally uncensored roleplay model that can run on an i5-10400F / RTX 3060 Ti 8GB box. Commenters quickly steer the user toward HauhauCS's Qwen3.5-4B Uncensored GGUF, with smaller 8B-and-below variants also in the mix.
// ANALYSIS
This thread is a good reminder that local LLM "uncensored" requests are mostly hardware-fitting problems, not secret-model-hunting problems. For an 8GB card, the 4B Qwen variant is the first suggestion that looks genuinely sane.
- –The model card lists GGUF quants as low as 2.6GB for Q4_K_M and 3.3GB for Q6_K, so it leaves real headroom for context and runtime overhead on a 3060 Ti.
- –It is natively multimodal with 262K context, which is more capability than this use case strictly needs, but a nice bonus if the user wants image inputs later.
- –The "aggressive" uncensored variant is explicitly tuned to strip refusals, which suits roleplay but also means looser guardrails by design.
- –Bigger Qwen3.5-35B-class suggestions in the thread are more aspirational than practical unless the user accepts a lot of offload and latency.
// TAGS
qwen3.5-4b-uncensored-hauhaucs-aggressivellmopen-sourceself-hostedmultimodalchatbot
DISCOVERED
12d ago
2026-03-30
PUBLISHED
12d ago
2026-03-30
RELEVANCE
8/ 10
AUTHOR
Opening-Ad6258