How to run Qwen 3.5 locally
Curiositry
26 points
5 comments
March 07, 2026
Related Discussions
Found 5 related stories in 53.5ms across 3,471 title embeddings via pgvector HNSW
- Qwen3.5 Fine-Tuning Guide bilsbie · 311 pts · March 04, 2026 · 66% similar
- Qwen3.6-Plus: Towards real world agents pretext · 482 pts · April 02, 2026 · 64% similar
- Qwen3.6-Plus: Towards Real World Agents meetpateltech · 20 pts · April 02, 2026 · 64% similar
- Qwen3.5-Omni meetpateltech · 18 pts · March 30, 2026 · 62% similar
- The Qwen 3.5 Small Model Series armcat · 11 pts · March 02, 2026 · 61% similar
Discussion Highlights (2 comments)
Twirrim
I've been finding it very practical to run the 35B-A3B model on an 8GB RTX 3050, it's pretty responsive and doing a good job of the coding tasks I've thrown at it. I need to grab the freshly updated models, the older one seems to occasionally get stuck in a loop with tool use, which they suggest they've fixed.
Curiositry
Qwen3.5 9b seems to be fairly competent at text manipulation and OCR running in llama.cpp on CPU, albeit slow. However, I have compiled it umpteen ways and still haven't gotten GPU offloading working properly (which I had with Ollama), on an old 1650 Ti with 4GB VRAM (it tries to allocate too much memory).