HELP - What settings do you use? Qwen3.5-35B-A3B
Sentiment Mix
Geography
Expert Signals
uber-linny
author • 1 mention
r/LocalLLaMA
source • 1 mention
AI-Generated Claims
Generated from linked receipts; click sources for full context.
I have a 16GB 9070xt , what settings do you use and what quant size for Qwen3.5-35B-A3B?
Supported by 1 story
I see every alot of people giving love to Qwen3.5-35B-A3B, but i feel like im setting it up incorrectly.
Supported by 1 story
cmd: C:\llamaROCM\llama-server.exe --port ${PORT} -m "C:\llamaROCM\models\Huihui-Qwen3.5-35B-A3B-abliterated.i1-IQ4_XS.gguf" -c 8192 -np 1 -ngl 99 -ncmoe 16 -fa on --temp 0.7 --top-k 20 --top-p 0.95 --min-p 0.00 --flash-attn on --cache-type-k f16 --cache-type-v f16 --threads 12 --context-shift --sleep-idle-seconds 300 -b 4096 -ub 2048
Supported by 1 story
Related Events
Qwen3.5 27B and 35B with 2x AMD 7900 XTX vLLM bench serve results
Hardware • 3/21/2026
Qwen 3.5 397B is the best local coder I have used until now
Uncategorized • 3/21/2026
RTX 5060 Ti 16GB vs Context Window Size
Uncategorized • 3/21/2026
Alibaba releases Qwen3-Coder-Next to rival OpenAI, Anthropic
LLMs • 3/21/2026
M5 Max 128G Performance tests. I just got my new toy, and here's what it can do.
Uncategorized • 3/21/2026