Qwen3-Coder: improved performance on RTX 5090 with llama.cpp
A user reported a significant throughput increase, up to 26 tokens/second, using the Qwen3-Coder-Next-Q4_K_S model with llama.cpp on an RTX 5090. The optimization was achieved by offloading MoE expert tensors to the CPU and quantizing the KV cache.