A new version of the Qwen 3.5 language model is now available.

The availability of a smaller version of the Qwen 3.5 model could facilitate its use in scenarios with limited computational resources. This is particularly relevant for those wishing to make on-premise or edge deployments, where resource optimization is key.

For those evaluating on-premise deployments, there are trade-offs to consider. AI-RADAR offers analytical frameworks on /llm-onpremise to evaluate these aspects.