by AlibabaRank #80Score 79.4
The Qwen3.5 native vision-language Flash models are built on a hybrid architecture that integrates a linear attention mechanism with a sparse mixture-of-experts model, achieving higher inference efficiency. Compared to the 3 series, these models deliver a leap forward in performance for both pure text and multimodal tasks, offering fast response times while balancing inference speed and overall performance.
| Signal | Normalized | Weight | Contribution | Freshness |
|---|---|---|---|---|
Capabilities capability | 83.3 | 20% | 16.7 | 2026-03-23T03:03:05.893Z |
Benchmarks benchmark | 66.7 | 30% | 20.0 | 2026-03-23T03:03:05.893Z |
Pricing pricing_tier | 0.3 | 15% | 0.0 | 2026-03-23T03:03:05.893Z |
Context Window context_window | 95.2 | 10% | 9.5 | 2026-03-23T03:03:05.893Z |
Recency recency | 100.0 | 15% | 15.0 | 2026-03-23T03:03:05.893Z |
Output Capacity output_capacity | 80.3 | 10% | 8.0 | 2026-03-23T03:03:05.893Z |
| Capability | Supported |
|---|---|
| Vision | Yes |
| Reasoning | Yes |
| JSON Mode | Yes |
| Streaming | Yes |
| Function Calling | Yes |
| Web Search | No |