Alibaba
✓ VerifiedCreator of the Qwen open-weight model family.
At a glance
Alibaba's AI arm
Alibaba's AI work centres on its DAMO Academy research lab and the cloud-services arm Alibaba Cloud, China's largest cloud provider. Alibaba's flagship LLM family is Qwen (通义千问, Tongyi Qianwen).
The Qwen series
Qwen has emerged as one of the most prominent open-weights LLM families in the world:
- Qwen 1 (2023) — initial release
- Qwen 2 (Jun 2024) — major architectural overhaul
- Qwen 2.5 (Sep 2024) — code, math, and long-context variants
- Qwen 3 (2025) — current flagship including 235B-parameter MoE
Qwen models are released under Apache 2.0 (with some exceptions for the largest sizes) and are widely used across the global open-source ecosystem, particularly in benchmarks and fine-tuning research.
Strategic role
Alibaba has been one of the most aggressive open-weights publishers in China, alongside DeepSeek. Its US$53 billion three-year AI capex commitment (announced Feb 2025) signals continued prioritisation of generative AI over its traditional commerce business.
Latest activity
Full history →- 0
Qwen2.5 72B Instruct released
Qwen2.5 72B Instruct discovered via OpenRouter. Provider: Alibaba.
Qwen2.5 72B Instruct - 0
Qwen: Qwen2.5 7B Instruct released
Qwen: Qwen2.5 7B Instruct discovered via OpenRouter. Provider: Alibaba.
Qwen2.5 7B Instruct - 0
Releases timeline
Showing 12 most recent- Qwen3.6 Plus
Qwen 3.6 Plus builds on a hybrid architecture that combines efficient linear attention with sparse mixture-of-experts routing, enabling strong scalability and high-performance inference. Compared to the 3.5 series, it delivers...
- Qwen3.5-9B
Qwen3.5-9B is a multimodal foundation model from the Qwen3.5 family, designed to deliver strong reasoning, coding, and visual understanding in an efficient 9B-parameter architecture. It uses a unified vision-language design...
- Qwen3.5-35B-A3B
The Qwen3.5 Series 35B-A3B is a native vision-language model designed with a hybrid architecture that integrates linear attention mechanisms and a sparse mixture-of-experts model, achieving higher inference efficiency. Its overall...
- Qwen3.5-27B
The Qwen3.5 27B native vision-language Dense model incorporates a linear attention mechanism, delivering fast response times while balancing inference speed and performance. Its overall capabilities are comparable to those of...
Active models
- Qwen3.6 PlusActive
Qwen 3.6 Plus builds on a hybrid architecture that combines efficient linear attention with sparse mixture-of-experts routing, enabling strong scalability and high-performance inference. Compared to the 3.5 series, it delivers...
1M ctx - Qwen3.5-9BActive
Qwen3.5-9B is a multimodal foundation model from the Qwen3.5 family, designed to deliver strong reasoning, coding, and visual understanding in an efficient 9B-parameter architecture. It uses a unified vision-language design...
262K ctx - Qwen3.5-35B-A3BActive
The Qwen3.5 Series 35B-A3B is a native vision-language model designed with a hybrid architecture that integrates linear attention mechanisms and a sparse mixture-of-experts model, achieving higher inference efficiency. Its overall...
Community ratings
Rate Alibaba
Sign in to rate and review.
Comments
Sign in to leave a comment.