Alibaba logo

Qwen3-235B-A22B-Thinking-2507

Alibaba

Qwen3-235B-A22B-Thinking-2507 is an advanced thinking-mode model on the Mixture-of-Experts architecture with 235B total parameters (22B active). Features 94 layers, 128 experts (8 active), and supports a native context length of 262K. This version is significantly improved in reasoning, achieving leading results among open models with thinking mode in logic, math, science, coding, and academic benchmarks.

Key Specifications

Parameters
235.0B
Context
256.0K
Release Date
July 24, 2025
Average Score
91.3%

Timeline

Key dates in the model's history
Announcement
July 24, 2025
Last Update
January 22, 2026
Today
March 25, 2026

Technical Specifications

Parameters
235.0B
Training Tokens
-
Knowledge Cutoff
-
Family
-
Capabilities
MultimodalZeroEval

Pricing & Availability

Input (per 1M tokens)
$0.30
Output (per 1M tokens)
$3.00
Max Input Tokens
256.0K
Max Output Tokens
131.1K
Supported Features
Function CallingStructured OutputCode ExecutionWeb SearchBatch InferenceFine-tuning

Benchmark Results

Model performance metrics across various tests and benchmarks

Other Tests

Specialized benchmarks
MMLU-Redux
Self-reported
94.0%
AIME 2025
Self-reported
92.0%
IFEval
Self-reported
88.0%

License & Metadata

License
apache-2.0
Announcement Date
July 24, 2025
Last Updated
January 22, 2026

Compare Qwen3-235B-A22B-Thinking-2507

All comparisons

Similar Models

All Models

Recommendations are based on similarity of characteristics: developer organization, multimodality, parameter size, and benchmark performance. Choose a model to compare or go to the full catalog to browse all available AI models.