Loading...
Loading...
Qwen3-235B-A22B-Instruct-2507 is a multilingual, instruction-tuned mixture-of-experts language model based on the Qwen3-235B architecture, with 22B active parameters per forward pass. It is optimized for general-purpose text generation, including instruction following, logical reasoning, math, code, and tool usage. The model supports a native 262K context length and does not implement "thinking mode" (<think> blocks). Compared to its base variant, this version delivers significant gains in knowledge coverage, long-context reasoning, coding benchmarks, and alignment with open-ended tasks. It is particularly strong on multilingual understanding, math reasoning (e.g., AIME, HMMT), and alignment evaluations like Arena-Hard and WritingBench.
Price/1M
$0.08
214th cheapest
75% below median
Top 32%
Context Window
262K
61st largest
Top 25%
Input
$0.07
per 1M tokens
Output
$0.10
per 1M tokens
Blended
$0.08
per 1M tokens
Cheaper than 68% of models. Median price is $0.31/1M tokens.
Daily
$0.08
Monthly
$2.35
Context Window
262K
tokens
Larger than 75% of models
Context Window Comparison