Qwen3 235B A22B Thinking 2507 pricing
Qwen3-235B-A22B-Thinking-2507 is a high-performance, open-weight Mixture-of-Experts (MoE) language model optimized for complex reasoning tasks. It activates 22B of its 235B parameters per forward pass and natively supports up to 262,144 tokens of context. This "thinking-only" variant enhances structured logical reasoning, mathematics, science, and long-form generation, showing strong benchmark performance across AIME, SuperGPQA, LiveCodeBench, and MMLU-Redux. It enforces a special reasoning mode. Live index: 7 priced offers. Best input $0.110 per million tokens from Openrouter. Best output $0.600 per million tokens from Openrouter.
Pricing across providers
Every row is a seller of Qwen3 235B A22B Thinking 2507 with token pricing we track. The cheapest input in this snapshot is from Openrouter. The bar chart shows the same input and output dollars per million for a quick scan.
| Provider | Input / 1M | Output / 1M | Cached input | Batch |
|---|---|---|---|---|
O Openrouter | $0.110 | $0.600 | — | — |
O Openrouter | $0.110 | $0.600 | — | — |
W Wandb | $10000.00 | $10000.00 | — | — |
N Novita | $0.300 | $3.00 | — | — |
D DeepInfra | $0.300 | $2.90 | — | — |
FA Fireworks AI | $0.220 | $0.880 | — | — |
TA Together AI | $0.650 | $3.00 | — | — |
Input vs output · per provider
Cost calculator
The calculator uses the same dollars per million tokens as the table. Adjust sliders to see how Qwen3 235B A22B Thinking 2507 cost scales with traffic.
Provider
0.011000¢ / req
0.030000¢ / req
Model specifications
Context length, caps, and capability flags for Qwen3 235B A22B Thinking 2507. Family: Qwen. Values follow the main provider (Alibaba) record in our index.
- Context window
- 262,144 tokens
- Max output
- 262,144 tokens
- Vision (images)
- No
- Tool / function calling
- Yes
- Streaming
- No
- Released
- Jul 2025
- Primary provider
- Alibaba
- Model family
- Qwen
Compare Qwen3 235B A22B Thinking 2507
Jump into a comparison when you want one table for two models instead of two tabs. 6 curated matches for Qwen3 235B A22B Thinking 2507.
- Qwen3 235B A22B Thinking 2507 vs GPT-4o
Compare pricing side by side
- Qwen3 235B A22B Thinking 2507 vs GPT-4o mini
Compare pricing side by side
- Qwen3 235B A22B Thinking 2507 vs Claude Sonnet 4.6
Compare pricing side by side
- Qwen3 235B A22B Thinking 2507 vs Gemini 2.0 Flash
Compare pricing side by side
- Qwen3 235B A22B Thinking 2507 vs o3
Compare pricing side by side
- Qwen3 235B A22B Thinking 2507 vs Llama 3.1 70B
Compare pricing side by side
Frequently asked questions
Answers pull from the same numbers you see on this page. The short model note from our index: Qwen3-235B-A22B-Thinking-2507 is a high-performance, open-weight Mixture-of-Experts (MoE) language model optimized for complex reasoning tasks. It activates 22B of its 235B parameters per forward pass and natively suppor...
Also from Alibaba
Other models by Alibaba with live pricing in our catalog.