Qwen3 235B A22B Thinking 2507 FP8
qwen
Qwen3-235B-A22B-Thinking-2507 is a high-performance, open-weight Mixture-of-Experts (MoE) language model optimized for complex reasoning tasks. It activates 22B of its 235B parameters per forward pass and natively supports up to 262,144...
- Context window
- 262,144 tokens
- Input cost
- $0.65 / 1M
- Output cost
- $3.00 / 1M
- Latency (p50)
- —
