Qwen: Qwen-Max
qwen
Qwen-Max, based on Qwen2.5, provides the best inference performance among [Qwen models](/qwen), especially for complex multi-step tasks. It's a large-scale MoE model that has been pretrained on over 20 trillion...
- Context window
- 32,768 tokens
- Input cost
- $1.04 / 1M
- Output cost
- $4.16 / 1M
- Latency (p50)
- —
