modelstop.top

Compare Models

Run side-by-side checks for pricing, context window, and latency.

Z.ai: GLM 4.6

z-ai

Compared with GLM-4.5, this generation brings several key improvements: Longer context window: The context window has been expanded from 128K to 200K tokens, enabling the model to handle more complex...

Context window
204,800 tokens
Input cost
$0.39 / 1M
Output cost
$1.90 / 1M
Latency (p50)