modelstop.top

Compare Models

Run side-by-side checks for pricing, context window, and latency.

gpt-oss:120b

ollama

gpt-oss:120b — available to run locally via Ollama on CPU and GPU hardware.

Context window
131,072 tokens
Input cost
Output cost
Latency (p50)