π All Models
408 models Β· Page 2 of 12
Kimi K2 Thinking
Kimi K2 Thinking is Moonshot AIβs most advanced open reasoning model to date, extending the K2 series into agentic, long-horizon reasoning. Built on the trillion-parameter Mixture-of-Experts (MoE) architecture introduced in...
Deepseek V3.2
Qwen3.5 9B Fp8
Nvidia Nemotron 3 Super 120B A12b Bf16
Meta Llama 3.2 1B Instruct
Llama 4 Maverick 17B 128E
Facebook CWM
Holo3 35B A3b
Qwen2.5 32B
Qwen2.5 32B β Alibaba's Qwen series language model with strong multilingual and coding capabilities.
Qwen3 235B A22B Thinking 2507 FP8
Qwen3-235B-A22B-Thinking-2507 is a high-performance, open-weight Mixture-of-Experts (MoE) language model optimized for complex reasoning tasks. It activates 22B of its 235B parameters per forward pass and natively supports up to 262,144...
Qwen3-VL-8B-Instruct
Qwen3-VL-8B-Instruct is a multimodal vision-language model from the Qwen3-VL series, built for high-fidelity understanding and reasoning across text, images, and video. It features improved multimodal fusion with Interleaved-MRoPE for long-horizon...
Cogito V1 Preview Llama 8B
Cogito V1 Preview Llama 8B β Meta's Llama open-source language model, one of the most widely deployed open models.
DeepSeek R1 Distill Qwen 1.5B
Cogito V1 Preview Qwen 32B
Deepseek V3
DeepSeek R1 Distill Qwen 14B
DeepSeek R1 Distill Qwen 7B
Cogito V1 Preview Qwen 14B
Deepseek V3.1 Base
Cogito V1 Preview Llama 70B
DeepSeek R1 Distill Llama 70B
GLM 5 Fp4
Cogito V1 Preview Llama 70B Turbo
Qwen3 Next 80B A3b Thinking
Qwen3-Next-80B-A3B-Thinking is a reasoning-first chat model in the Qwen3-Next line that outputs structured βthinkingβ traces by default. Itβs designed for hard multi-step problems; math proofs, code synthesis/debugging, logic, and agentic...
Llama Guard 4 12B
Llama Guard 4 is a Llama 4 Scout-derived multimodal pretrained model, fine-tuned for content safety classification. Similar to previous versions, it can be used to classify content in both LLM...
Qwen3-VL-32B-Instruct
Qwen3-VL-32B-Instruct is a large-scale multimodal vision-language model designed for high-precision understanding and reasoning across text, images, and video. With 32 billion parameters, it combines deep visual perception with advanced text...
Meta Llama 3.1 70B Instruct Turbo
GLM 4.7 Fp8
Qwen3 Coder 30B A3b Instruct
Qwen3-Coder-30B-A3B-Instruct is a 30.5B parameter Mixture-of-Experts (MoE) model with 128 experts (8 active per forward pass), designed for advanced code generation, repository-scale understanding, and agentic tool use. Built on the...
Cogito v2.1 671B
meta-llama/Llama-Guard-4-12B
Llama Guard 4 is a Llama 4 Scout-derived multimodal pretrained model, fine-tuned for content safety classification. Similar to previous versions, it can be used to classify content in both LLM...
google/gemma-4-26B-A4B-it
Gemma 4 26B A4B IT is an instruction-tuned Mixture-of-Experts (MoE) model from Google DeepMind. Despite 25.2B total parameters, only 3.8B activate per token during inference β delivering near-31B quality at...
Qwen/Qwen3-Max
Qwen3-Max is an updated release built on the Qwen3 series, offering major improvements in reasoning, instruction following, multilingual support, and long-tail knowledge coverage compared to the January 2025 version. It...
Qwen/Qwen3-VL-235B-A22B-Instruct
Qwen3-VL-235B-A22B Instruct is an open-weight multimodal model that unifies strong text generation with visual understanding across images and video. The Instruct model targets general vision-language use (VQA, document parsing, chart/table...
openai/gpt-oss-120b
gpt-oss-120b is an open-weight, 117B-parameter Mixture-of-Experts (MoE) language model from OpenAI designed for high-reasoning, agentic, and general-purpose production use cases. It activates 5.1B parameters per forward pass and is optimized...
nvidia/Nemotron-3-Nano-30B-A3B
NVIDIA Nemotron 3 Nano 30B A3B is a small language MoE model with highest compute efficiency and accuracy for developers to build specialized agentic AI systems. The model is fully...
