π All Models
287 models Β· Page 1 of 8
Ministral 3 14B Instruct 2512
Meta Llama 3.1 8B
Salesforce Llama Rank V1 (8B)
Mistral (7B) Instruct v0.3
Gemma-2 Instruct (27B)
Gemma 2 27B by Google is an open model built from the same research and technology used to create the [Gemini models](/models?q=gemini). Gemma models are well-suited for a variety of...
Qwen 2.5 Coder 32B Instruct
Meta Llama 3 8B Instruct
Meta Llama 3 8B Instruct β Meta's Llama open-source language model, one of the most widely deployed open models.
Qwen3 Next 80B A3b Instruct
Qwen3-Next-80B-A3B-Instruct is an instruction-tuned chat model in the Qwen3-Next series optimized for fast, stable responses without βthinkingβ traces. It targets complex tasks across reasoning, code generation, knowledge QA, and multilingual...
Meta Llama 3 70B Instruct Turbo
Meta Llama 3 70B Instruct Turbo β Meta's Llama open-source language model, one of the most widely deployed open models.
Meta Llama 3.1 8B Instruct Turbo
Mistral (7B) Instruct v0.1
Llama 4 Maverick Instruct (17Bx128E) FP8
Meta Llama 3.2 1B Instruct
Nvidia Nemotron Nano 9B V2
GLM 4.6 Fp8
Deepseek Coder 33B Instruct
Llama 3.1 Nemotron 70B Instruct HF
Llama 3.1 Nemotron 70B Instruct HF β Meta's Llama open-source language model, one of the most widely deployed open models.
Qwen 2.5 14B Instruct
Qwen 2 Instruct (1.5B)
Qwen3 30B A3b
Qwen3, the latest generation in the Qwen large language model series, features both dense and mixture-of-experts (MoE) architectures to excel in reasoning, multilingual support, and advanced agent tasks. Its unique...
Glm 4.5 Air Fp8
Qwen3 8B
Qwen3-8B is a dense 8.2B parameter causal language model from the Qwen3 series, designed for both reasoning-heavy tasks and efficient dialogue. It supports seamless switching between "thinking" mode for math,...
Meta Llama 3 8B Instruct Reference
Rime Labs Arcana v2
DeepSeek R1 Distill Qwen 1.5B
Qwen3 235B A22B Thinking 2507 FP8
Qwen3-235B-A22B-Thinking-2507 is a high-performance, open-weight Mixture-of-Experts (MoE) language model optimized for complex reasoning tasks. It activates 22B of its 235B parameters per forward pass and natively supports up to 262,144...
Multilingual E5 Large Instruct
Qwen3-VL-8B-Instruct
Qwen3-VL-8B-Instruct is a multimodal vision-language model from the Qwen3-VL series, built for high-fidelity understanding and reasoning across text, images, and video. It features improved multimodal fusion with Interleaved-MRoPE for long-horizon...
Meta Llama 3.1 70B Instruct Turbo
Nous Hermes 2 Mixtral 8X7B Dpo
Llama Guard 4 12B
Llama Guard 4 is a Llama 4 Scout-derived multimodal pretrained model, fine-tuned for content safety classification. Similar to previous versions, it can be used to classify content in both LLM...
Qwen3 Coder 30B A3b Instruct
Qwen3-Coder-30B-A3B-Instruct is a 30.5B parameter Mixture-of-Experts (MoE) model with 128 experts (8 active per forward pass), designed for advanced code generation, repository-scale understanding, and agentic tool use. Built on the...
Mixtral-8x7B Instruct v0.1
LFM2-24B-A2B
Qwen3 Next 80B A3b Thinking
Qwen3-Next-80B-A3B-Thinking is a reasoning-first chat model in the Qwen3-Next line that outputs structured βthinkingβ traces by default. Itβs designed for hard multi-step problems; math proofs, code synthesis/debugging, logic, and agentic...
