π All Models
105 models Β· Page 1 of 3
Qwen3 Next 80B A3b Instruct
Qwen3-Next-80B-A3B-Instruct is an instruction-tuned chat model in the Qwen3-Next series optimized for fast, stable responses without βthinkingβ traces. It targets complex tasks across reasoning, code generation, knowledge QA, and multilingual...
Deepseek Coder 33B Instruct
Qwen 2.5 Coder 32B Instruct
Qwen3 Coder 30B A3b Instruct
Qwen3-Coder-30B-A3B-Instruct is a 30.5B parameter Mixture-of-Experts (MoE) model with 128 experts (8 active per forward pass), designed for advanced code generation, repository-scale understanding, and agentic tool use. Built on the...
Qwen3 Next 80B A3b Thinking
Qwen3-Next-80B-A3B-Thinking is a reasoning-first chat model in the Qwen3-Next line that outputs structured βthinkingβ traces by default. Itβs designed for hard multi-step problems; math proofs, code synthesis/debugging, logic, and agentic...
Qwen/Qwen3-Coder-480B-A35B-Instruct-Turbo
Qwen/Qwen3-Coder-480B-A35B-Instruct
deepseek-ai/DeepSeek-V3
DeepSeek V3 β 671B MoE model with exceptional coding and math performance at very low cost.
Qwen/Qwen2.5-72B-Instruct
Alibaba Qwen2.5 72B Instruct on DeepInfra β top open-source model with multilingual and coding strengths.
codestral-2508
Our cutting-edge language model for coding released August 2025.
codestral-2508
Our cutting-edge language model for coding released August 2025.
devstral-medium-2507
Our medium code-agentic model.
devstral-small-2507
Our small open-source code-agentic model.
claude-opus-4.6
Anthropic's most intelligent model with state-of-the-art coding, reasoning, and agentic capabilities
gpt-5.4
OpenAI's most capable frontier model for complex professional work, coding, and multi-step reasoning.
Qwen2.5-Coder-7B-Instruct
Qwen2.5-Coder-7B-Instruct β Alibaba's Qwen series language model with strong multilingual and coding capabilities.
Qwen2.5-Coder-7B-Instruct-GPTQ-Int4
Open-source Qwen2.5-Coder-7B-Instruct-GPTQ-Int4 model from qwen β available for download and self-hosting on Hugging Face.
qwen3-coder-next
qwen3-coder-next β available to run locally via Ollama on CPU and GPU hardware.
qwen3-coder:480b
qwen3-coder:480b β available to run locally via Ollama on CPU and GPU hardware.
Qwen3-Coder-Next-FP8
Open-source Qwen3-Coder-Next-FP8 model from qwen β available for download and self-hosting on Hugging Face.
Qwen2.5-Coder-32B-Instruct
Open-source Qwen2.5-Coder-32B-Instruct model from qwen β available for download and self-hosting on Hugging Face.
Qwen2.5-Coder-14B-Instruct
Open-source Qwen2.5-Coder-14B-Instruct model from qwen β available for download and self-hosting on Hugging Face.
Qwen2.5-Coder-7B
Qwen2.5-Coder-7B β Alibaba's Qwen series language model with strong multilingual and coding capabilities.
Qwen/Qwen2.5-Coder-32B-Instruct
Qwen/Qwen2.5-Coder-32B-Instruct is a text generation model on Hugging Face with ~1,107,488 monthly downloads. Open access.
Qwen/Qwen2.5-Coder-7B-Instruct-GPTQ-Int4
Qwen/Qwen2.5-Coder-7B-Instruct-GPTQ-Int4 is a text generation model on Hugging Face with ~1,124,729 monthly downloads. Open access.
Qwen3-Coder-30B-A3B-Instruct
Open-source Qwen3-Coder-30B-A3B-Instruct model from qwen β available for download and self-hosting on Hugging Face.
Qwen/Qwen2.5-Coder-7B-Instruct
Qwen/Qwen2.5-Coder-7B-Instruct is a text generation model on Hugging Face with ~2,225,884 monthly downloads. Open access.
Samsung Gauss 2 54B Instruct
Samsung Gauss 2 is Samsung's large language model optimized for on-device and cloud workloads. Trained on multilingual data with a focus on Korean and English, covering general conversation, summarization, and code assistance.
Qwen3 235B A22B
Qwen3 235B A22B is Alibaba's flagship mixture-of-experts model with 235B total parameters and 22B active per token. Delivers frontier-level performance on coding, reasoning, and multilingual tasks at significantly lower inference cost.
Databricks DBRX Instruct
DBRX Instruct is an open, general-purpose LLM from Databricks. Built with a fine-grained mixture-of-experts (MoE) architecture, it was the most capable open LLM at launch and excels at code, math, and language tasks.
Microsoft Phi-4 Mini
Microsoft Phi-4 Mini is a 3.8B parameter compact model from Microsoft. Delivers impressive reasoning capabilities for edge and mobile deployment scenarios, with strong performance on math and coding tasks relative to its size.
IBM Granite 3.0 8B Instruct
IBM Granite 3.0 8B Instruct is a lightweight enterprise-grade language model trained on a carefully curated enterprise corpus and optimized for RAG, summarization, classification, and code generation in business contexts.
OpenAI: GPT-3.5 Turbo
GPT-3.5 Turbo is OpenAI's fastest model. It can understand and generate natural language or code, and is optimized for chat and traditional completion tasks. Training data up to Sep 2021.
OpenAI: GPT-3.5 Turbo (older v0613)
GPT-3.5 Turbo is OpenAI's fastest model. It can understand and generate natural language or code, and is optimized for chat and traditional completion tasks. Training data up to Sep 2021.
Mistral Large
This is Mistral AI's flagship model, Mistral Large 2 (version `mistral-large-2407`). It's a proprietary weights-available model and excels at reasoning, code, JSON, chat, and more. Read the launch announcement [here](https://mistral.ai/news/mistral-large-2407/)....
Mistral: Mixtral 8x22B Instruct
Mistral's official instruct fine-tuned version of [Mixtral 8x22B](/models/mistralai/mixtral-8x22b). It uses 39B active parameters out of 141B, offering unparalleled cost efficiency for its size. Its strengths include: - strong math, coding,...
