๐ All Models
1,316 models ยท Page 8 of 37
nvidia/Llama-3.1-Nemotron-70B-Instruct
NVIDIA's Llama 3.1 Nemotron 70B Instruct โ fine-tuned for helpfulness and aligned with human preferences.
ByteDance/Seedream-4
ByteDance/Seed-2.0-mini
Qwen/Qwen3-Max
Qwen3-Max is an updated release built on the Qwen3 series, offering major improvements in reasoning, instruction following, multilingual support, and long-tail knowledge coverage compared to the January 2025 version. It...
google/gemma-4-26B-A4B-it
Gemma 4 26B A4B IT is an instruction-tuned Mixture-of-Experts (MoE) model from Google DeepMind. Despite 25.2B total parameters, only 3.8B activate per token during inference โ delivering near-31B quality at...
deepseek-ai/DeepSeek-V3-0324
Qwen/Qwen3-30B-A3B
Qwen3, the latest generation in the Qwen large language model series, features both dense and mixture-of-experts (MoE) architectures to excel in reasoning, multilingual support, and advanced agent tasks. Its unique...
meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo
Qwen/Qwen3-Embedding-4B
black-forest-labs/FLUX-2-max
PrunaAI/p-image
Qwen/Qwen3-Embedding-0.6B
PaddlePaddle/PaddleOCR-VL-0.9B
black-forest-labs/FLUX-pro
Qwen/Qwen3-Embedding-0.6B-batch
google/gemma-3-27b-it
Gemma 3 introduces multimodality, supporting vision-language input and text outputs. It handles context windows up to 128k tokens, understands over 140 languages, and offers improved math, reasoning, and chat capabilities,...
sentence-transformers/all-MiniLM-L12-v2
zai-org/GLM-4.7-Flash
mistralai/Mistral-Small-3.2-24B-Instruct-2506
BAAI/bge-m3-multi
thenlper/gte-large
black-forest-labs/FLUX-2-dev
Wan-AI/Wan2.6-Image-Edit
Qwen/Qwen3.5-9B
Qwen/Qwen3-Coder-480B-A35B-Instruct
Qwen/Qwen2.5-VL-32B-Instruct
intfloat/multilingual-e5-large-instruct
deepseek-ai/DeepSeek-OCR
Bria/fibo_edit
Qwen/Qwen3.5-2B
nvidia/NVIDIA-Nemotron-3-Super-120B-A12B
Sao10K/L3-8B-Lunaris-v1-Turbo
zai-org/GLM-4.7
openai/gpt-oss-120b
gpt-oss-120b is an open-weight, 117B-parameter Mixture-of-Experts (MoE) language model from OpenAI designed for high-reasoning, agentic, and general-purpose production use cases. It activates 5.1B parameters per forward pass and is optimized...
