π All Models
33 models Β· Page 1 of 1
Qwen3 Next 80B A3b Instruct
Qwen3-Next-80B-A3B-Instruct is an instruction-tuned chat model in the Qwen3-Next series optimized for fast, stable responses without βthinkingβ traces. It targets complex tasks across reasoning, code generation, knowledge QA, and multilingual...
Qwen3 30B A3b
Qwen3, the latest generation in the Qwen large language model series, features both dense and mixture-of-experts (MoE) architectures to excel in reasoning, multilingual support, and advanced agent tasks. Its unique...
Multilingual E5 Large Instruct
Deepgram Nova 3 Multilingual
sentence-transformers/clip-ViT-B-32-multilingual-v1
Qwen/Qwen3-30B-A3B
Qwen3, the latest generation in the Qwen large language model series, features both dense and mixture-of-experts (MoE) architectures to excel in reasoning, multilingual support, and advanced agent tasks. Its unique...
Qwen/Qwen3-Max
Qwen3-Max is an updated release built on the Qwen3 series, offering major improvements in reasoning, instruction following, multilingual support, and long-tail knowledge coverage compared to the January 2025 version. It...
intfloat/multilingual-e5-large
mistralai/Mixtral-8x7B-Instruct-v0.1
Mixtral 8Γ7B Instruct on DeepInfra β popular MoE model with 32K context and strong multilingual performance.
Qwen/Qwen2.5-72B-Instruct
Alibaba Qwen2.5 72B Instruct on DeepInfra β top open-source model with multilingual and coding strengths.
embed-multilingual-light-v3.0
embed-multilingual-light-v3.0-image
embed-multilingual-v3.0-image
embed-multilingual-v3.0
Cohere's multilingual embeddings supporting 100+ languages.
intfloat/multilingual-e5-large-instruct
open-mistral-nemo
Our best multilingual open source model released July 2024.
open-mistral-nemo
Our best multilingual open source model released July 2024.
open-mistral-nemo
Our best multilingual open source model released July 2024.
open-mistral-nemo
Our best multilingual open source model released July 2024.
speech-2.8-hd
Minimax Speech 2.8 HD focuses on high-fidelity audio generation with features like studio-grade quality, flexible emotion control, multilingual support, and voice cloning capabilities
Samsung Gauss 2 54B Instruct
Samsung Gauss 2 is Samsung's large language model optimized for on-device and cloud workloads. Trained on multilingual data with a focus on Korean and English, covering general conversation, summarization, and code assistance.
Qwen3 235B A22B
Qwen3 235B A22B is Alibaba's flagship mixture-of-experts model with 235B total parameters and 22B active per token. Delivers frontier-level performance on coding, reasoning, and multilingual tasks at significantly lower inference cost.
Mistral: Mistral Nemo
A 12B parameter model with a 128k token context length built by Mistral in collaboration with NVIDIA. The model is multilingual, supporting English, French, German, Spanish, Italian, Portuguese, Chinese, Japanese,...
Cohere: Command R (08-2024)
command-r-08-2024 is an update of the [Command R](/models/cohere/command-r) with improved performance for multilingual retrieval-augmented generation (RAG) and tool use. More broadly, it is better at math, code and reasoning and...
Meta: Llama 3.2 1B Instruct
Llama 3.2 1B is a 1-billion-parameter language model focused on efficiently performing natural language tasks, such as summarization, dialogue, and multilingual text analysis. Its smaller size allows it to operate...
Meta: Llama 3.2 3B Instruct (free)
Llama 3.2 3B is a 3-billion-parameter multilingual large language model, optimized for advanced natural language processing tasks like dialogue generation, reasoning, and summarization. Designed with the latest transformer architecture, it...
Meta: Llama 3.3 70B Instruct (free)
The Meta Llama 3.3 multilingual large language model (LLM) is a pretrained and instruction tuned generative model in 70B (text in/text out). The Llama 3.3 instruction tuned text only model...
Cohere: Command A
Command A is an open-weights 111B parameter model with a 256k context window focused on delivering great performance across agentic, multilingual, and coding use cases. Compared to other leading proprietary...
Qwen: Qwen3 30B A3B
Qwen3, the latest generation in the Qwen large language model series, features both dense and mixture-of-experts (MoE) architectures to excel in reasoning, multilingual support, and advanced agent tasks. Its unique...
Qwen: Qwen3 235B A22B Instruct 2507
Qwen3-235B-A22B-Instruct-2507 is a multilingual, instruction-tuned mixture-of-experts language model based on the Qwen3-235B architecture, with 22B active parameters per forward pass. It is optimized for general-purpose text generation, including instruction following,...
Qwen: Qwen3 30B A3B Instruct 2507
Qwen3-30B-A3B-Instruct-2507 is a 30.5B-parameter mixture-of-experts language model from Qwen, with 3.3B active parameters per inference. It operates in non-thinking mode and is designed for high-quality instruction following, multilingual understanding, and...
Qwen: Qwen3 Next 80B A3B Instruct (free)
Qwen3-Next-80B-A3B-Instruct is an instruction-tuned chat model in the Qwen3-Next series optimized for fast, stable responses without βthinkingβ traces. It targets complex tasks across reasoning, code generation, knowledge QA, and multilingual...
Qwen: Qwen3 Max
Qwen3-Max is an updated release built on the Qwen3 series, offering major improvements in reasoning, instruction following, multilingual support, and long-tail knowledge coverage compared to the January 2025 version. It...
