๐ฌ Text Generation
850 models ยท Page 24 of 24
Qwen3-4B
Open-source Qwen3-4B model from qwen โ available for download and self-hosting on Hugging Face.
Qwen3-8B
Open-source Qwen3-8B model from qwen โ available for download and self-hosting on Hugging Face.
Qwen/Qwen2.5-3B-Instruct
Qwen/Qwen2.5-3B-Instruct is a text generation model on Hugging Face with ~8,750,507 monthly downloads. Open access.
meta-llama/Llama-3.1-8B-Instruct
meta-llama/Llama-3.1-8B-Instruct is a text generation model on Hugging Face with ~9,266,275 monthly downloads. (Gated access โ requires HuggingFace login.)
Qwen/Qwen2.5-1.5B-Instruct
Qwen/Qwen2.5-1.5B-Instruct is a text generation model on Hugging Face with ~9,915,572 monthly downloads. Open access.
Qwen/Qwen2.5-7B-Instruct
Qwen/Qwen2.5-7B-Instruct is a text generation model on Hugging Face with ~12,936,213 monthly downloads. Open access.
gpt2
Open-source gpt2 model from openai-community โ available for download and self-hosting on Hugging Face.
Qwen/Qwen3-0.6B
Qwen/Qwen3-0.6B is a text generation model on Hugging Face with ~15,133,638 monthly downloads. Open access.
Samsung Gauss 2 54B Instruct
Samsung Gauss 2 is Samsung's large language model optimized for on-device and cloud workloads. Trained on multilingual data with a focus on Korean and English, covering general conversation, summarization, and code assistance.
Qwen3 235B A22B
Qwen3 235B A22B is Alibaba's flagship mixture-of-experts model with 235B total parameters and 22B active per token. Delivers frontier-level performance on coding, reasoning, and multilingual tasks at significantly lower inference cost.
Falcon 180B
Falcon 180B is one of the largest openly available language models, trained on 3.5 trillion tokens with TII's custom RefinedWeb dataset. Excels at reasoning, summarization, and generation tasks at state-of-the-art quality for open models.
Databricks DBRX Instruct
DBRX Instruct is an open, general-purpose LLM from Databricks. Built with a fine-grained mixture-of-experts (MoE) architecture, it was the most capable open LLM at launch and excels at code, math, and language tasks.
NVIDIA Nemotron-4 340B Instruct
NVIDIA Nemotron-4 340B Instruct is a large open language model trained to generate diverse synthetic data for training other LLMs. Strong at following instructions, classification, and generating reward model training data.
AI21 Jamba 1.6 Mini
AI21 Jamba 1.6 Mini is a lightweight Mamba-Transformer hybrid optimized for cost-effective, high-throughput inference with an impressive 256K context window. An excellent choice for document-heavy workloads on a budget.
AI21 Jamba 1.6 Large
AI21 Jamba 1.6 Large uses a hybrid Mamba-Transformer architecture offering low memory footprint and high throughput compared to equivalent Transformer models. Features 256K context at a fraction of the inference cost.
Microsoft Phi-4 Mini
Microsoft Phi-4 Mini is a 3.8B parameter compact model from Microsoft. Delivers impressive reasoning capabilities for edge and mobile deployment scenarios, with strong performance on math and coding tasks relative to its size.
IBM Granite 3.0 2B Instruct
IBM Granite 3.0 2B Instruct is an ultra-compact enterprise model excelling at summarization, extraction, and classification. The smallest model in the Granite family, suitable for edge deployments and constrained environments.
IBM Granite 3.0 8B Instruct
IBM Granite 3.0 8B Instruct is a lightweight enterprise-grade language model trained on a carefully curated enterprise corpus and optimized for RAG, summarization, classification, and code generation in business contexts.
Amazon Titan Text Express
Amazon Titan Text Express is a generative LLM for summarization, text generation, classification, open-ended Q&A, and information extraction. Optimized for enterprise workloads via AWS Bedrock.
Amazon Nova Pro
Amazon Nova Pro is a highly capable multimodal model with the best combination of accuracy, speed, and cost across a wide range of tasks. Supports text, image, and video inputs.
Amazon Nova Lite
Amazon Nova Lite is a very low-cost multimodal model that can process image, video, and text inputs. Fast and accurate for a wide range of tasks requiring visual and language understanding.
Amazon Nova Micro
Amazon Nova Micro is the fastest and most cost-effective text-only model in the Nova family, optimized for speed and low latency. Ideal for customer service, summarization, and translation at scale.
