Use a language model
These large language models understand and generate natural language. They power chatbots, search engines, writing aids, and more.
Use these for:
- Conversational AI: Chat and engage in natural dialogue. Get an AI assistant.
- Question answering: Provide informative answers to questions. Build a knowledge base.
- Text generation: Generate fluent continuations of text. Autocomplete your writing.
- Summarization: Summarize long passages of text. Get key points quickly.
- Translation: Translate between languages. Communicate across language barriers.
Language models keep getting bigger and better at these tasks. The largest models today exhibit impressive reasoning skills. But you can get great results from smaller, faster, cheaper models too.
Our Pick: Meta Llama 3 8B Instruct
Meta’s new Llama 3 8B Instruct is the clear choice for most applications. With 8B parameters, an 8K context window, and advanced instruction tuning on 15T+ tokens, it achieves state-of-the-art performance on a wide range of tasks. A fast, affordable and flexible language model.
Upgrade Pick: Meta Llama 3 70B Instruct
For the most demanding applications, Llama 3 70B Instruct is the top performer. Its massive 70B parameters and training on 15T+ tokens deliver unparalleled accuracy and nuance across complex language tasks.
The 70B model shares the same efficiency benefits and safety features as the 8B version. But with greater capacity, it excels at applications like content creation, conversational AI, and code generation.
Budget Pick: Flan-T5 XL
For latency-sensitive, cost-constrained applications, Flan-T5 XL remains a strong choice. While it can’t match Llama 3’s overall performance, its lean 3B parameter size makes it fast and economical for focused tasks.
If speed and cost are critical and your use case is well-defined, like classification or summarization, Flan-T5 XL delivers reliable results quickly and affordably.
Recommended models
![](https://tjzk.replicate.delivery/models_models_featured_image/3dcb020b-1fad-4101-84cf-88af9b20ac21/meta-logo.png)
meta/meta-llama-3-70b-instruct
A 70 billion parameter language model from Meta, fine tuned for chat completions
![](https://tjzk.replicate.delivery/models_models_featured_image/68b7dc1a-4767-4353-b066-212b0126b5de/meta-logo.png)
meta/meta-llama-3-8b-instruct
An 8 billion parameter language model from Meta, fine tuned for chat completions
![](https://tjzk.replicate.delivery/models_models_cover_image/dd9ac11a-edda-4d33-b413-6a721c44dfb0/meta-logo.png)
meta/meta-llama-3-8b
Base version of Llama 3, an 8 billion parameter language model from Meta.
![](https://tjzk.replicate.delivery/models_models_featured_image/b9867b79-28e4-4e8f-b2b3-0f669111869c/mixtral.png)
mistralai/mixtral-8x7b-instruct-v0.1
The Mixtral-8x7B-instruct-v0.1 Large Language Model (LLM) is a pretrained generative Sparse Mixture of Experts tuned to be a helpful assistant.
![](https://tjzk.replicate.delivery/models_models_featured_image/e18de70b-6147-4d42-a730-4777d297737d/llama-logo.png)
meta/llama-2-7b-chat
A 7 billion parameter language model from Meta, fine tuned for chat completions
![](https://tjzk.replicate.delivery/models_models_featured_image/70238cb4-e24b-43ed-96e0-5f7d98f9cc3f/future-llama-70b-chat.jpg)
meta/llama-2-70b-chat
A 70 billion parameter language model from Meta, fine tuned for chat completions
![](https://tjzk.replicate.delivery/models_models_featured_image/e3fa9c03-7993-49d1-b6f3-351ead5e5dfc/high_tech_llama.png)
meta/llama-2-13b-chat
A 13 billion parameter language model from Meta, fine tuned for chat completions
![](https://tjzk.replicate.delivery/models_models_cover_image/7553cbf4-563d-4140-a51d-7386b8d13941/mistral-7b-instruct-cover.png)
mistralai/mistral-7b-instruct-v0.2
The Mistral-7B-Instruct-v0.2 Large Language Model (LLM) is an improved instruct fine-tuned version of Mistral-7B-Instruct-v0.1.
![](https://tjzk.replicate.delivery/models_models_cover_image/c56ddfb4-2193-4a65-862b-c274da78f07d/mistral7b_image2.jpeg)
mistralai/mistral-7b-v0.1
A 7 billion parameter language model from Mistral.
![](https://tjzk.replicate.delivery/models_models_featured_image/ad668055-b12b-4f67-92a7-62e9c2f87df7/mistral7b_image1.jpeg)
mistralai/mistral-7b-instruct-v0.1
An instruction-tuned 7 billion parameter language model from Mistral
![](https://tjzk.replicate.delivery/models_models_featured_image/bb398430-da26-49ef-aefc-283212538c39/dolly_sheep2.jpg)
replicate/dolly-v2-12b
An open source instruction-tuned large language model developed by Databricks
![](https://tjzk.replicate.delivery/models_models_cover_image/8e044b4c-0b20-4717-83bd-a94d89fb0dbe/meta-logo.png)
meta/meta-llama-3-70b
Base version of Llama 3, a 70 billion parameter language model from Meta.
![](https://tjzk.replicate.delivery/models_models_featured_image/b2b74ffa-1b41-4b80-a210-11a996f0a4b0/yi_thumbnail_1.png)
01-ai/yi-34b-chat
The Yi series models are large language models trained from scratch by developers at 01.AI.
![](https://tjzk.replicate.delivery/models_models_featured_image/e90f0bd4-10f1-4317-9e42-fbf76224e950/robot_llama.png)
replicate/vicuna-13b
A large language model that's been fine-tuned on ChatGPT interactions
![](https://tjzk.replicate.delivery/models_models_cover_image/d08ef0b6-2b80-491a-984a-d4287d6325c0/yi_thumbnail.png)
01-ai/yi-6b
The Yi series models are large language models trained from scratch by developers at 01.AI.
![](https://tjzk.replicate.delivery/models_models_featured_image/8186a063-0dc9-41eb-a941-3dc6dd9a14fc/flan.png)
replicate/flan-t5-xl
A language model by Google for tasks like classification, summarization, and more
![](https://tjzk.replicate.delivery/models_models_featured_image/12546857-d49a-4fd9-873f-14dd9800ddf8/mascot.png)
stability-ai/stablelm-tuned-alpha-7b
7 billion parameter version of Stability AI's language model
![](https://tjzk.replicate.delivery/models_models_featured_image/a2541b38-6df1-4fa5-9485-9b3068502515/space_llama.png)
replicate/llama-7b
Transformers implementation of the LLaMA language model
![](https://tjzk.replicate.delivery/models_models_cover_image/1b5053da-932d-4e62-b957-f9f33d166902/gemma.jpg)
google-deepmind/gemma-2b-it
2B instruct version of Google’s Gemma model
![](https://tjzk.replicate.delivery/models_models_featured_image/d2f84619-a92c-41ab-a88d-10faf493a0f4/gemma.jpg)
google-deepmind/gemma-7b-it
7B instruct version of Google’s Gemma model
![](https://tjzk.replicate.delivery/models_models_cover_image/ee2669a2-ebc5-4f3b-8365-9a3790050adb/nous-hermes-2.png)
nateraw/nous-hermes-2-solar-10.7b
Nous Hermes 2 - SOLAR 10.7B is the flagship Nous Research model on the SOLAR 10.7B base model..
![](https://tjzk.replicate.delivery/models_models_cover_image/67eed8a1-ba88-4251-89b3-a8b05bc225f1/nice-robot.png)
replicate/oasst-sft-1-pythia-12b
An open source instruction-tuned large language model developed by Open-Assistant
![](https://tjzk.replicate.delivery/models_models_cover_image/44b97533-9749-44c6-bea9-2e50b5527174/eleuther-ai-logo.png)
replicate/gpt-j-6b
A large language model by EleutherAI
![](https://tjzk.replicate.delivery/models_models_cover_image/1f2b3f7a-e583-43db-9f9c-a363c167e5df/hermes.png)
kcaverly/nous-hermes-2-yi-34b-gguf
Nous Hermes 2 - Yi-34B is a state of the art Yi Fine-tune, fine tuned on GPT-4 generated synthetic data
nateraw/nous-hermes-llama2-awq
TheBloke/Nous-Hermes-Llama2-AWQ served with vLLM
![](https://tjzk.replicate.delivery/models_models_cover_image/f1e31126-4ba2-441e-a62f-80f492b5d403/gemma.jpg)
google-deepmind/gemma-7b
7B base version of Google’s Gemma model
![](https://tjzk.replicate.delivery/models_models_cover_image/30bfbd81-5028-4553-be44-dc063ccae993/yi_thumbnail.png)
01-ai/yi-6b-chat
The Yi series models are large language models trained from scratch by developers at 01.AI.
![](https://tjzk.replicate.delivery/models_models_cover_image/1b725996-846c-47cc-af38-90c7fc6568f1/qwen.webp)
lucataco/qwen1.5-72b
Qwen1.5 is the beta version of Qwen2, a transformer-based decoder-only language model pretrained on a large amount of data
![](https://tjzk.replicate.delivery/models_models_cover_image/d074c7f8-0f6e-495a-8b7b-b3cbe0f3fb1b/microsoft.webp)
lucataco/phi-2
Phi-2 by Microsoft
![](https://tjzk.replicate.delivery/models_models_cover_image/a56da6ac-1bac-4722-8cc2-69c605e90a6a/replit.png)
replit/replit-code-v1-3b
Generate code with Replit's replit-code-v1-3b large language model
![](https://tjzk.replicate.delivery/models_models_cover_image/2ecc3f85-ba02-433c-b0e5-c72ab1d88fbf/gemma.jpg)
google-deepmind/gemma-2b
2B base version of Google’s Gemma model
![](https://tjzk.replicate.delivery/models_models_cover_image/ed04609c-5f64-4b27-9dd2-dccad8da7256/qwen.webp)
lucataco/qwen1.5-14b
Qwen1.5 is the beta version of Qwen2, a transformer-based decoder-only language model pretrained on a large amount of data
![](https://tjzk.replicate.delivery/models_models_cover_image/c46544ab-db95-429b-a131-2b5c811d3184/out-1_3.png)
adirik/mamba-2.8b
Base version of Mamba 2.8B, a 2.8 billion parameter state space language model
![](https://tjzk.replicate.delivery/models_models_cover_image/3c2d374c-5674-44fd-88fc-192e370fbe78/UOb2fvh.jpg)
lucataco/phixtral-2x2_8
phixtral-2x2_8 is the first Mixure of Experts (MoE) made with two microsoft/phi-2 models, inspired by the mistralai/Mixtral-8x7B-v0.1 architecture
![](https://tjzk.replicate.delivery/models_models_cover_image/e6699590-1255-4fbd-ae01-309dff015930/qwen.webp)
lucataco/qwen1.5-7b
Qwen1.5 is the beta version of Qwen2, a transformer-based decoder-only language model pretrained on a large amount of data
![](https://tjzk.replicate.delivery/models_models_cover_image/8d336548-0146-460e-a250-c2e3ae2980bb/out-1_3.png)
adirik/mamba-130m
Base version of Mamba 130M, a 130 million parameter state space language model
![](https://tjzk.replicate.delivery/models_models_cover_image/b2df520a-fda9-43b9-a792-55f4ccb3a79e/ai2.webp)
lucataco/olmo-7b
OLMo is a series of Open Language Models designed to enable the science of language models
![](https://tjzk.replicate.delivery/models_models_cover_image/82ea9cfb-95bf-48b6-8b06-9e5a9a2e27e1/out-1_3.png)
adirik/mamba-1.4b
Base version of Mamba 1.4B, a 1.4 billion parameter state space language model
![](https://tjzk.replicate.delivery/models_models_cover_image/4973f105-c7ab-435b-8c7c-0dde92cc6349/out-1_3.png)
adirik/mamba-2.8b-slimpj
Base version of Mamba 2.8B Slim Pyjama, a 2.8 billion parameter state space language model
![](https://tjzk.replicate.delivery/models_models_cover_image/bb41e380-ac7a-41df-96a9-6030cfb46eb7/out-1_3.png)
adirik/mamba-370m
Base version of Mamba 370M, a 370 million parameter state space language model
![](https://tjzk.replicate.delivery/models_models_cover_image/bbec284b-b3cb-43dc-8bc7-0908b07288a5/out-1_3.png)
adirik/mamba-790m
Base version of Mamba 790M, a 790 million parameter state space language model