Collections

Use a language model

These large language models understand and generate natural language. Key capabilities:

Language models power chatbots, search engines, and other natural language applications. Continued scale-up aims to improve reasoning skills.

Models we recommend

meta/llama-2-70b-chat

A 70 billion parameter language model from Meta, fine tuned for chat completions

4.5M runs

meta/llama-2-13b-chat

A 13 billion parameter language model from Meta, fine tuned for chat completions

3.8M runs

meta/llama-2-7b-chat

A 7 billion parameter language model from Meta, fine tuned for chat completions

3.4M runs

mistralai/mixtral-8x7b-instruct-v0.1

The Mixtral-8x7B-instruct-v0.1 Large Language Model (LLM) is a pretrained generative Sparse Mixture of Experts tuned to be a helpful assistant.

1.6M runs

mistralai/mistral-7b-instruct-v0.2

The Mistral-7B-Instruct-v0.2 Large Language Model (LLM) is an improved instruct fine-tuned version of Mistral-7B-Instruct-v0.1.

837.2K runs

mistralai/mistral-7b-instruct-v0.1

An instruction-tuned 7 billion parameter language model from Mistral

783.4K runs

replicate/dolly-v2-12b

An open source instruction-tuned large language model developed by Databricks

453.1K runs

replicate/vicuna-13b

A large language model that's been fine-tuned on ChatGPT interactions

235.7K runs

mistralai/mistral-7b-v0.1

A 7 billion parameter language model from Mistral.

199.4K runs

01-ai/yi-34b-chat

The Yi series models are large language models trained from scratch by developers at 01.AI.

186.6K runs

01-ai/yi-6b

The Yi series models are large language models trained from scratch by developers at 01.AI.

157.6K runs

replicate/flan-t5-xl

A language model by Google for tasks like classification, summarization, and more

118.3K runs

stability-ai/stablelm-tuned-alpha-7b

7 billion parameter version of Stability AI's language model

100.5K runs

replicate/llama-7b

Transformers implementation of the LLaMA language model

96.7K runs

replicate/oasst-sft-1-pythia-12b

An open source instruction-tuned large language model developed by Open-Assistant

32.4K runs

nateraw/nous-hermes-2-solar-10.7b

Nous Hermes 2 - SOLAR 10.7B is the flagship Nous Research model on the SOLAR 10.7B base model..

21.2K runs

kcaverly/nous-hermes-2-yi-34b-gguf

Nous Hermes 2 - Yi-34B is a state of the art Yi Fine-tune, fine tuned on GPT-4 generated synthetic data

8.3K runs

nateraw/nous-hermes-llama2-awq

TheBloke/Nous-Hermes-Llama2-AWQ served with vLLM

7.1K runs

replicate/gpt-j-6b

A large language model by EleutherAI

6.1K runs

01-ai/yi-6b-chat

The Yi series models are large language models trained from scratch by developers at 01.AI.

2.5K runs

replit/replit-code-v1-3b

Generate code with Replit's replit-code-v1-3b large language model

1.9K runs

lucataco/phi-2

Phi-2 by Microsoft

1.8K runs

lucataco/qwen1.5-72b

Qwen1.5 is the beta version of Qwen2, a transformer-based decoder-only language model pretrained on a large amount of data

143 runs

lucataco/qwen1.5-14b

Qwen1.5 is the beta version of Qwen2, a transformer-based decoder-only language model pretrained on a large amount of data

139 runs

adirik/mamba-2.8b

Base version of Mamba 2.8B, a 2.8 billion parameter state space language model

115 runs

lucataco/phixtral-2x2_​8

phixtral-2x2_8 is the first Mixure of Experts (MoE) made with two microsoft/phi-2 models, inspired by the mistralai/Mixtral-8x7B-v0.1 architecture

70 runs

lucataco/qwen1.5-7b

Qwen1.5 is the beta version of Qwen2, a transformer-based decoder-only language model pretrained on a large amount of data

33 runs

lucataco/olmo-7b

OLMo is a series of Open Language Models designed to enable the science of language models

32 runs

adirik/mamba-130m

Base version of Mamba 130M, a 130 million parameter state space language model

29 runs

adirik/mamba-2.8b-slimpj

Base version of Mamba 2.8B Slim Pyjama, a 2.8 billion parameter state space language model

26 runs

adirik/mamba-1.4b

Base version of Mamba 1.4B, a 1.4 billion parameter state space language model

15 runs

adirik/mamba-790m

Base version of Mamba 790M, a 790 million parameter state space language model

11 runs

adirik/mamba-370m

Base version of Mamba 370M, a 370 million parameter state space language model

9 runs