Collections

Streaming language models

Language models that support streaming responses. See https://replicate.com/docs/streaming

Recommended models

bytedance / sdxl-lightning-4step

SDXL-Lightning by ByteDance: a fast text-to-image model that makes high-quality images in 4 steps

466.9M runs

meta / meta-llama-3-70b-instruct

A 70 billion parameter language model from Meta, fine tuned for chat completions

120.9M runs

meta / meta-llama-3-8b-instruct

An 8 billion parameter language model from Meta, fine tuned for chat completions

110.6M runs

stability-ai / stable-diffusion

A latent text-to-image diffusion model capable of generating photo-realistic images given any text input

109.2M runs

black-forest-labs / flux-schnell

The fastest image generation model tailored for local development and personal use

66.8M runs

stability-ai / sdxl

A text-to-image generative AI model that creates beautiful images

66.8M runs

meta / meta-llama-3-8b

Base version of Llama 3, an 8 billion parameter language model from Meta.

50.6M runs

stability-ai / stable-diffusion-inpainting

Fill in masked parts of images with Stable Diffusion

18.4M runs

yorickvp / llava-13b

Visual instruction tuning towards large language and vision models with GPT-4 level capabilities

17.5M runs

mistralai / mixtral-8x7b-instruct-v0.1

The Mixtral-8x7B-instruct-v0.1 Large Language Model (LLM) is a pretrained generative Sparse Mixture of Experts tuned to be a helpful assistant.

12.8M runs

meta / llama-2-7b-chat

A 7 billion parameter language model from Meta, fine tuned for chat completions

12.4M runs

ai-forever / kandinsky-2.2

multilingual text2image latent diffusion model

10M runs

meta / llama-2-70b-chat

A 70 billion parameter language model from Meta, fine tuned for chat completions

8.5M runs

datacte / proteus-v0.2

Proteus v0.2 shows subtle yet significant improvements over Version 0.1. It demonstrates enhanced prompt understanding that surpasses MJ6, while also approaching its stylistic capabilities.

7.6M runs

ai-forever / kandinsky-2

text2img model trained on LAION HighRes and fine-tuned on internal datasets

6.2M runs

fofr / sdxl-emoji

An SDXL fine-tune based on Apple Emojis

6.1M runs

yorickvp / llava-v1.6-mistral-7b

LLaVA v1.6: Large Language and Vision Assistant (Mistral-7B)

4.8M runs

meta / llama-2-13b-chat

A 13 billion parameter language model from Meta, fine tuned for chat completions

4.7M runs

mistralai / mistral-7b-instruct-v0.2

The Mistral-7B-Instruct-v0.2 Large Language Model (LLM) is an improved instruct fine-tuned version of Mistral-7B-Instruct-v0.1.

3.5M runs

yorickvp / llava-v1.6-vicuna-13b

LLaVA v1.6: Large Language and Vision Assistant (Vicuna-13B)

3.3M runs

meta / meta-llama-3.1-405b-instruct

Meta's flagship 405 billion parameter language model, fine-tuned for chat completions

3.1M runs

tstramer / material-diffusion

Stable diffusion fork for generating tileable outputs using v1.5 model

2.2M runs

fofr / prompt-classifier

Determines the toxicity of text to image prompts, llama-13b fine-tune. [SAFETY_RANKING] between 0 (safe) and 10 (toxic)

1.8M runs

mistralai / mistral-7b-v0.1

A 7 billion parameter language model from Mistral.

1.8M runs

datacte / proteus-v0.3

ProteusV0.3: The Anime Update

1.7M runs

playgroundai / playground-v2.5-1024px-aesthetic

Playground v2.5 is the state-of-the-art open-source model in aesthetic quality

1.7M runs

yorickvp / llava-v1.6-34b

LLaVA v1.6: Large Language and Vision Assistant (Nous-Hermes-2-34B)

1.6M runs

stability-ai / stable-diffusion-3

A text-to-image model with greatly improved performance in image quality, typography, complex prompt understanding, and resource-efficiency

1.3M runs

snowflake / snowflake-arctic-instruct

An efficient, intelligent, and truly open-source language model

1.3M runs

fofr / latent-consistency-model

Super-fast, 0.6s per image. LCM with img2img, large batching and canny controlnet

1.1M runs

lucataco / ssd-1b

Segmind Stable Diffusion Model (SSD-1B) is a distilled 50% smaller version of SDXL, offering a 60% speedup while maintaining high-quality text-to-image generation capabilities

976.4K runs

fofr / realvisxl-v3-multi-controlnet-lora

RealVisXl V3 with multi-controlnet, lora loading, img2img, inpainting

786.3K runs

meta / meta-llama-3-70b

Base version of Llama 3, a 70 billion parameter language model from Meta.

699.4K runs

batouresearch / sdxl-controlnet-lora

'''Last update: Now supports img2img.''' SDXL Canny controlnet with LoRA support.

665.3K runs

meta / llama-2-7b

Base version of Llama 2 7B, a 7 billion parameter language model

644.1K runs

fofr / any-comfyui-workflow

Run any ComfyUI workflow. Guide: https://github.com/fofr/cog-comfyui

612K runs

fofr / sticker-maker

Make stickers with AI. Generates graphics with transparent backgrounds.

521.8K runs

spuuntries / flatdolphinmaid-8x7b-gguf

Undi95's FlatDolphinMaid 8x7B Mixtral Merge, GGUF Q5_K_M quantized by TheBloke.

412.8K runs

meta / llama-2-70b

Base version of Llama 2, a 70 billion parameter language model from Meta.

341.1K runs

lucataco / realvisxl2-lcm

RealvisXL-v2.0 with LCM LoRA - requires fewer steps (4 to 8 instead of the original 40 to 50)

289.3K runs

lucataco / realvisxl-v2.0

Implementation of SDXL RealVisXL_V2.0

278.9K runs

01-ai / yi-34b-chat

The Yi series models are large language models trained from scratch by developers at 01.AI.

270.3K runs

nateraw / goliath-120b

An auto-regressive causal LM created by combining 2x finetuned Llama-2 70B into one.

236.9K runs

fofr / sdxl-multi-controlnet-lora

Multi-controlnet, lora loading, img2img, inpainting

202.1K runs

antoinelyset / openhermes-2-mistral-7b-awq

199.2K runs

lucataco / moondream2

moondream2 is a small vision language model designed to run efficiently on edge devices

199.1K runs

meta / llama-2-13b

Base version of Llama 2 13B, a 13 billion parameter language model

193.5K runs

lucataco / dreamshaper-xl-turbo

DreamShaper is a general purpose SD model that aims at doing everything well, photos, art, anime, manga. It's designed to match Midjourney and DALL-E.

180.8K runs

01-ai / yi-6b

The Yi series models are large language models trained from scratch by developers at 01.AI.

159K runs

replicate / flan-t5-xl

A language model by Google for tasks like classification, summarization, and more

144.1K runs

meta / codellama-34b-instruct

A 34 billion parameter Llama tuned for coding and conversation

120.1K runs

stability-ai / stablelm-tuned-alpha-7b

7 billion parameter version of Stability AI's language model

112.5K runs

meta / codellama-13b

A 13 billion parameter Llama tuned for code completion

112.4K runs

lucataco / open-dalle-v1.1

A unique fusion that showcases exceptional prompt adherence and semantic understanding, it seems to be a step above base SDXL and a step closer to DALLE-3 in terms of prompt comprehension

108.8K runs

nateraw / openchat_​3.5-awq

OpenChat: Advancing Open-source Language Models with Mixed-Quality Data

102.6K runs

replicate / llama-7b

Transformers implementation of the LLaMA language model

98.9K runs

adirik / realvisxl-v3.0-turbo

Photorealism with RealVisXL V3.0 Turbo based on SDXL

97K runs

google-deepmind / gemma-2b-it

2B instruct version of Google’s Gemma model

86.8K runs

ai-forever / kandinsky-2-1

Kandinsky 2.1 Diffusion Model

83.9K runs

microsoft / phi-3-mini-4k-instruct

Phi-3-Mini-4K-Instruct is a 3.8B parameters, lightweight, state-of-the-art open model trained with the Phi-3 datasets

81.8K runs

antoinelyset / openhermes-2.5-mistral-7b-awq

68.3K runs

google-deepmind / gemma-7b-it

7B instruct version of Google’s Gemma model

66.1K runs

nateraw / mistral-7b-openorca

Mistral-7B-v0.1 fine tuned for chat with the OpenOrca dataset.

65.8K runs

nightmareai / disco-diffusion

Generate images using a variety of techniques - Powered by Discoart

64.3K runs

nateraw / nous-hermes-2-solar-10.7b

Nous Hermes 2 - SOLAR 10.7B is the flagship Nous Research model on the SOLAR 10.7B base model..

64K runs

yorickvp / llava-v1.6-vicuna-7b

LLaVA v1.6: Large Language and Vision Assistant (Vicuna-7B)

61.6K runs

fofr / image-prompts

Generate image prompts for Midjourney. Prefix inputs with "Image: "

52.9K runs

cuuupid / glm-4v-9b

GLM-4V is a multimodal model released by Tsinghua University that is competitive with GPT-4o and establishes a new SOTA on several benchmarks, including OCR.

52K runs

lucataco / pixart-xl-2

PixArt-Alpha 1024px is a transformer-based text-to-image diffusion system trained on text embeddings from T5

50.4K runs

meta / codellama-13b-instruct

A 13 billion parameter Llama tuned for coding and conversation

46.6K runs

meta / codellama-7b-instruct

A 7 billion parameter Llama tuned for coding and conversation

42.7K runs

adirik / realvisxl-v4.0

Photorealism with RealVisXL V4.0

42.1K runs

joehoover / falcon-40b-instruct

A 40 billion parameter language model trained to follow human instructions.

41.1K runs

lucataco / dolphin-2.2.1-mistral-7b

Mistral-7B-v0.1 fine tuned for chat with the Dolphin dataset (an open-source implementation of Microsoft's Orca)

33.3K runs

fofr / expression-editor

Quickly edit the expression of a face

33.1K runs

google-deepmind / gemma-2-2b

Gemma2 2b by Google

33.1K runs

kcaverly / openchat-3.5-1210-gguf

The "Overall Best Performing Open Source 7B Model" for Coding + Generalization or Mathematical Reasoning

26.2K runs

lucataco / realistic-vision-v5

Realistic Vision v5.0 with VAE

23.3K runs

meta / codellama-70b-instruct

A 70 billion parameter Llama tuned for coding and conversation

21.7K runs

nateraw / defog-sqlcoder-7b-2

A capable large language model for natural language to SQL generation.

21.3K runs

nousresearch / hermes-2-theta-llama-8b

Hermes-2 Θ (Theta) is the first experimental merged model released by Nous Research, in collaboration with Charles Goddard at Arcee, the team behind MergeKit.

20.5K runs

uwulewd / airoboros-llama-2-70b

Inference Airoboros L2 70B 2.1 - GPTQ using ExLlama.

17.6K runs

lucataco / wizardcoder-33b-v1.1-gguf

WizardCoder: Empowering Code Large Language Models with Evol-Instruct

16.6K runs

meta / codellama-7b

A 7 billion parameter Llama tuned for coding and conversation

15.3K runs

nomagick / chatglm3-6b

A 6B parameter open bilingual chat LLM | 开源双语对话语言模型

15.3K runs

mikeei / dolphin-2.9-llama3-70b-gguf

Dolphin is uncensored. I have filtered the dataset to remove alignment and bias. This makes the model more compliant.

15.2K runs

lucataco / wizard-vicuna-13b-uncensored

This is wizard-vicuna-13b trained with a subset of the dataset - responses that contained alignment / moralizing were removed

15K runs

spuuntries / miqumaid-v1-70b-gguf

NeverSleep's MiquMaid v1 70B Miqu Finetune, GGUF Q3_K_M quantized by NeverSleep.

14.1K runs

gregwdata / defog-sqlcoder-q8

Defog's SQLCoder is a state-of-the-art LLM for converting natural language questions to SQL queries. SQLCoder is a 15B parameter fine-tuned on a base StarCoder model.

12.4K runs

lucataco / dolphin-2.1-mistral-7b

Mistral-7B-v0.1 fine tuned for chat with the Dolphin dataset (an open-source implementation of Microsoft's Orca)

12.4K runs

kcaverly / neuralbeagle14-7b-gguf

NeuralBeagle14-7B is (probably) the best 7B model you can find!

12.2K runs

antoinelyset / openhermes-2.5-mistral-7b

12K runs

nomagick / chatglm2-6b

ChatGLM2-6B: An Open Bilingual Chat LLM | 开源双语对话语言模型

10.6K runs

hayooucom / vision-model

This is phi-3-vision model , cost by time ,have fun~

10.6K runs

lucataco / moondream1

(Research only) Moondream1 is a vision language model that performs on par with models twice its size

10.4K runs

meta / codellama-34b

A 34 billion parameter Llama tuned for coding and conversation

10K runs

kcaverly / nous-hermes-2-yi-34b-gguf

Nous Hermes 2 - Yi-34B is a state of the art Yi Fine-tune, fine tuned on GPT-4 generated synthetic data

9.6K runs

replicate / gpt-j-6b

A large language model by EleutherAI

9.2K runs

microsoft / phi-3-mini-128k-instruct

Phi-3-Mini-128K-Instruct is a 3.8 billion-parameter, lightweight, state-of-the-art open model trained using the Phi-3 datasets

8.1K runs

nateraw / nous-hermes-llama2-awq

TheBloke/Nous-Hermes-Llama2-AWQ served with vLLM

7.2K runs

google-deepmind / gemma-7b

7B base version of Google’s Gemma model

7.2K runs

joehoover / zephyr-7b-alpha

A high-performing language model trained to act as a helpful assistant

7K runs

hikikomori-haven / solar-uncensored

6.7K runs

nateraw / zephyr-7b-beta

Zephyr-7B-beta, an LLM trained to act as a helpful assistant.

5.7K runs

01-ai / yi-6b-chat

The Yi series models are large language models trained from scratch by developers at 01.AI.

5.1K runs

nomagick / qwen-14b-chat

Qwen-14B-Chat is a Transformer-based large language model, which is pretrained on a large volume of data, including web texts, books, codes, etc.

4.8K runs

mikeei / dolphin-2.9-llama3-8b-gguf

Dolphin is uncensored. I have filtered the dataset to remove alignment and bias. This makes the model more compliant.

4.5K runs

kcaverly / nous-hermes-2-solar-10.7b-gguf

Nous Hermes 2 - SOLAR 10.7B is the flagship Nous Research model on the SOLAR 10.7B base model.

4.5K runs

lucataco / qwen1.5-72b

Qwen1.5 is the beta version of Qwen2, a transformer-based decoder-only language model pretrained on a large amount of data

3.8K runs

meta / codellama-13b-python

A 13 billion parameter Llama tuned for coding with Python

3.8K runs

meta / codellama-7b-python

A 7 billion parameter Llama tuned for coding with Python

3.8K runs

joehoover / sql-generator

3.6K runs

01-ai / yi-34b

The Yi series models are large language models trained from scratch by developers at 01.AI.

3K runs

kcaverly / deepseek-coder-33b-instruct-gguf

A quantized 33B parameter language model from Deepseek for SOTA repository level code completion

2.8K runs

lucataco / qwen1.5-110b

Qwen1.5 is the beta version of Qwen2, a transformer-based decoder-only language model pretrained on a large amount of data

2.5K runs

organisciak / ocsai-llama2-7b

2.3K runs

deepseek-ai / deepseek-math-7b-base

Pushing the Limits of Mathematical Reasoning in Open Language Models - Base model

2K runs

deepseek-ai / deepseek-vl-7b-base

DeepSeek-VL: An open-source Vision-Language Model designed for real-world vision and language understanding applications

1.9K runs

01-ai / yi-34b-200k

The Yi series models are large language models trained from scratch by developers at 01.AI.

1.7K runs

google-deepmind / gemma-2b

2B base version of Google’s Gemma model

1.6K runs

zsxkib / qwen2-7b-instruct

Qwen 2: A 7 billion parameter language model from Alibaba Cloud, fine tuned for chat completions

1.5K runs

mattt / orca-2-13b

1.5K runs

daanelson / flan-t5-large

A language model for tasks like classification, summarization, and more.

1.4K runs

deepseek-ai / deepseek-math-7b-instruct

Pushing the Limits of Mathematical Reasoning in Open Language Models - Instruct model

1.4K runs

niron1 / openorca-platypus2-13b

OpenOrca-Platypus2-13B is a merge of garage-bAInd/Platypus2-13B and Open-Orca/OpenOrcaxOpenChat-Preview2-13B.

1.3K runs

lucataco / hermes-2-pro-llama-3-8b

Hermes 2 Pro is an updated and cleaned version of the OpenHermes 2.5 Dataset, as well as a newly introduced Function Calling and JSON Mode dataset developed in-house

1.3K runs

google-deepmind / gemma2-27b-it

Google's Gemma2 27b instruct model

1.2K runs

anotherjesse / sdxl-recur

explore img2img zooming sdxl

1.1K runs

meta / codellama-70b-python

A 70 billion parameter Llama tuned for coding with Python

1.1K runs

nwhitehead / llama2-7b-chat-gptq

1.1K runs

google-deepmind / gemma2-9b-it

Google's Gemma2 9b instruct model

1K runs

nomagick / qwen-vl-chat

Qwen-VL-Chat but with raw ChatML prompt interface and streaming

1K runs

niron1 / qwen-7b-chat

Qwen-7B is the 7B-parameter version of the large language model series, Qwen (abbr. Tongyi Qianwen), proposed by Aibaba Cloud. Qwen-7B`is a Transformer-based large language model, which is pretrained on a large volume of data, including web texts, books,

992 runs

lucataco / qwen2-57b-a14b-instruct

Qwen2 57 billion parameter language model from Alibaba Cloud, fine tuned for chat completions

973 runs

andreasjansson / codellama-34b-instruct-gguf

CodeLlama-34B-instruct with support for grammars and jsonschema

962 runs

kcaverly / nous-capybara-34b-gguf

A SOTA Nous Research finetune of 200k Yi-34B fine tuned on the Capybara dataset.

927 runs

nateraw / samsum-llama-2-13b

843 runs

spuuntries / miqumaid-v2-2x70b-dpo-gguf

NeverSleep's MiquMaid v2 2x70B Miqu-Mixtral MoE DPO Finetune, GGUF Q2_K quantized by NeverSleep.

836 runs

andreasjansson / wizardcoder-python-34b-v1-gguf

WizardCoder-python-34B-v1.0 with support for grammars and jsonschema

732 runs

andreasjansson / llama-2-13b-gguf

Llama-2 13B with support for grammars and jsonschema

693 runs

lucataco / qwen1.5-14b

Qwen1.5 is the beta version of Qwen2, a transformer-based decoder-only language model pretrained on a large amount of data

672 runs

moinnadeem / vllm-engine-llama-7b

669 runs

charles-dyfis-net / llama-2-13b-hf--lmtp-8bit

662 runs

nateraw / sqlcoder-70b-alpha

572 runs

mikeei / dolphin-2.9.1-llama3-8b-gguf

Dolphin is uncensored. I have filtered the dataset to remove alignment and bias. This makes the model more compliant.

565 runs

spuuntries / erosumika-7b-v3-0.2-gguf

localfultonextractor's Erosumika 7B Mistral Merge, GGUF Q4_K_S-imat quantized by Lewdiculous.

564 runs

papermoose / llama-pajama

547 runs

fofr / llama2-prompter

Llama2 13b base model fine-tuned on text to image prompts

505 runs

kcaverly / deepseek-coder-6.7b-instruct

A ~7B parameter language model from Deepseek for SOTA repository level code completion

468 runs

fofr / star-trek-gpt-j-6b

gpt-j-6b trained on the Memory Alpha Star Trek Wiki

431 runs

andreasjansson / plasma

Generate plasma shader equations

425 runs

stability-ai / stablelm-base-alpha-3b

3B parameter base version of Stability AI's language model

416 runs

lucataco / ollama-reflection-70b

Ollama Reflection 70b

385 runs

lucataco / ollama-deepseek-coder-v2-236b

Cog wrapper for Ollama deepseek-coder-v2:236b

373 runs

lucataco / tinyllama-1.1b-chat-v1.0

This is the chat model finetuned on top of TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T

363 runs

nomagick / chatglm3-6b-32k

A 6B parameter open bilingual chat LLM (optimized for 8k+ context) | 开源双语对话语言模型

327 runs

lucataco / hermes-2-theta-llama-3-70b

Hermes-2 Θ (Theta) 70B is the continuation of our experimental merged model released by Nous Research

306 runs

nateraw / axolotl-llama-2-7b-english-to-hinglish

295 runs

ignaciosgithub / pllava

292 runs

peter65374 / openbuddy-llemma-34b-gguf

This is a cog implementation of "openbuddy-llemma-34b" 4-bit quantization model.

272 runs

niron1 / llama-2-7b-chat

LLAMA-2 7b chat version by Meta. Stream support. Unaltered prompt. Temperature working properly. Economical hardware.

266 runs

cbh123 / dylan-lyrics

Llama 2 13B fine-tuned on Bob Dylan lyrics

249 runs

antoinelyset / openhermes-2-mistral-7b

Simple version of https://huggingface.co/teknium/OpenHermes-2-Mistral-7B

241 runs

camenduru / mixtral-8x22b-v0.1-instruct-oh

Mixtral-8x22b-v0.1-Instruct-Open-Hermes

234 runs

cjwbw / starcoder2-15b

Language Models for Code

233 runs

adirik / mamba-2.8b

Base version of Mamba 2.8B, a 2.8 billion parameter state space language model

230 runs

kcaverly / phind-codellama-34b-v2-gguf

A quantized 34B parameter language model from Phind for code completion

225 runs

hayooucom / llm-60k

llm model ,for CN

213 runs

lucataco / phixtral-2x2_​8

phixtral-2x2_8 is the first Mixure of Experts (MoE) made with two microsoft/phi-2 models, inspired by the mistralai/Mixtral-8x7B-v0.1 architecture

212 runs

lucataco / qwen1.5-7b

Qwen1.5 is the beta version of Qwen2, a transformer-based decoder-only language model pretrained on a large amount of data

210 runs

nomagick / chatglm2-6b-int4

ChatGLM2-6B: An Open Bilingual Chat LLM | 开源双语对话语言模型 (int4)

207 runs

deepseek-ai / deepseek-coder-v2-lite-instruct

DeepSeek-Coder-V2: Breaking the Barrier of Closed-Source Models in Code Intelligence

201 runs

zeke / nyu-llama-2-7b-chat-training-test

A test model for fine-tuning Llama 2

198 runs

lucataco / dolphin-2.9-llama3-8b

Dolphin-2.9 has a variety of instruction, conversational, and coding skills. It also has initial agentic abilities and supports function calling

197 runs

zsxkib / qwen2-1.5b-instruct

Qwen 2: A 1.5 billion parameter language model from Alibaba Cloud, fine tuned for chat completions

190 runs

justmalhar / meta-llama-3.2-3b

Meta Llama 3.2 1B

187 runs

zallesov / super-real-llama2

186 runs

xrunda / med

181 runs

hamelsmu / honeycomb-2

Honeycomb NLQ Generator

180 runs

zsxkib / qwen2-0.5b-instruct

Qwen 2: A 0.5 billion parameter language model from Alibaba Cloud, fine tuned for chat completions

178 runs

kcaverly / nexus-raven-v2-13b-gguf

A quantized 13B parameter language model from NexusFlow for SOTA zero-shot function calling

177 runs

fofr / star-trek-adventure

159 runs

nateraw / stablecode-completion-alpha-3b-4k

156 runs

fofr / neuromancer-13b

llama-13b-base fine-tuned on Neuromancer style

146 runs

m1guelpf / mario-gpt

Using language models to generate Super Mario Bros levels

137 runs

camenduru / zephyr-orpo-141b-a35b-v0.1

Mixtral 8x22b v0.1 Zephyr Orpo 141b A35b v0.1

134 runs

nateraw / samsum-llama-7b

llama-2-7b fine-tuned on the samsum dataset for dialogue summarization

131 runs

fofr / star-trek-flan

flan-t5-xl trained on the Memory Alpha Star Trek Wiki

131 runs

cuuupid / minicpm-llama3-v-2.5

MiniCPM LLama3-V 2.5, a new SOTA open-source VLM that surpasses GPT-4V-1106 and Phi-128k on a number of benchmarks.

127 runs

fofr / star-trek-llama

llama-7b trained on the Memory Alpha Star Trek Wiki

125 runs

titocosta / notus-7b-v1

Notus-7b-v1 model

124 runs

lucataco / hermes-2-pro-llama-3-70b

Hermes 2 Pro is an updated and cleaned version of the OpenHermes 2.5 Dataset, as well as a newly introduced Function Calling and JSON Mode dataset developed in-house

121 runs

nateraw / llama-2-7b-paraphrase-v1

112 runs

adirik / mamba-130m

Base version of Mamba 130M, a 130 million parameter state space language model

110 runs

cbh123 / samsum

103 runs

cjwbw / opencodeinterpreter-ds-6.7b

OpenCodeInterpreter: Integrating Code Generation with Execution and Refinement

100 runs

rybens92 / una-cybertron-7b-v2--lmtp-8bit

92 runs

nateraw / wizardcoder-python-34b-v1.0

88 runs

crowdy / line-lang-3.6b

an implementation of 3.6b Japanese large language model

86 runs

nateraw / aidc-ai-business-marcoroni-13b

85 runs

nateraw / llama-2-7b-chat-hf

83 runs

moinnadeem / codellama-34b-instruct-vllm

77 runs

lucataco / qwen1.5-1.8b

Qwen1.5 is the beta version of Qwen2, a transformer-based decoder-only language model pretrained on a large amount of data

76 runs

spuuntries / borealis-10.7b-dpo-gguf

Undi95's Borealis 10.7B Mistral DPO Finetune, GGUF Q5_K_M quantized by Undi95.

73 runs

lucataco / nous-hermes-2-mixtral-8x7b-dpo

Nous Hermes 2 Mixtral 8x7B DPO is a Nous Research model trained over the Mixtral 8x7B MoE LLM

73 runs

nateraw / codellama-7b-instruct-hf

70 runs

justmalhar / meta-llama-3.2-1b

Meta Llama 3.2 1B

68 runs

peter65374 / openbuddy-mistral-7b

Openbuddy finetuned mistral-7b in GPTQ quantization in 4bits by TheBloke

66 runs

lidarbtc / kollava-v1.5

korean version of llava-v1.5

63 runs

martintmv-git / moondream2

small vision language model

62 runs

chigozienri / llava-birds

61 runs

lucataco / yi-1.5-6b

Yi-1.5 is continuously pre-trained on Yi with a high-quality corpus of 500B tokens and fine-tuned on 3M diverse fine-tuning samples

60 runs

adirik / mamba-1.4b

Base version of Mamba 1.4B, a 1.4 billion parameter state space language model

59 runs

lucataco / qwen1.5-0.5b

Qwen1.5 is the beta version of Qwen2, a transformer-based decoder-only language model pretrained on a large amount of data

58 runs

dsingal0 / mixtral-single-gpu

Runs Mixtral 8x7B on a single A40 GPU

55 runs

microsoft / phi-3-medium-4k-instruct

A 14B parameter, lightweight, state-of-the-art open model trained with the Phi-3 datasets that includes both synthetic data and the filtered publicly available websites data with a focus on high-quality and reasoning dense pro

55 runs

deepseek-ai / deepseek-67b-base

DeepSeek LLM, an advanced language model comprising 67 billion parameters. Trained from scratch on a vast dataset of 2 trillion tokens in both English and Chinese

52 runs

lucataco / ollama-qwen2.5-72b

Ollama Qwen2.5 72b

52 runs

adirik / mamba-2.8b-slimpj

Base version of Mamba 2.8B Slim Pyjama, a 2.8 billion parameter state space language model

51 runs

cbh123 / homerbot

50 runs

lucataco / internlm2_​5-7b-chat

InternLM2.5 has open-sourced a 7 billion parameter base model and a chat model tailored for practical scenarios.

48 runs

lucataco / ollama-llama3-70b

Cog wrapper for Ollama llama3:70b

48 runs

titocosta / starling

Starling-LM-7B-alpha

46 runs

adirik / mamba-790m

Base version of Mamba 790M, a 790 million parameter state space language model

43 runs

adirik / mamba-370m

Base version of Mamba 370M, a 370 million parameter state space language model

41 runs

nateraw / llama-2-7b-samsum

36 runs

hamelsmu / honeycomb

Honeycomb NLQ Generator

36 runs

lucataco / qwen1.5-32b

Qwen1.5 is the beta version of Qwen2, a transformer-based decoder-only language model pretrained on a large amount of data

34 runs

nateraw / gairmath-abel-7b

31 runs

google-deepmind / gemma-2-2b-it

Gemma2 2b Instruction-tuned variant by Google

28 runs

seanoliver / bob-dylan-fun-tuning

Llama fine-tune-athon project training llama2 on bob dylan lyrics.

27 runs

intentface / poro-34b-gguf-checkpoint

Try out akx/Poro-34B-gguf, Q5_K, This is 1000B checkpoint model

23 runs

interact-brands / llava-13b-spotter-creator

Fine-tuned LLaVa model for youtube thumbnail classification

21 runs

lucataco / qwen1.5-4b

Qwen1.5 is the beta version of Qwen2, a transformer-based decoder-only language model pretrained on a large amount of data

19 runs

nateraw / codellama-7b-instruct

18 runs

ydideh810 / cosmo-speak

A chat-bot that specialises in Space/Aeronautics knowledge.

17 runs

lucataco / numinamath-7b-tir

NuminaMath is a series of language models that are trained to solve math problems using tool-integrated reasoning (TIR)

17 runs

nateraw / codellama-34b

16 runs

charles-dyfis-net / llama-2-7b-hf--lmtp-4bit

16 runs

johnnyoshika / llama2-combine-numbers

16 runs

msamogh / iiu-generator-llama2-7b-2

14 runs

divyavanmahajan / my-pet-llama

14 runs

lucataco / hermes-2-theta-llama-3-8b

Hermes-2 Θ (Theta) is the first experimental merged model released by Nous Research

14 runs

deniyes / dolly-v2-12b-demo

dolly-v2-12b, just for testing

14 runs

nateraw / codellama-7b

13 runs

hayooucom / vision-llama3

for test

12 runs

lucataco / ollama-llama3-8b

Cog wrapper for Ollama llama3:8b

12 runs

halevi / sandbox1

11 runs

charles-dyfis-net / llama-2-13b-hf--lmtp

11 runs

aodianyun / minicpm-v-26

10 runs

nateraw / codellama-13b

8 runs

aodianyun / minicpm-v-26-int4

8 runs

nateraw / codellama-13b-instruct

7 runs

charles-dyfis-net / llama-2-13b-hf--lmtp-4bit

7 runs