Streaming language models
Language models that support streaming responses. See https://replicate.com/docs/streaming
Recommended models
bytedance / sdxl-lightning-4step
SDXL-Lightning by ByteDance: a fast text-to-image model that makes high-quality images in 4 steps
meta / meta-llama-3-70b-instruct
A 70 billion parameter language model from Meta, fine tuned for chat completions
meta / meta-llama-3-8b-instruct
An 8 billion parameter language model from Meta, fine tuned for chat completions
stability-ai / stable-diffusion
A latent text-to-image diffusion model capable of generating photo-realistic images given any text input
black-forest-labs / flux-schnell
The fastest image generation model tailored for local development and personal use
stability-ai / sdxl
A text-to-image generative AI model that creates beautiful images
meta / meta-llama-3-8b
Base version of Llama 3, an 8 billion parameter language model from Meta.
stability-ai / stable-diffusion-inpainting
Fill in masked parts of images with Stable Diffusion
yorickvp / llava-13b
Visual instruction tuning towards large language and vision models with GPT-4 level capabilities
mistralai / mixtral-8x7b-instruct-v0.1
The Mixtral-8x7B-instruct-v0.1 Large Language Model (LLM) is a pretrained generative Sparse Mixture of Experts tuned to be a helpful assistant.
meta / llama-2-7b-chat
A 7 billion parameter language model from Meta, fine tuned for chat completions
ai-forever / kandinsky-2.2
multilingual text2image latent diffusion model
meta / llama-2-70b-chat
A 70 billion parameter language model from Meta, fine tuned for chat completions
datacte / proteus-v0.2
Proteus v0.2 shows subtle yet significant improvements over Version 0.1. It demonstrates enhanced prompt understanding that surpasses MJ6, while also approaching its stylistic capabilities.
ai-forever / kandinsky-2
text2img model trained on LAION HighRes and fine-tuned on internal datasets
fofr / sdxl-emoji
An SDXL fine-tune based on Apple Emojis
yorickvp / llava-v1.6-mistral-7b
LLaVA v1.6: Large Language and Vision Assistant (Mistral-7B)
meta / llama-2-13b-chat
A 13 billion parameter language model from Meta, fine tuned for chat completions
mistralai / mistral-7b-instruct-v0.2
The Mistral-7B-Instruct-v0.2 Large Language Model (LLM) is an improved instruct fine-tuned version of Mistral-7B-Instruct-v0.1.
yorickvp / llava-v1.6-vicuna-13b
LLaVA v1.6: Large Language and Vision Assistant (Vicuna-13B)
meta / meta-llama-3.1-405b-instruct
Meta's flagship 405 billion parameter language model, fine-tuned for chat completions
tstramer / material-diffusion
Stable diffusion fork for generating tileable outputs using v1.5 model
fofr / prompt-classifier
Determines the toxicity of text to image prompts, llama-13b fine-tune. [SAFETY_RANKING] between 0 (safe) and 10 (toxic)
mistralai / mistral-7b-v0.1
A 7 billion parameter language model from Mistral.
datacte / proteus-v0.3
ProteusV0.3: The Anime Update
playgroundai / playground-v2.5-1024px-aesthetic
Playground v2.5 is the state-of-the-art open-source model in aesthetic quality
yorickvp / llava-v1.6-34b
LLaVA v1.6: Large Language and Vision Assistant (Nous-Hermes-2-34B)
stability-ai / stable-diffusion-3
A text-to-image model with greatly improved performance in image quality, typography, complex prompt understanding, and resource-efficiency
snowflake / snowflake-arctic-instruct
An efficient, intelligent, and truly open-source language model
fofr / latent-consistency-model
Super-fast, 0.6s per image. LCM with img2img, large batching and canny controlnet
lucataco / ssd-1b
Segmind Stable Diffusion Model (SSD-1B) is a distilled 50% smaller version of SDXL, offering a 60% speedup while maintaining high-quality text-to-image generation capabilities
fofr / realvisxl-v3-multi-controlnet-lora
RealVisXl V3 with multi-controlnet, lora loading, img2img, inpainting
meta / meta-llama-3-70b
Base version of Llama 3, a 70 billion parameter language model from Meta.
batouresearch / sdxl-controlnet-lora
'''Last update: Now supports img2img.''' SDXL Canny controlnet with LoRA support.
meta / llama-2-7b
Base version of Llama 2 7B, a 7 billion parameter language model
fofr / any-comfyui-workflow
Run any ComfyUI workflow. Guide: https://github.com/fofr/cog-comfyui
fofr / sticker-maker
Make stickers with AI. Generates graphics with transparent backgrounds.
spuuntries / flatdolphinmaid-8x7b-gguf
Undi95's FlatDolphinMaid 8x7B Mixtral Merge, GGUF Q5_K_M quantized by TheBloke.
meta / llama-2-70b
Base version of Llama 2, a 70 billion parameter language model from Meta.
lucataco / realvisxl2-lcm
RealvisXL-v2.0 with LCM LoRA - requires fewer steps (4 to 8 instead of the original 40 to 50)
lucataco / realvisxl-v2.0
Implementation of SDXL RealVisXL_V2.0
01-ai / yi-34b-chat
The Yi series models are large language models trained from scratch by developers at 01.AI.
nateraw / goliath-120b
An auto-regressive causal LM created by combining 2x finetuned Llama-2 70B into one.
fofr / sdxl-multi-controlnet-lora
Multi-controlnet, lora loading, img2img, inpainting
antoinelyset / openhermes-2-mistral-7b-awq
lucataco / moondream2
moondream2 is a small vision language model designed to run efficiently on edge devices
meta / llama-2-13b
Base version of Llama 2 13B, a 13 billion parameter language model
lucataco / dreamshaper-xl-turbo
DreamShaper is a general purpose SD model that aims at doing everything well, photos, art, anime, manga. It's designed to match Midjourney and DALL-E.
01-ai / yi-6b
The Yi series models are large language models trained from scratch by developers at 01.AI.
replicate / flan-t5-xl
A language model by Google for tasks like classification, summarization, and more
meta / codellama-34b-instruct
A 34 billion parameter Llama tuned for coding and conversation
stability-ai / stablelm-tuned-alpha-7b
7 billion parameter version of Stability AI's language model
meta / codellama-13b
A 13 billion parameter Llama tuned for code completion
lucataco / open-dalle-v1.1
A unique fusion that showcases exceptional prompt adherence and semantic understanding, it seems to be a step above base SDXL and a step closer to DALLE-3 in terms of prompt comprehension
nateraw / openchat_3.5-awq
OpenChat: Advancing Open-source Language Models with Mixed-Quality Data
replicate / llama-7b
Transformers implementation of the LLaMA language model
adirik / realvisxl-v3.0-turbo
Photorealism with RealVisXL V3.0 Turbo based on SDXL
google-deepmind / gemma-2b-it
2B instruct version of Google’s Gemma model
ai-forever / kandinsky-2-1
Kandinsky 2.1 Diffusion Model
microsoft / phi-3-mini-4k-instruct
Phi-3-Mini-4K-Instruct is a 3.8B parameters, lightweight, state-of-the-art open model trained with the Phi-3 datasets
antoinelyset / openhermes-2.5-mistral-7b-awq
google-deepmind / gemma-7b-it
7B instruct version of Google’s Gemma model
nateraw / mistral-7b-openorca
Mistral-7B-v0.1 fine tuned for chat with the OpenOrca dataset.
nightmareai / disco-diffusion
Generate images using a variety of techniques - Powered by Discoart
nateraw / nous-hermes-2-solar-10.7b
Nous Hermes 2 - SOLAR 10.7B is the flagship Nous Research model on the SOLAR 10.7B base model..
yorickvp / llava-v1.6-vicuna-7b
LLaVA v1.6: Large Language and Vision Assistant (Vicuna-7B)
fofr / image-prompts
Generate image prompts for Midjourney. Prefix inputs with "Image: "
cuuupid / glm-4v-9b
GLM-4V is a multimodal model released by Tsinghua University that is competitive with GPT-4o and establishes a new SOTA on several benchmarks, including OCR.
lucataco / pixart-xl-2
PixArt-Alpha 1024px is a transformer-based text-to-image diffusion system trained on text embeddings from T5
meta / codellama-13b-instruct
A 13 billion parameter Llama tuned for coding and conversation
meta / codellama-7b-instruct
A 7 billion parameter Llama tuned for coding and conversation
adirik / realvisxl-v4.0
Photorealism with RealVisXL V4.0
joehoover / falcon-40b-instruct
A 40 billion parameter language model trained to follow human instructions.
lucataco / dolphin-2.2.1-mistral-7b
Mistral-7B-v0.1 fine tuned for chat with the Dolphin dataset (an open-source implementation of Microsoft's Orca)
fofr / expression-editor
Quickly edit the expression of a face
google-deepmind / gemma-2-2b
Gemma2 2b by Google
kcaverly / openchat-3.5-1210-gguf
The "Overall Best Performing Open Source 7B Model" for Coding + Generalization or Mathematical Reasoning
lucataco / realistic-vision-v5
Realistic Vision v5.0 with VAE
meta / codellama-70b-instruct
A 70 billion parameter Llama tuned for coding and conversation
nateraw / defog-sqlcoder-7b-2
A capable large language model for natural language to SQL generation.
nousresearch / hermes-2-theta-llama-8b
Hermes-2 Θ (Theta) is the first experimental merged model released by Nous Research, in collaboration with Charles Goddard at Arcee, the team behind MergeKit.
uwulewd / airoboros-llama-2-70b
Inference Airoboros L2 70B 2.1 - GPTQ using ExLlama.
lucataco / wizardcoder-33b-v1.1-gguf
WizardCoder: Empowering Code Large Language Models with Evol-Instruct
meta / codellama-7b
A 7 billion parameter Llama tuned for coding and conversation
nomagick / chatglm3-6b
A 6B parameter open bilingual chat LLM | 开源双语对话语言模型
mikeei / dolphin-2.9-llama3-70b-gguf
Dolphin is uncensored. I have filtered the dataset to remove alignment and bias. This makes the model more compliant.
lucataco / wizard-vicuna-13b-uncensored
This is wizard-vicuna-13b trained with a subset of the dataset - responses that contained alignment / moralizing were removed
spuuntries / miqumaid-v1-70b-gguf
NeverSleep's MiquMaid v1 70B Miqu Finetune, GGUF Q3_K_M quantized by NeverSleep.
gregwdata / defog-sqlcoder-q8
Defog's SQLCoder is a state-of-the-art LLM for converting natural language questions to SQL queries. SQLCoder is a 15B parameter fine-tuned on a base StarCoder model.
lucataco / dolphin-2.1-mistral-7b
Mistral-7B-v0.1 fine tuned for chat with the Dolphin dataset (an open-source implementation of Microsoft's Orca)
kcaverly / neuralbeagle14-7b-gguf
NeuralBeagle14-7B is (probably) the best 7B model you can find!
antoinelyset / openhermes-2.5-mistral-7b
nomagick / chatglm2-6b
ChatGLM2-6B: An Open Bilingual Chat LLM | 开源双语对话语言模型
hayooucom / vision-model
This is phi-3-vision model , cost by time ,have fun~
lucataco / moondream1
(Research only) Moondream1 is a vision language model that performs on par with models twice its size
meta / codellama-34b
A 34 billion parameter Llama tuned for coding and conversation
kcaverly / nous-hermes-2-yi-34b-gguf
Nous Hermes 2 - Yi-34B is a state of the art Yi Fine-tune, fine tuned on GPT-4 generated synthetic data
replicate / gpt-j-6b
A large language model by EleutherAI
microsoft / phi-3-mini-128k-instruct
Phi-3-Mini-128K-Instruct is a 3.8 billion-parameter, lightweight, state-of-the-art open model trained using the Phi-3 datasets
nateraw / nous-hermes-llama2-awq
TheBloke/Nous-Hermes-Llama2-AWQ served with vLLM
google-deepmind / gemma-7b
7B base version of Google’s Gemma model
joehoover / zephyr-7b-alpha
A high-performing language model trained to act as a helpful assistant
hikikomori-haven / solar-uncensored
nateraw / zephyr-7b-beta
Zephyr-7B-beta, an LLM trained to act as a helpful assistant.
01-ai / yi-6b-chat
The Yi series models are large language models trained from scratch by developers at 01.AI.
nomagick / qwen-14b-chat
Qwen-14B-Chat is a Transformer-based large language model, which is pretrained on a large volume of data, including web texts, books, codes, etc.
mikeei / dolphin-2.9-llama3-8b-gguf
Dolphin is uncensored. I have filtered the dataset to remove alignment and bias. This makes the model more compliant.
kcaverly / nous-hermes-2-solar-10.7b-gguf
Nous Hermes 2 - SOLAR 10.7B is the flagship Nous Research model on the SOLAR 10.7B base model.
lucataco / qwen1.5-72b
Qwen1.5 is the beta version of Qwen2, a transformer-based decoder-only language model pretrained on a large amount of data
meta / codellama-13b-python
A 13 billion parameter Llama tuned for coding with Python
meta / codellama-7b-python
A 7 billion parameter Llama tuned for coding with Python
joehoover / sql-generator
01-ai / yi-34b
The Yi series models are large language models trained from scratch by developers at 01.AI.
kcaverly / deepseek-coder-33b-instruct-gguf
A quantized 33B parameter language model from Deepseek for SOTA repository level code completion
lucataco / qwen1.5-110b
Qwen1.5 is the beta version of Qwen2, a transformer-based decoder-only language model pretrained on a large amount of data
organisciak / ocsai-llama2-7b
deepseek-ai / deepseek-math-7b-base
Pushing the Limits of Mathematical Reasoning in Open Language Models - Base model
deepseek-ai / deepseek-vl-7b-base
DeepSeek-VL: An open-source Vision-Language Model designed for real-world vision and language understanding applications
01-ai / yi-34b-200k
The Yi series models are large language models trained from scratch by developers at 01.AI.
google-deepmind / gemma-2b
2B base version of Google’s Gemma model
zsxkib / qwen2-7b-instruct
Qwen 2: A 7 billion parameter language model from Alibaba Cloud, fine tuned for chat completions
mattt / orca-2-13b
daanelson / flan-t5-large
A language model for tasks like classification, summarization, and more.
deepseek-ai / deepseek-math-7b-instruct
Pushing the Limits of Mathematical Reasoning in Open Language Models - Instruct model
niron1 / openorca-platypus2-13b
OpenOrca-Platypus2-13B is a merge of garage-bAInd/Platypus2-13B and Open-Orca/OpenOrcaxOpenChat-Preview2-13B.
lucataco / hermes-2-pro-llama-3-8b
Hermes 2 Pro is an updated and cleaned version of the OpenHermes 2.5 Dataset, as well as a newly introduced Function Calling and JSON Mode dataset developed in-house
google-deepmind / gemma2-27b-it
Google's Gemma2 27b instruct model
anotherjesse / sdxl-recur
explore img2img zooming sdxl
meta / codellama-70b-python
A 70 billion parameter Llama tuned for coding with Python
nwhitehead / llama2-7b-chat-gptq
google-deepmind / gemma2-9b-it
Google's Gemma2 9b instruct model
nomagick / qwen-vl-chat
Qwen-VL-Chat but with raw ChatML prompt interface and streaming
niron1 / qwen-7b-chat
Qwen-7B is the 7B-parameter version of the large language model series, Qwen (abbr. Tongyi Qianwen), proposed by Aibaba Cloud. Qwen-7B`is a Transformer-based large language model, which is pretrained on a large volume of data, including web texts, books,
lucataco / qwen2-57b-a14b-instruct
Qwen2 57 billion parameter language model from Alibaba Cloud, fine tuned for chat completions
andreasjansson / codellama-34b-instruct-gguf
CodeLlama-34B-instruct with support for grammars and jsonschema
kcaverly / nous-capybara-34b-gguf
A SOTA Nous Research finetune of 200k Yi-34B fine tuned on the Capybara dataset.
nateraw / samsum-llama-2-13b
spuuntries / miqumaid-v2-2x70b-dpo-gguf
NeverSleep's MiquMaid v2 2x70B Miqu-Mixtral MoE DPO Finetune, GGUF Q2_K quantized by NeverSleep.
andreasjansson / wizardcoder-python-34b-v1-gguf
WizardCoder-python-34B-v1.0 with support for grammars and jsonschema
andreasjansson / llama-2-13b-gguf
Llama-2 13B with support for grammars and jsonschema
lucataco / qwen1.5-14b
Qwen1.5 is the beta version of Qwen2, a transformer-based decoder-only language model pretrained on a large amount of data
moinnadeem / vllm-engine-llama-7b
charles-dyfis-net / llama-2-13b-hf--lmtp-8bit
nateraw / sqlcoder-70b-alpha
mikeei / dolphin-2.9.1-llama3-8b-gguf
Dolphin is uncensored. I have filtered the dataset to remove alignment and bias. This makes the model more compliant.
spuuntries / erosumika-7b-v3-0.2-gguf
localfultonextractor's Erosumika 7B Mistral Merge, GGUF Q4_K_S-imat quantized by Lewdiculous.
papermoose / llama-pajama
fofr / llama2-prompter
Llama2 13b base model fine-tuned on text to image prompts
kcaverly / deepseek-coder-6.7b-instruct
A ~7B parameter language model from Deepseek for SOTA repository level code completion
fofr / star-trek-gpt-j-6b
gpt-j-6b trained on the Memory Alpha Star Trek Wiki
andreasjansson / plasma
Generate plasma shader equations
stability-ai / stablelm-base-alpha-3b
3B parameter base version of Stability AI's language model
lucataco / ollama-reflection-70b
Ollama Reflection 70b
lucataco / ollama-deepseek-coder-v2-236b
Cog wrapper for Ollama deepseek-coder-v2:236b
lucataco / tinyllama-1.1b-chat-v1.0
This is the chat model finetuned on top of TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T
nomagick / chatglm3-6b-32k
A 6B parameter open bilingual chat LLM (optimized for 8k+ context) | 开源双语对话语言模型
lucataco / hermes-2-theta-llama-3-70b
Hermes-2 Θ (Theta) 70B is the continuation of our experimental merged model released by Nous Research
nateraw / axolotl-llama-2-7b-english-to-hinglish
ignaciosgithub / pllava
peter65374 / openbuddy-llemma-34b-gguf
This is a cog implementation of "openbuddy-llemma-34b" 4-bit quantization model.
niron1 / llama-2-7b-chat
LLAMA-2 7b chat version by Meta. Stream support. Unaltered prompt. Temperature working properly. Economical hardware.
cbh123 / dylan-lyrics
Llama 2 13B fine-tuned on Bob Dylan lyrics
antoinelyset / openhermes-2-mistral-7b
Simple version of https://huggingface.co/teknium/OpenHermes-2-Mistral-7B
camenduru / mixtral-8x22b-v0.1-instruct-oh
Mixtral-8x22b-v0.1-Instruct-Open-Hermes
cjwbw / starcoder2-15b
Language Models for Code
adirik / mamba-2.8b
Base version of Mamba 2.8B, a 2.8 billion parameter state space language model
kcaverly / phind-codellama-34b-v2-gguf
A quantized 34B parameter language model from Phind for code completion
hayooucom / llm-60k
llm model ,for CN
lucataco / phixtral-2x2_8
phixtral-2x2_8 is the first Mixure of Experts (MoE) made with two microsoft/phi-2 models, inspired by the mistralai/Mixtral-8x7B-v0.1 architecture
lucataco / qwen1.5-7b
Qwen1.5 is the beta version of Qwen2, a transformer-based decoder-only language model pretrained on a large amount of data
nomagick / chatglm2-6b-int4
ChatGLM2-6B: An Open Bilingual Chat LLM | 开源双语对话语言模型 (int4)
deepseek-ai / deepseek-coder-v2-lite-instruct
DeepSeek-Coder-V2: Breaking the Barrier of Closed-Source Models in Code Intelligence
zeke / nyu-llama-2-7b-chat-training-test
A test model for fine-tuning Llama 2
lucataco / dolphin-2.9-llama3-8b
Dolphin-2.9 has a variety of instruction, conversational, and coding skills. It also has initial agentic abilities and supports function calling
zsxkib / qwen2-1.5b-instruct
Qwen 2: A 1.5 billion parameter language model from Alibaba Cloud, fine tuned for chat completions
justmalhar / meta-llama-3.2-3b
Meta Llama 3.2 1B
zallesov / super-real-llama2
xrunda / med
hamelsmu / honeycomb-2
Honeycomb NLQ Generator
zsxkib / qwen2-0.5b-instruct
Qwen 2: A 0.5 billion parameter language model from Alibaba Cloud, fine tuned for chat completions
kcaverly / nexus-raven-v2-13b-gguf
A quantized 13B parameter language model from NexusFlow for SOTA zero-shot function calling
fofr / star-trek-adventure
nateraw / stablecode-completion-alpha-3b-4k
fofr / neuromancer-13b
llama-13b-base fine-tuned on Neuromancer style
m1guelpf / mario-gpt
Using language models to generate Super Mario Bros levels
camenduru / zephyr-orpo-141b-a35b-v0.1
Mixtral 8x22b v0.1 Zephyr Orpo 141b A35b v0.1
nateraw / samsum-llama-7b
llama-2-7b fine-tuned on the samsum dataset for dialogue summarization
fofr / star-trek-flan
flan-t5-xl trained on the Memory Alpha Star Trek Wiki
cuuupid / minicpm-llama3-v-2.5
MiniCPM LLama3-V 2.5, a new SOTA open-source VLM that surpasses GPT-4V-1106 and Phi-128k on a number of benchmarks.
fofr / star-trek-llama
llama-7b trained on the Memory Alpha Star Trek Wiki
titocosta / notus-7b-v1
Notus-7b-v1 model
lucataco / hermes-2-pro-llama-3-70b
Hermes 2 Pro is an updated and cleaned version of the OpenHermes 2.5 Dataset, as well as a newly introduced Function Calling and JSON Mode dataset developed in-house
nateraw / llama-2-7b-paraphrase-v1
adirik / mamba-130m
Base version of Mamba 130M, a 130 million parameter state space language model
cbh123 / samsum
cjwbw / opencodeinterpreter-ds-6.7b
OpenCodeInterpreter: Integrating Code Generation with Execution and Refinement
rybens92 / una-cybertron-7b-v2--lmtp-8bit
nateraw / wizardcoder-python-34b-v1.0
crowdy / line-lang-3.6b
an implementation of 3.6b Japanese large language model
nateraw / aidc-ai-business-marcoroni-13b
nateraw / llama-2-7b-chat-hf
moinnadeem / codellama-34b-instruct-vllm
lucataco / qwen1.5-1.8b
Qwen1.5 is the beta version of Qwen2, a transformer-based decoder-only language model pretrained on a large amount of data
spuuntries / borealis-10.7b-dpo-gguf
Undi95's Borealis 10.7B Mistral DPO Finetune, GGUF Q5_K_M quantized by Undi95.
lucataco / nous-hermes-2-mixtral-8x7b-dpo
Nous Hermes 2 Mixtral 8x7B DPO is a Nous Research model trained over the Mixtral 8x7B MoE LLM
nateraw / codellama-7b-instruct-hf
justmalhar / meta-llama-3.2-1b
Meta Llama 3.2 1B
peter65374 / openbuddy-mistral-7b
Openbuddy finetuned mistral-7b in GPTQ quantization in 4bits by TheBloke
lidarbtc / kollava-v1.5
korean version of llava-v1.5
martintmv-git / moondream2
small vision language model
chigozienri / llava-birds
lucataco / yi-1.5-6b
Yi-1.5 is continuously pre-trained on Yi with a high-quality corpus of 500B tokens and fine-tuned on 3M diverse fine-tuning samples
adirik / mamba-1.4b
Base version of Mamba 1.4B, a 1.4 billion parameter state space language model
lucataco / qwen1.5-0.5b
Qwen1.5 is the beta version of Qwen2, a transformer-based decoder-only language model pretrained on a large amount of data
dsingal0 / mixtral-single-gpu
Runs Mixtral 8x7B on a single A40 GPU
microsoft / phi-3-medium-4k-instruct
A 14B parameter, lightweight, state-of-the-art open model trained with the Phi-3 datasets that includes both synthetic data and the filtered publicly available websites data with a focus on high-quality and reasoning dense pro
deepseek-ai / deepseek-67b-base
DeepSeek LLM, an advanced language model comprising 67 billion parameters. Trained from scratch on a vast dataset of 2 trillion tokens in both English and Chinese
lucataco / ollama-qwen2.5-72b
Ollama Qwen2.5 72b
adirik / mamba-2.8b-slimpj
Base version of Mamba 2.8B Slim Pyjama, a 2.8 billion parameter state space language model
cbh123 / homerbot
lucataco / internlm2_5-7b-chat
InternLM2.5 has open-sourced a 7 billion parameter base model and a chat model tailored for practical scenarios.
lucataco / ollama-llama3-70b
Cog wrapper for Ollama llama3:70b
titocosta / starling
Starling-LM-7B-alpha
adirik / mamba-790m
Base version of Mamba 790M, a 790 million parameter state space language model
adirik / mamba-370m
Base version of Mamba 370M, a 370 million parameter state space language model
nateraw / llama-2-7b-samsum
hamelsmu / honeycomb
Honeycomb NLQ Generator
lucataco / qwen1.5-32b
Qwen1.5 is the beta version of Qwen2, a transformer-based decoder-only language model pretrained on a large amount of data
nateraw / gairmath-abel-7b
google-deepmind / gemma-2-2b-it
Gemma2 2b Instruction-tuned variant by Google
seanoliver / bob-dylan-fun-tuning
Llama fine-tune-athon project training llama2 on bob dylan lyrics.
intentface / poro-34b-gguf-checkpoint
Try out akx/Poro-34B-gguf, Q5_K, This is 1000B checkpoint model
interact-brands / llava-13b-spotter-creator
Fine-tuned LLaVa model for youtube thumbnail classification
lucataco / qwen1.5-4b
Qwen1.5 is the beta version of Qwen2, a transformer-based decoder-only language model pretrained on a large amount of data
nateraw / codellama-7b-instruct
ydideh810 / cosmo-speak
A chat-bot that specialises in Space/Aeronautics knowledge.
lucataco / numinamath-7b-tir
NuminaMath is a series of language models that are trained to solve math problems using tool-integrated reasoning (TIR)
nateraw / codellama-34b
charles-dyfis-net / llama-2-7b-hf--lmtp-4bit
johnnyoshika / llama2-combine-numbers
msamogh / iiu-generator-llama2-7b-2
divyavanmahajan / my-pet-llama
lucataco / hermes-2-theta-llama-3-8b
Hermes-2 Θ (Theta) is the first experimental merged model released by Nous Research
deniyes / dolly-v2-12b-demo
dolly-v2-12b, just for testing
nateraw / codellama-7b
hayooucom / vision-llama3
for test
lucataco / ollama-llama3-8b
Cog wrapper for Ollama llama3:8b
halevi / sandbox1
charles-dyfis-net / llama-2-13b-hf--lmtp
aodianyun / minicpm-v-26
nateraw / codellama-13b
aodianyun / minicpm-v-26-int4
nateraw / codellama-13b-instruct
charles-dyfis-net / llama-2-13b-hf--lmtp-4bit