Models that generate embeddings from inputs
andreasjansson / clip-features
Return CLIP features for the clip-vit-large-patch14 model
replicate / all-mpnet-base-v2
This is a language model that can be used to obtain document embeddings suitable for downstream tasks like semantic search and clustering.
daanelson / imagebind
A model for text, audio, and image embeddings in one space
nateraw / bge-large-en-v1.5
BAAI's bge-en-large-v1.5 for embedding text sequences
nateraw / jina-embeddings-v2-base-en
An 8k context text embedding model served FAST with ONNX on GPU. Check the examples tab to see different ways to run it.