nateraw / zephyr-7b-beta

Zephyr-7B-beta, an LLM trained to act as a helpful assistant.

  • Public
  • 5.7K runs
  • GitHub
  • Paper
  • License

Run time and cost

This model costs approximately $0.098 to run on Replicate, or 10 runs per $1, but this varies depending on your inputs. It is also open source and you can run it on your own computer with Docker.

This model runs on Nvidia A40 (Large) GPU hardware. Predictions typically complete within 135 seconds. The predict time for this model varies significantly based on the inputs.

Readme

Zephyr is a series of language models that are trained to act as helpful assistants. Zephyr-7B-β is the second model in the series, and is a fine-tuned version of mistralai/Mistral-7B-v0.1 that was trained on on a mix of publicly available, synthetic datasets using Direct Preference Optimization (DPO). We found that removing the in-built alignment of these datasets boosted performance on MT Bench and made the model more helpful. However, this means that model is likely to generate problematic text when prompted to do so and should only be used for educational and research purposes. You can find more details in the technical report.

See the full model card for more details: https://huggingface.co/HuggingFaceH4/zephyr-7b-beta

Thank you to TheBloke for releasing the AWQ version of the model, which is served here. Link: https://huggingface.co/TheBloke/zephyr-7B-beta-AWQ