stability-ai / stablelm-tuned-alpha-7b

7 billion parameter version of Stability AI's language model

  • Public
  • 140.5K runs
  • A100 (80GB)
  • GitHub
  • License

Input

string
Shift + Return to add a new line

Input Prompt.

Default: "What's your mood today?"

integer
(minimum: 1)

Maximum number of tokens to generate. A word is generally 2-3 tokens

Default: 100

number
(minimum: 0.01, maximum: 1)

Valid if you choose top_p decoding. When decoding text, samples from the top p percentage of most likely tokens; lower to ignore less likely tokens

Default: 1

number
(minimum: 0.01, maximum: 5)

Adjusts randomness of outputs, greater than 1 is random and 0 is deterministic, 0.75 is a good starting value.

Default: 0.75

number
(minimum: 0.01, maximum: 5)

Penalty for repeated words in generated text; 1 is no penalty, values greater than 1 discourage repetition, less than 1 encourage it.

Default: 1.2

Output

First start with some lovely vegetables like eggplant or zucchini - whatever you fancy! Then cook them up nice and tender. Next put on some herbs if you want something special, such as thyme or rosemary - maybe a sprinkling of smoked paprika for extra flavor. Once everything's cooked through stir it all together well so each spoonful contains delicious little bits of veggies, herbs and flavourings. And serve over plenty of good quality white rice – either keep it simple with basmati or let your guests choose their own pasta from amongst several available options... The end-result should look delectable too!
Generated in

This example was created by a different version, stability-ai/stablelm-tuned-alpha-7b:4a9a32b4.

Run time and cost

This model costs approximately $0.0026 to run on Replicate, or 384 runs per $1, but this varies depending on your inputs. It is also open source and you can run it on your own computer with Docker.

This model runs on Nvidia A100 (80GB) GPU hardware. Predictions typically complete within 2 seconds.

Readme

StableLM-Tuned-Alpha-7B is a 7B parameter decoder-only language model built on top of the StableLM-Base-Alpha models and further fine-tuned on various chat and instruction-following datasets.

The StableLM-Alpha models are trained on the new dataset that builds on The Pile, which contains 1.5 trillion tokens, roughly 3x the size of The Pile. These models will be trained on up to 1.5 trillion tokens. The context length for these models is 4096 tokens.

An upcoming technical report will document the model specifications and the training settings.

As a proof-of-concept, we also fine-tuned the model with Stanford Alpaca’s procedure using a combination of five recent datasets for conversational agents: Stanford’s Alpaca, Nomic-AI’s gpt4all, RyokoAI’s ShareGPT52K datasets, Databricks labs’ Dolly, and Anthropic’s HH. We will be releasing these models as StableLM-Tuned-Alpha.

Acknowledgements

  • StableLM-Tuned-Alpha would not have been possible without the helpful hand of Dakota Mahan @dmayhem93.

Licenses

  • Base model checkpoints (StableLM-Base-Alpha) are licensed under the Creative Commons license (CC BY-SA-4.0). Under the license, you must give credit to Stability AI, provide a link to the license, and indicate if changes were made. You may do so in any reasonable manner, but not in any way that suggests the Stability AI endorses you or your use.

  • Fine-tuned checkpoints (StableLM-Tuned-Alpha) are licensed under the Non-Commercial Creative Commons license (CC BY-NC-SA-4.0), in-line with the original non-commercial license specified by Stanford Alpaca.

  • All code in this repository is licensed under the Apache License 2.0 license.