Open-Assistant SST SFT-1 Pythia-12B or
oasst-sft-1-pythia-12b is a GPT-style large language model that has been trained to follow human instructions. Based on EleutherAI’s pythia-12b model, oasst-sft-1-pythia-12b was trained by Open-Assistant on ~22k human demonstrations of assistant conversations. This data was collected through the https://open-assistant.io/ human feedback web app before March 7, 2023.
oasst-sft-1-pythia-12b is a fine-tuned version of EleutherAI’s pythia-12b, a GPT-style causal language model trained on the Pile. See here for more information about the model architecture.
oasst-sft-1-pythia-12b was developed by fine-tuning the
pythia-12b checkpoint on roughly 22,000 instruction examples that were generated by the Open-Assistant community.
This is an experimental model that has been trained to act as an assistant, such that it responds to user queries with helpful answers. However, it has not been stress-tested and there are no guaranteed protections from malfunctions, inaccuracies, or harmful responses.
This model is not designed to avoid harmful or undesirable behavior and its output should not be unconditionally trusted in contexts where there are risks or costs of inaccuracy.