nwhitehead / llama2-70b-oasst-sft-v10

This model is the Open-Assistant fine-tuning of Meta's Llama2 70B LLM.

  • Public
  • 2 runs
  • License

Run nwhitehead/llama2-70b-oasst-sft-v10 with an API

Use one of our client libraries to get started quickly. Clicking on a library will take you to the Playground tab where you can tweak different inputs, see the results, and copy the corresponding code to use in your own project.

Input schema

The fields you can use to run this model with an API. If you don't give a value for a field its default value will be used.

Field Type Default value Description
prompt
string
USER: Hello, who are you? ASSISTANT:
Text prompt for the model
temperature
number
0.5

Min: 0.01

Max: 2

Temperature of the output, it's best to keep it below 1
top_p
number
1

Min: 0.01

Max: 1

Top cumulative probability to filter candidates
top_k
integer
20

Min: 1

Max: 100

Number of top candidates to keep
repetition_penalty
number
1

Min: 1

Max: 1.5

Penalty for repeated tokens in the model's output
max_tokens
integer
50

Min: 1

Max: 4096

Maximum tokens to generate
min_tokens
integer
1

Max: 4096

Minimum tokens to generate
seed
integer
-1

Min: -2147483648

Max: 2147483647

Seed for reproducibility, -1 for random seed

Output schema

The shape of the response you’ll get when you run this model with an API.

Schema
{
  "type": "array",
  "items": {
    "type": "string"
  },
  "title": "Output",
  "x-cog-array-type": "iterator",
  "x-cog-array-display": "concatenate"
}