You're looking at a specific version of this model. Jump to the model overview.
nwhitehead /llama2-7b-chat-gptq:8c1f632f
Input schema
The fields you can use to run this model with an API. If you don’t give a value for a field its default value will be used.
Field | Type | Default value | Description |
---|---|---|---|
prompt |
string
|
USER: Hello, who are you?
ASSISTANT:
|
Text prompt for the model
|
temperature |
number
|
0.5
Min: 0.01 Max: 2 |
Temperature of the output, it's best to keep it below 1
|
top_p |
number
|
1
Min: 0.01 Max: 1 |
Top cumulative probability to filter candidates
|
top_k |
integer
|
20
Min: 1 Max: 100 |
Number of top candidates to keep
|
repetition_penalty |
number
|
1
Min: 1 Max: 1.5 |
Penalty for repeated tokens in the model's output
|
max_tokens |
integer
|
50
Min: 1 Max: 4096 |
Maximum tokens to generate
|
min_tokens |
integer
|
1
Max: 4096 |
Minimum tokens to generate
|
seed |
integer
|
-1
Min: -2147483648 Max: 2147483647 |
Seed for reproducibility, -1 for random seed
|
Output schema
The shape of the response you’ll get when you run this model with an API.
Schema
{'items': {'type': 'string'},
'title': 'Output',
'type': 'array',
'x-cog-array-display': 'concatenate',
'x-cog-array-type': 'iterator'}