You're looking at a specific version of this model. Jump to the model overview.

technillogue /llama-2-7b-chat-hf-mlc:147a3c83

Input

*string
Shift + Return to add a new line

Prompt to send to the model.

string
Shift + Return to add a new line

System prompt to send to the model. This is prepended to the prompt and helps guide system behavior. Should not be blank.

Default: "You are a helpful, respectful and honest assistant."

integer
(minimum: 1)

Maximum number of tokens to generate. A word is generally 2-3 tokens

Default: 128

integer
(minimum: -1)

Minimum number of tokens to generate. To disable, set to -1. A word is generally 2-3 tokens.

Default: -1

number
(minimum: 0.01, maximum: 5)

Adjusts randomness of outputs, greater than 1 is random and 0 is deterministic, 0.75 is a good starting value.

Default: 0.7

number
(minimum: 0, maximum: 1)

When decoding text, samples from the top p percentage of most likely tokens; lower to ignore less likely tokens

Default: 0.95

number
(minimum: 0)

A parameter that controls how repetitive text can be. Lower means more repetitive, while higher means less repetitive. Set to 1.0 to disable.

Default: 1.15

string
Shift + Return to add a new line

A comma-separated list of sequences to stop generation at. For example, '<end>,<stop>' will stop generation at the first instance of 'end' or '<stop>'.

integer

Random seed. Leave blank to randomize the seed

boolean

provide debugging output in logs

Default: false

string
Shift + Return to add a new line

Template for formatting the prompt

Default: "[INST] <<SYS>>\n{system_prompt}\n<</SYS>>\n\n{prompt} [/INST]"

boolean

for debugging, do not use

Default: false

Output

No output yet! Press "Submit" to start a prediction.