You're looking at a specific version of this model. Jump to the model overview.
ibm-granite /granite-3.3-8b-instruct:8afd11cc
Input schema
The fields you can use to run this model with an API. If you don’t give a value for a field its default value will be used.
| Field | Type | Default value | Description |
|---|---|---|---|
| prompt |
string
|
Completion API user prompt.
|
|
| messages |
array
|
[]
|
Chat completion API messages.
|
| documents |
array
|
[]
|
Documents for request. Passed to the chat template.
|
| tools |
array
|
[]
|
Tools for request. Passed to the chat template.
|
| system_prompt |
string
|
Completion API system prompt. The chat template provides a good default.
|
|
| chat_template |
string
|
A template to format the prompt with. If not specified, the chat template provided by the model will be used.
|
|
| add_generation_prompt |
boolean
|
True
|
Add generation prompt. Passed to the chat template. Defaults to True.
|
| chat_template_kwargs |
object
|
{}
|
Additional arguments to be passed to the chat template.
|
| min_tokens |
integer
|
0
|
The minimum number of tokens the model should generate as output.
|
| max_tokens |
integer
|
512
|
The maximum number of tokens the model should generate as output.
|
| temperature |
number
|
0.6
|
The value used to modulate the next token probabilities.
|
| top_p |
number
|
0.9
|
A probability threshold for generating the output. If < 1.0, only keep the top tokens with cumulative probability >= top_p (nucleus filtering). Nucleus filtering is described in Holtzman et al. (http://arxiv.org/abs/1904.09751).
|
| top_k |
integer
|
50
|
The number of highest probability tokens to consider for generating the output. If > 0, only keep the top k tokens with highest probability (top-k filtering).
|
| presence_penalty |
number
|
0
|
Presence penalty
|
| frequency_penalty |
number
|
0
|
Frequency penalty
|
| stop_sequences |
string
|
A comma-separated list of sequences to stop generation at. For example, '<end>,<stop>' will stop generation at the first instance of '<end>' or '<stop>'.
|
|
| seed |
integer
|
Random seed. Leave unspecified to randomize the seed.
|
|
| stream |
boolean
|
False
|
Request streaming response. Defaults to False.
|
Output schema
The shape of the response you’ll get when you run this model with an API.
Schema
{'items': {'type': 'string'},
'title': 'Output',
'type': 'array',
'x-cog-array-display': 'concatenate',
'x-cog-array-type': 'iterator'}