You're looking at a specific version of this model. Jump to the model overview.

prunaai /gpt-oss-120b-fast:e994aeeb

Input schema

The fields you can use to run this model with an API. If you don’t give a value for a field its default value will be used.

Field Type Default value Description
message
string
Explain vLLM in one sentence
The user message to send to the model
system_prompt
string
Optional system prompt to set the model's behavior
max_tokens
integer
2048

Min: 1

Max: 16384

Maximum number of tokens to generate
temperature
number
0.7

Max: 2

Sampling temperature (higher = more creative, lower = more deterministic)
top_p
number
0.95

Max: 1

Nucleus sampling: only consider tokens with cumulative probability up to this value

Output schema

The shape of the response you’ll get when you run this model with an API.

Schema
{'items': {'type': 'string'},
 'title': 'Output',
 'type': 'array',
 'x-cog-array-display': 'concatenate',
 'x-cog-array-type': 'iterator'}