technillogue / mixtral-nix

  • Public
  • 33 runs
  • 4x A100 (80GB)
Iterate in playground

Input

*string
Shift + Return to add a new line
integer

The maximum number of tokens the model should generate as output.

Default: 512

number

The value used to modulate the next token probabilities.

Default: 0.6

number

A probability threshold for generating the output. If < 1.0, only keep the top tokens with cumulative probability >= top_p (nucleus filtering). Nucleus filtering is described in Holtzman et al. (http://arxiv.org/abs/1904.09751).

Default: 0.9

string
Shift + Return to add a new line

conversation messages as json

Default: "[]"

string
Shift + Return to add a new line

instead of a single prediction, handle a WebRTC offer as json, optionally with an ice_server key of ICE servers to use for connecting

Output

No output yet! Press "Submit" to start a prediction.

Run time and cost

This model runs on 4x Nvidia A100 (80GB) GPU hardware. We don't yet have enough runs of this model to provide performance information.

Readme

This model doesn't have a readme.