arboreal-ai / llama-2-7b-chat

Llama-2-7b-Chat (GPTQ) with additional generation parameters

  • Public
  • 4.7K runs
  • GitHub
  • License

Readme

This model exposes 2 additional parameters that assist in reduced verbosity of the LLama2B 7B Chat model: (1) exponential_decay_start - Number of tokens after which it starts to become increasingly more likely that the model will stop generation. (2) exponential decay - Decay factor for exponential decay of probability of continued generation. In the range of 1 (no decay) and above.