nateraw / yarn-mistral-7b-128k

Nous-Yarn-Mistral-7b-128k is a state-of-the-art language model for long context

  • Public
  • 2 runs
  • GitHub
  • Paper
  • License

Nous-Yarn-Mistral-7b-128k is a state-of-the-art language model for long context, further pretrained on long context data for 1500 steps using the YaRN extension method. It is an extension of Mistral-7B-v0.1 and supports a 128k token context window.

See the full model card here