nateraw / yarn-mistral-7b-128k

Nous-Yarn-Mistral-7b-128k is a state-of-the-art language model for long context

This model has no enabled versions.

Readme

Nous-Yarn-Mistral-7b-128k is a state-of-the-art language model for long context, further pretrained on long context data for 1500 steps using the YaRN extension method. It is an extension of Mistral-7B-v0.1 and supports a 128k token context window.

See the full model card here