lucataco / yi-1.5-6b

Yi-1.5 is continuously pre-trained on Yi with a high-quality corpus of 500B tokens and fine-tuned on 3M diverse fine-tuning samples

  • Public
  • 60 runs
  • GitHub
  • Weights
  • Paper
  • License

Run time and cost

This model runs on Nvidia A40 (Large) GPU hardware. We don't yet have enough runs of this model to provide performance information.

Readme

Intro

Yi-1.5 is an upgraded version of Yi. It is continuously pre-trained on Yi with a high-quality corpus of 500B tokens and fine-tuned on 3M diverse fine-tuning samples.

Compared with Yi, Yi-1.5 delivers stronger performance in coding, math, reasoning, and instruction-following capability, while still maintaining excellent capabilities in language understanding, commonsense reasoning, and reading comprehension.

Model | Context Length | Pre-trained Tokens | :------------: | :------------: | :------------: | | Yi-1.5 | 4K | 3.6T

Models

  • Chat models
Name Download
Yi-1.5-34B-Chat 🤗 Hugging Face🤖 ModelScope
Yi-1.5-9B-Chat 🤗 Hugging Face🤖 ModelScope
Yi-1.5-6B-Chat 🤗 Hugging Face🤖 ModelScope
  • Base models
Name Download
Yi-1.5-34B 🤗 Hugging Face🤖 ModelScope
Yi-1.5-9B 🤗 Hugging Face🤖 ModelScope
Yi-1.5-6B 🤗 Hugging Face🤖 ModelScope

Benchmarks

  • Chat models

Yi-1.5-34B-Chat is on par with or excels beyond larger models in most benchmarks.

image/png

Yi-1.5-9B-Chat is the top performer among similarly sized open-source models.

image/png

  • Base models

Yi-1.5-34B is on par with or excels beyond larger models in some benchmarks.

image/png

Yi-1.5-9B is the top performer among similarly sized open-source models.

image/png

Quick Start

For getting up and running with Yi-1.5 models quickly, see README.