An efficient, intelligent, and truly open-source language model
This model is warm. You'll get a fast response if the model is warm and already running, and a slower response if the model is cold and starting up.
This model is priced per input token and output token. View more