Readme
CodeLlama is a family of fine-tuned Llama 2 models for coding. This is CodeLlama-70b-Instruct, a 70 billion parameter Llama model tuned for chatting about code.
A 70 billion parameter Llama tuned for coding and conversation
Run this model in Node.js with one line of code:
npm install replicate
REPLICATE_API_TOKEN
environment variable:export REPLICATE_API_TOKEN=<paste-your-token-here>
Find your API token in your account settings.
import Replicate from "replicate";
const replicate = new Replicate({
auth: process.env.REPLICATE_API_TOKEN,
});
Run meta/codellama-70b-instruct using Replicate’s API. Check out the model's schema for an overview of inputs and outputs.
const output = await replicate.run(
"meta/codellama-70b-instruct:a279116fe47a0f65701a8817188601e2fe8f4b9e04a518789655ea7b995851bf",
{
input: {
top_k: 10,
top_p: 0.95,
prompt: "In Bash, how do I list all text files in the current directory (excluding subdirectories) that have been modified in the last month?",
max_tokens: 500,
temperature: 0.8,
system_prompt: "",
repeat_penalty: 1.1,
presence_penalty: 0,
frequency_penalty: 0
}
}
);
console.log(output);
To learn more, take a look at the guide on getting started with Node.js.
This model costs approximately $0.038 to run on Replicate, or 26 runs per $1, but this varies depending on your inputs. It is also open source and you can run it on your own computer with Docker.
This model runs on Nvidia A100 (80GB) GPU hardware. Predictions typically complete within 28 seconds.
CodeLlama is a family of fine-tuned Llama 2 models for coding. This is CodeLlama-70b-Instruct, a 70 billion parameter Llama model tuned for chatting about code.