Run meta/llama-2-13b-chat using Replicate’s API. Check out the model's schema for an overview of inputs and outputs.
output = replicate.run(
"meta/llama-2-13b-chat:56acad22679f6b95d6e45c78309a2b50a670d5ed29a37dd73d182e89772c02f1",
input={
"debug": False,
"top_p": 1,
"max_length": 500,
"temperature": 0.75,
"repetition_penalty": 1
}
)
# The meta/llama-2-13b-chat model can stream output as it's running.# The predict method returns an iterator, and you can iterate over that output.for item in output:
# https://replicate.com/meta/llama-2-13b-chat/api#output-schemaprint(item, end="")