Run meta/llama-2-13b-chat using Replicate’s API. Check out the model's schema for an overview of inputs and outputs.
output = replicate.run(
"meta/llama-2-13b-chat:df7690f1994d94e96ad9d568eac121aecf50684a0b0963b25a41cc40061269e5",
input={
"debug": False,
"top_p": 1,
"max_length": 500,
"temperature": 0.75,
"repetition_penalty": 1
}
)
# The meta/llama-2-13b-chat model can stream output as it's running.# The predict method returns an iterator, and you can iterate over that output.for item in output:
# https://replicate.com/meta/llama-2-13b-chat/api#output-schemaprint(item, end="")