Run technillogue/llama-89 using Replicate’s API. Check out the model's schema for an overview of inputs and outputs.
output = replicate.run(
"technillogue/llama-89:1b8a0ce6eb03080cc97a8c0ef5939e9e9a0ea7e871f70212b15bf085d3233965",
input={
"top_k": 0,
"top_p": 0.95,
"max_tokens": 512,
"temperature": 0.7,
"system_prompt": "",
"length_penalty": 1,
"prompt_template": "{prompt}",
"presence_penalty": 0,
"log_performance_metrics": False
}
)
# The technillogue/llama-89 model can stream output as it's running.# The predict method returns an iterator, and you can iterate over that output.for item in output:
# https://replicate.com/technillogue/llama-89/api#output-schemaprint(item, end="")