Run incomingflyingbrick/llm-test using Replicate’s API. Check out the model's schema for an overview of inputs and outputs.
output = replicate.run(
"incomingflyingbrick/llm-test:60f67a102d94a742b3981dfeab2d239d3c90ad58e6a7b4ac5e1c8bbbf306b823",
input={}
)
# The incomingflyingbrick/llm-test model can stream output as it's running.# The predict method returns an iterator, and you can iterate over that output.for item in output:
# https://replicate.com/incomingflyingbrick/llm-test/api#output-schemaprint(item, end="")
This model is not yet booted but ready for API calls. Your first API call will boot the model and may take longer, but after that subsequent responses will be fast.
This model runs on CPU hardware which costs $0.0001 per second