technillogue/llama2-summarizer

Public
44 runs

Input

Set the REPLICATE_API_TOKEN environment variable:
export REPLICATE_API_TOKEN=<paste-your-token-here>

Find your API token in your account settings.

Run technillogue/llama2-summarizer using Replicate’s API. Check out the model's schema for an overview of inputs and outputs.

curl -s -X POST \
  -H "Authorization: Bearer $REPLICATE_API_TOKEN" \
  -H "Content-Type: application/json" \
  -H "Prefer: wait" \
  -d $'{
    "version": "technillogue/llama2-summarizer:d2744ec63ed9cd13530b2187863340eb4e129efcc06fdcc34ee5243afcde9617",
    "input": {
      "debug": false,
      "top_k": 50,
      "top_p": 0.9,
      "temperature": 0.75,
      "max_new_tokens": 128,
      "min_new_tokens": -1
    }
  }' \
  https://api.replicate.com/v1/predictions

To learn more, take a look at Replicate’s HTTP API reference docs.

Output

No output yet! Press "Submit" to start a prediction.

Run time and cost

This model runs on Nvidia L40S GPU hardware. We don't yet have enough runs of this model to provide performance information.

Readme

This model doesn't have a readme.