nateraw / llama-2-70b-chat-awq

llama-2-70b-chat quantized with AWQ and served with vLLM

  • Public
  • 77 runs
  • GitHub
  • Paper
  • License