fofr / star-trek-gpt-j-6b

gpt-j-6b trained on the Memory Alpha Star Trek Wiki

  • Public
  • 424 runs
  • GitHub
  • License



Run time and cost

This model runs on Nvidia A100 (40GB) GPU hardware. Predictions typically complete within 3 seconds. The predict time for this model varies significantly based on the inputs.


A GPT-J-6B large language model fine tuned on the Star Trek Wiki ‘Memory Alpha’.

This model is an experiment to see how the Wiki information translates after fine tuning.

Content from Memory Alpha is licensed under a creative commons non-commercial license:

Intended use

An experiment in fine tuning.

Ethical considerations

This model is not suitable for commercial purposes.

Caveats and recommendations

The model has deep knowledge of the Star Trek universe, but sometimes it acts like the facts have been shuffled.