upvote
There’s an art to GPT sampling. You have to use temperature 0.7. People never believe it makes such a massive difference, but it does.
reply
Probably a much better prompt, too. I just literally pasted in the top part of my comment and let fly to see what would happen.
reply
Here is the XL model. 20x the size of the medium model. Still just 2B parameters, but on the bright side it was trained pre-wordslop.

https://huggingface.co/openai-community/gpt2-xl

reply