Here is the XL model. 20x the size of the medium model. Still just 2B parameters, but on the bright side it was trained pre-wordslop.
https://huggingface.co/openai-community/gpt2-xl