A100’s are Nvidia GPU’s. You can rent them from providers like AWS or LamdaLabs. The readme has instructions for downloading the original GPT2 weights from OpenAI. You can also train a very simple version on a smaller dataset from your laptop as described in the README.
Should I use this or minGPT?
It says it needs 8XA100 40GB node. What is that and where do I acquire it?
Could someone else train this and then send me the model? What would be required to run it as opposed to training it?