Hacker News new | past | comments | ask | show | jobs | submit login

Really cool. Can anyone answer these questions:

Should I use this or minGPT?

It says it needs 8XA100 40GB node. What is that and where do I acquire it?

Could someone else train this and then send me the model? What would be required to run it as opposed to training it?




A100’s are Nvidia GPU’s. You can rent them from providers like AWS or LamdaLabs. The readme has instructions for downloading the original GPT2 weights from OpenAI. You can also train a very simple version on a smaller dataset from your laptop as described in the README.

If you just want to play with a similar but much better model goto https://chat.openai.com




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: