Hacker News new | past | comments | ask | show | jobs | submit login

I'm pretty sure he suggested it was a 16 way 110 MoE



The exact quote: "Sam Altman won’t tell you that GPT 4 has 220 billion parameters and is a 16 way mixture model with eight sets of weights."




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: