Hacker News new | past | comments | ask | show | jobs | submit login

The bandwidth would end up costing way more than the compute would.



Depends how you do it. LLM are huge, but their input and output is miniscule bits of text. If you find a way to put "narrow paths" in the hidden layers, basically to subdivide a model into smaller interconnected models, then the bandwidth will be similarly massively reduced.

This is not without precedent, look up how your brain hemispheres and the regions within are connected.




Consider applying for YC's Spring batch! Applications are open till Feb 11.

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: