Hacker News new | past | comments | ask | show | jobs | submit login

We might see that with the inference ASICs later this year I guess?



Ooh, what are these ASICs you're talking about? My understanding was that we'll see AMD/Nvidia gpus continue to be pushed and very competitive as well as have new system architectures like cerebras or grok. I haven't heard about new compute platforms framed as ASICs.


Cerebras has ridiculously large LLM ASICs that can hit crazy speeds. You can try it with llama 8B and 70B:

https://inference.cerebras.ai/

It's pretty fast, but my understanding is that it is still too expensive even accounting for the speed-up.


Is Cerebras an integrated circuit or more an integrated wafer? :-)

And yeah their cost is ridiculous, on the order for high 6 to low 7 figures per wafer. The rack alone looks several times more expensive than the 8x NVIDIA pods [1]

[1] https://web.archive.org/web/20230812020202/https://www.youtu...


https://www.etched.com/announcing-etched

I think there's another one but I can't remember the name of it.

Also a bit further out is https://spectrum.ieee.org/superconducting-computer

"Instead of the transistor, the basic element in superconducting logic is the Josephson-junction."




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: