Hacker News new | past | comments | ask | show | jobs | submit login

He’s not trying or claiming to improve attention. He’s trying to reduce outliers to improve the ability to quantize the parameters.



He refers all over the blog post to an "error" in attention. specifically says

The problem with using softmax is that it forces each attention head to make an annotation, even if it has no information to add to the output vector. Using softmax to choose among discrete alternatives is great; using it for optional annotation (i.e. as input into addition) is, like, not cool, man.

I'm saying it uses the current position to do this, that if it was a significant error I would expect it to improve the training loss. I sort of interpreted the blog post as being a bit more positive on the idea than just being about improving the quantization


I agree that he used the term error somewhat incorrectly. But he seems mainly to just be making the point that sumac introduces a large outlier which in turn is only now an issue that the community is now aggressively trying to quantize models




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: