Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

I still can not understand the name of "attention is what you need" study.


The "attention is all you need" paper did not invent attention mechanisms. It showed that existing models that were already using attention could have their non-attention parts removed and still worked. So those other parts were unnecessary and only attention was needed.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: