this post was submitted on 17 Jul 2023
2 points (100.0% liked)
Machine Learning
1765 readers
1 users here now
founded 4 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
I am reading 'attention is not all you need' https://arxiv.org/abs/2103.03404
I had read this paper in the past but felt the need to refresh my memory and look at self attention with mildly critical lens. Afaik, this paper talks about attention networks without surrounding structures like MLP, skip connections etc and its behaviour.