Connectionists: attention mechanisms

Baldi,Pierre pfbaldi at ics.uci.edu
Tue Mar 21 17:35:24 EDT 2023


On a less exciting note than the GPT discussion, let me bring to your 
attention this article that just came out:

https://doi.org/10.1016/j.artint.2023.103901

Basically it identifies the basic building blocks of attention in deep 
learning architectures and shows why these are

computationally efficient. Caution: this is just a beginning, not a full 
theory of transformers.

--Pierre



More information about the Connectionists mailing list