Connectionists: attention mechanisms
Baldi,Pierre
pfbaldi at ics.uci.edu
Tue Mar 21 17:35:24 EDT 2023
On a less exciting note than the GPT discussion, let me bring to your
attention this article that just came out:
https://doi.org/10.1016/j.artint.2023.103901
Basically it identifies the basic building blocks of attention in deep
learning architectures and shows why these are
computationally efficient. Caution: this is just a beginning, not a full
theory of transformers.
--Pierre
More information about the Connectionists
mailing list