r-attention 0.4.0
Channel: guix-cran
Home page: https://cran.r-project.org/package=attention
Licenses: GPL 3+
Synopsis: Self-Attention Algorithm
Description:
Self-Attention algorithm helper functions and demonstration vignettes of increasing depth on how to construct the Self-Attention algorithm, this is based on Vaswani et al. (2017) <doi:10.48550/arXiv.1706.03762>
, Dan Jurafsky and James H. Martin (2022, ISBN:978-0131873216) <https://web.stanford.edu/~jurafsky/slp3/> "Speech and Language Processing (3rd ed.)" and Alex Graves (2020) <https://www.youtube.com/watch?v=AIiwuClvH6k>
"Attention and Memory in Deep Learning".
Total results: 1