gittymutt / scaled_dot_product_attention Goto Github PK
View Code? Open in Web Editor NEWThis project forked from brandontrabucco/scaled_dot_product_attention
Implements the scaled dot product attention mechamism from the transformer. Vaswani, A. et al. https://arxiv.org/pdf/1706.03762.pdf
License: MIT License