Skip to content

MayankPalan2004/Memorizing-Transformers

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

4 Commits
 
 
 
 

Repository files navigation

Memorizing-Transformers

Pytorch Implementation of Memorizing Transformers by Yuhuai Wu, Markus N. Rabe, DeLesley Hutchins, Christian Szegedy [https://arxiv.org/abs/2203.08913] The "Memorizing Transformers" paper introduces the Memory-Augmented Transformer (MAT) architecture, integrating a Memory Attention Module (MAM) utilizing k-nearest neighbors (KNN) search for efficient retrieval of relevant information from memory. This enhances the transformer's ability to handle tasks requiring memorization by facilitating effective storage and retrieval of key information during decoding.

2744B704-37A6-4619-9F4B-9E12B6F0AA96

Results

Trained for 160 epochs where the loss dropped from 5.03 to 2.39

744F0981-FA6C-488E-8D56-5E58D1AEDFF5_4_5005_c

286CF399-8674-4791-B3EF-5E95B74B60D4_4_5005_c

About

Pytorch Implementation of Memorizing Transformers

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published