Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

New Layers, aux. functions and built-in model Vision Transformer #11

Open
wants to merge 4 commits into
base: master
Choose a base branch
from

Conversation

carloshnpa
Copy link

In a final course project, guided by Dr. Talles Medeiros (@tallesmedeiros), we invested in building new layers and auxiliary functions that enabled the modeling of a Transformer architecture applied to computer vision problems, using the equations from Devito. The new layers only implement the forward phase due to time constraints in completing the work. The weights and biases were transferred from an equivalent network using the PyTorch framework.

Added layers and functions:

  • 3D FullyConnected
  • Einsun function
  • Dropout 1, 2, 3 and 4 dimensions
  • Norm 2D
  • Norm 3D
  • Softmax 3D and 4D function

Running model exemple:

  • Vision Transformer (pre-trained with MNIST 28px)

@tallesmedeiros
Copy link

@navjotk

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants