Skip to content

Latest commit

 

History

History
29 lines (22 loc) · 606 Bytes

README.md

File metadata and controls

29 lines (22 loc) · 606 Bytes

vit-pytorch

TODOs

  • Add patch embeddings
  • Add transformer encoder layer
  • Add transformer encoder (multiple layers)
  • Why repeat class token?
  • Attention dropout
  • Embedding dropout
  • MLP dropout (in encoder)
  • Add classification head
  • Complete ViT-Base
  • Make named layers to make torchvision compatible
  • Add training scripts

Install

pip install vit_pytorch

How to use

Load a config.yml file and pass to ViT module to modify architecture parameters.