A GLUE Codespace(NLP). All the code writed with transformers(4.18.0).
- L0Bert LEARNING SPARSE NEURAL NETWORKS THROUGH L0 REGULARIZATION
- GDASBert Searching for A Robust Neural Architecture in Four GPU Hours
- LadderTuning LST: Ladder Side-Tuning for Parameter and Memory Efficient Transfer Learning
- PyraidBert(TokenDropping) Pyramid-BERT: Reducing Complexity via Successive Core-set based Token Selection