🔥Accepted in TMLR (08/2024) OpenReview
Traditional deep learning models are trained and tested on relatively low-resolution images (
Create a conda envirnoment:
conda create -n pgd python=3.12
conda activate pgd
Install requirements using the following command:
pip install -r requirements.txt
Experiments mentioned in the paper use the following datasets:
PANDA and UltraMNIST dataset processing scripts are included in the utility_codes directory where folds for PANDA and full dataset for UltraMNIST can be generated.
For ImageNet and TCGA(LUAD and LUSC), the dataset can be downloaded from Kaggle (for ImageNet) and (LUAD & LUSC with setup instructions listed here) the splits can be made from the dataset.
- baselines directory contains the code to run baseline experiments mentioned in the paper for PANDA and UltraMNIST
- HAR_1d_example directory contains code to run experiments on the Human Activity Recognition dataset (1-d generalization of PatchGD)
- patch_gd directory contains the code to run experiments using PatchGD algorithm for PANDA and UltraMNIST
- utility_codes directory contains utitlity codes for PANDA and UltraMNIST including creating dataset and folds, calculation of stats, running multiple experiments on multiple gpus etc.
Please cite using the following citation:
@article{gupta2023patch,
title={Patch gradient descent: Training neural networks on very large images},
author={Gupta, Deepak K and Mago, Gowreesh and Chavan, Arnav and Prasad, Dilip K},
journal={arXiv preprint arXiv:2301.13817},
year={2023}
}