You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
It would be quite helpful to know approx memory requirements and training times for different configs. I could not find such information in the Mogrifier paper or anywhere in this repo.
I've started training the Mogrifier with one of the provided Wikitext-2 configs. Each "turn" took 4-5 minutes on a V100 GPU, so taking 1000 turns as specified in the config would take ~3 days. It filled almost all of the V100's memory, so 16GB is probably required for the provided configs? Or would 12GB cut it? (8GB did not, I've tried).
Are the figures above typical for all the datasets and provided configs, or do they differ significantly?
(or perhaps my setup was wrong and performance should be higher?)
Finally, is there any chance of the trained models getting published?
The text was updated successfully, but these errors were encountered:
On PTB, training took about 15 hours for the base LSTM. Three days for the Mogrifier on Wikitext-2 sounds about right.
Whether 12GB is enough depends on the model size. All I can say without trying is that 16GB is enough. If you have a gpu with 8GB, you can still probably train the model by adding a line like this to train.sh:
accum_batch_size=32
where 32 must be a divisor of the real batch size (which is 64, if I recall correctly).
It would be quite helpful to know approx memory requirements and training times for different configs. I could not find such information in the Mogrifier paper or anywhere in this repo.
I've started training the Mogrifier with one of the provided Wikitext-2 configs. Each "turn" took 4-5 minutes on a V100 GPU, so taking 1000 turns as specified in the config would take ~3 days. It filled almost all of the V100's memory, so 16GB is probably required for the provided configs? Or would 12GB cut it? (8GB did not, I've tried).
Are the figures above typical for all the datasets and provided configs, or do they differ significantly?
(or perhaps my setup was wrong and performance should be higher?)
Finally, is there any chance of the trained models getting published?
The text was updated successfully, but these errors were encountered: