You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I am trying to reproduce result for ReCoRD task by fine-tuning bert-base-uncased model but it takes days for 1 GPU (Tesla V100) because the training set is quite big (~1.13M examples).
Issue by ThangPM
Saturday Jun 27, 2020 at 16:53 GMT
Originally opened as nyu-mll/jiant#1099
Hello,
I am trying to reproduce result for ReCoRD task by fine-tuning bert-base-uncased model but it takes days for 1 GPU (Tesla V100) because the training set is quite big (~1.13M examples).
It takes 10 secs for (340603 - 340556) = 47 steps
I decided to speed up this process by using 8 GPUs (still Tesla V100) and update batch_size from 8 to 128 but it seems to take longer than 1 GPU.
Now it takes around 14 secs for only 15 steps. Am I doing anything wrong or is this an issue?
Any comments would be appreciated.
The text was updated successfully, but these errors were encountered: