You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I think caching the data returned from prepare_data() to disk refer to saving these huge extracted pixel representations to files in some specified format (e.g. .npy file) on your disk. Later on you can load these files and read the data out for the training of MLPs.
But I have another issue here, which is the extracted pixel representations are so huge that before the prepare_data() function return the data, the process I'm running the script on will crash. For example, the process will be killed when it tries to run the following two lines to faltten the pixel representations.
X = X.transpose(1,0,2,3).reshape(d,-1).transpose(1,0)
y = y.flatten()
My solution is to write another prepare_data() function, which processes one image instead of all labelled training images at one time. Then during the training of the pixel classifier, at each epoch I will create a dataloader for one image and iterate through all training images. But the results I get this way are not quite the same as the results reported in the paper. Wondering if anyone has encountered the same problem or knows how to fix it?
Hi @arieling,
Great job with this repository, it is awesome.
On the ReadMe page you state that:
"One can cache the data returned from prepare_data function to disk but it will increase trianing time due to I/O burden."
How would I implement this?
Thank You!
The text was updated successfully, but these errors were encountered: