-
Notifications
You must be signed in to change notification settings - Fork 212
related to #347 - Examples And Performance Results #419
Comments
Has this site gone dead already?? |
During conference submission crunch time, yes unfortunately :) but we're now back Reg. benchmarking, we have been making progress in #423 during this week, best is to wait until we land it early next week. I had to resort to some build hacks following the caffe2 source of truth moving to pytorch that I will need to clean up. Reg. your previous questions:
I have only run myself in sbatch mode, it should run in srun mode with minor modifications there is nothing magical about it. The current script uses SLURM_ARRAY_JOB_ID to set a path but you can easily adapt and set the path to whatever.
Essentially yes. We have reduced to 100 by default but essentially yes. If you want to run 1K times pass
Yes if you checkout the branch from #423 you can then just run
GPU only for now, I'm prioritizing CPU starting this week, I would say it will take about 1 month to get things in a decent state.
TC + pytorch is still in an extremely alpha state right now, I haven't had the chance to benchmark it myself yet. For the benchmarks we report perf on, it's C++ only atm. |
Thanks very much for your detailed answers. I'll then try to install with caffe2 and see if I can run the kernels. I assume the overall performance will be better after building from source though... if you are willing to help me that, that would be very greatly appreciated. Hopefully I can contribute in some shape or form... |
Hi,
so I had to slightly modify the autotuner_parallel.sh script, and I still have a few questions:
Cheers, Kevin
The text was updated successfully, but these errors were encountered: