-
Notifications
You must be signed in to change notification settings - Fork 93
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Nightly bigint conversation performance is different than I reported #2146
Comments
I'm using the gnu backend instead of llvm, but I'd be surprised if that had such a large impact. I'm also using minimal modules, but not clear why that would matter:
@bmcdonald3 do you have some time to see if you can reproduce the better numbers I'm seeing and chase down the difference between that and nightly? |
Sure, can take a look today. I have some speculations, but I will refrain from voicing them until I have proof since I have been fighting with the CS all morning. |
Hmm, ya, I am seeing numbers closer to nightly with both LLVM and C backend, though, the C backend numbers are slightly better. This may be a stupid question, but are you doing CPU specialization? (and do oyu usually when running on CS? I've got my fingers crossed that this is where my problems are coming from) LLVM:
C Backend:
printchplenv outputLLVM env:
C-backend env:
|
No, not doing CPU specialization. Ah.. it turns out I had a patch for parallel deinit applied (for chapel-lang/chapel#15215 / #2088 (comment)) and that was resulting in my better performance. Without that I see performance more in line with you and nightly:
Given how big the performance benefit is from the parallel deinit, we'll definitely want to make sure that gets into our next release. |
Closing, we're tracking this internally. |
Nightly 16-node-cs-hdr bigint conversation performance is ~8 and ~9 GiB/s, but I reported ~66 and ~113 GiB/s in #2140.
That was for a larger problem size, but building as myself with today's master I still better performance for the default problem size:
The text was updated successfully, but these errors were encountered: