-
Notifications
You must be signed in to change notification settings - Fork 3.5k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[BYOC] Enable bfloat16 in DNNL BYOC #11111
Conversation
be5f228
to
e2fcfd1
Compare
41e9720
to
4caede4
Compare
Relative modifications since the original PR:
All the CI checks have passed now, what a long journey. |
@@ -37,6 +37,8 @@ | |||
ids=["compile", "run"], | |||
) | |||
|
|||
bf16_supported = "avx512" in open("/proc/cpuinfo", "r").read() |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Probably need more precise detection, but ok.
cc @AndrewZhaoLuo this is cool (the first e2e run of AMP + bf16!!) |
Thanks a lot. |
for simplicity in DNNL run-time; we need to remove TR, and maybe move to apache#11111
Enable bfloat16 in DNNL BYOC following the path:
Main work include:
With those improvements, a float32 graph could be converted to bfloat16 through AMP, and then be lowered by native codegen or consumed by oneDNN and finally inference in bfloat16 mode now.