-
Notifications
You must be signed in to change notification settings - Fork 3.5k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
TVM v0.3 Release Note #854
Comments
Tianqi, 1-bit looks a cool feature. However, in our internal experiments, 1-bit weight(with 4-bit activation) still has non-negligible accuracy degradation(the best result still has around 3 percentage gap). So may I ask why we want to invest resource on this low-bit feature? Internally we are also leveraging TVM for boosting our productivity and already submit small patch to the community. We would like to align our engineering efforts with the community. Model compression is right now on our radars, and we would like to leverage TVM for fully exploiting the potential of model compressions. Thanks. |
We want to keep doors open for different kinds of optimizations, low bits(not necessarily 1bit) is among one of them. We are also thinking about more general low-bits(e.g. 2, 3 or 4 bits) operations. they allow interesting tradeoffs to be made(not necessarily).This direction can also be of particular interest to folks who works on hardware optimizations. |
8bit inference is the most important thing 😄 |
Support for tensor comprehension would be great: optimized c code generation with openmp pragmas. Support for ONNX interoperability |
@danilopau Just so that you know that many things you mentioned are already supported. Include high performance optimized cpu code generation, ONNX interoperability, 32 FP to 16 FP, part of binary computation |
custom hardware backend example .... please look https://blog.st.com/orlando-neural-network-iot/. |
Thanks for everyone who have pushed to last release cycle in the past three month. I would like to propose release of v0.3 on May 21th. As usual, the current checklist will go into the release note, and we will move the unfinished ones into the roadmap of next release cycle. Main contributors to the past release, see complete list here : Code Reviewers
TOPI:
Compiler:
|
Please reply to see if there is something we would like to merge in before tag v0.3 |
v0.3 is now tagged, next cycle roadmap issue is available at #1170 |
v0.3 is now tagged, next cycle roadmap issue is available at #1170
Release Note
This release features numerous improvements in TOPI and backends. We make the first step toward object detection support in TOPI, featuring operators necessary for YOLO and SSDs. The topi now supports numpy-style API and operator overloading. RPC is significantly improved to support resource allocation and using a pool of devices. We are adding two new backends: WebGL for running GPUs on the browser, and Vulkan for running on next-generation graphics API.
Change List
Main Contributors
See complete list here. Thanks to all the contributors to contribute to this release.
Code Reviewers
TOPI:
Compiler:
The text was updated successfully, but these errors were encountered: