Skip to content
This repository has been archived by the owner on Nov 17, 2023. It is now read-only.

Huge performance decrease by quantization #13720

Closed
kice opened this issue Dec 22, 2018 · 9 comments
Closed

Huge performance decrease by quantization #13720

kice opened this issue Dec 22, 2018 · 9 comments
Labels
Operator Performance Quantization Issues/Feature Requests related to Quantization

Comments

@kice
Copy link
Contributor

kice commented Dec 22, 2018

I use the code from PR #13715, and I got a huge performance decrease by doing quantization on my model. I tested on Windows 10 with CUDA 10 and cudnn7 on Titan X (Pascal), using pre-release build from pip mxnet-cu100.

Alought by this issue #10897, it claimed that INT8 quantization can save GPU memory during usage, I got almose 2x more VRAM usage by quantization.

Do we excepted that INT8 quantization is super slow and use more memory on GPU?

And I may assume that UINT8 quantization is not yet supported since the UINT8 quantizated parameters is signed integer.

So, do we have any plan for improving INT8 quantization in the near future?

@marcoabreu
Copy link
Contributor

@DickJC123

@pengzhao-intel
Copy link
Contributor

@kice
Copy link
Contributor Author

kice commented Dec 23, 2018

@pengzhao-intel

Form your link

Performance is expected to decrease with GPU, however the memory footprint of a quantized model is smaller.

But in my case, it double the GPU memory usage. I don't think it can be considered as "smaller".

@pengzhao-intel
Copy link
Contributor

FYI again, #13145 (comment)

@kice
Copy link
Contributor Author

kice commented Dec 24, 2018

I couldn't find any information about why GPU memory usage increased.

@pengzhao-intel
Copy link
Contributor

@ThomasDelteil do you have some data to show the memory changes of INT8 flow?

@reminisce to comment your question.

@vrakesh
Copy link
Contributor

vrakesh commented Dec 25, 2018

@mxnet-label-bot add [Operator, Performance]

@vrakesh
Copy link
Contributor

vrakesh commented Dec 25, 2018

@mxnet-label-bot add [Quantization]

@marcoabreu marcoabreu added the Quantization Issues/Feature Requests related to Quantization label Dec 25, 2018
@pengzhao-intel
Copy link
Contributor

#9552 (comment)

Since there is no plan till now, closing this issue.
Welcome to file PR to fix related issues.

Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
Operator Performance Quantization Issues/Feature Requests related to Quantization
Projects
None yet
Development

No branches or pull requests

4 participants