Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[luci] Support Transpose in InsertQuantizeOpOnDTypeMismatch #14488

Merged

Conversation

jinevening
Copy link
Contributor

This supports Transpose Op in InsertQuantizeOpOnDTypeMismatch.

ONE-DCO-1.0-Signed-off-by: Hyukjin Jeong hj1.jeong@samsung.com

This supports Transpose Op in InsertQuantizeOpOnDTypeMismatch.

ONE-DCO-1.0-Signed-off-by: Hyukjin Jeong <hj1.jeong@samsung.com>
@jinevening
Copy link
Contributor Author

jinevening commented Dec 20, 2024

To reviewers: This PR is for Llama3.2 quantization.

Comment on lines +146 to +150
if (not is_quantized(input))
return;

if (not is_quantized(node))
return;
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I think these lines follows other implementations. But, just curiosity, can this case really happen?

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I expect that it should not happen for our target model, but I'm not 100% sure. These branches are useful because I don't have to consider whether this case will never happen or not. We can just apply this pass to safe cases.

Copy link
Contributor

@mhs4670go mhs4670go left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM

@jinevening jinevening merged commit 615fa81 into Samsung:master Dec 24, 2024
9 checks passed
@jinevening jinevening deleted the luci/insert_quantize_transpose branch December 24, 2024 02:19
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants