Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Quantized distillgpt2 accuracy issue when quant params are inputs #3612

Open
shivadbhavsar opened this issue Nov 11, 2024 · 0 comments
Open
Assignees
Labels
bug Something isn't working

Comments

@shivadbhavsar
Copy link
Contributor

Issue originating from pytorch-quantized model where the scales and zero points are passed as inputs rather than being embedded in the model as literals.

MXR for small problematic block can be found in nas at: /migraphx/models/torch_exports/distilgpt2_block_torch.mxr

to reproduce use:
migraphx-driver verify distilgpt2_block_torch.mxr --fill1 arg0_1 --fill0 arg4_1 --fill0 arg10_1 --fill0 arg12_1 --fill0 arg14_1

This verification only fails after #3362

@shivadbhavsar shivadbhavsar added the bug Something isn't working label Nov 11, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working
Projects
None yet
Development

No branches or pull requests

2 participants