Skip to content

shards llama3 model fail #7439

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Closed
AndreaChiChengdu opened this issue Dec 26, 2024 · 1 comment
Closed

shards llama3 model fail #7439

AndreaChiChengdu opened this issue Dec 26, 2024 · 1 comment

Comments

@AndreaChiChengdu
Copy link

🐛 Describe the bug

When I was using the Qualcomm-quantized LLaMa 3 8B model, I found that if I used the sharding number, corresponding to lines 779 and 797 in the file example/model/export_llama_lib.py, I needed to import 'canonicalize_program' from 'executorch.backends.qualcomm.utils.utils'. However, there is no module named 'canonicalize_program' in the corresponding Qualcomm backend utils.py. Where could the issue be?
https://github.com/pytorch/executorch/blob/main/backends/qualcomm/utils/utils.py
thanks and Merry Christmas!

Versions

v0.5 main branch

@AndreaChiChengdu
Copy link
Author

I found that this bug has been discussed in #6955 ,but not fixed

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant