-
Notifications
You must be signed in to change notification settings - Fork 259
Issue #998 Fix #1074
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Issue #998 Fix #1074
Conversation
🔗 Helpful Links🧪 See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/ao/1074
Note: Links to docs will display an error until the docs builds have been completed. ❌ 4 New Failures, 1 Unrelated FailureAs of commit 19b014c with merge base e7b33bc ( NEW FAILURES - The following jobs have failed:
BROKEN TRUNK - The following job failed but were present on the merge base:👉 Rebase onto the `viable/strict` branch to avoid these failures
This comment was automatically generated by Dr. CI and updates every 15 minutes. |
Hi @DevyRuxpin! Thank you for your pull request and welcome to our community. Action RequiredIn order to merge any pull request (code, docs, etc.), we require contributors to sign our Contributor License Agreement, and we don't seem to have one on file for you. ProcessIn order for us to review and merge your suggested changes, please sign at https://code.facebook.com/cla. If you are contributing on behalf of someone else (eg your employer), the individual CLA may not be sufficient and your employer may need to sign the corporate CLA. Once the CLA is signed, our tooling will perform checks and validations. Afterwards, the pull request will be tagged with If you have received this in error or have any questions, please contact us at [email protected]. Thanks! |
Thank you for signing our Contributor License Agreement. We can now accept your code for this (and any) Meta Open Source project. Thanks! |
Might need to also add some tests so CI signal is helpful and just to be clear are we only supporting bf16 moving forward or both fp16 and bf16? |
I think we shouldn't remove FP16 support, since it is the original dtype that Quant-LLM support. Not sure how we can structure it nicely (i.e. support both FP16 and BF16), but you can come up with a way first. Ideally it shouldn't have intrusive changes to the original code, so if there are upstream updates, it's easier for us to pull the new updates in. Maybe some kind of templates or just have a separate folder for BF16? |
I will work on this further. Thank you for feedback, i will draw things back some on the changes and incorporate your suggestions. i will work on this tonight/tomorrow. |
@gau-nernst @msaroufim I have already been working on a templated solution that supports both FP16 and BF16, and it is almost finished (see the diff here). I think this would be a better solution than removing FP16 all together. I'm just smoothing out some final bugs and it should be as good as ready. I didn't make a PR yet because I wanted to make sure that everything worked properly first. In hindsight, it might have been better to already create a PR to avoid the duplicate work done here (although I did say I was working on this feature in the original issue). |
Possible solution to issue #998