-
Notifications
You must be signed in to change notification settings - Fork 617
Fix KleidiAI FP16 #3769
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Fix KleidiAI FP16 #3769
Conversation
✅ Deploy Preview for pytorch-fbgemm-docs ready!
To edit notification comments on pull requests, go to your Netlify site configuration. |
This pull request was exported from Phabricator. Differential Revision: D70606808 |
This pull request was exported from Phabricator. Differential Revision: D70606808 |
3cccf41
to
426d983
Compare
Summary: X-link: facebookresearch/FBGEMM#849 FP32Test was failing when KleidiAI was enabled. It turns out FPCommon.h and PackMatrixB.h were not acconditioned to handle using kleidAI for FP16 and not for FP32 Differential Revision: D70606808
This pull request was exported from Phabricator. Differential Revision: D70606808 |
Summary: Pull Request resolved: pytorch#3769 X-link: facebookresearch/FBGEMM#849 FP32Test was failing when KleidiAI was enabled. It turns out FPCommon.h and PackMatrixB.h were not acconditioned to handle using kleidAI for FP16 and not for FP32 Differential Revision: D70606808
9df54cc
to
bf5b9aa
Compare
Summary: X-link: facebookresearch/FBGEMM#849 FP32Test was failing when KleidiAI was enabled. It turns out FPCommon.h and PackMatrixB.h were not acconditioned to handle using kleidAI for FP16 and not for FP32 Differential Revision: D70606808
This pull request was exported from Phabricator. Differential Revision: D70606808 |
Summary: Pull Request resolved: pytorch#3769 X-link: facebookresearch/FBGEMM#849 FP32Test was failing when KleidiAI was enabled. It turns out FPCommon.h and PackMatrixB.h were not acconditioned to handle using kleidAI for FP16 and not for FP32 Differential Revision: D70606808
eb22db6
to
2400c83
Compare
Summary: X-link: facebookresearch/FBGEMM#849 FP32Test was failing when KleidiAI was enabled. It turns out FPCommon.h and PackMatrixB.h were not acconditioned to handle using kleidAI for FP16 and not for FP32 PackedGemmMatrixFP16 constructors were moved to a .cc file, compiled with the rest of fbgemm. This ensures the KleidiAI flag is set when compiling such code. Previously, consumers of the library would include FbgemmPackMatrixB.h from their .cpp files. Because the KleidiAI flag was not set when compiling their code, the intended case for the float16 was never run. Additionally, we have ingested a change in KleidiAI's inline assembly, which handles nan beta values as 0. Reviewed By: psaab Differential Revision: D70606808
This pull request was exported from Phabricator. Differential Revision: D70606808 |
2400c83
to
7d3cdf6
Compare
Summary: X-link: facebookresearch/FBGEMM#849 FP32Test was failing when KleidiAI was enabled. It turns out FPCommon.h and PackMatrixB.h were not acconditioned to handle using kleidAI for FP16 and not for FP32 PackedGemmMatrixFP16 constructors were moved to a .cc file, compiled with the rest of fbgemm. This ensures the KleidiAI flag is set when compiling such code. Previously, consumers of the library would include FbgemmPackMatrixB.h from their .cpp files. Because the KleidiAI flag was not set when compiling their code, the intended case for the float16 was never run. Additionally, we have ingested a change in KleidiAI's inline assembly, which handles nan beta values as 0. Reviewed By: psaab Differential Revision: D70606808
This pull request was exported from Phabricator. Differential Revision: D70606808 |
Summary: Pull Request resolved: pytorch#3769 X-link: facebookresearch/FBGEMM#849 FP32Test was failing when KleidiAI was enabled. It turns out FPCommon.h and PackMatrixB.h were not acconditioned to handle using kleidAI for FP16 and not for FP32 PackedGemmMatrixFP16 constructors were moved to a .cc file, compiled with the rest of fbgemm. This ensures the KleidiAI flag is set when compiling such code. Previously, consumers of the library would include FbgemmPackMatrixB.h from their .cpp files. Because the KleidiAI flag was not set when compiling their code, the intended case for the float16 was never run. Additionally, we have ingested a change in KleidiAI's inline assembly, which handles nan beta values as 0. Reviewed By: psaab Differential Revision: D70606808
c2d560a
to
eb3c685
Compare
Summary: X-link: facebookresearch/FBGEMM#849 FP32Test was failing when KleidiAI was enabled. It turns out FPCommon.h and PackMatrixB.h were not acconditioned to handle using kleidAI for FP16 and not for FP32 PackedGemmMatrixFP16 constructors were moved to a .cc file, compiled with the rest of fbgemm. This ensures the KleidiAI flag is set when compiling such code. Previously, consumers of the library would include FbgemmPackMatrixB.h from their .cpp files. Because the KleidiAI flag was not set when compiling their code, the intended case for the float16 was never run. Additionally, we have ingested a change in KleidiAI's inline assembly, which handles nan beta values as 0. Reviewed By: psaab Differential Revision: D70606808
This pull request was exported from Phabricator. Differential Revision: D70606808 |
Summary: Pull Request resolved: pytorch#3769 X-link: facebookresearch/FBGEMM#849 FP32Test was failing when KleidiAI was enabled. It turns out FPCommon.h and PackMatrixB.h were not acconditioned to handle using kleidAI for FP16 and not for FP32 PackedGemmMatrixFP16 constructors were moved to a .cc file, compiled with the rest of fbgemm. This ensures the KleidiAI flag is set when compiling such code. Previously, consumers of the library would include FbgemmPackMatrixB.h from their .cpp files. Because the KleidiAI flag was not set when compiling their code, the intended case for the float16 was never run. Additionally, we have ingested a change in KleidiAI's inline assembly, which handles nan beta values as 0. Reviewed By: psaab Differential Revision: D70606808
eb3c685
to
111bf81
Compare
Summary: X-link: facebookresearch/FBGEMM#849 FP32Test was failing when KleidiAI was enabled. It turns out FPCommon.h and PackMatrixB.h were not acconditioned to handle using kleidAI for FP16 and not for FP32 PackedGemmMatrixFP16 constructors were moved to a .cc file, compiled with the rest of fbgemm. This ensures the KleidiAI flag is set when compiling such code. Previously, consumers of the library would include FbgemmPackMatrixB.h from their .cpp files. Because the KleidiAI flag was not set when compiling their code, the intended case for the float16 was never run. Additionally, we have ingested a change in KleidiAI's inline assembly, which handles nan beta values as 0. Reviewed By: psaab Differential Revision: D70606808
111bf81
to
27d6cbe
Compare
Summary: Pull Request resolved: pytorch#3769 X-link: facebookresearch/FBGEMM#849 FP32Test was failing when KleidiAI was enabled. It turns out FPCommon.h and PackMatrixB.h were not acconditioned to handle using kleidAI for FP16 and not for FP32 PackedGemmMatrixFP16 constructors were moved to a .cc file, compiled with the rest of fbgemm. This ensures the KleidiAI flag is set when compiling such code. Previously, consumers of the library would include FbgemmPackMatrixB.h from their .cpp files. Because the KleidiAI flag was not set when compiling their code, the intended case for the float16 was never run. Additionally, we have ingested a change in KleidiAI's inline assembly, which handles nan beta values as 0. Reviewed By: psaab Differential Revision: D70606808
This pull request was exported from Phabricator. Differential Revision: D70606808 |
27d6cbe
to
fdc8224
Compare
This pull request has been merged in d17faaa. |
Summary: Pull Request resolved: pytorch#3769 X-link: https://github.com/facebookresearch/FBGEMM/pull/849 FP32Test was failing when KleidiAI was enabled. It turns out FPCommon.h and PackMatrixB.h were not acconditioned to handle using kleidAI for FP16 and not for FP32 PackedGemmMatrixFP16 constructors were moved to a .cc file, compiled with the rest of fbgemm. This ensures the KleidiAI flag is set when compiling such code. Previously, consumers of the library would include FbgemmPackMatrixB.h from their .cpp files. Because the KleidiAI flag was not set when compiling their code, the intended case for the float16 was never run. Additionally, we have ingested a change in KleidiAI's inline assembly, which handles nan beta values as 0. Reviewed By: psaab Differential Revision: D70606808 fbshipit-source-id: b5ada2519eee65e826fa5647967a23db30911f23
Summary: X-link: pytorch#3769 Pull Request resolved: facebookresearch/FBGEMM#849 FP32Test was failing when KleidiAI was enabled. It turns out FPCommon.h and PackMatrixB.h were not acconditioned to handle using kleidAI for FP16 and not for FP32 PackedGemmMatrixFP16 constructors were moved to a .cc file, compiled with the rest of fbgemm. This ensures the KleidiAI flag is set when compiling such code. Previously, consumers of the library would include FbgemmPackMatrixB.h from their .cpp files. Because the KleidiAI flag was not set when compiling their code, the intended case for the float16 was never run. Additionally, we have ingested a change in KleidiAI's inline assembly, which handles nan beta values as 0. Reviewed By: psaab Differential Revision: D70606808 fbshipit-source-id: b5ada2519eee65e826fa5647967a23db30911f23
Summary:
FP32Test was failing when KleidiAI was enabled.
It turns out FPCommon.h and PackMatrixB.h were not acconditioned to handle using kleidAI for FP16 and not for FP32
Differential Revision: D70606808