-
Notifications
You must be signed in to change notification settings - Fork 2.9k
[GPU] Micro sdpa #24656
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[GPU] Micro sdpa #24656
Conversation
0f1fafb to
343d560
Compare
51b63e8 to
511ad70
Compare
5dbb9c7 to
d149dd4
Compare
5cc640c to
6c596a9
Compare
6c596a9 to
80b247c
Compare
979a0d4 to
9525423
Compare
| return false; | ||
|
|
||
| // For platforms with DPAS support we don't have any shape-based limitations | ||
| if (device_info.supports_immad && cldnn::query_microkernels_supported(m_context->get_engine(), config)) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
According to kernel selector, micro SDPA kernel supports only f16 data type and bfyx 4D format - should we relax these limitations for kernel, or restrict such cases here? Also, dynamic num_heads dimension is not supported and then sdpa_ref kernel will be used
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Adjusted callback to decompose those cases
9525423 to
452e837
Compare
sshlyapn
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Looks good to me
|
|
||
| #ifdef ENABLE_ONEDNN_FOR_GPU | ||
|
|
||
| #pragma once |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
nit: could be moved to the top of the file probably
Details:
Tickets: