Skip to content

Fix quantized embedding export logic #3095

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Closed
wants to merge 4 commits into from
Closed

Conversation

larryliu0820
Copy link
Contributor

@larryliu0820 larryliu0820 commented Apr 17, 2024

Add patches to make 4bit quantized embedding work for export. Fixed:

  • Schema mismatch between functional embedding_4bit and out variant
  • Set packed=True for 4bit quantization

Copy link

pytorch-bot bot commented Apr 17, 2024

🔗 Helpful Links

🧪 See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/executorch/3095

Note: Links to docs will display an error until the docs builds have been completed.

✅ No Failures

As of commit a28e73b with merge base 06beace (image):
💚 Looks good so far! There are no failures yet. 💚

This comment was automatically generated by Dr. CI and updates every 15 minutes.

@facebook-github-bot facebook-github-bot added the CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed. label Apr 17, 2024
@larryliu0820 larryliu0820 force-pushed the larryliu0820-patch branch 2 times, most recently from 7977cc2 to 1dc7a5c Compare April 18, 2024 16:39
kimishpatel and others added 3 commits April 18, 2024 20:42
Summary: This diff adds support for multi query attention for sdpa with kv cache

Reviewed By: iseeyuan

Differential Revision: D56212419
Summary:

4b embedding quantizer

Reviewed By: larryliu0820

Differential Revision: D56229021
Summary:

Test Plan:

Reviewers:

Subscribers:

Tasks:

Tags:
@larryliu0820 larryliu0820 force-pushed the larryliu0820-patch branch 2 times, most recently from d504a61 to 4b7050d Compare April 19, 2024 05:25
@larryliu0820 larryliu0820 changed the title Larryliu0820 patch Fix quantized embedding export logic Apr 19, 2024
@facebook-github-bot
Copy link
Contributor

@larryliu0820 has imported this pull request. If you are a Meta employee, you can view this diff on Phabricator.

Summary:

Test Plan:

Reviewers:

Subscribers:

Tasks:

Tags:
@facebook-github-bot
Copy link
Contributor

@larryliu0820 has imported this pull request. If you are a Meta employee, you can view this diff on Phabricator.

Copy link

@mikekgfb mikekgfb left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thank you!

@facebook-github-bot
Copy link
Contributor

@larryliu0820 merged this pull request in 2c467dd.

facebook-github-bot pushed a commit that referenced this pull request Apr 19, 2024
Summary: In  #3095 there's an issue with the embedding_4bit schema which causes mismatch between functional and out variant. P1217884556

Differential Revision: D56357762
larryliu0820 added a commit that referenced this pull request Apr 19, 2024
Summary:

In  #3095 there's an issue with the embedding_4bit schema which causes mismatch between functional and out variant. P1217884556

Differential Revision: D56357762
facebook-github-bot pushed a commit that referenced this pull request Apr 19, 2024
Summary:
Pull Request resolved: #3151

In  #3095 there's an issue with the embedding_4bit schema which causes mismatch between functional and out variant. P1217884556

Reviewed By: mergennachin, digantdesai

Differential Revision: D56357762

fbshipit-source-id: e8a1c249a02bfb4db295a1a933a8b3054e11099a
@mergennachin mergennachin mentioned this pull request Apr 26, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed. Merged
Projects
None yet
Development

Successfully merging this pull request may close these issues.

4 participants