Skip to content

Conversation

@mengniwang95
Copy link
Contributor

@mengniwang95 mengniwang95 commented Dec 18, 2025

  • LLMC uses quantize_block API to quantize block, where iters maybe 0, add process for dump_info in _quantize_block
  • LLMC passes decoding layer into quantize_block, which doesn't have config attribute

@yiliu30
Copy link
Contributor

yiliu30 commented Dec 19, 2025

@XuehaoSun @chensuyue Is it possible to cherry‑pick this into 0.9.3, so we can use it when we upgrade AutoRound on the LLMC side?
cc @thuang6

@yiliu30 yiliu30 requested review from n1ck-guo and yiliu30 December 19, 2025 00:33
Copy link
Contributor

@yiliu30 yiliu30 left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

code change LGTM

@mengniwang95
Copy link
Contributor Author

@n1ck-guo please take a look for the template part code change

@chensuyue chensuyue added this to the 0.9.3 milestone Dec 22, 2025
@chensuyue chensuyue merged commit 3c88b3b into main Dec 23, 2025
28 checks passed
@chensuyue chensuyue deleted the mengni/llmc_llama4 branch December 23, 2025 06:16
chensuyue pushed a commit that referenced this pull request Dec 23, 2025
Signed-off-by: Mengni Wang <[email protected]>
(cherry picked from commit 3c88b3b)
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

7 participants