Migrate export_llama to new ao quantize API #8422
Labels
module: examples
Issues related to demos under examples/
module: llm
Issues related to LLM examples and apps, and to the extensions/llm/ code
triaged
This issue has been looked at a team member, and triaged and prioritized into an appropriate module
🚀 The feature, motivation and pitch
Int8DynActInt4WeightQuantizer
for-qmode 8da4w
is no longer being developed by ao and doesn't support bias. Migrate to the newquantize_
api which can take inint8_dynamic_activation_int4_weight
.Alternatives
No response
Additional context
No response
RFC (Optional)
No response
cc @mergennachin @iseeyuan @lucylq @helunwencser @tarun292 @kimishpatel @cccclai
The text was updated successfully, but these errors were encountered: