Skip to content

Commit be9b90c

Browse files
committed
Added info on lowering
1 parent 9d80b62 commit be9b90c

File tree

1 file changed

+8
-1
lines changed

1 file changed

+8
-1
lines changed

docs/source/quantization-overview.md

+8-1
Original file line numberDiff line numberDiff line change
@@ -24,8 +24,15 @@ This scheme represents 4-bit weight quantization with 8-bit dynamic quantization
2424
Imported with ``from torchao.quantization.quant_api import Int8DynActInt4WeightQuantizer``, this class uses a quantization instance constructed with a specified dtype precision and groupsize, to mutate a provided ``nn.Module``.
2525

2626
```
27+
# Source Quant
2728
from torchao.quantization.quant_api import Int8DynActInt4WeightQuantizer
2829
2930
model = Int8DynActInt4WeightQuantizer(precision=torch_dtype, groupsize=group_size).quantize(model)
30-
```
3131
32+
# Export to ExecuTorch
33+
from executorch.exir import to_edge
34+
from torch.export import export
35+
36+
exported_model = export(model, ...)
37+
et_program = to_edge(exported_model, ...).to_executorch(...)
38+
```

0 commit comments

Comments
 (0)