Skip to content

Commit aa3f22c

Browse files
pytorchbotyifan_shen3
and
yifan_shen3
authored
Core ML Has Added Index_Put Support, No Need to Skip Anymore (#2975) (#3157)
Summary: It was a workaround to skip `aten.index_put` op in Core ML delegation, at the cost of partitioning the Llama model into 13 pieces. For better performance, we prefer to delegate the whole model to Core ML. Since Core ML has added the [necessary support](apple/coremltools#2190), it is time to revert this workaround Pull Request resolved: #2975 Reviewed By: kirklandsign Differential Revision: D56002979 Pulled By: cccclai fbshipit-source-id: e7a7c8c43706cb57eba3e6f720b3d713bec5065b (cherry picked from commit 7d4bafc) Co-authored-by: yifan_shen3 <[email protected]>
1 parent 27e1a62 commit aa3f22c

File tree

1 file changed

+0
-3
lines changed

1 file changed

+0
-3
lines changed

examples/models/llama2/export_llama_lib.py

-3
Original file line numberDiff line numberDiff line change
@@ -663,9 +663,6 @@ def _export_llama(modelname, args) -> str: # noqa: C901
663663
partitioners.append(
664664
# pyre-ignore: Undefined attribute [16]: Module `executorch.backends` has no attribute `apple`
665665
CoreMLPartitioner(
666-
skip_ops_for_coreml_delegation=[
667-
"aten.index_put.default",
668-
],
669666
compile_specs=compile_specs,
670667
)
671668
)

0 commit comments

Comments
 (0)