Skip to content

Commit

Permalink
Core ML Has Added Index_Put Support, No Need to Skip Anymore (#2975)
Browse files Browse the repository at this point in the history
Summary:
It was a workaround to skip `aten.index_put` op in Core ML delegation, at the cost of partitioning the Llama model into 13 pieces.

For better performance, we prefer to delegate the whole model to Core ML. Since Core ML has added the [necessary support](apple/coremltools#2190), it is time to revert this workaround

Pull Request resolved: #2975

Reviewed By: kirklandsign

Differential Revision: D56002979

Pulled By: cccclai

fbshipit-source-id: e7a7c8c43706cb57eba3e6f720b3d713bec5065b
  • Loading branch information
yifan_shen3 authored and facebook-github-bot committed Apr 11, 2024
1 parent d761f99 commit 7d4bafc
Showing 1 changed file with 0 additions and 3 deletions.
3 changes: 0 additions & 3 deletions examples/models/llama2/export_llama_lib.py
Original file line number Diff line number Diff line change
Expand Up @@ -605,9 +605,6 @@ def _export_llama(modelname, args) -> str: # noqa: C901
partitioners.append(
# pyre-ignore: Undefined attribute [16]: Module `executorch.backends` has no attribute `apple`
CoreMLPartitioner(
skip_ops_for_coreml_delegation=[
"aten.index_put.default",
],
compile_specs=compile_specs,
)
)
Expand Down

0 comments on commit 7d4bafc

Please sign in to comment.