Qualcomm AI Engine Direct - Delegate mutable buffer and fix the mutable buffer issue#11782
Conversation
…le buffer issue Summary: - Add a parameter to support mutable buffer delegation in QNN Backend - Set the same memory address for I/O of mutable buffer at runtime - Avoid annotating the input node because mutable buffers will be folded during the convert_pt2e process. - Deprecated use_legacy_export in executorch llama
🔗 Helpful Links🧪 See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/executorch/11782
Note: Links to docs will display an error until the docs builds have been completed. ✅ You can merge normally! (1 Unrelated Failure)As of commit 5fde193 with merge base 44d2643 ( BROKEN TRUNK - The following job failed but were present on the merge base:👉 Rebase onto the `viable/strict` branch to avoid these failures
This comment was automatically generated by Dr. CI and updates every 15 minutes. |
This PR needs a
|
Is the input node still folded after we land pytorch/ao#2345? |
Yes, unless we apply run_decomposition after export. I think we can wait until run_decomposition becomes a pass and doesn't require re-tracing. After that we can change it back to annotate mutable buffer. What do you think? |
|
BTW, in previous, we have submitted a PR to deprecated convert_bmm_to_matmul pass. It will result in multiple partitions for Meta's llama due to not using |
|
@cccclai has imported this pull request. If you are a Meta employee, you can view this diff on Phabricator. |
|
I ran into this error for some internal use cases. Can we turn if off in this PR, and I will have another PR to remove this arg? |
|
@cccclai has imported this pull request. If you are a Meta employee, you can view this diff on Phabricator. |
…le buffer issue (pytorch#11782) Summary: - Add a parameter to support mutable buffer delegation in QNN Backend - Set the same memory address for I/O of mutable buffer at runtime - Ref: pytorch#6727 - Avoid annotating the input node because mutable buffers will be folded during the convert_pt2e process. - Deprecated use_legacy_export in executorch llama cc @cccclai @winskuo-quic @cbilgin
Summary: As title, try to see if we can get rid of the legacy export. It should be fixed with pytorch#11782 Differential Revision: D77761473
Summary:
cc @cccclai @winskuo-quic @cbilgin