Partition Mutable Buffer as Core ML State#5165
Conversation
🔗 Helpful Links🧪 See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/executorch/5165
Note: Links to docs will display an error until the docs builds have been completed. ❌ 1 New Failure, 1 Unrelated FailureAs of commit 0bfa422 with merge base 13da62b ( NEW FAILURE - The following job has failed:
FLAKY - The following job failed but was likely due to flakiness present on trunk:
This comment was automatically generated by Dr. CI and updates every 15 minutes. |
|
@cccclai has imported this pull request. If you are a Meta employee, you can view this diff on Phabricator. |
|
Thanks! lintrunner seems failing. Could you help fixing it? |
Fixed |
|
The coreml end to end CI breaks in this PR, is it expected? Does it mean it needs @cymbalrush pr for the CI job? |
|
Llama runner runs well on my local. Is CI machine MacOS 14? State is a new feature in MacOS 15, so failure on older MacOS is expected. I can make the stateful llama no longer the default option. |
3dc37e6 to
2d9728f
Compare
Yeah CI machine is still MacOS 14. Looks like it's passing now. Thanks! |
|
@cccclai has imported this pull request. If you are a Meta employee, you can view this diff on Phabricator. |
cccclai
left a comment
There was a problem hiding this comment.
Looks great. Thanks! Just some minor comments
|
Looks like there are still lint error: |
…eful llama until CI machine upgraded to MacOS 15
2d9728f to
1735a20
Compare
|
Lint error fixed, comments addressed |
|
@cccclai has imported this pull request. If you are a Meta employee, you can view this diff on Phabricator. |
|
The test failing in |
…, that does not support numpy 2.0
|
Fixed, it's due to coremltools 8.0b2 starts to support numpy 2.0, but the mobilebert test is using an old transformers that requires numpy 1.x Temporarily added a numpy downgrade in coreml install_requirements.sh. Will remove it once executorch migrates to numpy 2.0 |
|
@cccclai has imported this pull request. If you are a Meta employee, you can view this diff on Phabricator. |
#4830 opens the door toward delegate mutable buffer. Now in this PR, we tag the mutable buffers in Core ML partitioner, to delegate them as Core ML state
With #5143, we are able to run the stateful Core ML delegate