-
Notifications
You must be signed in to change notification settings - Fork 790
lpai e2e & minimum inference runtime support #16421
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: main
Are you sure you want to change the base?
Conversation
🔗 Helpful Links🧪 See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/executorch/16421
Note: Links to docs will display an error until the docs builds have been completed. ❗ 1 Active SEVsThere are 1 currently active SEVs. If your PR is affected, please view them below: ❌ 9 New Failures, 1 Unrelated FailureAs of commit dadf883 with merge base 47dc1de ( NEW FAILURES - The following jobs have failed:
UNSTABLE - The following job is marked as unstable, possibly due to flakiness on trunk:
This comment was automatically generated by Dr. CI and updates every 15 minutes. |
This PR needs a
|
|
Hi @cccclai, this PoC is for LPAI backend & minimum inference runtime on hexagon scalar core. Currently only aDSP is supported with 2 models given from RL team are verified. We'll also work on general solution in parallel (e.g. support cDSP(HTP), backend awareness quantizer, etc.). |
|
tagging @mohankumarkumar to review |
61b71d5 to
dadf883
Compare
Test plan
Tested with QNN version 2.41.0.251128
[Initialization]
[aarch64-android]
cd path/to/executorch mkdir lpai_artifacts python backends/qualcomm/tests/test_qnn_delegate.py TestQNNQuantizedOperator.test_qnn_backend_lpai -b build-android -s f3c0531 -m SM8850 --backend lpai -a lpai_artifacts[hexagon direct]
cd path/to/executorch backends/qualcomm/scripts/lpai_utils.sh --direct --lpai v6 --hexagon v81 --workspace /data/local/tmp/lpai_direct --artifact lpai_artifacts/ --serial f3c0531