Skip to content

Commit a1acef7

Browse files
cccclaifacebook-github-bot
authored andcommitted
Instruct users to run llama for qnn to the active repro (#10231)
Summary: Many users are trying to export llama with this flow https://github.com/pytorch/executorch/tree/main/examples/models/llama and end up with non performant, or different isssues, like #10226. Instruct users to use the qcom version Reviewed By: kirklandsign Differential Revision: D73125467
1 parent 2ff9abd commit a1acef7

File tree

1 file changed

+4
-0
lines changed

1 file changed

+4
-0
lines changed

examples/models/llama/export_llama_lib.py

+4
Original file line numberDiff line numberDiff line change
@@ -816,6 +816,10 @@ def _to_edge_and_lower_llama( # noqa: C901
816816
modelname = f"coreml_{modelname}"
817817

818818
if args.qnn:
819+
logging.warning(
820+
f"The model definition in current repro is not performant, please refer to the instruction"
821+
" in https://github.com/pytorch/executorch/tree/main/examples/qualcomm/oss_scripts/llama/README.md for better performance."
822+
)
819823
from executorch.extension.llm.custom_ops import model_sharding
820824

821825
partitioners.append(

0 commit comments

Comments
 (0)