Skip to content

Commit f911567

Browse files
authored
Instruct users to run llama for qnn to the active repro
Differential Revision: D73125467 Pull Request resolved: #10231
1 parent 075a619 commit f911567

File tree

1 file changed

+4
-0
lines changed

1 file changed

+4
-0
lines changed

examples/models/llama/export_llama_lib.py

+4
Original file line numberDiff line numberDiff line change
@@ -816,6 +816,10 @@ def _to_edge_and_lower_llama( # noqa: C901
816816
modelname = f"coreml_{modelname}"
817817

818818
if args.qnn:
819+
logging.warning(
820+
"The model definition in current repro is not performant, please refer to the instruction"
821+
" in https://github.com/pytorch/executorch/tree/main/examples/qualcomm/oss_scripts/llama/README.md for better performance."
822+
)
819823
from executorch.extension.llm.custom_ops import model_sharding
820824

821825
partitioners.append(

0 commit comments

Comments
 (0)