We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
1 parent 075a619 commit f911567Copy full SHA for f911567
examples/models/llama/export_llama_lib.py
@@ -816,6 +816,10 @@ def _to_edge_and_lower_llama( # noqa: C901
816
modelname = f"coreml_{modelname}"
817
818
if args.qnn:
819
+ logging.warning(
820
+ "The model definition in current repro is not performant, please refer to the instruction"
821
+ " in https://github.com/pytorch/executorch/tree/main/examples/qualcomm/oss_scripts/llama/README.md for better performance."
822
+ )
823
from executorch.extension.llm.custom_ops import model_sharding
824
825
partitioners.append(
0 commit comments