Skip to content

Commit

Permalink
Merge pull request #19 from anandhu-eng/cm_readme_inference_update
Browse files Browse the repository at this point in the history
updated neural magic documentation
  • Loading branch information
arjunsuresh authored Sep 3, 2024
2 parents 5c73d16 + d078534 commit edbaf90
Showing 1 changed file with 2 additions and 0 deletions.
2 changes: 2 additions & 0 deletions main.py
Original file line number Diff line number Diff line change
Expand Up @@ -383,6 +383,7 @@ def mlperf_inference_run_command(spaces, model, implementation, framework, categ
elif implementation == "neuralmagic":
docker_cmd_suffix += f" \\\n{pre_space} --api_server=http://localhost:8000"
docker_cmd_suffix += f" \\\n{pre_space} --vllm_model_name=nm-testing/Llama-2-70b-chat-hf-FP8"
docker_cmd_suffix += f" \\\n{pre_space} --adr.mlperf-implementation.tags=_repo.https://github.com/neuralmagic/inference,_branch.vllm"

if "dlrm-v2" in model.lower() and implementation == "nvidia":
docker_cmd_suffix += f" \\\n{pre_space} --criteo_day23_raw_data_path=<PATH_TO_CRITEO_DAY23_RAW_DATA>"
Expand Down Expand Up @@ -415,6 +416,7 @@ def mlperf_inference_run_command(spaces, model, implementation, framework, categ
elif implementation == "neuralmagic":
cmd_suffix += f" \\\n{pre_space} --api_server=http://localhost:8000"
cmd_suffix += f" \\\n{pre_space} --vllm_model_name=nm-testing/Llama-2-70b-chat-hf-FP8"
cmd_suffix += f" \\\n{pre_space} --adr.mlperf-implementation.tags=_repo.https://github.com/neuralmagic/inference,_branch.vllm"

if "dlrm-v2" in model and implementation == "nvidia":
cmd_suffix += f" \\\n{pre_space} --criteo_day23_raw_data_path=<PATH_TO_CRITEO_DAY23_RAW_DATA>"
Expand Down

0 comments on commit edbaf90

Please sign in to comment.