Skip to content

Commit

Permalink
Fix publish dir of llama tests, update xpass 8b test and 405b fp8 tes…
Browse files Browse the repository at this point in the history
…t failures (#580)

Fixes publish dir of llama tests to `out/llm/llama/benchmarks`, update
xpass 8b test (`testBenchmark8B_f16_Non_Decomposed_Prefill`) and 405b
fp8 test failures (`testBenchmark405B_fp8_TP8_Decomposed` and
`testBenchmark405B_fp8_TP8_Non_Decomposed`).

---------

Signed-off-by: aviator19941 <[email protected]>
Co-authored-by: saienduri <[email protected]>
  • Loading branch information
aviator19941 and saienduri authored Nov 22, 2024
1 parent a6cb442 commit fd15aa5
Show file tree
Hide file tree
Showing 2 changed files with 13 additions and 7 deletions.
7 changes: 4 additions & 3 deletions .github/workflows/ci-llama-large-tests.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -8,6 +8,7 @@ name: Llama Benchmarking Tests

on:
workflow_dispatch:
pull_request:
schedule:
# Weekdays at 4:00 AM UTC = 9:00 PM PST.
- cron: "0 4 * * 1-5"
Expand Down Expand Up @@ -76,14 +77,14 @@ jobs:
iree-base-runtime
- name: Run llama tests
run: pytest sharktank/tests/models/llama/benchmark_amdgpu_test.py -v -s --run-nightly-llama-tests --iree-hip-target=gfx942 --html=out/index.html
run: pytest sharktank/tests/models/llama/benchmark_amdgpu_test.py -v -s --run-nightly-llama-tests --iree-hip-target=gfx942 --html=out/llm/llama/benchmark/index.html

- name: Deploy to GitHub Pages
uses: peaceiris/actions-gh-pages@4f9cc6602d3f66b9c108549d475ec49e8ef4d45e # v4.0.0
with:
github_token: ${{ secrets.SHARK_PLATFORM_GH_TOKEN }}
publish_dir: ./out/llm/llama/benchmarks
destination_dir: ./llm/llama/benchmarks
publish_dir: ./out/llm/llama/benchmark
destination_dir: ./llm/llama/benchmark
keep_files: true

- name: Upload llama executable files
Expand Down
13 changes: 9 additions & 4 deletions sharktank/tests/models/llama/benchmark_amdgpu_test.py
Original file line number Diff line number Diff line change
Expand Up @@ -197,7 +197,6 @@ def testBenchmark8B_f16_Decomposed(self):
)

@skipif_run_quick_llama_test
@pytest.mark.xfail(reason="Compile Error", strict=True, raises=IreeCompileException)
def testBenchmark8B_f16_Non_Decomposed_Prefill(self):
output_file_name = self.dir_path_8b / "f16_torch_prefill"
output_mlir = self.llama8b_f16_torch_sdpa_artifacts.create_file(
Expand Down Expand Up @@ -780,7 +779,9 @@ def testBenchmark405B_f16_TP8_Decomposed(self):
cwd=self.repo_root,
)

@pytest.mark.xfail(reason="Compile Error", strict=True, raises=IreeCompileException)
@pytest.mark.xfail(
reason="Benchmarking Error", strict=True, raises=IreeBenchmarkException
)
def testBenchmark405B_f16_TP8_Non_Decomposed(self):
output_file_name = self.dir_path_405b / "f16_torch"
output_mlir = self.llama405b_f16_torch_sdpa_artifacts.create_file(
Expand Down Expand Up @@ -828,7 +829,9 @@ def testBenchmark405B_f16_TP8_Non_Decomposed(self):
cwd=self.repo_root,
)

@pytest.mark.xfail(reason="Compile Error", strict=True, raises=IreeCompileException)
@pytest.mark.xfail(
reason="KeyError in theta.py", strict=True, raises=ExportMlirException
)
def testBenchmark405B_fp8_TP8_Decomposed(self):
output_file_name = self.dir_path_405b / "fp8_decomposed"
output_mlir = self.llama405b_fp8_decomposed_artifacts.create_file(
Expand Down Expand Up @@ -874,7 +877,9 @@ def testBenchmark405B_fp8_TP8_Decomposed(self):
cwd=self.repo_root,
)

@pytest.mark.xfail(reason="Compile Error", strict=True, raises=IreeCompileException)
@pytest.mark.xfail(
reason="KeyError in theta.py", strict=True, raises=ExportMlirException
)
def testBenchmark405B_fp8_TP8_Non_Decomposed(self):
output_file_name = self.dir_path_405b / "fp8_torch"
output_mlir = self.llama405b_fp8_torch_sdpa_artifacts.create_file(
Expand Down

0 comments on commit fd15aa5

Please sign in to comment.