Skip to content

Commit 2c6b191

Browse files
Merge branch 'perplexity-pre-submit' of https://github.com/nod-ai/shark-ai into perplexity-pre-submit
2 parents 1335166 + 71a520d commit 2c6b191

File tree

3 files changed

+16
-9
lines changed

3 files changed

+16
-9
lines changed

.github/workflows/ci-llama-large-tests.yaml

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -76,14 +76,14 @@ jobs:
7676
iree-base-runtime
7777
7878
- name: Run llama tests
79-
run: pytest sharktank/tests/models/llama/benchmark_amdgpu_test.py -v -s --run-nightly-llama-tests --iree-hip-target=gfx942 --html=out/index.html
79+
run: pytest sharktank/tests/models/llama/benchmark_amdgpu_test.py -v -s --run-nightly-llama-tests --iree-hip-target=gfx942 --html=out/llm/llama/benchmark/index.html
8080

8181
- name: Deploy to GitHub Pages
8282
uses: peaceiris/actions-gh-pages@4f9cc6602d3f66b9c108549d475ec49e8ef4d45e # v4.0.0
8383
with:
8484
github_token: ${{ secrets.SHARK_PLATFORM_GH_TOKEN }}
85-
publish_dir: ./out/llm/llama/benchmarks
86-
destination_dir: ./llm/llama/benchmarks
85+
publish_dir: ./out/llm/llama/benchmark
86+
destination_dir: ./llm/llama/benchmark
8787
keep_files: true
8888

8989
- name: Upload llama executable files

docs/developer_guide.md

Lines changed: 4 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -15,15 +15,17 @@ sudo apt update && sudo apt install -y clang lld
1515

1616
Install:
1717

18-
```
19-
python-is-python3 python3-venv python3-dev
18+
```bash
19+
sudo apt install python-is-python3 python3-venv python3-dev
2020
```
2121

2222
<details>
2323

2424
<summary> Or, alternatively, use `pyenv` to manage a separate python installation for more control over its version: </summary>
2525

2626

27+
The following instructions are taken from pyenv's guide here: https://github.com/pyenv/pyenv?tab=readme-ov-file#a-getting-pyenv
28+
2729
First, install pyenv and its dependencies.
2830

2931
```bash

sharktank/tests/models/llama/benchmark_amdgpu_test.py

Lines changed: 9 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -197,7 +197,6 @@ def testBenchmark8B_f16_Decomposed(self):
197197
)
198198

199199
@skipif_run_quick_llama_test
200-
@pytest.mark.xfail(reason="Compile Error", strict=True, raises=IreeCompileException)
201200
def testBenchmark8B_f16_Non_Decomposed_Prefill(self):
202201
output_file_name = self.dir_path_8b / "f16_torch_prefill"
203202
output_mlir = self.llama8b_f16_torch_sdpa_artifacts.create_file(
@@ -780,7 +779,9 @@ def testBenchmark405B_f16_TP8_Decomposed(self):
780779
cwd=self.repo_root,
781780
)
782781

783-
@pytest.mark.xfail(reason="Compile Error", strict=True, raises=IreeCompileException)
782+
@pytest.mark.xfail(
783+
reason="Benchmarking Error", strict=True, raises=IreeBenchmarkException
784+
)
784785
def testBenchmark405B_f16_TP8_Non_Decomposed(self):
785786
output_file_name = self.dir_path_405b / "f16_torch"
786787
output_mlir = self.llama405b_f16_torch_sdpa_artifacts.create_file(
@@ -828,7 +829,9 @@ def testBenchmark405B_f16_TP8_Non_Decomposed(self):
828829
cwd=self.repo_root,
829830
)
830831

831-
@pytest.mark.xfail(reason="Compile Error", strict=True, raises=IreeCompileException)
832+
@pytest.mark.xfail(
833+
reason="KeyError in theta.py", strict=True, raises=ExportMlirException
834+
)
832835
def testBenchmark405B_fp8_TP8_Decomposed(self):
833836
output_file_name = self.dir_path_405b / "fp8_decomposed"
834837
output_mlir = self.llama405b_fp8_decomposed_artifacts.create_file(
@@ -874,7 +877,9 @@ def testBenchmark405B_fp8_TP8_Decomposed(self):
874877
cwd=self.repo_root,
875878
)
876879

877-
@pytest.mark.xfail(reason="Compile Error", strict=True, raises=IreeCompileException)
880+
@pytest.mark.xfail(
881+
reason="KeyError in theta.py", strict=True, raises=ExportMlirException
882+
)
878883
def testBenchmark405B_fp8_TP8_Non_Decomposed(self):
879884
output_file_name = self.dir_path_405b / "fp8_torch"
880885
output_mlir = self.llama405b_fp8_torch_sdpa_artifacts.create_file(

0 commit comments

Comments
 (0)