Skip to content

Commit

Permalink
Merge branch 'main' into perplexity-pre-submit
Browse files Browse the repository at this point in the history
  • Loading branch information
archana-ramalingam authored Nov 22, 2024
2 parents 33d25e1 + fd15aa5 commit 453b4d2
Show file tree
Hide file tree
Showing 3 changed files with 17 additions and 9 deletions.
7 changes: 4 additions & 3 deletions .github/workflows/ci-llama-large-tests.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -8,6 +8,7 @@ name: Llama Benchmarking Tests

on:
workflow_dispatch:
pull_request:
schedule:
# Weekdays at 4:00 AM UTC = 9:00 PM PST.
- cron: "0 4 * * 1-5"
Expand Down Expand Up @@ -76,14 +77,14 @@ jobs:
iree-base-runtime
- name: Run llama tests
run: pytest sharktank/tests/models/llama/benchmark_amdgpu_test.py -v -s --run-nightly-llama-tests --iree-hip-target=gfx942 --html=out/index.html
run: pytest sharktank/tests/models/llama/benchmark_amdgpu_test.py -v -s --run-nightly-llama-tests --iree-hip-target=gfx942 --html=out/llm/llama/benchmark/index.html

- name: Deploy to GitHub Pages
uses: peaceiris/actions-gh-pages@4f9cc6602d3f66b9c108549d475ec49e8ef4d45e # v4.0.0
with:
github_token: ${{ secrets.SHARK_PLATFORM_GH_TOKEN }}
publish_dir: ./out/llm/llama/benchmarks
destination_dir: ./llm/llama/benchmarks
publish_dir: ./out/llm/llama/benchmark
destination_dir: ./llm/llama/benchmark
keep_files: true

- name: Upload llama executable files
Expand Down
6 changes: 4 additions & 2 deletions docs/developer_guide.md
Original file line number Diff line number Diff line change
Expand Up @@ -15,15 +15,17 @@ sudo apt update && sudo apt install -y clang lld

Install:

```
python-is-python3 python3-venv python3-dev
```bash
sudo apt install python-is-python3 python3-venv python3-dev
```

<details>

<summary> Or, alternatively, use `pyenv` to manage a separate python installation for more control over its version: </summary>


The following instructions are taken from pyenv's guide here: https://github.com/pyenv/pyenv?tab=readme-ov-file#a-getting-pyenv

First, install pyenv and its dependencies.

```bash
Expand Down
13 changes: 9 additions & 4 deletions sharktank/tests/models/llama/benchmark_amdgpu_test.py
Original file line number Diff line number Diff line change
Expand Up @@ -197,7 +197,6 @@ def testBenchmark8B_f16_Decomposed(self):
)

@skipif_run_quick_llama_test
@pytest.mark.xfail(reason="Compile Error", strict=True, raises=IreeCompileException)
def testBenchmark8B_f16_Non_Decomposed_Prefill(self):
output_file_name = self.dir_path_8b / "f16_torch_prefill"
output_mlir = self.llama8b_f16_torch_sdpa_artifacts.create_file(
Expand Down Expand Up @@ -780,7 +779,9 @@ def testBenchmark405B_f16_TP8_Decomposed(self):
cwd=self.repo_root,
)

@pytest.mark.xfail(reason="Compile Error", strict=True, raises=IreeCompileException)
@pytest.mark.xfail(
reason="Benchmarking Error", strict=True, raises=IreeBenchmarkException
)
def testBenchmark405B_f16_TP8_Non_Decomposed(self):
output_file_name = self.dir_path_405b / "f16_torch"
output_mlir = self.llama405b_f16_torch_sdpa_artifacts.create_file(
Expand Down Expand Up @@ -828,7 +829,9 @@ def testBenchmark405B_f16_TP8_Non_Decomposed(self):
cwd=self.repo_root,
)

@pytest.mark.xfail(reason="Compile Error", strict=True, raises=IreeCompileException)
@pytest.mark.xfail(
reason="KeyError in theta.py", strict=True, raises=ExportMlirException
)
def testBenchmark405B_fp8_TP8_Decomposed(self):
output_file_name = self.dir_path_405b / "fp8_decomposed"
output_mlir = self.llama405b_fp8_decomposed_artifacts.create_file(
Expand Down Expand Up @@ -874,7 +877,9 @@ def testBenchmark405B_fp8_TP8_Decomposed(self):
cwd=self.repo_root,
)

@pytest.mark.xfail(reason="Compile Error", strict=True, raises=IreeCompileException)
@pytest.mark.xfail(
reason="KeyError in theta.py", strict=True, raises=ExportMlirException
)
def testBenchmark405B_fp8_TP8_Non_Decomposed(self):
output_file_name = self.dir_path_405b / "fp8_torch"
output_mlir = self.llama405b_fp8_torch_sdpa_artifacts.create_file(
Expand Down

0 comments on commit 453b4d2

Please sign in to comment.