Skip to content

Commit

Permalink
scenarios -> fixed_scenarios in docs
Browse files Browse the repository at this point in the history
  • Loading branch information
arjunsuresh committed Oct 22, 2024
1 parent 8967ecf commit fe99768
Show file tree
Hide file tree
Showing 2 changed files with 3 additions and 3 deletions.
2 changes: 1 addition & 1 deletion docs/benchmarks/language/reproducibility/indyscc24-bert.md
Original file line number Diff line number Diff line change
Expand Up @@ -27,7 +27,7 @@ All the needed files are automatically pushed to the GitHub repository if you ma
=== "MLCommons-Python"
## MLPerf Reference Implementation in Python

{{ mlperf_inference_implementation_readme (4, "bert-99", "reference", extra_variation_tags="", scenarios=["Offline"],categories=["Edge"], setup_tips=False) }}
{{ mlperf_inference_implementation_readme (4, "bert-99", "reference", extra_variation_tags="", fixed_scenarios=["Offline"],categories=["Edge"], setup_tips=False) }}


## Submission Commands
Expand Down
4 changes: 2 additions & 2 deletions docs/benchmarks/text_to_image/reproducibility/scc24.md
Original file line number Diff line number Diff line change
Expand Up @@ -46,11 +46,11 @@ or supporting multi-node execution) useful for the community and [MLCommons](htt
=== "MLCommons-Python"
## MLPerf Reference Implementation in Python

{{ mlperf_inference_implementation_readme (4, "sdxl", "reference", extra_variation_tags=",_short,_scc24-base", devices=["ROCm", "CUDA"],scenarios=["Offline"],categories=["Datacenter"], setup_tips=False, skip_test_query_count=True, extra_input_string="--precision=float16") }}
{{ mlperf_inference_implementation_readme (4, "sdxl", "reference", extra_variation_tags=",_short,_scc24-base", devices=["ROCm", "CUDA"],fixed_scenarios=["Offline"],categories=["Datacenter"], setup_tips=False, skip_test_query_count=True, extra_input_string="--precision=float16") }}

=== "Nvidia"
## Nvidia MLPerf Implementation
{{ mlperf_inference_implementation_readme (4, "sdxl", "nvidia", extra_variation_tags=",_short,_scc24-base", scenarios=["Offline"],categories=["Datacenter"], setup_tips=False, implementation_tips=False, skip_test_query_count=True) }}
{{ mlperf_inference_implementation_readme (4, "sdxl", "nvidia", extra_variation_tags=",_short,_scc24-base", fixed_scenarios=["Offline"],categories=["Datacenter"], setup_tips=False, implementation_tips=False, skip_test_query_count=True) }}

!!! info
Once the above run is successful, you can change `_scc24-base` to `_scc24-main` to run the main variant.
Expand Down

0 comments on commit fe99768

Please sign in to comment.