Skip to content

Commit

Permalink
Update README_nvidia.md
Browse files Browse the repository at this point in the history
  • Loading branch information
arjunsuresh authored Feb 15, 2024
1 parent 624bfb5 commit 2ec5974
Showing 1 changed file with 4 additions and 0 deletions.
4 changes: 4 additions & 0 deletions docs/mlperf/inference/gpt-j/README_nvidia.md
Original file line number Diff line number Diff line change
Expand Up @@ -17,6 +17,8 @@ cmr "generate-run-cmds inference _find-performance _all-scenarios" \
```
* Use `--division=closed` to run all scenarios for the closed division.
* Use `--category=datacenter` to run datacenter scenarios
* Use `--model=gptj-99.9` to run the high-accuracy model
* Use `--rerun` to force a rerun even when result files (from a previous run) exist

### Do full accuracy and performance runs for all the scenarios

Expand All @@ -30,6 +32,8 @@ cmr "generate-run-cmds inference _submission _all-scenarios" --model=gptj-99 \
* Use `--power=yes` for measuring power. It is ignored for accuracy and compliance runs
* Use `--division=closed` to run all scenarios for the closed division. No compliance runs are there for gptj.
* `--offline_target_qps`, `--server_target_qps`, and `--singlestream_target_latency` can be used to override the determined performance numbers
* Use `--model=gptj-99.9` to run the high-accuracy model
* Use `--rerun` to force a rerun even when result files (from a previous run) exist


### Generate and upload MLPerf submission
Expand Down

0 comments on commit 2ec5974

Please sign in to comment.