Skip to content

Commit 3153b89

Browse files
committed
Results from self hosted Github actions - NVIDIARTX4090
1 parent aa502f5 commit 3153b89

File tree

14 files changed

+419
-419
lines changed

14 files changed

+419
-419
lines changed
Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -1,3 +1,3 @@
1-
| Model | Scenario | Accuracy | Throughput | Latency (in ms) |
2-
|---------------------|------------|------------|--------------|-------------------|
3-
| stable-diffusion-xl | offline | () | 0.352 | - |
1+
| Model | Scenario | Accuracy | Throughput | Latency (in ms) |
2+
|---------|------------|------------|--------------|-------------------|
3+
| gptj-99 | offline | 264 | 49.207 | - |

open/MLCommons/measurements/gh_action-reference-gpu-pytorch_v2.5.1-cu124/gptj-99/offline/README.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -16,7 +16,7 @@ pip install -U cmind
1616

1717
cm rm cache -f
1818

19-
cm pull repo gateoverflow@mlperf-automations --checkout=37a8fb04148356c3b7c35014ca75477580652179
19+
cm pull repo gateoverflow@mlperf-automations --checkout=ee7be856e5def9e46bc4535b6128ec342f6931db
2020

2121
cm run script \
2222
--tags=app,mlperf,inference,generic,_reference,_gptj-99,_pytorch,_cuda,_test,_r5.0-dev_default,_float16,_offline \
@@ -105,4 +105,4 @@ Model Precision: fp32
105105
`GEN_LEN`: `264.0`, Required accuracy for closed division `>= 42.55663`
106106

107107
### Performance Results
108-
`Samples per second`: `48.2391`
108+
`Samples per second`: `49.2068`

open/MLCommons/measurements/gh_action-reference-gpu-pytorch_v2.5.1-cu124/gptj-99/offline/accuracy_console.out

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -2,17 +2,17 @@ Constructing QSL
22
Encoding Samples
33
Finished constructing QSL.
44
Loading PyTorch model...
5-
Loading checkpoint shards: 0%| | 0/3 [00:00<?, ?it/s]Loading checkpoint shards: 33%|███▎ | 1/3 [00:00<00:01, 1.41it/s]Loading checkpoint shards: 67%|██████▋ | 2/3 [00:01<00:00, 1.44it/s]Loading checkpoint shards: 100%|██████████| 3/3 [00:01<00:00, 1.97it/s]Loading checkpoint shards: 100%|██████████| 3/3 [00:01<00:00, 1.79it/s]
5+
Loading checkpoint shards: 0%| | 0/3 [00:00<?, ?it/s]Loading checkpoint shards: 33%|███▎ | 1/3 [00:00<00:01, 1.71it/s]Loading checkpoint shards: 67%|██████▋ | 2/3 [00:01<00:00, 1.76it/s]Loading checkpoint shards: 100%|██████████| 3/3 [00:01<00:00, 2.30it/s]Loading checkpoint shards: 100%|██████████| 3/3 [00:01<00:00, 2.11it/s]
66
Some weights of the model checkpoint at /home/cmuser/CM/repos/local/cache/31767c21a8f149e5/checkpoint/checkpoint-final were not used when initializing GPTJForCausalLM: ['transformer.h.0.attn.bias', 'transformer.h.0.attn.masked_bias', 'transformer.h.1.attn.bias', 'transformer.h.1.attn.masked_bias', 'transformer.h.10.attn.bias', 'transformer.h.10.attn.masked_bias', 'transformer.h.11.attn.bias', 'transformer.h.11.attn.masked_bias', 'transformer.h.12.attn.bias', 'transformer.h.12.attn.masked_bias', 'transformer.h.13.attn.bias', 'transformer.h.13.attn.masked_bias', 'transformer.h.14.attn.bias', 'transformer.h.14.attn.masked_bias', 'transformer.h.15.attn.bias', 'transformer.h.15.attn.masked_bias', 'transformer.h.16.attn.bias', 'transformer.h.16.attn.masked_bias', 'transformer.h.17.attn.bias', 'transformer.h.17.attn.masked_bias', 'transformer.h.18.attn.bias', 'transformer.h.18.attn.masked_bias', 'transformer.h.19.attn.bias', 'transformer.h.19.attn.masked_bias', 'transformer.h.2.attn.bias', 'transformer.h.2.attn.masked_bias', 'transformer.h.20.attn.bias', 'transformer.h.20.attn.masked_bias', 'transformer.h.21.attn.bias', 'transformer.h.21.attn.masked_bias', 'transformer.h.22.attn.bias', 'transformer.h.22.attn.masked_bias', 'transformer.h.23.attn.bias', 'transformer.h.23.attn.masked_bias', 'transformer.h.24.attn.bias', 'transformer.h.24.attn.masked_bias', 'transformer.h.25.attn.bias', 'transformer.h.25.attn.masked_bias', 'transformer.h.26.attn.bias', 'transformer.h.26.attn.masked_bias', 'transformer.h.27.attn.bias', 'transformer.h.27.attn.masked_bias', 'transformer.h.3.attn.bias', 'transformer.h.3.attn.masked_bias', 'transformer.h.4.attn.bias', 'transformer.h.4.attn.masked_bias', 'transformer.h.5.attn.bias', 'transformer.h.5.attn.masked_bias', 'transformer.h.6.attn.bias', 'transformer.h.6.attn.masked_bias', 'transformer.h.7.attn.bias', 'transformer.h.7.attn.masked_bias', 'transformer.h.8.attn.bias', 'transformer.h.8.attn.masked_bias', 'transformer.h.9.attn.bias', 'transformer.h.9.attn.masked_bias']
77
- This IS expected if you are initializing GPTJForCausalLM from the checkpoint of a model trained on another task or with another architecture (e.g. initializing a BertForSequenceClassification model from a BertForPreTraining model).
88
- This IS NOT expected if you are initializing GPTJForCausalLM from the checkpoint of a model that you expect to be exactly identical (initializing a BertForSequenceClassification model from a BertForSequenceClassification model).
99
Casting models to GPU...
10-
0%| | 0/285 [00:00<?, ?it/s]100%|██████████| 285/285 [00:00<00:00, 1934266.41it/s]
10+
0%| | 0/285 [00:00<?, ?it/s]100%|██████████| 285/285 [00:00<00:00, 1962851.63it/s]
1111
Running LoadGen test...
1212
Number of Samples in query_samples : 1
1313
0%| | 0/1 [00:00<?, ?it/s]/home/cmuser/venv/cm/lib/python3.10/site-packages/transformers/generation/configuration_utils.py:676: UserWarning: `num_beams` is set to 1. However, `early_stopping` is set to `True` -- this flag is only used in beam-based generation modes. You should set `num_beams>1` or unset `early_stopping`.
1414
warnings.warn(
15-
100%|██████████| 1/1 [00:01<00:00, 1.22s/it]100%|██████████| 1/1 [00:01<00:00, 1.22s/it]
15+
100%|██████████| 1/1 [00:01<00:00, 1.21s/it]100%|██████████| 1/1 [00:01<00:00, 1.21s/it]
1616

1717
No warnings encountered during test.
1818

open/MLCommons/measurements/gh_action-reference-gpu-pytorch_v2.5.1-cu124/gptj-99/offline/cm-deps.mmd

Lines changed: 19 additions & 19 deletions
Original file line numberDiff line numberDiff line change
@@ -31,22 +31,22 @@ graph TD
3131
get-generic-python-lib,94b62a682bc44791_(_package.numpy_) --> get,generic-python-lib,_pip
3232
get-cuda-devices,7a3ede4d3558427a_(_with-pycuda_) --> get,generic-python-lib,_package.numpy
3333
app-mlperf-inference,d775cac873ee4231_(_reference,_gptj-99,_pytorch,_cuda,_test,_r5.0-dev_default,_float16,_offline_) --> get,cuda-devices,_with-pycuda
34-
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_pytorch,_cuda,_offline,_gptj-99,_float16_) --> detect,os
34+
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_cuda,_pytorch,_offline,_gptj-99,_float16_) --> detect,os
3535
detect-cpu,586c8a43320142f7 --> detect,os
36-
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_pytorch,_cuda,_offline,_gptj-99,_float16_) --> detect,cpu
37-
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_pytorch,_cuda,_offline,_gptj-99,_float16_) --> get,sys-utils-cm
38-
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_pytorch,_cuda,_offline,_gptj-99,_float16_) --> get,python
39-
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_pytorch,_cuda,_offline,_gptj-99,_float16_) --> get,cuda,_cudnn
36+
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_cuda,_pytorch,_offline,_gptj-99,_float16_) --> detect,cpu
37+
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_cuda,_pytorch,_offline,_gptj-99,_float16_) --> get,sys-utils-cm
38+
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_cuda,_pytorch,_offline,_gptj-99,_float16_) --> get,python
39+
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_cuda,_pytorch,_offline,_gptj-99,_float16_) --> get,cuda,_cudnn
4040
get-generic-python-lib,94b62a682bc44791_(_torch_cuda_) --> get,python3
41-
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_pytorch,_cuda,_offline,_gptj-99,_float16_) --> get,generic-python-lib,_torch_cuda
41+
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_cuda,_pytorch,_offline,_gptj-99,_float16_) --> get,generic-python-lib,_torch_cuda
4242
get-generic-python-lib,94b62a682bc44791_(_torchvision_cuda_) --> get,python3
43-
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_pytorch,_cuda,_offline,_gptj-99,_float16_) --> get,generic-python-lib,_torchvision_cuda
43+
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_cuda,_pytorch,_offline,_gptj-99,_float16_) --> get,generic-python-lib,_torchvision_cuda
4444
get-generic-python-lib,94b62a682bc44791_(_transformers_) --> get,python3
45-
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_pytorch,_cuda,_offline,_gptj-99,_float16_) --> get,generic-python-lib,_transformers
45+
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_cuda,_pytorch,_offline,_gptj-99,_float16_) --> get,generic-python-lib,_transformers
4646
download-and-extract,c67e81a4ce2649f5_(_rclone,_url.mlc-inference:mlcommons-inference-wg-public/gpt-j_) --> download,file,_rclone,_url.mlc-inference:mlcommons-inference-wg-public/gpt-j
4747
get-ml-model-gptj,a41166210f294fbf_(_pytorch_) --> download-and-extract,_rclone,_url.mlc-inference:mlcommons-inference-wg-public/gpt-j
48-
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_pytorch,_cuda,_offline,_gptj-99,_float16_) --> get,ml-model,large-language-model,gptj,raw,_pytorch
49-
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_pytorch,_cuda,_offline,_gptj-99,_float16_) --> get,dataset,cnndm,_validation
48+
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_cuda,_pytorch,_offline,_gptj-99,_float16_) --> get,ml-model,large-language-model,gptj,raw,_pytorch
49+
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_cuda,_pytorch,_offline,_gptj-99,_float16_) --> get,dataset,cnndm,_validation
5050
generate-mlperf-inference-user-conf,3af4475745964b93 --> detect,os
5151
detect-cpu,586c8a43320142f7 --> detect,os
5252
generate-mlperf-inference-user-conf,3af4475745964b93 --> detect,cpu
@@ -57,22 +57,22 @@ graph TD
5757
generate-mlperf-inference-user-conf,3af4475745964b93 --> get,mlcommons,inference,src
5858
get-mlperf-inference-sut-configs,c2fbf72009e2445b --> get,cache,dir,_name.mlperf-inference-sut-configs
5959
generate-mlperf-inference-user-conf,3af4475745964b93 --> get,sut,configs
60-
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_pytorch,_cuda,_offline,_gptj-99,_float16_) --> generate,user-conf,mlperf,inference
61-
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_pytorch,_cuda,_offline,_gptj-99,_float16_) --> get,loadgen
60+
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_cuda,_pytorch,_offline,_gptj-99,_float16_) --> generate,user-conf,mlperf,inference
61+
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_cuda,_pytorch,_offline,_gptj-99,_float16_) --> get,loadgen
6262
get-mlperf-inference-src,4b57186581024797 --> detect,os
6363
get-mlperf-inference-src,4b57186581024797 --> get,python3
6464
get-mlperf-inference-src,4b57186581024797 --> get,git,repo,_branch.master,_repo.https://github.com/mlcommons/inference
65-
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_pytorch,_cuda,_offline,_gptj-99,_float16_) --> get,mlcommons,inference,src
66-
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_pytorch,_cuda,_offline,_gptj-99,_float16_) --> get,mlcommons,inference,src
65+
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_cuda,_pytorch,_offline,_gptj-99,_float16_) --> get,mlcommons,inference,src
66+
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_cuda,_pytorch,_offline,_gptj-99,_float16_) --> get,mlcommons,inference,src
6767
get-generic-python-lib,94b62a682bc44791_(_package.psutil_) --> get,python3
68-
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_pytorch,_cuda,_offline,_gptj-99,_float16_) --> get,generic-python-lib,_package.psutil
68+
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_cuda,_pytorch,_offline,_gptj-99,_float16_) --> get,generic-python-lib,_package.psutil
6969
get-generic-python-lib,94b62a682bc44791_(_package.datasets_) --> get,python3
70-
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_pytorch,_cuda,_offline,_gptj-99,_float16_) --> get,generic-python-lib,_package.datasets
70+
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_cuda,_pytorch,_offline,_gptj-99,_float16_) --> get,generic-python-lib,_package.datasets
7171
get-generic-python-lib,94b62a682bc44791_(_package.attrs_) --> get,python3
72-
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_pytorch,_cuda,_offline,_gptj-99,_float16_) --> get,generic-python-lib,_package.attrs
72+
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_cuda,_pytorch,_offline,_gptj-99,_float16_) --> get,generic-python-lib,_package.attrs
7373
get-generic-python-lib,94b62a682bc44791_(_package.accelerate_) --> get,python3
74-
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_pytorch,_cuda,_offline,_gptj-99,_float16_) --> get,generic-python-lib,_package.accelerate
74+
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_cuda,_pytorch,_offline,_gptj-99,_float16_) --> get,generic-python-lib,_package.accelerate
7575
detect-cpu,586c8a43320142f7 --> detect,os
7676
benchmark-program,19f369ef47084895 --> detect,cpu
7777
benchmark-program-mlperf,cfff0132a8aa4018 --> benchmark-program,program
78-
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_pytorch,_cuda,_offline,_gptj-99,_float16_) --> benchmark-mlperf
78+
app-mlperf-inference-mlcommons-python,ff149e9781fc4b65_(_cuda,_pytorch,_offline,_gptj-99,_float16_) --> benchmark-mlperf
814 Bytes
Loading

0 commit comments

Comments
 (0)