Skip to content

Commit 4ff08c5

Browse files
committed
Results on system pku 2
1 parent ff5c466 commit 4ff08c5

File tree

12 files changed

+431
-416
lines changed

12 files changed

+431
-416
lines changed
Original file line numberDiff line numberDiff line change
@@ -1,3 +1,3 @@
11
| Model | Scenario | Accuracy | Throughput | Latency (in ms) |
22
|---------------------|------------|----------------------|--------------|-------------------|
3-
| stable-diffusion-xl | offline | (14.02827, 84.33062) | 8.001 | - |
3+
| stable-diffusion-xl | offline | (14.02827, 84.33062) | 8.281 | - |

open/pku/measurements/mlperf_inference_lry_40-nvidia_original-gpu-tensorrt-vdefault-scc24-main/stable-diffusion-xl/offline/README.md

+2-2
Original file line numberDiff line numberDiff line change
@@ -26,7 +26,7 @@ cm run script \
2626
--model=sdxl \
2727
--implementation=nvidia \
2828
--max_query_count=5000 \
29-
--min_query_count=72 \
29+
--min_query_count=504 \
3030
--framework=tensorrt \
3131
--category=datacenter \
3232
--scenario=Offline \
@@ -57,4 +57,4 @@ Model Precision: int8
5757
`FID_SCORE`: `84.33062`, Required accuracy for closed division `>= 23.01086` and `<= 23.95008`
5858

5959
### Performance Results
60-
`Samples per second`: `8.00134`
60+
`Samples per second`: `8.2807`

open/pku/measurements/mlperf_inference_lry_40-nvidia_original-gpu-tensorrt-vdefault-scc24-main/stable-diffusion-xl/offline/accuracy_console.out

+56-51
Large diffs are not rendered by default.

open/pku/measurements/mlperf_inference_lry_40-nvidia_original-gpu-tensorrt-vdefault-scc24-main/stable-diffusion-xl/offline/cm-version-info.json

+37-37
Original file line numberDiff line numberDiff line change
@@ -117,7 +117,7 @@
117117
"script_tags": "detect-os,detect,os,info",
118118
"script_variations": "",
119119
"version": "",
120-
"parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_cuda,_sdxl,_tensorrt,_offline,_gpu_memory.80 )"
120+
"parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_tensorrt,_cuda,_sdxl,_offline,_gpu_memory.80 )"
121121
}
122122
},
123123
{
@@ -137,7 +137,7 @@
137137
"script_tags": "detect,cpu,detect-cpu,info",
138138
"script_variations": "",
139139
"version": "",
140-
"parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_cuda,_sdxl,_tensorrt,_offline,_gpu_memory.80 )"
140+
"parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_tensorrt,_cuda,_sdxl,_offline,_gpu_memory.80 )"
141141
}
142142
},
143143
{
@@ -157,7 +157,7 @@
157157
"script_tags": "get,mlperf,inference,nvidia,scratch,space",
158158
"script_variations": "",
159159
"version": "",
160-
"parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_cuda,_sdxl,_tensorrt,_offline,_gpu_memory.80 )"
160+
"parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_tensorrt,_cuda,_sdxl,_offline,_gpu_memory.80 )"
161161
}
162162
},
163163
{
@@ -167,7 +167,7 @@
167167
"script_tags": "get,install,generic,generic-python-lib",
168168
"script_variations": "mlperf_logging",
169169
"version": "0.0.0",
170-
"parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_cuda,_sdxl,_tensorrt,_offline,_gpu_memory.80 )"
170+
"parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_tensorrt,_cuda,_sdxl,_offline,_gpu_memory.80 )"
171171
}
172172
},
173173
{
@@ -177,7 +177,7 @@
177177
"script_tags": "get,src,source,inference,inference-src,inference-source,mlperf,mlcommons",
178178
"script_variations": "",
179179
"version": "master-git-4763c6b5e7c3ad904c461bc1053674c448771ec4",
180-
"parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_cuda,_sdxl,_tensorrt,_offline,_gpu_memory.80 )"
180+
"parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_tensorrt,_cuda,_sdxl,_offline,_gpu_memory.80 )"
181181
}
182182
},
183183
{
@@ -187,7 +187,7 @@
187187
"script_tags": "get,nvidia,mlperf,inference,common-code",
188188
"script_variations": "mlcommons",
189189
"version": "r4.0",
190-
"parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_cuda,_sdxl,_tensorrt,_offline,_gpu_memory.80 )"
190+
"parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_tensorrt,_cuda,_sdxl,_offline,_gpu_memory.80 )"
191191
}
192192
},
193193
{
@@ -267,7 +267,7 @@
267267
"script_tags": "generate,mlperf,inference,user-conf,inference-user-conf",
268268
"script_variations": "",
269269
"version": "",
270-
"parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_cuda,_sdxl,_tensorrt,_offline,_gpu_memory.80 )"
270+
"parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_tensorrt,_cuda,_sdxl,_offline,_gpu_memory.80 )"
271271
}
272272
},
273273
{
@@ -277,7 +277,7 @@
277277
"script_tags": "get,install,generic,generic-python-lib",
278278
"script_variations": "package.pycuda",
279279
"version": "2022.2.2",
280-
"parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_cuda,_sdxl,_tensorrt,_offline,_gpu_memory.80 )"
280+
"parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_tensorrt,_cuda,_sdxl,_offline,_gpu_memory.80 )"
281281
}
282282
},
283283
{
@@ -287,7 +287,7 @@
287287
"script_tags": "get,nvidia,mitten,nvidia-mitten",
288288
"script_variations": "",
289289
"version": "master",
290-
"parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_cuda,_sdxl,_tensorrt,_offline,_gpu_memory.80 )"
290+
"parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_tensorrt,_cuda,_sdxl,_offline,_gpu_memory.80 )"
291291
}
292292
},
293293
{
@@ -297,7 +297,7 @@
297297
"script_tags": "get,cuda,cuda-compiler,cuda-lib,toolkit,lib,nvcc,get-nvcc,get-cuda,46d133d9ef92422d",
298298
"script_variations": "cudnn",
299299
"version": "12.2",
300-
"parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_cuda,_sdxl,_tensorrt,_offline,_gpu_memory.80 )"
300+
"parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_tensorrt,_cuda,_sdxl,_offline,_gpu_memory.80 )"
301301
}
302302
},
303303
{
@@ -307,7 +307,7 @@
307307
"script_tags": "get,tensorrt,nvidia",
308308
"script_variations": "",
309309
"version": "vdetected",
310-
"parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_cuda,_sdxl,_tensorrt,_offline,_gpu_memory.80 )"
310+
"parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_tensorrt,_cuda,_sdxl,_offline,_gpu_memory.80 )"
311311
}
312312
},
313313
{
@@ -317,27 +317,27 @@
317317
"script_tags": "build,mlcommons,mlperf,inference,inference-server,server,nvidia-harness,nvidia",
318318
"script_variations": "mlcommons",
319319
"version": "r4.0",
320-
"parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_cuda,_sdxl,_tensorrt,_offline,_gpu_memory.80 )"
320+
"parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_tensorrt,_cuda,_sdxl,_offline,_gpu_memory.80 )"
321321
}
322322
},
323323
{
324-
"reproduce,mlperf,inference,nvidia,harness,_build_engine,_cuda,_sdxl,_tensorrt,_offline,_batch_size.8,_v4.1-dev": {
324+
"reproduce,mlperf,inference,nvidia,harness,_build_engine,_tensorrt,_cuda,_sdxl,_offline,_batch_size.8,_v4.1-dev": {
325325
"script_uid": "bc3b17fb430f4732",
326326
"script_alias": "app-mlperf-inference-nvidia",
327327
"script_tags": "reproduce,mlcommons,mlperf,inference,harness,nvidia-harness,nvidia",
328-
"script_variations": "build_engine,cuda,sdxl,tensorrt,offline,batch_size.8,v4.1-dev",
328+
"script_variations": "build_engine,tensorrt,cuda,sdxl,offline,batch_size.8,v4.1-dev",
329329
"version": "",
330-
"parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_cuda,_sdxl,_tensorrt,_offline,_gpu_memory.80 )"
330+
"parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_tensorrt,_cuda,_sdxl,_offline,_gpu_memory.80 )"
331331
}
332332
},
333333
{
334-
"reproduce,mlperf,inference,nvidia,harness,_preprocess_data,_cuda,_sdxl,_tensorrt,_v4.1-dev": {
334+
"reproduce,mlperf,inference,nvidia,harness,_preprocess_data,_tensorrt,_cuda,_sdxl,_v4.1-dev": {
335335
"script_uid": "bc3b17fb430f4732",
336336
"script_alias": "app-mlperf-inference-nvidia",
337337
"script_tags": "reproduce,mlcommons,mlperf,inference,harness,nvidia-harness,nvidia",
338-
"script_variations": "preprocess_data,cuda,sdxl,tensorrt,v4.1-dev",
338+
"script_variations": "preprocess_data,tensorrt,cuda,sdxl,v4.1-dev",
339339
"version": "",
340-
"parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_cuda,_sdxl,_tensorrt,_offline,_gpu_memory.80 )"
340+
"parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_tensorrt,_cuda,_sdxl,_offline,_gpu_memory.80 )"
341341
}
342342
},
343343
{
@@ -347,7 +347,7 @@
347347
"script_tags": "get,install,generic,generic-python-lib",
348348
"script_variations": "package.diffusers",
349349
"version": "0.21.2",
350-
"parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_cuda,_sdxl,_tensorrt,_offline,_gpu_memory.80 )"
350+
"parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_tensorrt,_cuda,_sdxl,_offline,_gpu_memory.80 )"
351351
}
352352
},
353353
{
@@ -357,7 +357,7 @@
357357
"script_tags": "get,install,generic,generic-python-lib",
358358
"script_variations": "package.transformers",
359359
"version": "4.33.1",
360-
"parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_cuda,_sdxl,_tensorrt,_offline,_gpu_memory.80 )"
360+
"parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_tensorrt,_cuda,_sdxl,_offline,_gpu_memory.80 )"
361361
}
362362
},
363363
{
@@ -367,7 +367,7 @@
367367
"script_tags": "get,install,generic,generic-python-lib",
368368
"script_variations": "package.accelerate",
369369
"version": "0.23.0",
370-
"parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_cuda,_sdxl,_tensorrt,_offline,_gpu_memory.80 )"
370+
"parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_tensorrt,_cuda,_sdxl,_offline,_gpu_memory.80 )"
371371
}
372372
},
373373
{
@@ -377,7 +377,7 @@
377377
"script_tags": "get,install,generic,generic-python-lib",
378378
"script_variations": "package.torchmetrics",
379379
"version": "0.11.0",
380-
"parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_cuda,_sdxl,_tensorrt,_offline,_gpu_memory.80 )"
380+
"parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_tensorrt,_cuda,_sdxl,_offline,_gpu_memory.80 )"
381381
}
382382
},
383383
{
@@ -387,7 +387,7 @@
387387
"script_tags": "get,install,generic,generic-python-lib",
388388
"script_variations": "package.torch-fidelity",
389389
"version": "0.3.0",
390-
"parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_cuda,_sdxl,_tensorrt,_offline,_gpu_memory.80 )"
390+
"parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_tensorrt,_cuda,_sdxl,_offline,_gpu_memory.80 )"
391391
}
392392
},
393393
{
@@ -397,7 +397,7 @@
397397
"script_tags": "get,install,generic,generic-python-lib",
398398
"script_variations": "package.open_clip_torch",
399399
"version": "2.7.0",
400-
"parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_cuda,_sdxl,_tensorrt,_offline,_gpu_memory.80 )"
400+
"parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_tensorrt,_cuda,_sdxl,_offline,_gpu_memory.80 )"
401401
}
402402
},
403403
{
@@ -407,7 +407,7 @@
407407
"script_tags": "get,install,generic,generic-python-lib",
408408
"script_variations": "package.opencv-python",
409409
"version": "4.10.0.84",
410-
"parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_cuda,_sdxl,_tensorrt,_offline,_gpu_memory.80 )"
410+
"parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_tensorrt,_cuda,_sdxl,_offline,_gpu_memory.80 )"
411411
}
412412
},
413413
{
@@ -417,7 +417,7 @@
417417
"script_tags": "get,install,generic,generic-python-lib",
418418
"script_variations": "package.polygraphy",
419419
"version": "0.47.1",
420-
"parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_cuda,_sdxl,_tensorrt,_offline,_gpu_memory.80 )"
420+
"parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_tensorrt,_cuda,_sdxl,_offline,_gpu_memory.80 )"
421421
}
422422
},
423423
{
@@ -427,7 +427,7 @@
427427
"script_tags": "get,install,generic,generic-python-lib",
428428
"script_variations": "package.nvtx",
429429
"version": "0.2.8",
430-
"parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_cuda,_sdxl,_tensorrt,_offline,_gpu_memory.80 )"
430+
"parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_tensorrt,_cuda,_sdxl,_offline,_gpu_memory.80 )"
431431
}
432432
},
433433
{
@@ -437,7 +437,7 @@
437437
"script_tags": "get,install,generic,generic-python-lib",
438438
"script_variations": "package.cuda-python",
439439
"version": "12.1.0",
440-
"parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_cuda,_sdxl,_tensorrt,_offline,_gpu_memory.80 )"
440+
"parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_tensorrt,_cuda,_sdxl,_offline,_gpu_memory.80 )"
441441
}
442442
},
443443
{
@@ -447,7 +447,7 @@
447447
"script_tags": "get,install,generic,generic-python-lib",
448448
"script_variations": "package.ninja",
449449
"version": "1.11.1",
450-
"parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_cuda,_sdxl,_tensorrt,_offline,_gpu_memory.80 )"
450+
"parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_tensorrt,_cuda,_sdxl,_offline,_gpu_memory.80 )"
451451
}
452452
},
453453
{
@@ -457,7 +457,7 @@
457457
"script_tags": "get,install,generic,generic-python-lib",
458458
"script_variations": "package.onnxruntime",
459459
"version": "1.18.1",
460-
"parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_cuda,_sdxl,_tensorrt,_offline,_gpu_memory.80 )"
460+
"parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_tensorrt,_cuda,_sdxl,_offline,_gpu_memory.80 )"
461461
}
462462
},
463463
{
@@ -467,7 +467,7 @@
467467
"script_tags": "get,install,generic,generic-python-lib",
468468
"script_variations": "package.colored",
469469
"version": "1.4.4",
470-
"parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_cuda,_sdxl,_tensorrt,_offline,_gpu_memory.80 )"
470+
"parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_tensorrt,_cuda,_sdxl,_offline,_gpu_memory.80 )"
471471
}
472472
},
473473
{
@@ -477,7 +477,7 @@
477477
"script_tags": "get,install,generic,generic-python-lib",
478478
"script_variations": "package.nvidia-ammo",
479479
"version": "0.7.4",
480-
"parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_cuda,_sdxl,_tensorrt,_offline,_gpu_memory.80 )"
480+
"parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_tensorrt,_cuda,_sdxl,_offline,_gpu_memory.80 )"
481481
}
482482
},
483483
{
@@ -487,7 +487,7 @@
487487
"script_tags": "get,install,generic,generic-python-lib",
488488
"script_variations": "package.optimum",
489489
"version": "1.16.0",
490-
"parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_cuda,_sdxl,_tensorrt,_offline,_gpu_memory.80 )"
490+
"parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_tensorrt,_cuda,_sdxl,_offline,_gpu_memory.80 )"
491491
}
492492
},
493493
{
@@ -497,7 +497,7 @@
497497
"script_tags": "get,install,generic,generic-python-lib",
498498
"script_variations": "package.onnx",
499499
"version": "1.14.0",
500-
"parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_cuda,_sdxl,_tensorrt,_offline,_gpu_memory.80 )"
500+
"parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_tensorrt,_cuda,_sdxl,_offline,_gpu_memory.80 )"
501501
}
502502
},
503503
{
@@ -507,7 +507,7 @@
507507
"script_tags": "get,install,generic,generic-python-lib",
508508
"script_variations": "package.scipy",
509509
"version": "1.10.1",
510-
"parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_cuda,_sdxl,_tensorrt,_offline,_gpu_memory.80 )"
510+
"parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_tensorrt,_cuda,_sdxl,_offline,_gpu_memory.80 )"
511511
}
512512
},
513513
{
@@ -517,7 +517,7 @@
517517
"script_tags": "get,install,generic,generic-python-lib",
518518
"script_variations": "package.numpy",
519519
"version": "1.22.0",
520-
"parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_cuda,_sdxl,_tensorrt,_offline,_gpu_memory.80 )"
520+
"parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_tensorrt,_cuda,_sdxl,_offline,_gpu_memory.80 )"
521521
}
522522
},
523523
{
@@ -557,7 +557,7 @@
557557
"script_tags": "mlperf,benchmark-mlperf",
558558
"script_variations": "",
559559
"version": "",
560-
"parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_cuda,_sdxl,_tensorrt,_offline,_gpu_memory.80 )"
560+
"parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_tensorrt,_cuda,_sdxl,_offline,_gpu_memory.80 )"
561561
}
562562
}
563563
]

0 commit comments

Comments
 (0)