|
117 | 117 | "script_tags": "detect-os,detect,os,info",
|
118 | 118 | "script_variations": "",
|
119 | 119 | "version": "",
|
120 |
| - "parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_cuda,_sdxl,_tensorrt,_offline,_gpu_memory.80 )" |
| 120 | + "parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_tensorrt,_cuda,_sdxl,_offline,_gpu_memory.80 )" |
121 | 121 | }
|
122 | 122 | },
|
123 | 123 | {
|
|
137 | 137 | "script_tags": "detect,cpu,detect-cpu,info",
|
138 | 138 | "script_variations": "",
|
139 | 139 | "version": "",
|
140 |
| - "parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_cuda,_sdxl,_tensorrt,_offline,_gpu_memory.80 )" |
| 140 | + "parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_tensorrt,_cuda,_sdxl,_offline,_gpu_memory.80 )" |
141 | 141 | }
|
142 | 142 | },
|
143 | 143 | {
|
|
157 | 157 | "script_tags": "get,mlperf,inference,nvidia,scratch,space",
|
158 | 158 | "script_variations": "",
|
159 | 159 | "version": "",
|
160 |
| - "parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_cuda,_sdxl,_tensorrt,_offline,_gpu_memory.80 )" |
| 160 | + "parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_tensorrt,_cuda,_sdxl,_offline,_gpu_memory.80 )" |
161 | 161 | }
|
162 | 162 | },
|
163 | 163 | {
|
|
167 | 167 | "script_tags": "get,install,generic,generic-python-lib",
|
168 | 168 | "script_variations": "mlperf_logging",
|
169 | 169 | "version": "0.0.0",
|
170 |
| - "parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_cuda,_sdxl,_tensorrt,_offline,_gpu_memory.80 )" |
| 170 | + "parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_tensorrt,_cuda,_sdxl,_offline,_gpu_memory.80 )" |
171 | 171 | }
|
172 | 172 | },
|
173 | 173 | {
|
|
177 | 177 | "script_tags": "get,src,source,inference,inference-src,inference-source,mlperf,mlcommons",
|
178 | 178 | "script_variations": "",
|
179 | 179 | "version": "master-git-4763c6b5e7c3ad904c461bc1053674c448771ec4",
|
180 |
| - "parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_cuda,_sdxl,_tensorrt,_offline,_gpu_memory.80 )" |
| 180 | + "parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_tensorrt,_cuda,_sdxl,_offline,_gpu_memory.80 )" |
181 | 181 | }
|
182 | 182 | },
|
183 | 183 | {
|
|
187 | 187 | "script_tags": "get,nvidia,mlperf,inference,common-code",
|
188 | 188 | "script_variations": "mlcommons",
|
189 | 189 | "version": "r4.0",
|
190 |
| - "parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_cuda,_sdxl,_tensorrt,_offline,_gpu_memory.80 )" |
| 190 | + "parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_tensorrt,_cuda,_sdxl,_offline,_gpu_memory.80 )" |
191 | 191 | }
|
192 | 192 | },
|
193 | 193 | {
|
|
267 | 267 | "script_tags": "generate,mlperf,inference,user-conf,inference-user-conf",
|
268 | 268 | "script_variations": "",
|
269 | 269 | "version": "",
|
270 |
| - "parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_cuda,_sdxl,_tensorrt,_offline,_gpu_memory.80 )" |
| 270 | + "parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_tensorrt,_cuda,_sdxl,_offline,_gpu_memory.80 )" |
271 | 271 | }
|
272 | 272 | },
|
273 | 273 | {
|
|
277 | 277 | "script_tags": "get,install,generic,generic-python-lib",
|
278 | 278 | "script_variations": "package.pycuda",
|
279 | 279 | "version": "2022.2.2",
|
280 |
| - "parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_cuda,_sdxl,_tensorrt,_offline,_gpu_memory.80 )" |
| 280 | + "parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_tensorrt,_cuda,_sdxl,_offline,_gpu_memory.80 )" |
281 | 281 | }
|
282 | 282 | },
|
283 | 283 | {
|
|
287 | 287 | "script_tags": "get,nvidia,mitten,nvidia-mitten",
|
288 | 288 | "script_variations": "",
|
289 | 289 | "version": "master",
|
290 |
| - "parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_cuda,_sdxl,_tensorrt,_offline,_gpu_memory.80 )" |
| 290 | + "parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_tensorrt,_cuda,_sdxl,_offline,_gpu_memory.80 )" |
291 | 291 | }
|
292 | 292 | },
|
293 | 293 | {
|
|
297 | 297 | "script_tags": "get,cuda,cuda-compiler,cuda-lib,toolkit,lib,nvcc,get-nvcc,get-cuda,46d133d9ef92422d",
|
298 | 298 | "script_variations": "cudnn",
|
299 | 299 | "version": "12.2",
|
300 |
| - "parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_cuda,_sdxl,_tensorrt,_offline,_gpu_memory.80 )" |
| 300 | + "parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_tensorrt,_cuda,_sdxl,_offline,_gpu_memory.80 )" |
301 | 301 | }
|
302 | 302 | },
|
303 | 303 | {
|
|
307 | 307 | "script_tags": "get,tensorrt,nvidia",
|
308 | 308 | "script_variations": "",
|
309 | 309 | "version": "vdetected",
|
310 |
| - "parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_cuda,_sdxl,_tensorrt,_offline,_gpu_memory.80 )" |
| 310 | + "parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_tensorrt,_cuda,_sdxl,_offline,_gpu_memory.80 )" |
311 | 311 | }
|
312 | 312 | },
|
313 | 313 | {
|
|
317 | 317 | "script_tags": "build,mlcommons,mlperf,inference,inference-server,server,nvidia-harness,nvidia",
|
318 | 318 | "script_variations": "mlcommons",
|
319 | 319 | "version": "r4.0",
|
320 |
| - "parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_cuda,_sdxl,_tensorrt,_offline,_gpu_memory.80 )" |
| 320 | + "parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_tensorrt,_cuda,_sdxl,_offline,_gpu_memory.80 )" |
321 | 321 | }
|
322 | 322 | },
|
323 | 323 | {
|
324 |
| - "reproduce,mlperf,inference,nvidia,harness,_build_engine,_cuda,_sdxl,_tensorrt,_offline,_batch_size.8,_v4.1-dev": { |
| 324 | + "reproduce,mlperf,inference,nvidia,harness,_build_engine,_tensorrt,_cuda,_sdxl,_offline,_batch_size.8,_v4.1-dev": { |
325 | 325 | "script_uid": "bc3b17fb430f4732",
|
326 | 326 | "script_alias": "app-mlperf-inference-nvidia",
|
327 | 327 | "script_tags": "reproduce,mlcommons,mlperf,inference,harness,nvidia-harness,nvidia",
|
328 |
| - "script_variations": "build_engine,cuda,sdxl,tensorrt,offline,batch_size.8,v4.1-dev", |
| 328 | + "script_variations": "build_engine,tensorrt,cuda,sdxl,offline,batch_size.8,v4.1-dev", |
329 | 329 | "version": "",
|
330 |
| - "parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_cuda,_sdxl,_tensorrt,_offline,_gpu_memory.80 )" |
| 330 | + "parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_tensorrt,_cuda,_sdxl,_offline,_gpu_memory.80 )" |
331 | 331 | }
|
332 | 332 | },
|
333 | 333 | {
|
334 |
| - "reproduce,mlperf,inference,nvidia,harness,_preprocess_data,_cuda,_sdxl,_tensorrt,_v4.1-dev": { |
| 334 | + "reproduce,mlperf,inference,nvidia,harness,_preprocess_data,_tensorrt,_cuda,_sdxl,_v4.1-dev": { |
335 | 335 | "script_uid": "bc3b17fb430f4732",
|
336 | 336 | "script_alias": "app-mlperf-inference-nvidia",
|
337 | 337 | "script_tags": "reproduce,mlcommons,mlperf,inference,harness,nvidia-harness,nvidia",
|
338 |
| - "script_variations": "preprocess_data,cuda,sdxl,tensorrt,v4.1-dev", |
| 338 | + "script_variations": "preprocess_data,tensorrt,cuda,sdxl,v4.1-dev", |
339 | 339 | "version": "",
|
340 |
| - "parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_cuda,_sdxl,_tensorrt,_offline,_gpu_memory.80 )" |
| 340 | + "parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_tensorrt,_cuda,_sdxl,_offline,_gpu_memory.80 )" |
341 | 341 | }
|
342 | 342 | },
|
343 | 343 | {
|
|
347 | 347 | "script_tags": "get,install,generic,generic-python-lib",
|
348 | 348 | "script_variations": "package.diffusers",
|
349 | 349 | "version": "0.21.2",
|
350 |
| - "parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_cuda,_sdxl,_tensorrt,_offline,_gpu_memory.80 )" |
| 350 | + "parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_tensorrt,_cuda,_sdxl,_offline,_gpu_memory.80 )" |
351 | 351 | }
|
352 | 352 | },
|
353 | 353 | {
|
|
357 | 357 | "script_tags": "get,install,generic,generic-python-lib",
|
358 | 358 | "script_variations": "package.transformers",
|
359 | 359 | "version": "4.33.1",
|
360 |
| - "parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_cuda,_sdxl,_tensorrt,_offline,_gpu_memory.80 )" |
| 360 | + "parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_tensorrt,_cuda,_sdxl,_offline,_gpu_memory.80 )" |
361 | 361 | }
|
362 | 362 | },
|
363 | 363 | {
|
|
367 | 367 | "script_tags": "get,install,generic,generic-python-lib",
|
368 | 368 | "script_variations": "package.accelerate",
|
369 | 369 | "version": "0.23.0",
|
370 |
| - "parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_cuda,_sdxl,_tensorrt,_offline,_gpu_memory.80 )" |
| 370 | + "parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_tensorrt,_cuda,_sdxl,_offline,_gpu_memory.80 )" |
371 | 371 | }
|
372 | 372 | },
|
373 | 373 | {
|
|
377 | 377 | "script_tags": "get,install,generic,generic-python-lib",
|
378 | 378 | "script_variations": "package.torchmetrics",
|
379 | 379 | "version": "0.11.0",
|
380 |
| - "parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_cuda,_sdxl,_tensorrt,_offline,_gpu_memory.80 )" |
| 380 | + "parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_tensorrt,_cuda,_sdxl,_offline,_gpu_memory.80 )" |
381 | 381 | }
|
382 | 382 | },
|
383 | 383 | {
|
|
387 | 387 | "script_tags": "get,install,generic,generic-python-lib",
|
388 | 388 | "script_variations": "package.torch-fidelity",
|
389 | 389 | "version": "0.3.0",
|
390 |
| - "parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_cuda,_sdxl,_tensorrt,_offline,_gpu_memory.80 )" |
| 390 | + "parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_tensorrt,_cuda,_sdxl,_offline,_gpu_memory.80 )" |
391 | 391 | }
|
392 | 392 | },
|
393 | 393 | {
|
|
397 | 397 | "script_tags": "get,install,generic,generic-python-lib",
|
398 | 398 | "script_variations": "package.open_clip_torch",
|
399 | 399 | "version": "2.7.0",
|
400 |
| - "parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_cuda,_sdxl,_tensorrt,_offline,_gpu_memory.80 )" |
| 400 | + "parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_tensorrt,_cuda,_sdxl,_offline,_gpu_memory.80 )" |
401 | 401 | }
|
402 | 402 | },
|
403 | 403 | {
|
|
407 | 407 | "script_tags": "get,install,generic,generic-python-lib",
|
408 | 408 | "script_variations": "package.opencv-python",
|
409 | 409 | "version": "4.10.0.84",
|
410 |
| - "parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_cuda,_sdxl,_tensorrt,_offline,_gpu_memory.80 )" |
| 410 | + "parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_tensorrt,_cuda,_sdxl,_offline,_gpu_memory.80 )" |
411 | 411 | }
|
412 | 412 | },
|
413 | 413 | {
|
|
417 | 417 | "script_tags": "get,install,generic,generic-python-lib",
|
418 | 418 | "script_variations": "package.polygraphy",
|
419 | 419 | "version": "0.47.1",
|
420 |
| - "parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_cuda,_sdxl,_tensorrt,_offline,_gpu_memory.80 )" |
| 420 | + "parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_tensorrt,_cuda,_sdxl,_offline,_gpu_memory.80 )" |
421 | 421 | }
|
422 | 422 | },
|
423 | 423 | {
|
|
427 | 427 | "script_tags": "get,install,generic,generic-python-lib",
|
428 | 428 | "script_variations": "package.nvtx",
|
429 | 429 | "version": "0.2.8",
|
430 |
| - "parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_cuda,_sdxl,_tensorrt,_offline,_gpu_memory.80 )" |
| 430 | + "parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_tensorrt,_cuda,_sdxl,_offline,_gpu_memory.80 )" |
431 | 431 | }
|
432 | 432 | },
|
433 | 433 | {
|
|
437 | 437 | "script_tags": "get,install,generic,generic-python-lib",
|
438 | 438 | "script_variations": "package.cuda-python",
|
439 | 439 | "version": "12.1.0",
|
440 |
| - "parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_cuda,_sdxl,_tensorrt,_offline,_gpu_memory.80 )" |
| 440 | + "parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_tensorrt,_cuda,_sdxl,_offline,_gpu_memory.80 )" |
441 | 441 | }
|
442 | 442 | },
|
443 | 443 | {
|
|
447 | 447 | "script_tags": "get,install,generic,generic-python-lib",
|
448 | 448 | "script_variations": "package.ninja",
|
449 | 449 | "version": "1.11.1",
|
450 |
| - "parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_cuda,_sdxl,_tensorrt,_offline,_gpu_memory.80 )" |
| 450 | + "parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_tensorrt,_cuda,_sdxl,_offline,_gpu_memory.80 )" |
451 | 451 | }
|
452 | 452 | },
|
453 | 453 | {
|
|
457 | 457 | "script_tags": "get,install,generic,generic-python-lib",
|
458 | 458 | "script_variations": "package.onnxruntime",
|
459 | 459 | "version": "1.18.1",
|
460 |
| - "parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_cuda,_sdxl,_tensorrt,_offline,_gpu_memory.80 )" |
| 460 | + "parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_tensorrt,_cuda,_sdxl,_offline,_gpu_memory.80 )" |
461 | 461 | }
|
462 | 462 | },
|
463 | 463 | {
|
|
467 | 467 | "script_tags": "get,install,generic,generic-python-lib",
|
468 | 468 | "script_variations": "package.colored",
|
469 | 469 | "version": "1.4.4",
|
470 |
| - "parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_cuda,_sdxl,_tensorrt,_offline,_gpu_memory.80 )" |
| 470 | + "parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_tensorrt,_cuda,_sdxl,_offline,_gpu_memory.80 )" |
471 | 471 | }
|
472 | 472 | },
|
473 | 473 | {
|
|
477 | 477 | "script_tags": "get,install,generic,generic-python-lib",
|
478 | 478 | "script_variations": "package.nvidia-ammo",
|
479 | 479 | "version": "0.7.4",
|
480 |
| - "parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_cuda,_sdxl,_tensorrt,_offline,_gpu_memory.80 )" |
| 480 | + "parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_tensorrt,_cuda,_sdxl,_offline,_gpu_memory.80 )" |
481 | 481 | }
|
482 | 482 | },
|
483 | 483 | {
|
|
487 | 487 | "script_tags": "get,install,generic,generic-python-lib",
|
488 | 488 | "script_variations": "package.optimum",
|
489 | 489 | "version": "1.16.0",
|
490 |
| - "parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_cuda,_sdxl,_tensorrt,_offline,_gpu_memory.80 )" |
| 490 | + "parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_tensorrt,_cuda,_sdxl,_offline,_gpu_memory.80 )" |
491 | 491 | }
|
492 | 492 | },
|
493 | 493 | {
|
|
497 | 497 | "script_tags": "get,install,generic,generic-python-lib",
|
498 | 498 | "script_variations": "package.onnx",
|
499 | 499 | "version": "1.14.0",
|
500 |
| - "parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_cuda,_sdxl,_tensorrt,_offline,_gpu_memory.80 )" |
| 500 | + "parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_tensorrt,_cuda,_sdxl,_offline,_gpu_memory.80 )" |
501 | 501 | }
|
502 | 502 | },
|
503 | 503 | {
|
|
507 | 507 | "script_tags": "get,install,generic,generic-python-lib",
|
508 | 508 | "script_variations": "package.scipy",
|
509 | 509 | "version": "1.10.1",
|
510 |
| - "parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_cuda,_sdxl,_tensorrt,_offline,_gpu_memory.80 )" |
| 510 | + "parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_tensorrt,_cuda,_sdxl,_offline,_gpu_memory.80 )" |
511 | 511 | }
|
512 | 512 | },
|
513 | 513 | {
|
|
517 | 517 | "script_tags": "get,install,generic,generic-python-lib",
|
518 | 518 | "script_variations": "package.numpy",
|
519 | 519 | "version": "1.22.0",
|
520 |
| - "parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_cuda,_sdxl,_tensorrt,_offline,_gpu_memory.80 )" |
| 520 | + "parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_tensorrt,_cuda,_sdxl,_offline,_gpu_memory.80 )" |
521 | 521 | }
|
522 | 522 | },
|
523 | 523 | {
|
|
557 | 557 | "script_tags": "mlperf,benchmark-mlperf",
|
558 | 558 | "script_variations": "",
|
559 | 559 | "version": "",
|
560 |
| - "parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_cuda,_sdxl,_tensorrt,_offline,_gpu_memory.80 )" |
| 560 | + "parent": "app-mlperf-inference-nvidia,bc3b17fb430f4732 ( run_harness,_tensorrt,_cuda,_sdxl,_offline,_gpu_memory.80 )" |
561 | 561 | }
|
562 | 562 | }
|
563 | 563 | ]
|
|
0 commit comments