@@ -213,9 +213,7 @@ def parse(self, text: str) -> dict:
213
213
"metric.value" : value ,
214
214
}
215
215
)
216
- if maxrss := run ["metadata" ].get (
217
- "mem_max_rss" , js ["metadata" ].get ("mem_max_rss" , 0 )
218
- ):
216
+ if maxrss := run ["metadata" ].get ("mem_max_rss" , 0 ):
219
217
r .append (
220
218
{
221
219
"bench-suite" : self .suiteName ,
@@ -295,7 +293,7 @@ def parse(self, text: str) -> dict:
295
293
return r
296
294
297
295
298
- class PyPyJsonRule (mx_benchmark .Rule ):
296
+ class PyPyJsonRule (mx_benchmark .Rule , mx_benchmark . AveragingBenchmarkMixin ):
299
297
"""Parses a JSON file produced by the Unladen Swallow or PyPy benchmark harness and creates a measurement result."""
300
298
301
299
def __init__ (self , filename : str , suiteName : str ):
@@ -321,7 +319,7 @@ def parse(self, text: str) -> dict:
321
319
{
322
320
"bench-suite" : self .suiteName ,
323
321
"benchmark" : name ,
324
- "metric.name" : "time " ,
322
+ "metric.name" : "warmup " ,
325
323
"metric.unit" : "s" ,
326
324
"metric.score-function" : "id" ,
327
325
"metric.better" : "lower" ,
@@ -330,7 +328,7 @@ def parse(self, text: str) -> dict:
330
328
"metric.value" : value ,
331
329
}
332
330
)
333
-
331
+ self . addAverageAcrossLatestResults ( r )
334
332
return r
335
333
336
334
@@ -517,7 +515,6 @@ def _vmRun(self, vm, workdir, command, benchmarks, bmSuiteArgs):
517
515
"run" ,
518
516
"--inherit-environ" ,
519
517
"PIP_INDEX_URL,PIP_TRUSTED_HOST,PIP_TIMEOUT,PIP_RETRIES,LD_LIBRARY_PATH,LIBRARY_PATH,CPATH,PATH,PYPY_GC_MAX" ,
520
- "-m" ,
521
518
"-o" ,
522
519
json_file ,
523
520
* bms ,
0 commit comments