diff options
Diffstat (limited to 'notify/output-bmk-results.log')
-rw-r--r-- | notify/output-bmk-results.log | 132 |
1 files changed, 6 insertions, 126 deletions
diff --git a/notify/output-bmk-results.log b/notify/output-bmk-results.log index 8dbadd8..338a1f4 100644 --- a/notify/output-bmk-results.log +++ b/notify/output-bmk-results.log @@ -123,21 +123,6 @@ output-bmk-results.py(111): return default_threshold[(change_kind,metric,mod output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ output-bmk-results.py(213): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold)) output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ -DEBUG: checking exe.regression : 510.parest_r,parest_r_base.default : size=0% (threshold=1%) -output-bmk-results.py(216): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold): - --- modulename: output-bmk-results, funcname: is_entry_regression -output-bmk-results.py(167): if metric in metric_utils.higher_regress_metrics: -output-bmk-results.py(168): return (result - 100 > threshold) -output-bmk-results.py(217): continue -output-bmk-results.py(208): for index, row in out_df.iterrows(): -output-bmk-results.py(210): threshold = get_threshold(sym_type, metric, mode, row["benchmark"], row["symbol"]) - --- modulename: output-bmk-results, funcname: get_threshold -output-bmk-results.py(98): if metric == "sample": -output-bmk-results.py(108): if metric == "num_vect_loops" or metric == "num_sve_loops": -output-bmk-results.py(111): return default_threshold[(change_kind,metric,mode)] -output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ -output-bmk-results.py(213): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold)) -output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ DEBUG: checking exe.regression : 511.povray_r,povray_r_base.default : size=0% (threshold=1%) output-bmk-results.py(216): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold): --- modulename: output-bmk-results, funcname: is_entry_regression @@ -153,21 +138,6 @@ output-bmk-results.py(111): return default_threshold[(change_kind,metric,mod output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ output-bmk-results.py(213): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold)) output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ -DEBUG: checking exe.regression : 526.blender_r,blender_r_base.default : size=0% (threshold=1%) -output-bmk-results.py(216): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold): - --- modulename: output-bmk-results, funcname: is_entry_regression -output-bmk-results.py(167): if metric in metric_utils.higher_regress_metrics: -output-bmk-results.py(168): return (result - 100 > threshold) -output-bmk-results.py(217): continue -output-bmk-results.py(208): for index, row in out_df.iterrows(): -output-bmk-results.py(210): threshold = get_threshold(sym_type, metric, mode, row["benchmark"], row["symbol"]) - --- modulename: output-bmk-results, funcname: get_threshold -output-bmk-results.py(98): if metric == "sample": -output-bmk-results.py(108): if metric == "num_vect_loops" or metric == "num_sve_loops": -output-bmk-results.py(111): return default_threshold[(change_kind,metric,mode)] -output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ -output-bmk-results.py(213): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold)) -output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ DEBUG: checking exe.regression : 600.perlbench_s,perlbench_s_base.default : size=0% (threshold=1%) output-bmk-results.py(216): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold): --- modulename: output-bmk-results, funcname: is_entry_regression @@ -183,21 +153,6 @@ output-bmk-results.py(111): return default_threshold[(change_kind,metric,mod output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ output-bmk-results.py(213): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold)) output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ -DEBUG: checking exe.regression : 602.gcc_s,sgcc_base.default : size=0% (threshold=1%) -output-bmk-results.py(216): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold): - --- modulename: output-bmk-results, funcname: is_entry_regression -output-bmk-results.py(167): if metric in metric_utils.higher_regress_metrics: -output-bmk-results.py(168): return (result - 100 > threshold) -output-bmk-results.py(217): continue -output-bmk-results.py(208): for index, row in out_df.iterrows(): -output-bmk-results.py(210): threshold = get_threshold(sym_type, metric, mode, row["benchmark"], row["symbol"]) - --- modulename: output-bmk-results, funcname: get_threshold -output-bmk-results.py(98): if metric == "sample": -output-bmk-results.py(108): if metric == "num_vect_loops" or metric == "num_sve_loops": -output-bmk-results.py(111): return default_threshold[(change_kind,metric,mode)] -output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ -output-bmk-results.py(213): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold)) -output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ DEBUG: checking exe.regression : 605.mcf_s,mcf_s_base.default : size=0% (threshold=1%) output-bmk-results.py(216): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold): --- modulename: output-bmk-results, funcname: is_entry_regression @@ -258,21 +213,6 @@ output-bmk-results.py(111): return default_threshold[(change_kind,metric,mod output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ output-bmk-results.py(213): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold)) output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ -DEBUG: checking exe.regression : 623.xalancbmk_s,xalancbmk_s_base.default : size=0% (threshold=1%) -output-bmk-results.py(216): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold): - --- modulename: output-bmk-results, funcname: is_entry_regression -output-bmk-results.py(167): if metric in metric_utils.higher_regress_metrics: -output-bmk-results.py(168): return (result - 100 > threshold) -output-bmk-results.py(217): continue -output-bmk-results.py(208): for index, row in out_df.iterrows(): -output-bmk-results.py(210): threshold = get_threshold(sym_type, metric, mode, row["benchmark"], row["symbol"]) - --- modulename: output-bmk-results, funcname: get_threshold -output-bmk-results.py(98): if metric == "sample": -output-bmk-results.py(108): if metric == "num_vect_loops" or metric == "num_sve_loops": -output-bmk-results.py(111): return default_threshold[(change_kind,metric,mode)] -output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ -output-bmk-results.py(213): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold)) -output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ DEBUG: checking exe.regression : 625.x264_s,x264_s_base.default : size=0% (threshold=1%) output-bmk-results.py(216): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold): --- modulename: output-bmk-results, funcname: is_entry_regression @@ -396,21 +336,6 @@ output-bmk-results.py(111): return default_threshold[(change_kind,metric,mod output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ output-bmk-results.py(213): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold)) output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ -DEBUG: checking exe.improvement : 510.parest_r,parest_r_base.default : size=0% (threshold=1%) -output-bmk-results.py(216): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold): - --- modulename: output-bmk-results, funcname: is_entry_improvement -output-bmk-results.py(176): if metric in metric_utils.higher_regress_metrics: -output-bmk-results.py(177): return (100 - result > threshold) -output-bmk-results.py(217): continue -output-bmk-results.py(208): for index, row in out_df.iterrows(): -output-bmk-results.py(210): threshold = get_threshold(sym_type, metric, mode, row["benchmark"], row["symbol"]) - --- modulename: output-bmk-results, funcname: get_threshold -output-bmk-results.py(98): if metric == "sample": -output-bmk-results.py(108): if metric == "num_vect_loops" or metric == "num_sve_loops": -output-bmk-results.py(111): return default_threshold[(change_kind,metric,mode)] -output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ -output-bmk-results.py(213): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold)) -output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ DEBUG: checking exe.improvement : 511.povray_r,povray_r_base.default : size=0% (threshold=1%) output-bmk-results.py(216): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold): --- modulename: output-bmk-results, funcname: is_entry_improvement @@ -426,21 +351,6 @@ output-bmk-results.py(111): return default_threshold[(change_kind,metric,mod output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ output-bmk-results.py(213): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold)) output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ -DEBUG: checking exe.improvement : 526.blender_r,blender_r_base.default : size=0% (threshold=1%) -output-bmk-results.py(216): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold): - --- modulename: output-bmk-results, funcname: is_entry_improvement -output-bmk-results.py(176): if metric in metric_utils.higher_regress_metrics: -output-bmk-results.py(177): return (100 - result > threshold) -output-bmk-results.py(217): continue -output-bmk-results.py(208): for index, row in out_df.iterrows(): -output-bmk-results.py(210): threshold = get_threshold(sym_type, metric, mode, row["benchmark"], row["symbol"]) - --- modulename: output-bmk-results, funcname: get_threshold -output-bmk-results.py(98): if metric == "sample": -output-bmk-results.py(108): if metric == "num_vect_loops" or metric == "num_sve_loops": -output-bmk-results.py(111): return default_threshold[(change_kind,metric,mode)] -output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ -output-bmk-results.py(213): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold)) -output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ DEBUG: checking exe.improvement : 600.perlbench_s,perlbench_s_base.default : size=0% (threshold=1%) output-bmk-results.py(216): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold): --- modulename: output-bmk-results, funcname: is_entry_improvement @@ -456,21 +366,6 @@ output-bmk-results.py(111): return default_threshold[(change_kind,metric,mod output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ output-bmk-results.py(213): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold)) output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ -DEBUG: checking exe.improvement : 602.gcc_s,sgcc_base.default : size=0% (threshold=1%) -output-bmk-results.py(216): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold): - --- modulename: output-bmk-results, funcname: is_entry_improvement -output-bmk-results.py(176): if metric in metric_utils.higher_regress_metrics: -output-bmk-results.py(177): return (100 - result > threshold) -output-bmk-results.py(217): continue -output-bmk-results.py(208): for index, row in out_df.iterrows(): -output-bmk-results.py(210): threshold = get_threshold(sym_type, metric, mode, row["benchmark"], row["symbol"]) - --- modulename: output-bmk-results, funcname: get_threshold -output-bmk-results.py(98): if metric == "sample": -output-bmk-results.py(108): if metric == "num_vect_loops" or metric == "num_sve_loops": -output-bmk-results.py(111): return default_threshold[(change_kind,metric,mode)] -output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ -output-bmk-results.py(213): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold)) -output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ DEBUG: checking exe.improvement : 605.mcf_s,mcf_s_base.default : size=0% (threshold=1%) output-bmk-results.py(216): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold): --- modulename: output-bmk-results, funcname: is_entry_improvement @@ -531,21 +426,6 @@ output-bmk-results.py(111): return default_threshold[(change_kind,metric,mod output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ output-bmk-results.py(213): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold)) output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ -DEBUG: checking exe.improvement : 623.xalancbmk_s,xalancbmk_s_base.default : size=0% (threshold=1%) -output-bmk-results.py(216): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold): - --- modulename: output-bmk-results, funcname: is_entry_improvement -output-bmk-results.py(176): if metric in metric_utils.higher_regress_metrics: -output-bmk-results.py(177): return (100 - result > threshold) -output-bmk-results.py(217): continue -output-bmk-results.py(208): for index, row in out_df.iterrows(): -output-bmk-results.py(210): threshold = get_threshold(sym_type, metric, mode, row["benchmark"], row["symbol"]) - --- modulename: output-bmk-results, funcname: get_threshold -output-bmk-results.py(98): if metric == "sample": -output-bmk-results.py(108): if metric == "num_vect_loops" or metric == "num_sve_loops": -output-bmk-results.py(111): return default_threshold[(change_kind,metric,mode)] -output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ -output-bmk-results.py(213): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold)) -output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ DEBUG: checking exe.improvement : 625.x264_s,x264_s_base.default : size=0% (threshold=1%) output-bmk-results.py(216): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold): --- modulename: output-bmk-results, funcname: is_entry_improvement @@ -654,7 +534,7 @@ output-bmk-results.py(111): return default_threshold[(change_kind,metric,mod output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ output-bmk-results.py(213): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold)) output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ -DEBUG: checking symbol.regression : 607.cactuBSSN_s,libomp.so : size=-1% (threshold=10%) +DEBUG: checking symbol.regression : 607.cactuBSSN_s,libomp.so : size=0% (threshold=10%) output-bmk-results.py(216): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold): --- modulename: output-bmk-results, funcname: is_entry_regression output-bmk-results.py(167): if metric in metric_utils.higher_regress_metrics: @@ -669,7 +549,7 @@ output-bmk-results.py(111): return default_threshold[(change_kind,metric,mod output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ output-bmk-results.py(213): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold)) output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ -DEBUG: checking symbol.regression : 619.lbm_s,libomp.so : size=-1% (threshold=10%) +DEBUG: checking symbol.regression : 619.lbm_s,libomp.so : size=0% (threshold=10%) output-bmk-results.py(216): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold): --- modulename: output-bmk-results, funcname: is_entry_regression output-bmk-results.py(167): if metric in metric_utils.higher_regress_metrics: @@ -714,7 +594,7 @@ output-bmk-results.py(111): return default_threshold[(change_kind,metric,mod output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ output-bmk-results.py(213): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold)) output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ -DEBUG: checking symbol.regression : 644.nab_s,libomp.so : size=-1% (threshold=10%) +DEBUG: checking symbol.regression : 644.nab_s,libomp.so : size=0% (threshold=10%) output-bmk-results.py(216): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold): --- modulename: output-bmk-results, funcname: is_entry_regression output-bmk-results.py(167): if metric in metric_utils.higher_regress_metrics: @@ -747,7 +627,7 @@ output-bmk-results.py(111): return default_threshold[(change_kind,metric,mod output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ output-bmk-results.py(213): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold)) output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ -DEBUG: checking symbol.improvement : 607.cactuBSSN_s,libomp.so : size=-1% (threshold=10%) +DEBUG: checking symbol.improvement : 607.cactuBSSN_s,libomp.so : size=0% (threshold=10%) output-bmk-results.py(216): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold): --- modulename: output-bmk-results, funcname: is_entry_improvement output-bmk-results.py(176): if metric in metric_utils.higher_regress_metrics: @@ -762,7 +642,7 @@ output-bmk-results.py(111): return default_threshold[(change_kind,metric,mod output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ output-bmk-results.py(213): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold)) output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ -DEBUG: checking symbol.improvement : 619.lbm_s,libomp.so : size=-1% (threshold=10%) +DEBUG: checking symbol.improvement : 619.lbm_s,libomp.so : size=0% (threshold=10%) output-bmk-results.py(216): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold): --- modulename: output-bmk-results, funcname: is_entry_improvement output-bmk-results.py(176): if metric in metric_utils.higher_regress_metrics: @@ -807,7 +687,7 @@ output-bmk-results.py(111): return default_threshold[(change_kind,metric,mod output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ output-bmk-results.py(213): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold)) output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ -DEBUG: checking symbol.improvement : 644.nab_s,libomp.so : size=-1% (threshold=10%) +DEBUG: checking symbol.improvement : 644.nab_s,libomp.so : size=0% (threshold=10%) output-bmk-results.py(216): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold): --- modulename: output-bmk-results, funcname: is_entry_improvement output-bmk-results.py(176): if metric in metric_utils.higher_regress_metrics: |