diff options
author | TCWG BuildSlave <tcwg-buildslave@linaro.org> | 2023-09-03 08:55:44 +0000 |
---|---|---|
committer | TCWG BuildSlave <tcwg-buildslave@linaro.org> | 2023-09-03 08:56:08 +0000 |
commit | 602a1077adbc47a0c8bcac8eea4f498b24e1bff8 (patch) | |
tree | ef7ee6341cc669cbf490dc8c1b52d575bf8ab6a2 /notify/output-bmk-results.log | |
parent | 357e79b84eb2f52e3e733a5e8359e9c5df3d43eb (diff) |
onsuccess: #41: 1: [TCWG CI] https://ci.linaro.org/job/tcwg_bmk-code_size-spec2k6--llvm-aarch64-master-O2_LTO-build/41/
Results :
| # reset_artifacts:
| -10
| # build_bmk_llvm:
| -3
| # benchmark -- -O2_LTO:
| 1
check_regression status : 0
Diffstat (limited to 'notify/output-bmk-results.log')
-rw-r--r-- | notify/output-bmk-results.log | 150 |
1 files changed, 0 insertions, 150 deletions
diff --git a/notify/output-bmk-results.log b/notify/output-bmk-results.log index f2654f5..a5b8a4a 100644 --- a/notify/output-bmk-results.log +++ b/notify/output-bmk-results.log @@ -153,21 +153,6 @@ output-bmk-results.py(111): return default_threshold[(change_kind,metric,mod output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ output-bmk-results.py(213): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold)) output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ -DEBUG: checking exe.regression : 410.bwaves,bwaves_base.default : size=0% (threshold=1%) -output-bmk-results.py(216): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold): - --- modulename: output-bmk-results, funcname: is_entry_regression -output-bmk-results.py(167): if metric in metric_utils.higher_regress_metrics: -output-bmk-results.py(168): return (result - 100 > threshold) -output-bmk-results.py(217): continue -output-bmk-results.py(208): for index, row in out_df.iterrows(): -output-bmk-results.py(210): threshold = get_threshold(sym_type, metric, mode, row["benchmark"], row["symbol"]) - --- modulename: output-bmk-results, funcname: get_threshold -output-bmk-results.py(98): if metric == "sample": -output-bmk-results.py(108): if metric == "num_vect_loops" or metric == "num_sve_loops": -output-bmk-results.py(111): return default_threshold[(change_kind,metric,mode)] -output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ -output-bmk-results.py(213): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold)) -output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ DEBUG: checking exe.regression : 429.mcf,mcf_base.default : size=0% (threshold=1%) output-bmk-results.py(216): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold): --- modulename: output-bmk-results, funcname: is_entry_regression @@ -198,66 +183,6 @@ output-bmk-results.py(111): return default_threshold[(change_kind,metric,mod output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ output-bmk-results.py(213): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold)) output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ -DEBUG: checking exe.regression : 434.zeusmp,zeusmp_base.default : size=0% (threshold=1%) -output-bmk-results.py(216): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold): - --- modulename: output-bmk-results, funcname: is_entry_regression -output-bmk-results.py(167): if metric in metric_utils.higher_regress_metrics: -output-bmk-results.py(168): return (result - 100 > threshold) -output-bmk-results.py(217): continue -output-bmk-results.py(208): for index, row in out_df.iterrows(): -output-bmk-results.py(210): threshold = get_threshold(sym_type, metric, mode, row["benchmark"], row["symbol"]) - --- modulename: output-bmk-results, funcname: get_threshold -output-bmk-results.py(98): if metric == "sample": -output-bmk-results.py(108): if metric == "num_vect_loops" or metric == "num_sve_loops": -output-bmk-results.py(111): return default_threshold[(change_kind,metric,mode)] -output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ -output-bmk-results.py(213): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold)) -output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ -DEBUG: checking exe.regression : 435.gromacs,gromacs_base.default : size=0% (threshold=1%) -output-bmk-results.py(216): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold): - --- modulename: output-bmk-results, funcname: is_entry_regression -output-bmk-results.py(167): if metric in metric_utils.higher_regress_metrics: -output-bmk-results.py(168): return (result - 100 > threshold) -output-bmk-results.py(217): continue -output-bmk-results.py(208): for index, row in out_df.iterrows(): -output-bmk-results.py(210): threshold = get_threshold(sym_type, metric, mode, row["benchmark"], row["symbol"]) - --- modulename: output-bmk-results, funcname: get_threshold -output-bmk-results.py(98): if metric == "sample": -output-bmk-results.py(108): if metric == "num_vect_loops" or metric == "num_sve_loops": -output-bmk-results.py(111): return default_threshold[(change_kind,metric,mode)] -output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ -output-bmk-results.py(213): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold)) -output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ -DEBUG: checking exe.regression : 436.cactusADM,cactusADM_base.default : size=0% (threshold=1%) -output-bmk-results.py(216): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold): - --- modulename: output-bmk-results, funcname: is_entry_regression -output-bmk-results.py(167): if metric in metric_utils.higher_regress_metrics: -output-bmk-results.py(168): return (result - 100 > threshold) -output-bmk-results.py(217): continue -output-bmk-results.py(208): for index, row in out_df.iterrows(): -output-bmk-results.py(210): threshold = get_threshold(sym_type, metric, mode, row["benchmark"], row["symbol"]) - --- modulename: output-bmk-results, funcname: get_threshold -output-bmk-results.py(98): if metric == "sample": -output-bmk-results.py(108): if metric == "num_vect_loops" or metric == "num_sve_loops": -output-bmk-results.py(111): return default_threshold[(change_kind,metric,mode)] -output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ -output-bmk-results.py(213): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold)) -output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ -DEBUG: checking exe.regression : 437.leslie3d,leslie3d_base.default : size=0% (threshold=1%) -output-bmk-results.py(216): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold): - --- modulename: output-bmk-results, funcname: is_entry_regression -output-bmk-results.py(167): if metric in metric_utils.higher_regress_metrics: -output-bmk-results.py(168): return (result - 100 > threshold) -output-bmk-results.py(217): continue -output-bmk-results.py(208): for index, row in out_df.iterrows(): -output-bmk-results.py(210): threshold = get_threshold(sym_type, metric, mode, row["benchmark"], row["symbol"]) - --- modulename: output-bmk-results, funcname: get_threshold -output-bmk-results.py(98): if metric == "sample": -output-bmk-results.py(108): if metric == "num_vect_loops" or metric == "num_sve_loops": -output-bmk-results.py(111): return default_threshold[(change_kind,metric,mode)] -output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ -output-bmk-results.py(213): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold)) -output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ DEBUG: checking exe.regression : 444.namd,namd_base.default : size=0% (threshold=1%) output-bmk-results.py(216): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold): --- modulename: output-bmk-results, funcname: is_entry_regression @@ -531,21 +456,6 @@ output-bmk-results.py(111): return default_threshold[(change_kind,metric,mod output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ output-bmk-results.py(213): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold)) output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ -DEBUG: checking exe.improvement : 410.bwaves,bwaves_base.default : size=0% (threshold=1%) -output-bmk-results.py(216): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold): - --- modulename: output-bmk-results, funcname: is_entry_improvement -output-bmk-results.py(176): if metric in metric_utils.higher_regress_metrics: -output-bmk-results.py(177): return (100 - result > threshold) -output-bmk-results.py(217): continue -output-bmk-results.py(208): for index, row in out_df.iterrows(): -output-bmk-results.py(210): threshold = get_threshold(sym_type, metric, mode, row["benchmark"], row["symbol"]) - --- modulename: output-bmk-results, funcname: get_threshold -output-bmk-results.py(98): if metric == "sample": -output-bmk-results.py(108): if metric == "num_vect_loops" or metric == "num_sve_loops": -output-bmk-results.py(111): return default_threshold[(change_kind,metric,mode)] -output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ -output-bmk-results.py(213): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold)) -output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ DEBUG: checking exe.improvement : 429.mcf,mcf_base.default : size=0% (threshold=1%) output-bmk-results.py(216): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold): --- modulename: output-bmk-results, funcname: is_entry_improvement @@ -576,66 +486,6 @@ output-bmk-results.py(111): return default_threshold[(change_kind,metric,mod output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ output-bmk-results.py(213): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold)) output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ -DEBUG: checking exe.improvement : 434.zeusmp,zeusmp_base.default : size=0% (threshold=1%) -output-bmk-results.py(216): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold): - --- modulename: output-bmk-results, funcname: is_entry_improvement -output-bmk-results.py(176): if metric in metric_utils.higher_regress_metrics: -output-bmk-results.py(177): return (100 - result > threshold) -output-bmk-results.py(217): continue -output-bmk-results.py(208): for index, row in out_df.iterrows(): -output-bmk-results.py(210): threshold = get_threshold(sym_type, metric, mode, row["benchmark"], row["symbol"]) - --- modulename: output-bmk-results, funcname: get_threshold -output-bmk-results.py(98): if metric == "sample": -output-bmk-results.py(108): if metric == "num_vect_loops" or metric == "num_sve_loops": -output-bmk-results.py(111): return default_threshold[(change_kind,metric,mode)] -output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ -output-bmk-results.py(213): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold)) -output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ -DEBUG: checking exe.improvement : 435.gromacs,gromacs_base.default : size=0% (threshold=1%) -output-bmk-results.py(216): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold): - --- modulename: output-bmk-results, funcname: is_entry_improvement -output-bmk-results.py(176): if metric in metric_utils.higher_regress_metrics: -output-bmk-results.py(177): return (100 - result > threshold) -output-bmk-results.py(217): continue -output-bmk-results.py(208): for index, row in out_df.iterrows(): -output-bmk-results.py(210): threshold = get_threshold(sym_type, metric, mode, row["benchmark"], row["symbol"]) - --- modulename: output-bmk-results, funcname: get_threshold -output-bmk-results.py(98): if metric == "sample": -output-bmk-results.py(108): if metric == "num_vect_loops" or metric == "num_sve_loops": -output-bmk-results.py(111): return default_threshold[(change_kind,metric,mode)] -output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ -output-bmk-results.py(213): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold)) -output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ -DEBUG: checking exe.improvement : 436.cactusADM,cactusADM_base.default : size=0% (threshold=1%) -output-bmk-results.py(216): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold): - --- modulename: output-bmk-results, funcname: is_entry_improvement -output-bmk-results.py(176): if metric in metric_utils.higher_regress_metrics: -output-bmk-results.py(177): return (100 - result > threshold) -output-bmk-results.py(217): continue -output-bmk-results.py(208): for index, row in out_df.iterrows(): -output-bmk-results.py(210): threshold = get_threshold(sym_type, metric, mode, row["benchmark"], row["symbol"]) - --- modulename: output-bmk-results, funcname: get_threshold -output-bmk-results.py(98): if metric == "sample": -output-bmk-results.py(108): if metric == "num_vect_loops" or metric == "num_sve_loops": -output-bmk-results.py(111): return default_threshold[(change_kind,metric,mode)] -output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ -output-bmk-results.py(213): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold)) -output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ -DEBUG: checking exe.improvement : 437.leslie3d,leslie3d_base.default : size=0% (threshold=1%) -output-bmk-results.py(216): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold): - --- modulename: output-bmk-results, funcname: is_entry_improvement -output-bmk-results.py(176): if metric in metric_utils.higher_regress_metrics: -output-bmk-results.py(177): return (100 - result > threshold) -output-bmk-results.py(217): continue -output-bmk-results.py(208): for index, row in out_df.iterrows(): -output-bmk-results.py(210): threshold = get_threshold(sym_type, metric, mode, row["benchmark"], row["symbol"]) - --- modulename: output-bmk-results, funcname: get_threshold -output-bmk-results.py(98): if metric == "sample": -output-bmk-results.py(108): if metric == "num_vect_loops" or metric == "num_sve_loops": -output-bmk-results.py(111): return default_threshold[(change_kind,metric,mode)] -output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ -output-bmk-results.py(213): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold)) -output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ DEBUG: checking exe.improvement : 444.namd,namd_base.default : size=0% (threshold=1%) output-bmk-results.py(216): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold): --- modulename: output-bmk-results, funcname: is_entry_improvement |