summaryrefslogtreecommitdiff
path: root/notify/output-bmk-results.log
diff options
context:
space:
mode:
authorTCWG BuildSlave <tcwg-buildslave@linaro.org>2023-09-03 08:55:44 +0000
committerTCWG BuildSlave <tcwg-buildslave@linaro.org>2023-09-03 08:56:08 +0000
commit602a1077adbc47a0c8bcac8eea4f498b24e1bff8 (patch)
treeef7ee6341cc669cbf490dc8c1b52d575bf8ab6a2 /notify/output-bmk-results.log
parent357e79b84eb2f52e3e733a5e8359e9c5df3d43eb (diff)
onsuccess: #41: 1: [TCWG CI] https://ci.linaro.org/job/tcwg_bmk-code_size-spec2k6--llvm-aarch64-master-O2_LTO-build/41/
Results : | # reset_artifacts: | -10 | # build_bmk_llvm: | -3 | # benchmark -- -O2_LTO: | 1 check_regression status : 0
Diffstat (limited to 'notify/output-bmk-results.log')
-rw-r--r--notify/output-bmk-results.log150
1 files changed, 0 insertions, 150 deletions
diff --git a/notify/output-bmk-results.log b/notify/output-bmk-results.log
index f2654f5..a5b8a4a 100644
--- a/notify/output-bmk-results.log
+++ b/notify/output-bmk-results.log
@@ -153,21 +153,6 @@ output-bmk-results.py(111): return default_threshold[(change_kind,metric,mod
output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
output-bmk-results.py(213): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold))
output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
-DEBUG: checking exe.regression : 410.bwaves,bwaves_base.default : size=0% (threshold=1%)
-output-bmk-results.py(216): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold):
- --- modulename: output-bmk-results, funcname: is_entry_regression
-output-bmk-results.py(167): if metric in metric_utils.higher_regress_metrics:
-output-bmk-results.py(168): return (result - 100 > threshold)
-output-bmk-results.py(217): continue
-output-bmk-results.py(208): for index, row in out_df.iterrows():
-output-bmk-results.py(210): threshold = get_threshold(sym_type, metric, mode, row["benchmark"], row["symbol"])
- --- modulename: output-bmk-results, funcname: get_threshold
-output-bmk-results.py(98): if metric == "sample":
-output-bmk-results.py(108): if metric == "num_vect_loops" or metric == "num_sve_loops":
-output-bmk-results.py(111): return default_threshold[(change_kind,metric,mode)]
-output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
-output-bmk-results.py(213): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold))
-output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
DEBUG: checking exe.regression : 429.mcf,mcf_base.default : size=0% (threshold=1%)
output-bmk-results.py(216): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold):
--- modulename: output-bmk-results, funcname: is_entry_regression
@@ -198,66 +183,6 @@ output-bmk-results.py(111): return default_threshold[(change_kind,metric,mod
output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
output-bmk-results.py(213): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold))
output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
-DEBUG: checking exe.regression : 434.zeusmp,zeusmp_base.default : size=0% (threshold=1%)
-output-bmk-results.py(216): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold):
- --- modulename: output-bmk-results, funcname: is_entry_regression
-output-bmk-results.py(167): if metric in metric_utils.higher_regress_metrics:
-output-bmk-results.py(168): return (result - 100 > threshold)
-output-bmk-results.py(217): continue
-output-bmk-results.py(208): for index, row in out_df.iterrows():
-output-bmk-results.py(210): threshold = get_threshold(sym_type, metric, mode, row["benchmark"], row["symbol"])
- --- modulename: output-bmk-results, funcname: get_threshold
-output-bmk-results.py(98): if metric == "sample":
-output-bmk-results.py(108): if metric == "num_vect_loops" or metric == "num_sve_loops":
-output-bmk-results.py(111): return default_threshold[(change_kind,metric,mode)]
-output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
-output-bmk-results.py(213): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold))
-output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
-DEBUG: checking exe.regression : 435.gromacs,gromacs_base.default : size=0% (threshold=1%)
-output-bmk-results.py(216): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold):
- --- modulename: output-bmk-results, funcname: is_entry_regression
-output-bmk-results.py(167): if metric in metric_utils.higher_regress_metrics:
-output-bmk-results.py(168): return (result - 100 > threshold)
-output-bmk-results.py(217): continue
-output-bmk-results.py(208): for index, row in out_df.iterrows():
-output-bmk-results.py(210): threshold = get_threshold(sym_type, metric, mode, row["benchmark"], row["symbol"])
- --- modulename: output-bmk-results, funcname: get_threshold
-output-bmk-results.py(98): if metric == "sample":
-output-bmk-results.py(108): if metric == "num_vect_loops" or metric == "num_sve_loops":
-output-bmk-results.py(111): return default_threshold[(change_kind,metric,mode)]
-output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
-output-bmk-results.py(213): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold))
-output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
-DEBUG: checking exe.regression : 436.cactusADM,cactusADM_base.default : size=0% (threshold=1%)
-output-bmk-results.py(216): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold):
- --- modulename: output-bmk-results, funcname: is_entry_regression
-output-bmk-results.py(167): if metric in metric_utils.higher_regress_metrics:
-output-bmk-results.py(168): return (result - 100 > threshold)
-output-bmk-results.py(217): continue
-output-bmk-results.py(208): for index, row in out_df.iterrows():
-output-bmk-results.py(210): threshold = get_threshold(sym_type, metric, mode, row["benchmark"], row["symbol"])
- --- modulename: output-bmk-results, funcname: get_threshold
-output-bmk-results.py(98): if metric == "sample":
-output-bmk-results.py(108): if metric == "num_vect_loops" or metric == "num_sve_loops":
-output-bmk-results.py(111): return default_threshold[(change_kind,metric,mode)]
-output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
-output-bmk-results.py(213): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold))
-output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
-DEBUG: checking exe.regression : 437.leslie3d,leslie3d_base.default : size=0% (threshold=1%)
-output-bmk-results.py(216): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold):
- --- modulename: output-bmk-results, funcname: is_entry_regression
-output-bmk-results.py(167): if metric in metric_utils.higher_regress_metrics:
-output-bmk-results.py(168): return (result - 100 > threshold)
-output-bmk-results.py(217): continue
-output-bmk-results.py(208): for index, row in out_df.iterrows():
-output-bmk-results.py(210): threshold = get_threshold(sym_type, metric, mode, row["benchmark"], row["symbol"])
- --- modulename: output-bmk-results, funcname: get_threshold
-output-bmk-results.py(98): if metric == "sample":
-output-bmk-results.py(108): if metric == "num_vect_loops" or metric == "num_sve_loops":
-output-bmk-results.py(111): return default_threshold[(change_kind,metric,mode)]
-output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
-output-bmk-results.py(213): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold))
-output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
DEBUG: checking exe.regression : 444.namd,namd_base.default : size=0% (threshold=1%)
output-bmk-results.py(216): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold):
--- modulename: output-bmk-results, funcname: is_entry_regression
@@ -531,21 +456,6 @@ output-bmk-results.py(111): return default_threshold[(change_kind,metric,mod
output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
output-bmk-results.py(213): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold))
output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
-DEBUG: checking exe.improvement : 410.bwaves,bwaves_base.default : size=0% (threshold=1%)
-output-bmk-results.py(216): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold):
- --- modulename: output-bmk-results, funcname: is_entry_improvement
-output-bmk-results.py(176): if metric in metric_utils.higher_regress_metrics:
-output-bmk-results.py(177): return (100 - result > threshold)
-output-bmk-results.py(217): continue
-output-bmk-results.py(208): for index, row in out_df.iterrows():
-output-bmk-results.py(210): threshold = get_threshold(sym_type, metric, mode, row["benchmark"], row["symbol"])
- --- modulename: output-bmk-results, funcname: get_threshold
-output-bmk-results.py(98): if metric == "sample":
-output-bmk-results.py(108): if metric == "num_vect_loops" or metric == "num_sve_loops":
-output-bmk-results.py(111): return default_threshold[(change_kind,metric,mode)]
-output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
-output-bmk-results.py(213): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold))
-output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
DEBUG: checking exe.improvement : 429.mcf,mcf_base.default : size=0% (threshold=1%)
output-bmk-results.py(216): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold):
--- modulename: output-bmk-results, funcname: is_entry_improvement
@@ -576,66 +486,6 @@ output-bmk-results.py(111): return default_threshold[(change_kind,metric,mod
output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
output-bmk-results.py(213): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold))
output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
-DEBUG: checking exe.improvement : 434.zeusmp,zeusmp_base.default : size=0% (threshold=1%)
-output-bmk-results.py(216): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold):
- --- modulename: output-bmk-results, funcname: is_entry_improvement
-output-bmk-results.py(176): if metric in metric_utils.higher_regress_metrics:
-output-bmk-results.py(177): return (100 - result > threshold)
-output-bmk-results.py(217): continue
-output-bmk-results.py(208): for index, row in out_df.iterrows():
-output-bmk-results.py(210): threshold = get_threshold(sym_type, metric, mode, row["benchmark"], row["symbol"])
- --- modulename: output-bmk-results, funcname: get_threshold
-output-bmk-results.py(98): if metric == "sample":
-output-bmk-results.py(108): if metric == "num_vect_loops" or metric == "num_sve_loops":
-output-bmk-results.py(111): return default_threshold[(change_kind,metric,mode)]
-output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
-output-bmk-results.py(213): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold))
-output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
-DEBUG: checking exe.improvement : 435.gromacs,gromacs_base.default : size=0% (threshold=1%)
-output-bmk-results.py(216): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold):
- --- modulename: output-bmk-results, funcname: is_entry_improvement
-output-bmk-results.py(176): if metric in metric_utils.higher_regress_metrics:
-output-bmk-results.py(177): return (100 - result > threshold)
-output-bmk-results.py(217): continue
-output-bmk-results.py(208): for index, row in out_df.iterrows():
-output-bmk-results.py(210): threshold = get_threshold(sym_type, metric, mode, row["benchmark"], row["symbol"])
- --- modulename: output-bmk-results, funcname: get_threshold
-output-bmk-results.py(98): if metric == "sample":
-output-bmk-results.py(108): if metric == "num_vect_loops" or metric == "num_sve_loops":
-output-bmk-results.py(111): return default_threshold[(change_kind,metric,mode)]
-output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
-output-bmk-results.py(213): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold))
-output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
-DEBUG: checking exe.improvement : 436.cactusADM,cactusADM_base.default : size=0% (threshold=1%)
-output-bmk-results.py(216): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold):
- --- modulename: output-bmk-results, funcname: is_entry_improvement
-output-bmk-results.py(176): if metric in metric_utils.higher_regress_metrics:
-output-bmk-results.py(177): return (100 - result > threshold)
-output-bmk-results.py(217): continue
-output-bmk-results.py(208): for index, row in out_df.iterrows():
-output-bmk-results.py(210): threshold = get_threshold(sym_type, metric, mode, row["benchmark"], row["symbol"])
- --- modulename: output-bmk-results, funcname: get_threshold
-output-bmk-results.py(98): if metric == "sample":
-output-bmk-results.py(108): if metric == "num_vect_loops" or metric == "num_sve_loops":
-output-bmk-results.py(111): return default_threshold[(change_kind,metric,mode)]
-output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
-output-bmk-results.py(213): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold))
-output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
-DEBUG: checking exe.improvement : 437.leslie3d,leslie3d_base.default : size=0% (threshold=1%)
-output-bmk-results.py(216): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold):
- --- modulename: output-bmk-results, funcname: is_entry_improvement
-output-bmk-results.py(176): if metric in metric_utils.higher_regress_metrics:
-output-bmk-results.py(177): return (100 - result > threshold)
-output-bmk-results.py(217): continue
-output-bmk-results.py(208): for index, row in out_df.iterrows():
-output-bmk-results.py(210): threshold = get_threshold(sym_type, metric, mode, row["benchmark"], row["symbol"])
- --- modulename: output-bmk-results, funcname: get_threshold
-output-bmk-results.py(98): if metric == "sample":
-output-bmk-results.py(108): if metric == "num_vect_loops" or metric == "num_sve_loops":
-output-bmk-results.py(111): return default_threshold[(change_kind,metric,mode)]
-output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
-output-bmk-results.py(213): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold))
-output-bmk-results.py(212): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
DEBUG: checking exe.improvement : 444.namd,namd_base.default : size=0% (threshold=1%)
output-bmk-results.py(216): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold):
--- modulename: output-bmk-results, funcname: is_entry_improvement