summaryrefslogtreecommitdiff
path: root/notify
diff options
context:
space:
mode:
authorTCWG BuildSlave <tcwg-buildslave@linaro.org>2023-11-22 23:38:42 +0000
committerTCWG BuildSlave <tcwg-buildslave@linaro.org>2023-11-22 23:39:40 +0000
commit7167e8887de4b148e09752c76d68b053626d4bd7 (patch)
treef3163f3eb0fa0b128f5a7f46b82235d7a26bff12 /notify
parentc392062909627522c0da945e789af38801cf8e0f (diff)
onsuccess: #24: 1: [TCWG CI] https://ci.linaro.org/job/tcwg_bmk-code_speed-cpu2017rate--llvm-arm-master-O2-build/24/
Results : | # reset_artifacts: | -10 | # build_bmk_llvm: | -3 | # benchmark -- -O2_marm: | 1 check_regression status : 0
Diffstat (limited to 'notify')
-rw-r--r--notify/exe.improvement1
-rw-r--r--notify/extra-bisect-params2
-rw-r--r--notify/jira/comment-template.txt4
-rw-r--r--notify/lnt_report.json44
-rw-r--r--notify/mail-body.txt15
-rw-r--r--notify/mail-recipients.txt2
-rw-r--r--notify/mail-subject.txt2
-rw-r--r--notify/output-bmk-results.log185
-rw-r--r--notify/results.regressions2
-rw-r--r--notify/status.regression2
10 files changed, 86 insertions, 173 deletions
diff --git a/notify/exe.improvement b/notify/exe.improvement
deleted file mode 100644
index f0e5a5a..0000000
--- a/notify/exe.improvement
+++ /dev/null
@@ -1 +0,0 @@
-200,519.lbm_r,lbm_r_base.default,sped up by 200% - 519.lbm_r,sped up by 200% - 519.lbm_r - from 1 to -1 perf samples
diff --git a/notify/extra-bisect-params b/notify/extra-bisect-params
index b13b8ee..fa6c7c9 100644
--- a/notify/extra-bisect-params
+++ b/notify/extra-bisect-params
@@ -1 +1 @@
-extra_build_params=++benchmarks 519.lbm_r ++benchmarks 520.omnetpp_r
+extra_build_params=
diff --git a/notify/jira/comment-template.txt b/notify/jira/comment-template.txt
index a7af19d..f450703 100644
--- a/notify/jira/comment-template.txt
+++ b/notify/jira/comment-template.txt
@@ -1,3 +1,3 @@
[LLVM-651]
-sped up by 200% - 519.lbm_r
-Details: https://ci.linaro.org/job/tcwg_bmk-code_speed-cpu2017rate--llvm-arm-master-O2-build/23/artifact/artifacts/notify/mail-body.txt/*view*/
+No change
+Details: https://ci.linaro.org/job/tcwg_bmk-code_speed-cpu2017rate--llvm-arm-master-O2-build/24/artifact/artifacts/notify/mail-body.txt/*view*/
diff --git a/notify/lnt_report.json b/notify/lnt_report.json
index 12aaf55..2433925 100644
--- a/notify/lnt_report.json
+++ b/notify/lnt_report.json
@@ -6,14 +6,22 @@
"Run": {
"Info": {
"__report_version__": "1",
- "run_order": "llvmorg-17-init-08890-gefd71ebea4ba",
+ "run_order": "llvmorg-17-init-09356-g815eab2d3cbb",
"tag": "tcwg_bmk-code_speed-cpu2017rate"
},
- "Start Time": "2023-11-22 23:31:31"
+ "Start Time": "2023-11-22 23:38:36"
},
"Tests": [
{
"Data": [
+ 192787
+ ],
+ "Info": {},
+ "Name": "tcwg_bmk-code_speed-cpu2017rate.544.nab_r.code_size"
+ }
+ ,
+ {
+ "Data": [
29232
],
"Info": {},
@@ -22,42 +30,42 @@
,
{
"Data": [
- 3879608
+ 1719248
],
"Info": {},
- "Name": "tcwg_bmk-code_speed-cpu2017rate.523.xalancbmk_r.code_size"
+ "Name": "tcwg_bmk-code_speed-cpu2017rate.538.imagick_r.code_size"
}
,
{
"Data": [
- 91600
+ 135865
],
"Info": {},
- "Name": "tcwg_bmk-code_speed-cpu2017rate.531.deepsjeng_r.code_size"
+ "Name": "tcwg_bmk-code_speed-cpu2017rate.541.leela_r.code_size"
}
,
{
"Data": [
- 1719272
+ 177927
],
"Info": {},
- "Name": "tcwg_bmk-code_speed-cpu2017rate.538.imagick_r.code_size"
+ "Name": "tcwg_bmk-code_speed-cpu2017rate.557.xz_r.code_size"
}
,
{
"Data": [
- 135849
+ 91600
],
"Info": {},
- "Name": "tcwg_bmk-code_speed-cpu2017rate.541.leela_r.code_size"
+ "Name": "tcwg_bmk-code_speed-cpu2017rate.531.deepsjeng_r.code_size"
}
,
{
"Data": [
- 177783
+ 3879556
],
"Info": {},
- "Name": "tcwg_bmk-code_speed-cpu2017rate.557.xz_r.code_size"
+ "Name": "tcwg_bmk-code_speed-cpu2017rate.523.xalancbmk_r.code_size"
}
,
{
@@ -70,26 +78,26 @@
,
{
"Data": [
- 10112
+ 10286
],
"Info": {},
- "Name": "tcwg_bmk-code_speed-cpu2017rate.557.xz_r.exec"
+ "Name": "tcwg_bmk-code_speed-cpu2017rate.531.deepsjeng_r.exec"
}
,
{
"Data": [
- 10355
+ 14010
],
"Info": {},
- "Name": "tcwg_bmk-code_speed-cpu2017rate.531.deepsjeng_r.exec"
+ "Name": "tcwg_bmk-code_speed-cpu2017rate.505.mcf_r.exec"
}
,
{
"Data": [
- 14049
+ 10148
],
"Info": {},
- "Name": "tcwg_bmk-code_speed-cpu2017rate.505.mcf_r.exec"
+ "Name": "tcwg_bmk-code_speed-cpu2017rate.557.xz_r.exec"
}
]
}
diff --git a/notify/mail-body.txt b/notify/mail-body.txt
index 2017cd4..32ba936 100644
--- a/notify/mail-body.txt
+++ b/notify/mail-body.txt
@@ -2,10 +2,15 @@ Dear contributor, our automatic CI has detected problems related to your patch(e
In CI config tcwg_bmk-code_speed-cpu2017rate/llvm-arm-master-O2 after:
- | baseline build
+ | 466 commits in llvm
+ | 815eab2d3cbb [DebugLocEntry][nfc] Remove redundant cast
+ | fea8c073561f [Support][Parallel] Add sequential mode to TaskGroup::spawn().
+ | 329bfcc8df47 [LangRef] Improve wording DW_OP_LLVM_entry_value
+ | e23891a3823e [AMDGPU][Disassembler] Fix a spurious error message in an instruction comment.
+ | c08dc8b5fbd0 [AMDGPU][Disassembler] Pre-commit a test on an error message generated in an instruction comment.
+ | ... and 461 more commits in llvm
-the following benchmarks speeds up by more than 3%:
-- sped up by 200% - 519.lbm_r - from 1 to -1 perf samples
+No change
The configuration of this build is:
Below reproducer instructions can be used to re-build both "first_bad" and "last_good" cross-toolchains used in this bisection. Naturally, the scripts will fail when triggerring benchmarking jobs if you don\'t have access to Linaro TCWG CI.
@@ -23,6 +28,6 @@ This benchmarking CI is work-in-progress, and we welcome feedback and suggestion
-----------------8<--------------------------8<--------------------------8<--------------------------
The information below can be used to reproduce a debug environment:
-Current build : https://ci.linaro.org/job/tcwg_bmk-code_speed-cpu2017rate--llvm-arm-master-O2-build/23/artifact/artifacts
-Reference build : https://ci.linaro.org/job/tcwg_bmk-code_speed-cpu2017rate--llvm-arm-master-O2-build/21/artifact/artifacts
+Current build : https://ci.linaro.org/job/tcwg_bmk-code_speed-cpu2017rate--llvm-arm-master-O2-build/24/artifact/artifacts
+Reference build : https://ci.linaro.org/job/tcwg_bmk-code_speed-cpu2017rate--llvm-arm-master-O2-build/23/artifact/artifacts
diff --git a/notify/mail-recipients.txt b/notify/mail-recipients.txt
index aa219ef..56b65bb 100644
--- a/notify/mail-recipients.txt
+++ b/notify/mail-recipients.txt
@@ -1 +1 @@
-bcc:tcwg-validation@linaro.org
+bcc:tcwg-validation@linaro.org,maxim.kuvyrkov@linaro.org
diff --git a/notify/mail-subject.txt b/notify/mail-subject.txt
index f5b1b05..dacb113 100644
--- a/notify/mail-subject.txt
+++ b/notify/mail-subject.txt
@@ -1 +1 @@
-[Linaro-TCWG-CI] baseline build: sped up by 200% - 519.lbm_r on arm O2
+[Linaro-TCWG-CI] 466 commits in llvm: No change on arm O2
diff --git a/notify/output-bmk-results.log b/notify/output-bmk-results.log
index 77a9e4a..5a78195 100644
--- a/notify/output-bmk-results.log
+++ b/notify/output-bmk-results.log
@@ -109,8 +109,8 @@ output-bmk-results.py(258): print(results_df)
7 508.namd_r namd_r_base.default ... failed-to-run failed-to-run
8 510.parest_r parest_r_base.default ... failed-to-run failed-to-run
9 511.povray_r povray_r_base.default ... failed-to-run failed-to-run
-10 519.lbm_r lbm_r_base.default ... success failed-to-run
-11 520.omnetpp_r omnetpp_r_base.default ... success failed-to-run
+10 519.lbm_r lbm_r_base.default ... failed-to-run failed-to-run
+11 520.omnetpp_r omnetpp_r_base.default ... failed-to-run failed-to-run
12 523.xalancbmk_r cpuxalan_r_base.default ... failed-to-run failed-to-run
13 525.x264_r x264_r_base.default ... failed-to-run failed-to-run
14 526.blender_r blender_r_base.default ... failed-to-run failed-to-run
@@ -213,25 +213,12 @@ output-bmk-results.py(115): short_diag=""
output-bmk-results.py(116): classif=""
output-bmk-results.py(118): if row["status_x"]!="failed-to-build" and row["status_y"]=="failed-to-build":
output-bmk-results.py(121): elif row["status_x"]=="success" and row["status_y"]=="failed-to-run":
-output-bmk-results.py(122): short_diag = "{0} failed to run".format(bmk)
-output-bmk-results.py(123): classif="regression"
+output-bmk-results.py(124): elif row["status_x"]=="failed-to-build" and row["status_y"]=="failed-to-run":
+output-bmk-results.py(127): elif row["status_x"]=="failed-to-run" and row["status_y"]=="success":
+output-bmk-results.py(130): elif row["status_x"]=="failed-to-build" and row["status_y"]=="success":
output-bmk-results.py(134): return classif, short_diag
output-bmk-results.py(265): if classif != change_kind:
-output-bmk-results.py(268): print("DEBUG: *** {0},{1} : {2}".format(row["benchmark"], row["symbol"], short_diag))
-DEBUG: *** 519.lbm_r,lbm_r_base.default : 519.lbm_r failed to run
-output-bmk-results.py(270): f_out.write_csv((100, row["benchmark"], row["symbol"], short_diag, short_diag))
- --- modulename: output-bmk-results, funcname: write_csv
-output-bmk-results.py(41): if not self.predicate or not self.csvwriter:
-output-bmk-results.py(43): self.csvwriter.writerow(arr)
-output-bmk-results.py(271): if change_kind == "regression":
-output-bmk-results.py(272): f_regr.write("# {0},{1}\n".format(row["symbol"], short_diag))
- --- modulename: output-bmk-results, funcname: write
-output-bmk-results.py(36): if not self.predicate or not self.outf:
-output-bmk-results.py(38): self.outf.write(string)
-output-bmk-results.py(273): f_ebp.write("++benchmarks {0} ".format(row["benchmark"]))
- --- modulename: output-bmk-results, funcname: write
-output-bmk-results.py(36): if not self.predicate or not self.outf:
-output-bmk-results.py(38): self.outf.write(string)
+output-bmk-results.py(266): continue;
output-bmk-results.py(261): for index, row in results_df.iterrows():
output-bmk-results.py(263): classif, short_diag = get_status_diag(row)
--- modulename: output-bmk-results, funcname: get_status_diag
@@ -240,25 +227,12 @@ output-bmk-results.py(115): short_diag=""
output-bmk-results.py(116): classif=""
output-bmk-results.py(118): if row["status_x"]!="failed-to-build" and row["status_y"]=="failed-to-build":
output-bmk-results.py(121): elif row["status_x"]=="success" and row["status_y"]=="failed-to-run":
-output-bmk-results.py(122): short_diag = "{0} failed to run".format(bmk)
-output-bmk-results.py(123): classif="regression"
+output-bmk-results.py(124): elif row["status_x"]=="failed-to-build" and row["status_y"]=="failed-to-run":
+output-bmk-results.py(127): elif row["status_x"]=="failed-to-run" and row["status_y"]=="success":
+output-bmk-results.py(130): elif row["status_x"]=="failed-to-build" and row["status_y"]=="success":
output-bmk-results.py(134): return classif, short_diag
output-bmk-results.py(265): if classif != change_kind:
-output-bmk-results.py(268): print("DEBUG: *** {0},{1} : {2}".format(row["benchmark"], row["symbol"], short_diag))
-DEBUG: *** 520.omnetpp_r,omnetpp_r_base.default : 520.omnetpp_r failed to run
-output-bmk-results.py(270): f_out.write_csv((100, row["benchmark"], row["symbol"], short_diag, short_diag))
- --- modulename: output-bmk-results, funcname: write_csv
-output-bmk-results.py(41): if not self.predicate or not self.csvwriter:
-output-bmk-results.py(43): self.csvwriter.writerow(arr)
-output-bmk-results.py(271): if change_kind == "regression":
-output-bmk-results.py(272): f_regr.write("# {0},{1}\n".format(row["symbol"], short_diag))
- --- modulename: output-bmk-results, funcname: write
-output-bmk-results.py(36): if not self.predicate or not self.outf:
-output-bmk-results.py(38): self.outf.write(string)
-output-bmk-results.py(273): f_ebp.write("++benchmarks {0} ".format(row["benchmark"]))
- --- modulename: output-bmk-results, funcname: write
-output-bmk-results.py(36): if not self.predicate or not self.outf:
-output-bmk-results.py(38): self.outf.write(string)
+output-bmk-results.py(266): continue;
output-bmk-results.py(261): for index, row in results_df.iterrows():
output-bmk-results.py(263): classif, short_diag = get_status_diag(row)
--- modulename: output-bmk-results, funcname: get_status_diag
@@ -377,6 +351,7 @@ output-bmk-results.py(275): f_out.close()
output-bmk-results.py(29): if not self.outf:
output-bmk-results.py(31): self.outf.close()
output-bmk-results.py(32): if os.stat(self.filename).st_size == 0:
+output-bmk-results.py(33): os.remove(self.filename)
output-bmk-results.py(294): output_bmk_results_status(exe_df, "improvement", None, None, run_step_artifacts, details)
--- modulename: output-bmk-results, funcname: output_bmk_results_status
output-bmk-results.py(256): f_out = Outfile("{0}/status.{1}".format(run_step_artifacts, change_kind), "w", predicate=(details=="verbose"))
@@ -394,8 +369,8 @@ output-bmk-results.py(258): print(results_df)
7 508.namd_r namd_r_base.default ... failed-to-run failed-to-run
8 510.parest_r parest_r_base.default ... failed-to-run failed-to-run
9 511.povray_r povray_r_base.default ... failed-to-run failed-to-run
-10 519.lbm_r lbm_r_base.default ... success failed-to-run
-11 520.omnetpp_r omnetpp_r_base.default ... success failed-to-run
+10 519.lbm_r lbm_r_base.default ... failed-to-run failed-to-run
+11 520.omnetpp_r omnetpp_r_base.default ... failed-to-run failed-to-run
12 523.xalancbmk_r cpuxalan_r_base.default ... failed-to-run failed-to-run
13 525.x264_r x264_r_base.default ... failed-to-run failed-to-run
14 526.blender_r blender_r_base.default ... failed-to-run failed-to-run
@@ -498,8 +473,9 @@ output-bmk-results.py(115): short_diag=""
output-bmk-results.py(116): classif=""
output-bmk-results.py(118): if row["status_x"]!="failed-to-build" and row["status_y"]=="failed-to-build":
output-bmk-results.py(121): elif row["status_x"]=="success" and row["status_y"]=="failed-to-run":
-output-bmk-results.py(122): short_diag = "{0} failed to run".format(bmk)
-output-bmk-results.py(123): classif="regression"
+output-bmk-results.py(124): elif row["status_x"]=="failed-to-build" and row["status_y"]=="failed-to-run":
+output-bmk-results.py(127): elif row["status_x"]=="failed-to-run" and row["status_y"]=="success":
+output-bmk-results.py(130): elif row["status_x"]=="failed-to-build" and row["status_y"]=="success":
output-bmk-results.py(134): return classif, short_diag
output-bmk-results.py(265): if classif != change_kind:
output-bmk-results.py(266): continue;
@@ -511,8 +487,9 @@ output-bmk-results.py(115): short_diag=""
output-bmk-results.py(116): classif=""
output-bmk-results.py(118): if row["status_x"]!="failed-to-build" and row["status_y"]=="failed-to-build":
output-bmk-results.py(121): elif row["status_x"]=="success" and row["status_y"]=="failed-to-run":
-output-bmk-results.py(122): short_diag = "{0} failed to run".format(bmk)
-output-bmk-results.py(123): classif="regression"
+output-bmk-results.py(124): elif row["status_x"]=="failed-to-build" and row["status_y"]=="failed-to-run":
+output-bmk-results.py(127): elif row["status_x"]=="failed-to-run" and row["status_y"]=="success":
+output-bmk-results.py(130): elif row["status_x"]=="failed-to-build" and row["status_y"]=="success":
output-bmk-results.py(134): return classif, short_diag
output-bmk-results.py(265): if classif != change_kind:
output-bmk-results.py(266): continue;
@@ -684,30 +661,6 @@ output-bmk-results.py(60): var = specific_variability[ (specific_variability
output-bmk-results.py(61): if var.empty:
output-bmk-results.py(63): elif len(var)>1:
output-bmk-results.py(68): if var.iloc[0]['sample_variation_average']>0 :
-output-bmk-results.py(83): return np.nan
-output-bmk-results.py(100): if not np.isnan(spec_thr):
-output-bmk-results.py(107): if metric == "num_vect_loops" or metric == "num_sve_loops":
-output-bmk-results.py(110): return default_threshold[(change_kind,metric,mode)]
-output-bmk-results.py(228): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
-output-bmk-results.py(229): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold))
-output-bmk-results.py(228): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
-DEBUG: checking exe.regression : 519.lbm_r,lbm_r_base.default : sample=200% (threshold=3%)
-output-bmk-results.py(232): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold):
- --- modulename: output-bmk-results, funcname: is_entry_regression
-output-bmk-results.py(183): if metric in metric_utils.higher_regress_metrics:
-output-bmk-results.py(184): return (result - 100 > threshold)
-output-bmk-results.py(233): continue
-output-bmk-results.py(224): for index, row in out_df.iterrows():
-output-bmk-results.py(226): threshold = get_threshold(sym_type, metric, mode, row["benchmark"], row["symbol"])
- --- modulename: output-bmk-results, funcname: get_threshold
-output-bmk-results.py(98): if metric == "sample":
-output-bmk-results.py(99): spec_thr = get_specific_thresholds(metric, mode, bmk, symb)
- --- modulename: output-bmk-results, funcname: get_specific_thresholds
-output-bmk-results.py(57): if specific_variability is None:
-output-bmk-results.py(60): var = specific_variability[ (specific_variability['benchmark'] == bmk) & (specific_variability['symbol'].str.strip() == symb)]
-output-bmk-results.py(61): if var.empty:
-output-bmk-results.py(63): elif len(var)>1:
-output-bmk-results.py(68): if var.iloc[0]['sample_variation_average']>0 :
output-bmk-results.py(69): threshold = ( var.iloc[0]['sample_variation_average'] )
output-bmk-results.py(70): if mode == "build":
output-bmk-results.py(74): threshold *= 3
@@ -718,7 +671,7 @@ output-bmk-results.py(105): return spec_thr
output-bmk-results.py(228): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
output-bmk-results.py(229): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold))
output-bmk-results.py(228): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
-DEBUG: checking exe.regression : 531.deepsjeng_r,deepsjeng_r_base.default : sample=-1% (threshold=3%)
+DEBUG: checking exe.regression : 531.deepsjeng_r,deepsjeng_r_base.default : sample=1% (threshold=3%)
output-bmk-results.py(232): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold):
--- modulename: output-bmk-results, funcname: is_entry_regression
output-bmk-results.py(183): if metric in metric_utils.higher_regress_metrics:
@@ -807,56 +760,6 @@ output-bmk-results.py(60): var = specific_variability[ (specific_variability
output-bmk-results.py(61): if var.empty:
output-bmk-results.py(63): elif len(var)>1:
output-bmk-results.py(68): if var.iloc[0]['sample_variation_average']>0 :
-output-bmk-results.py(83): return np.nan
-output-bmk-results.py(100): if not np.isnan(spec_thr):
-output-bmk-results.py(107): if metric == "num_vect_loops" or metric == "num_sve_loops":
-output-bmk-results.py(110): return default_threshold[(change_kind,metric,mode)]
-output-bmk-results.py(228): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
-output-bmk-results.py(229): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold))
-output-bmk-results.py(228): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
-DEBUG: checking exe.improvement : 519.lbm_r,lbm_r_base.default : sample=200% (threshold=3%)
-output-bmk-results.py(232): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold):
- --- modulename: output-bmk-results, funcname: is_entry_improvement
-output-bmk-results.py(192): if metric in metric_utils.higher_regress_metrics:
-output-bmk-results.py(193): return (100 - result > threshold)
-output-bmk-results.py(235): percent_change, short_diag, long_diag = get_short_long_diag(row, metric, sym_type, change_kind)
- --- modulename: output-bmk-results, funcname: get_short_long_diag
-output-bmk-results.py(137): bmk = row["benchmark"]
-output-bmk-results.py(139): rel_value = row["rel_" + metric]
-output-bmk-results.py(140): prev_value = row[metric + "_x"]
-output-bmk-results.py(141): curr_value = row[metric + "_y"]
-output-bmk-results.py(142): if metric == "num_vect_loops" or metric == "num_sve_loops":
-output-bmk-results.py(152): suffix = ""
-output-bmk-results.py(153): if metric == "sample":
-output-bmk-results.py(154): prefix_regression = "slowed down by"
-output-bmk-results.py(155): prefix_improvement = "sped up by"
-output-bmk-results.py(156): suffix = "perf samples"
-output-bmk-results.py(167): if sym_type=="symbol":
-output-bmk-results.py(170): item=bmk
-output-bmk-results.py(172): short_diag = "{1} {2}% - {0}".format(item, locals()["prefix_" + change_kind], abs(rel_value - 100))
-output-bmk-results.py(173): long_diag = "{0} - from {1} to {2} {3}".format(short_diag, prev_value, curr_value, suffix)
-output-bmk-results.py(174): return abs(rel_value - 100), short_diag, long_diag
-output-bmk-results.py(239): if metric == "sample" \
-output-bmk-results.py(240): and row['symbol_md5sum_x'] == row['symbol_md5sum_y'] \
-output-bmk-results.py(241): and row['symbol_md5sum_x'] != "-1" \
-output-bmk-results.py(246): print("DEBUG: *** {0},{1} : {2}".format(row["benchmark"], row["symbol"], long_diag))
-DEBUG: *** 519.lbm_r,lbm_r_base.default : sped up by 200% - 519.lbm_r - from 1 to -1 perf samples
-output-bmk-results.py(248): f_out.write_csv((percent_change, row["benchmark"], row["symbol"], short_diag, long_diag))
- --- modulename: output-bmk-results, funcname: write_csv
-output-bmk-results.py(41): if not self.predicate or not self.csvwriter:
-output-bmk-results.py(43): self.csvwriter.writerow(arr)
-output-bmk-results.py(249): if change_kind == "regression":
-output-bmk-results.py(224): for index, row in out_df.iterrows():
-output-bmk-results.py(226): threshold = get_threshold(sym_type, metric, mode, row["benchmark"], row["symbol"])
- --- modulename: output-bmk-results, funcname: get_threshold
-output-bmk-results.py(98): if metric == "sample":
-output-bmk-results.py(99): spec_thr = get_specific_thresholds(metric, mode, bmk, symb)
- --- modulename: output-bmk-results, funcname: get_specific_thresholds
-output-bmk-results.py(57): if specific_variability is None:
-output-bmk-results.py(60): var = specific_variability[ (specific_variability['benchmark'] == bmk) & (specific_variability['symbol'].str.strip() == symb)]
-output-bmk-results.py(61): if var.empty:
-output-bmk-results.py(63): elif len(var)>1:
-output-bmk-results.py(68): if var.iloc[0]['sample_variation_average']>0 :
output-bmk-results.py(69): threshold = ( var.iloc[0]['sample_variation_average'] )
output-bmk-results.py(70): if mode == "build":
output-bmk-results.py(74): threshold *= 3
@@ -867,7 +770,7 @@ output-bmk-results.py(105): return spec_thr
output-bmk-results.py(228): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
output-bmk-results.py(229): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold))
output-bmk-results.py(228): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
-DEBUG: checking exe.improvement : 531.deepsjeng_r,deepsjeng_r_base.default : sample=-1% (threshold=3%)
+DEBUG: checking exe.improvement : 531.deepsjeng_r,deepsjeng_r_base.default : sample=1% (threshold=3%)
output-bmk-results.py(232): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold):
--- modulename: output-bmk-results, funcname: is_entry_improvement
output-bmk-results.py(192): if metric in metric_utils.higher_regress_metrics:
@@ -906,6 +809,7 @@ output-bmk-results.py(253): f_out.close()
output-bmk-results.py(29): if not self.outf:
output-bmk-results.py(31): self.outf.close()
output-bmk-results.py(32): if os.stat(self.filename).st_size == 0:
+output-bmk-results.py(33): os.remove(self.filename)
output-bmk-results.py(300): output_bmk_results_1(sym_df, "symbol", "regression", f_regr, f_skip, f_ebp, run_step_artifacts, metric, mode, details)
--- modulename: output-bmk-results, funcname: output_bmk_results_1
output-bmk-results.py(218): f_out = Outfile("{0}/{1}.{2}".format(run_step_artifacts, sym_type, change_kind), "w", predicate=(details=="verbose"))
@@ -938,7 +842,7 @@ output-bmk-results.py(105): return spec_thr
output-bmk-results.py(228): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
output-bmk-results.py(229): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold))
output-bmk-results.py(228): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
-DEBUG: checking symbol.regression : 505.mcf_r,[.] primal_bea_mpp : sample=3% (threshold=15%)
+DEBUG: checking symbol.regression : 505.mcf_r,[.] primal_bea_mpp : sample=-1% (threshold=15%)
output-bmk-results.py(232): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold):
--- modulename: output-bmk-results, funcname: is_entry_regression
output-bmk-results.py(183): if metric in metric_utils.higher_regress_metrics:
@@ -965,7 +869,7 @@ output-bmk-results.py(105): return spec_thr
output-bmk-results.py(228): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
output-bmk-results.py(229): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold))
output-bmk-results.py(228): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
-DEBUG: checking symbol.regression : 505.mcf_r,[.] price_out_impl : sample=2% (threshold=15%)
+DEBUG: checking symbol.regression : 505.mcf_r,[.] price_out_impl : sample=3% (threshold=15%)
output-bmk-results.py(232): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold):
--- modulename: output-bmk-results, funcname: is_entry_regression
output-bmk-results.py(183): if metric in metric_utils.higher_regress_metrics:
@@ -992,7 +896,7 @@ output-bmk-results.py(105): return spec_thr
output-bmk-results.py(228): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
output-bmk-results.py(229): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold))
output-bmk-results.py(228): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
-DEBUG: checking symbol.regression : 505.mcf_r,[.] cost_compare : sample=-1% (threshold=15%)
+DEBUG: checking symbol.regression : 505.mcf_r,[.] cost_compare : sample=0% (threshold=15%)
output-bmk-results.py(232): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold):
--- modulename: output-bmk-results, funcname: is_entry_regression
output-bmk-results.py(183): if metric in metric_utils.higher_regress_metrics:
@@ -1019,7 +923,7 @@ output-bmk-results.py(105): return spec_thr
output-bmk-results.py(228): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
output-bmk-results.py(229): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold))
output-bmk-results.py(228): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
-DEBUG: checking symbol.regression : 505.mcf_r,[.] replace_weaker_arc : sample=-3% (threshold=15%)
+DEBUG: checking symbol.regression : 505.mcf_r,[.] replace_weaker_arc : sample=-4% (threshold=15%)
output-bmk-results.py(232): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold):
--- modulename: output-bmk-results, funcname: is_entry_regression
output-bmk-results.py(183): if metric in metric_utils.higher_regress_metrics:
@@ -1046,7 +950,7 @@ output-bmk-results.py(105): return spec_thr
output-bmk-results.py(228): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
output-bmk-results.py(229): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold))
output-bmk-results.py(228): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
-DEBUG: checking symbol.regression : 531.deepsjeng_r,[.] _Z5fevalP7state_tiP12t_eval_comps : sample=2% (threshold=15%)
+DEBUG: checking symbol.regression : 531.deepsjeng_r,[.] _Z5fevalP7state_tiP12t_eval_comps : sample=-2% (threshold=15%)
output-bmk-results.py(232): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold):
--- modulename: output-bmk-results, funcname: is_entry_regression
output-bmk-results.py(183): if metric in metric_utils.higher_regress_metrics:
@@ -1073,7 +977,7 @@ output-bmk-results.py(105): return spec_thr
output-bmk-results.py(228): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
output-bmk-results.py(229): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold))
output-bmk-results.py(228): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
-DEBUG: checking symbol.regression : 531.deepsjeng_r,[.] _Z7ProbeTTP7state_tPiiiPjS1_S1_S1_S1_i : sample=6% (threshold=15%)
+DEBUG: checking symbol.regression : 531.deepsjeng_r,[.] _Z7ProbeTTP7state_tPiiiPjS1_S1_S1_S1_i : sample=-4% (threshold=15%)
output-bmk-results.py(232): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold):
--- modulename: output-bmk-results, funcname: is_entry_regression
output-bmk-results.py(183): if metric in metric_utils.higher_regress_metrics:
@@ -1100,7 +1004,7 @@ output-bmk-results.py(105): return spec_thr
output-bmk-results.py(228): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
output-bmk-results.py(229): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold))
output-bmk-results.py(228): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
-DEBUG: checking symbol.regression : 531.deepsjeng_r,[.] _Z4makeP7state_ti : sample=1% (threshold=15%)
+DEBUG: checking symbol.regression : 531.deepsjeng_r,[.] _Z6searchP7state_tiiiii : sample=2% (threshold=20.79%)
output-bmk-results.py(232): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold):
--- modulename: output-bmk-results, funcname: is_entry_regression
output-bmk-results.py(183): if metric in metric_utils.higher_regress_metrics:
@@ -1127,7 +1031,7 @@ output-bmk-results.py(105): return spec_thr
output-bmk-results.py(228): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
output-bmk-results.py(229): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold))
output-bmk-results.py(228): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
-DEBUG: checking symbol.regression : 531.deepsjeng_r,[.] _Z6searchP7state_tiiiii : sample=-3% (threshold=20.79%)
+DEBUG: checking symbol.regression : 531.deepsjeng_r,[.] _Z4makeP7state_ti : sample=9% (threshold=15%)
output-bmk-results.py(232): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold):
--- modulename: output-bmk-results, funcname: is_entry_regression
output-bmk-results.py(183): if metric in metric_utils.higher_regress_metrics:
@@ -1154,7 +1058,7 @@ output-bmk-results.py(105): return spec_thr
output-bmk-results.py(228): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
output-bmk-results.py(229): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold))
output-bmk-results.py(228): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
-DEBUG: checking symbol.regression : 557.xz_r,[.] lzma_mf_bt4_find : sample=0% (threshold=15%)
+DEBUG: checking symbol.regression : 557.xz_r,[.] lzma_mf_bt4_find : sample=-2% (threshold=15%)
output-bmk-results.py(232): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold):
--- modulename: output-bmk-results, funcname: is_entry_regression
output-bmk-results.py(183): if metric in metric_utils.higher_regress_metrics:
@@ -1181,7 +1085,7 @@ output-bmk-results.py(105): return spec_thr
output-bmk-results.py(228): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
output-bmk-results.py(229): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold))
output-bmk-results.py(228): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
-DEBUG: checking symbol.regression : 557.xz_r,[.] lzma_lzma_optimum_normal : sample=0% (threshold=15%)
+DEBUG: checking symbol.regression : 557.xz_r,[.] lzma_lzma_optimum_normal : sample=1% (threshold=15%)
output-bmk-results.py(232): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold):
--- modulename: output-bmk-results, funcname: is_entry_regression
output-bmk-results.py(183): if metric in metric_utils.higher_regress_metrics:
@@ -1208,7 +1112,7 @@ output-bmk-results.py(105): return spec_thr
output-bmk-results.py(228): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
output-bmk-results.py(229): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold))
output-bmk-results.py(228): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
-DEBUG: checking symbol.regression : 557.xz_r,[.] lzma_mf_bt4_skip : sample=1% (threshold=15%)
+DEBUG: checking symbol.regression : 557.xz_r,[.] lzma_mf_bt4_skip : sample=2% (threshold=15%)
output-bmk-results.py(232): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold):
--- modulename: output-bmk-results, funcname: is_entry_regression
output-bmk-results.py(183): if metric in metric_utils.higher_regress_metrics:
@@ -1253,7 +1157,7 @@ output-bmk-results.py(105): return spec_thr
output-bmk-results.py(228): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
output-bmk-results.py(229): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold))
output-bmk-results.py(228): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
-DEBUG: checking symbol.improvement : 505.mcf_r,[.] primal_bea_mpp : sample=3% (threshold=15%)
+DEBUG: checking symbol.improvement : 505.mcf_r,[.] primal_bea_mpp : sample=-1% (threshold=15%)
output-bmk-results.py(232): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold):
--- modulename: output-bmk-results, funcname: is_entry_improvement
output-bmk-results.py(192): if metric in metric_utils.higher_regress_metrics:
@@ -1280,7 +1184,7 @@ output-bmk-results.py(105): return spec_thr
output-bmk-results.py(228): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
output-bmk-results.py(229): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold))
output-bmk-results.py(228): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
-DEBUG: checking symbol.improvement : 505.mcf_r,[.] price_out_impl : sample=2% (threshold=15%)
+DEBUG: checking symbol.improvement : 505.mcf_r,[.] price_out_impl : sample=3% (threshold=15%)
output-bmk-results.py(232): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold):
--- modulename: output-bmk-results, funcname: is_entry_improvement
output-bmk-results.py(192): if metric in metric_utils.higher_regress_metrics:
@@ -1307,7 +1211,7 @@ output-bmk-results.py(105): return spec_thr
output-bmk-results.py(228): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
output-bmk-results.py(229): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold))
output-bmk-results.py(228): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
-DEBUG: checking symbol.improvement : 505.mcf_r,[.] cost_compare : sample=-1% (threshold=15%)
+DEBUG: checking symbol.improvement : 505.mcf_r,[.] cost_compare : sample=0% (threshold=15%)
output-bmk-results.py(232): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold):
--- modulename: output-bmk-results, funcname: is_entry_improvement
output-bmk-results.py(192): if metric in metric_utils.higher_regress_metrics:
@@ -1334,7 +1238,7 @@ output-bmk-results.py(105): return spec_thr
output-bmk-results.py(228): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
output-bmk-results.py(229): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold))
output-bmk-results.py(228): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
-DEBUG: checking symbol.improvement : 505.mcf_r,[.] replace_weaker_arc : sample=-3% (threshold=15%)
+DEBUG: checking symbol.improvement : 505.mcf_r,[.] replace_weaker_arc : sample=-4% (threshold=15%)
output-bmk-results.py(232): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold):
--- modulename: output-bmk-results, funcname: is_entry_improvement
output-bmk-results.py(192): if metric in metric_utils.higher_regress_metrics:
@@ -1361,7 +1265,7 @@ output-bmk-results.py(105): return spec_thr
output-bmk-results.py(228): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
output-bmk-results.py(229): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold))
output-bmk-results.py(228): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
-DEBUG: checking symbol.improvement : 531.deepsjeng_r,[.] _Z5fevalP7state_tiP12t_eval_comps : sample=2% (threshold=15%)
+DEBUG: checking symbol.improvement : 531.deepsjeng_r,[.] _Z5fevalP7state_tiP12t_eval_comps : sample=-2% (threshold=15%)
output-bmk-results.py(232): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold):
--- modulename: output-bmk-results, funcname: is_entry_improvement
output-bmk-results.py(192): if metric in metric_utils.higher_regress_metrics:
@@ -1388,7 +1292,7 @@ output-bmk-results.py(105): return spec_thr
output-bmk-results.py(228): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
output-bmk-results.py(229): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold))
output-bmk-results.py(228): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
-DEBUG: checking symbol.improvement : 531.deepsjeng_r,[.] _Z7ProbeTTP7state_tPiiiPjS1_S1_S1_S1_i : sample=6% (threshold=15%)
+DEBUG: checking symbol.improvement : 531.deepsjeng_r,[.] _Z7ProbeTTP7state_tPiiiPjS1_S1_S1_S1_i : sample=-4% (threshold=15%)
output-bmk-results.py(232): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold):
--- modulename: output-bmk-results, funcname: is_entry_improvement
output-bmk-results.py(192): if metric in metric_utils.higher_regress_metrics:
@@ -1415,7 +1319,7 @@ output-bmk-results.py(105): return spec_thr
output-bmk-results.py(228): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
output-bmk-results.py(229): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold))
output-bmk-results.py(228): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
-DEBUG: checking symbol.improvement : 531.deepsjeng_r,[.] _Z4makeP7state_ti : sample=1% (threshold=15%)
+DEBUG: checking symbol.improvement : 531.deepsjeng_r,[.] _Z6searchP7state_tiiiii : sample=2% (threshold=20.79%)
output-bmk-results.py(232): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold):
--- modulename: output-bmk-results, funcname: is_entry_improvement
output-bmk-results.py(192): if metric in metric_utils.higher_regress_metrics:
@@ -1442,7 +1346,7 @@ output-bmk-results.py(105): return spec_thr
output-bmk-results.py(228): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
output-bmk-results.py(229): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold))
output-bmk-results.py(228): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
-DEBUG: checking symbol.improvement : 531.deepsjeng_r,[.] _Z6searchP7state_tiiiii : sample=-3% (threshold=20.79%)
+DEBUG: checking symbol.improvement : 531.deepsjeng_r,[.] _Z4makeP7state_ti : sample=9% (threshold=15%)
output-bmk-results.py(232): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold):
--- modulename: output-bmk-results, funcname: is_entry_improvement
output-bmk-results.py(192): if metric in metric_utils.higher_regress_metrics:
@@ -1469,7 +1373,7 @@ output-bmk-results.py(105): return spec_thr
output-bmk-results.py(228): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
output-bmk-results.py(229): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold))
output-bmk-results.py(228): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
-DEBUG: checking symbol.improvement : 557.xz_r,[.] lzma_mf_bt4_find : sample=0% (threshold=15%)
+DEBUG: checking symbol.improvement : 557.xz_r,[.] lzma_mf_bt4_find : sample=-2% (threshold=15%)
output-bmk-results.py(232): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold):
--- modulename: output-bmk-results, funcname: is_entry_improvement
output-bmk-results.py(192): if metric in metric_utils.higher_regress_metrics:
@@ -1496,7 +1400,7 @@ output-bmk-results.py(105): return spec_thr
output-bmk-results.py(228): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
output-bmk-results.py(229): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold))
output-bmk-results.py(228): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
-DEBUG: checking symbol.improvement : 557.xz_r,[.] lzma_lzma_optimum_normal : sample=0% (threshold=15%)
+DEBUG: checking symbol.improvement : 557.xz_r,[.] lzma_lzma_optimum_normal : sample=1% (threshold=15%)
output-bmk-results.py(232): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold):
--- modulename: output-bmk-results, funcname: is_entry_improvement
output-bmk-results.py(192): if metric in metric_utils.higher_regress_metrics:
@@ -1523,7 +1427,7 @@ output-bmk-results.py(105): return spec_thr
output-bmk-results.py(228): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
output-bmk-results.py(229): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold))
output-bmk-results.py(228): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
-DEBUG: checking symbol.improvement : 557.xz_r,[.] lzma_mf_bt4_skip : sample=1% (threshold=15%)
+DEBUG: checking symbol.improvement : 557.xz_r,[.] lzma_mf_bt4_skip : sample=2% (threshold=15%)
output-bmk-results.py(232): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold):
--- modulename: output-bmk-results, funcname: is_entry_improvement
output-bmk-results.py(192): if metric in metric_utils.higher_regress_metrics:
@@ -1551,6 +1455,7 @@ output-bmk-results.py(306): f_regr.close()
output-bmk-results.py(29): if not self.outf:
output-bmk-results.py(31): self.outf.close()
output-bmk-results.py(32): if os.stat(self.filename).st_size == 0:
+output-bmk-results.py(33): os.remove(self.filename)
output-bmk-results.py(307): f_ebp.close()
--- modulename: output-bmk-results, funcname: close
output-bmk-results.py(29): if not self.outf:
diff --git a/notify/results.regressions b/notify/results.regressions
deleted file mode 100644
index c1eab88..0000000
--- a/notify/results.regressions
+++ /dev/null
@@ -1,2 +0,0 @@
-# lbm_r_base.default,519.lbm_r failed to run
-# omnetpp_r_base.default,520.omnetpp_r failed to run
diff --git a/notify/status.regression b/notify/status.regression
deleted file mode 100644
index 8bdb17f..0000000
--- a/notify/status.regression
+++ /dev/null
@@ -1,2 +0,0 @@
-100,519.lbm_r,lbm_r_base.default,519.lbm_r failed to run,519.lbm_r failed to run
-100,520.omnetpp_r,omnetpp_r_base.default,520.omnetpp_r failed to run,520.omnetpp_r failed to run