diff options
Diffstat (limited to 'notify')
-rw-r--r-- | notify/any.skipped | 2 | ||||
-rw-r--r-- | notify/extra-bisect-params | 2 | ||||
-rw-r--r-- | notify/jira/comment-template.txt | 2 | ||||
-rw-r--r-- | notify/lnt_report.json | 44 | ||||
-rw-r--r-- | notify/mail-body.txt | 12 | ||||
-rw-r--r-- | notify/mail-recipients.txt | 2 | ||||
-rw-r--r-- | notify/mail-subject.txt | 2 | ||||
-rw-r--r-- | notify/output-bmk-results.log | 303 | ||||
-rw-r--r-- | notify/results.regressions | 5 | ||||
-rw-r--r-- | notify/status.regression | 5 |
10 files changed, 93 insertions, 286 deletions
diff --git a/notify/any.skipped b/notify/any.skipped deleted file mode 100644 index 01142f7..0000000 --- a/notify/any.skipped +++ /dev/null @@ -1,2 +0,0 @@ -519.lbm_r,lbm_r_base.default,sped up by 200% - 519.lbm_r,sped up by 200% - 519.lbm_r - from 1 to -1 perf samples
-523.xalancbmk_r,cpuxalan_r_base.default,sped up by 100% - 523.xalancbmk_r,sped up by 100% - 523.xalancbmk_r - from 2505 to -1 perf samples
diff --git a/notify/extra-bisect-params b/notify/extra-bisect-params index 0ede073..fa6c7c9 100644 --- a/notify/extra-bisect-params +++ b/notify/extra-bisect-params @@ -1 +1 @@ -extra_build_params=++benchmarks 508.namd_r ++benchmarks 510.parest_r ++benchmarks 511.povray_r ++benchmarks 519.lbm_r ++benchmarks 520.omnetpp_r +extra_build_params= diff --git a/notify/jira/comment-template.txt b/notify/jira/comment-template.txt index 36384e1..4f0fc7b 100644 --- a/notify/jira/comment-template.txt +++ b/notify/jira/comment-template.txt @@ -1,3 +1,3 @@ [LLVM-651] No change -Details: https://ci.linaro.org/job/tcwg_bmk-code_speed-cpu2017rate--llvm-arm-master-O3_LTO-build/25/artifact/artifacts/notify/mail-body.txt/*view*/ +Details: https://ci.linaro.org/job/tcwg_bmk-code_speed-cpu2017rate--llvm-arm-master-O3_LTO-build/26/artifact/artifacts/notify/mail-body.txt/*view*/ diff --git a/notify/lnt_report.json b/notify/lnt_report.json index 0cb63d1..cb11385 100644 --- a/notify/lnt_report.json +++ b/notify/lnt_report.json @@ -6,71 +6,71 @@ "Run": { "Info": { "__report_version__": "1", - "run_order": "llvmorg-17-init-09617-g575fdea70a86", + "run_order": "llvmorg-17-init-10167-g1d8ab713adfe", "tag": "tcwg_bmk-code_speed-cpu2017rate" }, - "Start Time": "2023-11-22 21:18:50" + "Start Time": "2023-11-22 21:21:28" }, "Tests": [ { "Data": [ - 27668 + 1322807 ], "Info": {}, - "Name": "tcwg_bmk-code_speed-cpu2017rate.505.mcf_r.code_size" + "Name": "tcwg_bmk-code_speed-cpu2017rate.538.imagick_r.code_size" } , { "Data": [ - 138239 + 94203 ], "Info": {}, - "Name": "tcwg_bmk-code_speed-cpu2017rate.557.xz_r.code_size" + "Name": "tcwg_bmk-code_speed-cpu2017rate.541.leela_r.code_size" } , { "Data": [ - 94159 + 3645574 ], "Info": {}, - "Name": "tcwg_bmk-code_speed-cpu2017rate.541.leela_r.code_size" + "Name": "tcwg_bmk-code_speed-cpu2017rate.523.xalancbmk_r.code_size" } , { "Data": [ - 11062 + 138415 ], "Info": {}, - "Name": "tcwg_bmk-code_speed-cpu2017rate.519.lbm_r.code_size" + "Name": "tcwg_bmk-code_speed-cpu2017rate.557.xz_r.code_size" } , { "Data": [ - 3645542 + 111048 ], "Info": {}, - "Name": "tcwg_bmk-code_speed-cpu2017rate.523.xalancbmk_r.code_size" + "Name": "tcwg_bmk-code_speed-cpu2017rate.531.deepsjeng_r.code_size" } , { "Data": [ - 111016 + 27672 ], "Info": {}, - "Name": "tcwg_bmk-code_speed-cpu2017rate.531.deepsjeng_r.code_size" + "Name": "tcwg_bmk-code_speed-cpu2017rate.505.mcf_r.code_size" } , { "Data": [ - 11609 + 11062 ], "Info": {}, - "Name": "tcwg_bmk-code_speed-cpu2017rate.531.deepsjeng_r.exec" + "Name": "tcwg_bmk-code_speed-cpu2017rate.519.lbm_r.code_size" } , { "Data": [ - 10015 + 10066 ], "Info": {}, "Name": "tcwg_bmk-code_speed-cpu2017rate.557.xz_r.exec" @@ -78,10 +78,18 @@ , { "Data": [ - 12028 + 12047 ], "Info": {}, "Name": "tcwg_bmk-code_speed-cpu2017rate.505.mcf_r.exec" } + , + { + "Data": [ + 11680 + ], + "Info": {}, + "Name": "tcwg_bmk-code_speed-cpu2017rate.531.deepsjeng_r.exec" + } ] } diff --git a/notify/mail-body.txt b/notify/mail-body.txt index 600f1c2..f1b173e 100644 --- a/notify/mail-body.txt +++ b/notify/mail-body.txt @@ -2,7 +2,13 @@ Dear contributor, our automatic CI has detected problems related to your patch(e In CI config tcwg_bmk-code_speed-cpu2017rate/llvm-arm-master-O3_LTO after: - | baseline build + | 550 commits in llvm + | 1d8ab713adfe Revert "[DebugLine] save one debug line entry for empty prologue" + | f238a98e8447 [OpenMP][libomptarget][AMDGPU] Enable active HSA wait state + | f3dcd3ad992c [clang-format] Correctly limit formatted ranges when specifying qualifier alignment + | cb7e3da05452 [libc][rpc] Treat pointers as arrays consistently + | 9e9bf1e3ed9e [RISCV] Use setcc to truncate results in widenVectorOpsToi8 + | ... and 545 more commits in llvm No change @@ -22,6 +28,6 @@ This benchmarking CI is work-in-progress, and we welcome feedback and suggestion -----------------8<--------------------------8<--------------------------8<-------------------------- The information below can be used to reproduce a debug environment: -Current build : https://ci.linaro.org/job/tcwg_bmk-code_speed-cpu2017rate--llvm-arm-master-O3_LTO-build/25/artifact/artifacts -Reference build : https://ci.linaro.org/job/tcwg_bmk-code_speed-cpu2017rate--llvm-arm-master-O3_LTO-build/23/artifact/artifacts +Current build : https://ci.linaro.org/job/tcwg_bmk-code_speed-cpu2017rate--llvm-arm-master-O3_LTO-build/26/artifact/artifacts +Reference build : https://ci.linaro.org/job/tcwg_bmk-code_speed-cpu2017rate--llvm-arm-master-O3_LTO-build/25/artifact/artifacts diff --git a/notify/mail-recipients.txt b/notify/mail-recipients.txt index aa219ef..56b65bb 100644 --- a/notify/mail-recipients.txt +++ b/notify/mail-recipients.txt @@ -1 +1 @@ -bcc:tcwg-validation@linaro.org +bcc:tcwg-validation@linaro.org,maxim.kuvyrkov@linaro.org diff --git a/notify/mail-subject.txt b/notify/mail-subject.txt index 370e184..2318611 100644 --- a/notify/mail-subject.txt +++ b/notify/mail-subject.txt @@ -1 +1 @@ -[Linaro-TCWG-CI] baseline build: No change on arm O3 -LTO +[Linaro-TCWG-CI] 550 commits in llvm: No change on arm O3 -LTO diff --git a/notify/output-bmk-results.log b/notify/output-bmk-results.log index bc934ed..d73b5fb 100644 --- a/notify/output-bmk-results.log +++ b/notify/output-bmk-results.log @@ -106,12 +106,12 @@ output-bmk-results.py(258): print(results_df) 0 500.perlbench_r perlbench_r_base.default ... failed-to-run failed-to-run 1 502.gcc_r cpugcc_r_base.default ... failed-to-run failed-to-run 2 505.mcf_r mcf_r_base.default ... success success -3 508.namd_r namd_r_base.default ... success failed-to-run -4 510.parest_r parest_r_base.default ... success failed-to-run -5 511.povray_r povray_r_base.default ... success failed-to-run -6 519.lbm_r lbm_r_base.default ... success failed-to-run -7 520.omnetpp_r omnetpp_r_base.default ... success failed-to-run -8 523.xalancbmk_r cpuxalan_r_base.default ... -1 failed-to-run +3 508.namd_r namd_r_base.default ... failed-to-run failed-to-run +4 510.parest_r parest_r_base.default ... failed-to-run failed-to-run +5 511.povray_r povray_r_base.default ... failed-to-run failed-to-run +6 519.lbm_r lbm_r_base.default ... failed-to-run failed-to-run +7 520.omnetpp_r omnetpp_r_base.default ... failed-to-run failed-to-run +8 523.xalancbmk_r cpuxalan_r_base.default ... failed-to-run failed-to-run 9 525.x264_r x264_r_base.default ... failed-to-run failed-to-run 10 526.blender_r blender_r_base.default ... failed-to-run failed-to-run 11 531.deepsjeng_r deepsjeng_r_base.default ... success success @@ -171,25 +171,12 @@ output-bmk-results.py(115): short_diag="" output-bmk-results.py(116): classif="" output-bmk-results.py(118): if row["status_x"]!="failed-to-build" and row["status_y"]=="failed-to-build": output-bmk-results.py(121): elif row["status_x"]=="success" and row["status_y"]=="failed-to-run": -output-bmk-results.py(122): short_diag = "{0} failed to run".format(bmk) -output-bmk-results.py(123): classif="regression" +output-bmk-results.py(124): elif row["status_x"]=="failed-to-build" and row["status_y"]=="failed-to-run": +output-bmk-results.py(127): elif row["status_x"]=="failed-to-run" and row["status_y"]=="success": +output-bmk-results.py(130): elif row["status_x"]=="failed-to-build" and row["status_y"]=="success": output-bmk-results.py(134): return classif, short_diag output-bmk-results.py(265): if classif != change_kind: -output-bmk-results.py(268): print("DEBUG: *** {0},{1} : {2}".format(row["benchmark"], row["symbol"], short_diag)) -DEBUG: *** 508.namd_r,namd_r_base.default : 508.namd_r failed to run -output-bmk-results.py(270): f_out.write_csv((100, row["benchmark"], row["symbol"], short_diag, short_diag)) - --- modulename: output-bmk-results, funcname: write_csv -output-bmk-results.py(41): if not self.predicate or not self.csvwriter: -output-bmk-results.py(43): self.csvwriter.writerow(arr) -output-bmk-results.py(271): if change_kind == "regression": -output-bmk-results.py(272): f_regr.write("# {0},{1}\n".format(row["symbol"], short_diag)) - --- modulename: output-bmk-results, funcname: write -output-bmk-results.py(36): if not self.predicate or not self.outf: -output-bmk-results.py(38): self.outf.write(string) -output-bmk-results.py(273): f_ebp.write("++benchmarks {0} ".format(row["benchmark"])) - --- modulename: output-bmk-results, funcname: write -output-bmk-results.py(36): if not self.predicate or not self.outf: -output-bmk-results.py(38): self.outf.write(string) +output-bmk-results.py(266): continue; output-bmk-results.py(261): for index, row in results_df.iterrows(): output-bmk-results.py(263): classif, short_diag = get_status_diag(row) --- modulename: output-bmk-results, funcname: get_status_diag @@ -198,25 +185,12 @@ output-bmk-results.py(115): short_diag="" output-bmk-results.py(116): classif="" output-bmk-results.py(118): if row["status_x"]!="failed-to-build" and row["status_y"]=="failed-to-build": output-bmk-results.py(121): elif row["status_x"]=="success" and row["status_y"]=="failed-to-run": -output-bmk-results.py(122): short_diag = "{0} failed to run".format(bmk) -output-bmk-results.py(123): classif="regression" +output-bmk-results.py(124): elif row["status_x"]=="failed-to-build" and row["status_y"]=="failed-to-run": +output-bmk-results.py(127): elif row["status_x"]=="failed-to-run" and row["status_y"]=="success": +output-bmk-results.py(130): elif row["status_x"]=="failed-to-build" and row["status_y"]=="success": output-bmk-results.py(134): return classif, short_diag output-bmk-results.py(265): if classif != change_kind: -output-bmk-results.py(268): print("DEBUG: *** {0},{1} : {2}".format(row["benchmark"], row["symbol"], short_diag)) -DEBUG: *** 510.parest_r,parest_r_base.default : 510.parest_r failed to run -output-bmk-results.py(270): f_out.write_csv((100, row["benchmark"], row["symbol"], short_diag, short_diag)) - --- modulename: output-bmk-results, funcname: write_csv -output-bmk-results.py(41): if not self.predicate or not self.csvwriter: -output-bmk-results.py(43): self.csvwriter.writerow(arr) -output-bmk-results.py(271): if change_kind == "regression": -output-bmk-results.py(272): f_regr.write("# {0},{1}\n".format(row["symbol"], short_diag)) - --- modulename: output-bmk-results, funcname: write -output-bmk-results.py(36): if not self.predicate or not self.outf: -output-bmk-results.py(38): self.outf.write(string) -output-bmk-results.py(273): f_ebp.write("++benchmarks {0} ".format(row["benchmark"])) - --- modulename: output-bmk-results, funcname: write -output-bmk-results.py(36): if not self.predicate or not self.outf: -output-bmk-results.py(38): self.outf.write(string) +output-bmk-results.py(266): continue; output-bmk-results.py(261): for index, row in results_df.iterrows(): output-bmk-results.py(263): classif, short_diag = get_status_diag(row) --- modulename: output-bmk-results, funcname: get_status_diag @@ -225,25 +199,12 @@ output-bmk-results.py(115): short_diag="" output-bmk-results.py(116): classif="" output-bmk-results.py(118): if row["status_x"]!="failed-to-build" and row["status_y"]=="failed-to-build": output-bmk-results.py(121): elif row["status_x"]=="success" and row["status_y"]=="failed-to-run": -output-bmk-results.py(122): short_diag = "{0} failed to run".format(bmk) -output-bmk-results.py(123): classif="regression" +output-bmk-results.py(124): elif row["status_x"]=="failed-to-build" and row["status_y"]=="failed-to-run": +output-bmk-results.py(127): elif row["status_x"]=="failed-to-run" and row["status_y"]=="success": +output-bmk-results.py(130): elif row["status_x"]=="failed-to-build" and row["status_y"]=="success": output-bmk-results.py(134): return classif, short_diag output-bmk-results.py(265): if classif != change_kind: -output-bmk-results.py(268): print("DEBUG: *** {0},{1} : {2}".format(row["benchmark"], row["symbol"], short_diag)) -DEBUG: *** 511.povray_r,povray_r_base.default : 511.povray_r failed to run -output-bmk-results.py(270): f_out.write_csv((100, row["benchmark"], row["symbol"], short_diag, short_diag)) - --- modulename: output-bmk-results, funcname: write_csv -output-bmk-results.py(41): if not self.predicate or not self.csvwriter: -output-bmk-results.py(43): self.csvwriter.writerow(arr) -output-bmk-results.py(271): if change_kind == "regression": -output-bmk-results.py(272): f_regr.write("# {0},{1}\n".format(row["symbol"], short_diag)) - --- modulename: output-bmk-results, funcname: write -output-bmk-results.py(36): if not self.predicate or not self.outf: -output-bmk-results.py(38): self.outf.write(string) -output-bmk-results.py(273): f_ebp.write("++benchmarks {0} ".format(row["benchmark"])) - --- modulename: output-bmk-results, funcname: write -output-bmk-results.py(36): if not self.predicate or not self.outf: -output-bmk-results.py(38): self.outf.write(string) +output-bmk-results.py(266): continue; output-bmk-results.py(261): for index, row in results_df.iterrows(): output-bmk-results.py(263): classif, short_diag = get_status_diag(row) --- modulename: output-bmk-results, funcname: get_status_diag @@ -252,25 +213,12 @@ output-bmk-results.py(115): short_diag="" output-bmk-results.py(116): classif="" output-bmk-results.py(118): if row["status_x"]!="failed-to-build" and row["status_y"]=="failed-to-build": output-bmk-results.py(121): elif row["status_x"]=="success" and row["status_y"]=="failed-to-run": -output-bmk-results.py(122): short_diag = "{0} failed to run".format(bmk) -output-bmk-results.py(123): classif="regression" +output-bmk-results.py(124): elif row["status_x"]=="failed-to-build" and row["status_y"]=="failed-to-run": +output-bmk-results.py(127): elif row["status_x"]=="failed-to-run" and row["status_y"]=="success": +output-bmk-results.py(130): elif row["status_x"]=="failed-to-build" and row["status_y"]=="success": output-bmk-results.py(134): return classif, short_diag output-bmk-results.py(265): if classif != change_kind: -output-bmk-results.py(268): print("DEBUG: *** {0},{1} : {2}".format(row["benchmark"], row["symbol"], short_diag)) -DEBUG: *** 519.lbm_r,lbm_r_base.default : 519.lbm_r failed to run -output-bmk-results.py(270): f_out.write_csv((100, row["benchmark"], row["symbol"], short_diag, short_diag)) - --- modulename: output-bmk-results, funcname: write_csv -output-bmk-results.py(41): if not self.predicate or not self.csvwriter: -output-bmk-results.py(43): self.csvwriter.writerow(arr) -output-bmk-results.py(271): if change_kind == "regression": -output-bmk-results.py(272): f_regr.write("# {0},{1}\n".format(row["symbol"], short_diag)) - --- modulename: output-bmk-results, funcname: write -output-bmk-results.py(36): if not self.predicate or not self.outf: -output-bmk-results.py(38): self.outf.write(string) -output-bmk-results.py(273): f_ebp.write("++benchmarks {0} ".format(row["benchmark"])) - --- modulename: output-bmk-results, funcname: write -output-bmk-results.py(36): if not self.predicate or not self.outf: -output-bmk-results.py(38): self.outf.write(string) +output-bmk-results.py(266): continue; output-bmk-results.py(261): for index, row in results_df.iterrows(): output-bmk-results.py(263): classif, short_diag = get_status_diag(row) --- modulename: output-bmk-results, funcname: get_status_diag @@ -279,25 +227,12 @@ output-bmk-results.py(115): short_diag="" output-bmk-results.py(116): classif="" output-bmk-results.py(118): if row["status_x"]!="failed-to-build" and row["status_y"]=="failed-to-build": output-bmk-results.py(121): elif row["status_x"]=="success" and row["status_y"]=="failed-to-run": -output-bmk-results.py(122): short_diag = "{0} failed to run".format(bmk) -output-bmk-results.py(123): classif="regression" +output-bmk-results.py(124): elif row["status_x"]=="failed-to-build" and row["status_y"]=="failed-to-run": +output-bmk-results.py(127): elif row["status_x"]=="failed-to-run" and row["status_y"]=="success": +output-bmk-results.py(130): elif row["status_x"]=="failed-to-build" and row["status_y"]=="success": output-bmk-results.py(134): return classif, short_diag output-bmk-results.py(265): if classif != change_kind: -output-bmk-results.py(268): print("DEBUG: *** {0},{1} : {2}".format(row["benchmark"], row["symbol"], short_diag)) -DEBUG: *** 520.omnetpp_r,omnetpp_r_base.default : 520.omnetpp_r failed to run -output-bmk-results.py(270): f_out.write_csv((100, row["benchmark"], row["symbol"], short_diag, short_diag)) - --- modulename: output-bmk-results, funcname: write_csv -output-bmk-results.py(41): if not self.predicate or not self.csvwriter: -output-bmk-results.py(43): self.csvwriter.writerow(arr) -output-bmk-results.py(271): if change_kind == "regression": -output-bmk-results.py(272): f_regr.write("# {0},{1}\n".format(row["symbol"], short_diag)) - --- modulename: output-bmk-results, funcname: write -output-bmk-results.py(36): if not self.predicate or not self.outf: -output-bmk-results.py(38): self.outf.write(string) -output-bmk-results.py(273): f_ebp.write("++benchmarks {0} ".format(row["benchmark"])) - --- modulename: output-bmk-results, funcname: write -output-bmk-results.py(36): if not self.predicate or not self.outf: -output-bmk-results.py(38): self.outf.write(string) +output-bmk-results.py(266): continue; output-bmk-results.py(261): for index, row in results_df.iterrows(): output-bmk-results.py(263): classif, short_diag = get_status_diag(row) --- modulename: output-bmk-results, funcname: get_status_diag @@ -416,6 +351,7 @@ output-bmk-results.py(275): f_out.close() output-bmk-results.py(29): if not self.outf: output-bmk-results.py(31): self.outf.close() output-bmk-results.py(32): if os.stat(self.filename).st_size == 0: +output-bmk-results.py(33): os.remove(self.filename) output-bmk-results.py(294): output_bmk_results_status(exe_df, "improvement", None, None, run_step_artifacts, details) --- modulename: output-bmk-results, funcname: output_bmk_results_status output-bmk-results.py(256): f_out = Outfile("{0}/status.{1}".format(run_step_artifacts, change_kind), "w", predicate=(details=="verbose")) @@ -430,12 +366,12 @@ output-bmk-results.py(258): print(results_df) 0 500.perlbench_r perlbench_r_base.default ... failed-to-run failed-to-run 1 502.gcc_r cpugcc_r_base.default ... failed-to-run failed-to-run 2 505.mcf_r mcf_r_base.default ... success success -3 508.namd_r namd_r_base.default ... success failed-to-run -4 510.parest_r parest_r_base.default ... success failed-to-run -5 511.povray_r povray_r_base.default ... success failed-to-run -6 519.lbm_r lbm_r_base.default ... success failed-to-run -7 520.omnetpp_r omnetpp_r_base.default ... success failed-to-run -8 523.xalancbmk_r cpuxalan_r_base.default ... -1 failed-to-run +3 508.namd_r namd_r_base.default ... failed-to-run failed-to-run +4 510.parest_r parest_r_base.default ... failed-to-run failed-to-run +5 511.povray_r povray_r_base.default ... failed-to-run failed-to-run +6 519.lbm_r lbm_r_base.default ... failed-to-run failed-to-run +7 520.omnetpp_r omnetpp_r_base.default ... failed-to-run failed-to-run +8 523.xalancbmk_r cpuxalan_r_base.default ... failed-to-run failed-to-run 9 525.x264_r x264_r_base.default ... failed-to-run failed-to-run 10 526.blender_r blender_r_base.default ... failed-to-run failed-to-run 11 531.deepsjeng_r deepsjeng_r_base.default ... success success @@ -495,8 +431,9 @@ output-bmk-results.py(115): short_diag="" output-bmk-results.py(116): classif="" output-bmk-results.py(118): if row["status_x"]!="failed-to-build" and row["status_y"]=="failed-to-build": output-bmk-results.py(121): elif row["status_x"]=="success" and row["status_y"]=="failed-to-run": -output-bmk-results.py(122): short_diag = "{0} failed to run".format(bmk) -output-bmk-results.py(123): classif="regression" +output-bmk-results.py(124): elif row["status_x"]=="failed-to-build" and row["status_y"]=="failed-to-run": +output-bmk-results.py(127): elif row["status_x"]=="failed-to-run" and row["status_y"]=="success": +output-bmk-results.py(130): elif row["status_x"]=="failed-to-build" and row["status_y"]=="success": output-bmk-results.py(134): return classif, short_diag output-bmk-results.py(265): if classif != change_kind: output-bmk-results.py(266): continue; @@ -508,8 +445,9 @@ output-bmk-results.py(115): short_diag="" output-bmk-results.py(116): classif="" output-bmk-results.py(118): if row["status_x"]!="failed-to-build" and row["status_y"]=="failed-to-build": output-bmk-results.py(121): elif row["status_x"]=="success" and row["status_y"]=="failed-to-run": -output-bmk-results.py(122): short_diag = "{0} failed to run".format(bmk) -output-bmk-results.py(123): classif="regression" +output-bmk-results.py(124): elif row["status_x"]=="failed-to-build" and row["status_y"]=="failed-to-run": +output-bmk-results.py(127): elif row["status_x"]=="failed-to-run" and row["status_y"]=="success": +output-bmk-results.py(130): elif row["status_x"]=="failed-to-build" and row["status_y"]=="success": output-bmk-results.py(134): return classif, short_diag output-bmk-results.py(265): if classif != change_kind: output-bmk-results.py(266): continue; @@ -521,8 +459,9 @@ output-bmk-results.py(115): short_diag="" output-bmk-results.py(116): classif="" output-bmk-results.py(118): if row["status_x"]!="failed-to-build" and row["status_y"]=="failed-to-build": output-bmk-results.py(121): elif row["status_x"]=="success" and row["status_y"]=="failed-to-run": -output-bmk-results.py(122): short_diag = "{0} failed to run".format(bmk) -output-bmk-results.py(123): classif="regression" +output-bmk-results.py(124): elif row["status_x"]=="failed-to-build" and row["status_y"]=="failed-to-run": +output-bmk-results.py(127): elif row["status_x"]=="failed-to-run" and row["status_y"]=="success": +output-bmk-results.py(130): elif row["status_x"]=="failed-to-build" and row["status_y"]=="success": output-bmk-results.py(134): return classif, short_diag output-bmk-results.py(265): if classif != change_kind: output-bmk-results.py(266): continue; @@ -534,8 +473,9 @@ output-bmk-results.py(115): short_diag="" output-bmk-results.py(116): classif="" output-bmk-results.py(118): if row["status_x"]!="failed-to-build" and row["status_y"]=="failed-to-build": output-bmk-results.py(121): elif row["status_x"]=="success" and row["status_y"]=="failed-to-run": -output-bmk-results.py(122): short_diag = "{0} failed to run".format(bmk) -output-bmk-results.py(123): classif="regression" +output-bmk-results.py(124): elif row["status_x"]=="failed-to-build" and row["status_y"]=="failed-to-run": +output-bmk-results.py(127): elif row["status_x"]=="failed-to-run" and row["status_y"]=="success": +output-bmk-results.py(130): elif row["status_x"]=="failed-to-build" and row["status_y"]=="success": output-bmk-results.py(134): return classif, short_diag output-bmk-results.py(265): if classif != change_kind: output-bmk-results.py(266): continue; @@ -547,8 +487,9 @@ output-bmk-results.py(115): short_diag="" output-bmk-results.py(116): classif="" output-bmk-results.py(118): if row["status_x"]!="failed-to-build" and row["status_y"]=="failed-to-build": output-bmk-results.py(121): elif row["status_x"]=="success" and row["status_y"]=="failed-to-run": -output-bmk-results.py(122): short_diag = "{0} failed to run".format(bmk) -output-bmk-results.py(123): classif="regression" +output-bmk-results.py(124): elif row["status_x"]=="failed-to-build" and row["status_y"]=="failed-to-run": +output-bmk-results.py(127): elif row["status_x"]=="failed-to-run" and row["status_y"]=="success": +output-bmk-results.py(130): elif row["status_x"]=="failed-to-build" and row["status_y"]=="success": output-bmk-results.py(134): return classif, short_diag output-bmk-results.py(265): if classif != change_kind: output-bmk-results.py(266): continue; @@ -720,51 +661,7 @@ output-bmk-results.py(110): return default_threshold[(change_kind,metric,mod output-bmk-results.py(228): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ output-bmk-results.py(229): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold)) output-bmk-results.py(228): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ -DEBUG: checking exe.regression : 519.lbm_r,lbm_r_base.default : sample=200% (threshold=3%) -output-bmk-results.py(232): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold): - --- modulename: output-bmk-results, funcname: is_entry_regression -output-bmk-results.py(183): if metric in metric_utils.higher_regress_metrics: -output-bmk-results.py(184): return (result - 100 > threshold) -output-bmk-results.py(233): continue -output-bmk-results.py(224): for index, row in out_df.iterrows(): -output-bmk-results.py(226): threshold = get_threshold(sym_type, metric, mode, row["benchmark"], row["symbol"]) - --- modulename: output-bmk-results, funcname: get_threshold -output-bmk-results.py(98): if metric == "sample": -output-bmk-results.py(99): spec_thr = get_specific_thresholds(metric, mode, bmk, symb) - --- modulename: output-bmk-results, funcname: get_specific_thresholds -output-bmk-results.py(57): if specific_variability is None: -output-bmk-results.py(60): var = specific_variability[ (specific_variability['benchmark'] == bmk) & (specific_variability['symbol'].str.strip() == symb)] -output-bmk-results.py(61): if var.empty: -output-bmk-results.py(62): return np.nan -output-bmk-results.py(100): if not np.isnan(spec_thr): -output-bmk-results.py(107): if metric == "num_vect_loops" or metric == "num_sve_loops": -output-bmk-results.py(110): return default_threshold[(change_kind,metric,mode)] -output-bmk-results.py(228): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ -output-bmk-results.py(229): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold)) -output-bmk-results.py(228): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ -DEBUG: checking exe.regression : 523.xalancbmk_r,cpuxalan_r_base.default : sample=100% (threshold=3%) -output-bmk-results.py(232): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold): - --- modulename: output-bmk-results, funcname: is_entry_regression -output-bmk-results.py(183): if metric in metric_utils.higher_regress_metrics: -output-bmk-results.py(184): return (result - 100 > threshold) -output-bmk-results.py(233): continue -output-bmk-results.py(224): for index, row in out_df.iterrows(): -output-bmk-results.py(226): threshold = get_threshold(sym_type, metric, mode, row["benchmark"], row["symbol"]) - --- modulename: output-bmk-results, funcname: get_threshold -output-bmk-results.py(98): if metric == "sample": -output-bmk-results.py(99): spec_thr = get_specific_thresholds(metric, mode, bmk, symb) - --- modulename: output-bmk-results, funcname: get_specific_thresholds -output-bmk-results.py(57): if specific_variability is None: -output-bmk-results.py(60): var = specific_variability[ (specific_variability['benchmark'] == bmk) & (specific_variability['symbol'].str.strip() == symb)] -output-bmk-results.py(61): if var.empty: -output-bmk-results.py(62): return np.nan -output-bmk-results.py(100): if not np.isnan(spec_thr): -output-bmk-results.py(107): if metric == "num_vect_loops" or metric == "num_sve_loops": -output-bmk-results.py(110): return default_threshold[(change_kind,metric,mode)] -output-bmk-results.py(228): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ -output-bmk-results.py(229): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold)) -output-bmk-results.py(228): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ -DEBUG: checking exe.regression : 531.deepsjeng_r,deepsjeng_r_base.default : sample=1% (threshold=3%) +DEBUG: checking exe.regression : 531.deepsjeng_r,deepsjeng_r_base.default : sample=-1% (threshold=3%) output-bmk-results.py(232): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold): --- modulename: output-bmk-results, funcname: is_entry_regression output-bmk-results.py(183): if metric in metric_utils.higher_regress_metrics: @@ -786,7 +683,7 @@ output-bmk-results.py(110): return default_threshold[(change_kind,metric,mod output-bmk-results.py(228): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ output-bmk-results.py(229): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold)) output-bmk-results.py(228): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ -DEBUG: checking exe.regression : 557.xz_r,xz_r_base.default : sample=0% (threshold=3%) +DEBUG: checking exe.regression : 557.xz_r,xz_r_base.default : sample=-1% (threshold=3%) output-bmk-results.py(232): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold): --- modulename: output-bmk-results, funcname: is_entry_regression output-bmk-results.py(183): if metric in metric_utils.higher_regress_metrics: @@ -848,101 +745,7 @@ output-bmk-results.py(110): return default_threshold[(change_kind,metric,mod output-bmk-results.py(228): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ output-bmk-results.py(229): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold)) output-bmk-results.py(228): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ -DEBUG: checking exe.improvement : 519.lbm_r,lbm_r_base.default : sample=200% (threshold=3%) -output-bmk-results.py(232): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold): - --- modulename: output-bmk-results, funcname: is_entry_improvement -output-bmk-results.py(192): if metric in metric_utils.higher_regress_metrics: -output-bmk-results.py(193): return (100 - result > threshold) -output-bmk-results.py(235): percent_change, short_diag, long_diag = get_short_long_diag(row, metric, sym_type, change_kind) - --- modulename: output-bmk-results, funcname: get_short_long_diag -output-bmk-results.py(137): bmk = row["benchmark"] -output-bmk-results.py(139): rel_value = row["rel_" + metric] -output-bmk-results.py(140): prev_value = row[metric + "_x"] -output-bmk-results.py(141): curr_value = row[metric + "_y"] -output-bmk-results.py(142): if metric == "num_vect_loops" or metric == "num_sve_loops": -output-bmk-results.py(152): suffix = "" -output-bmk-results.py(153): if metric == "sample": -output-bmk-results.py(154): prefix_regression = "slowed down by" -output-bmk-results.py(155): prefix_improvement = "sped up by" -output-bmk-results.py(156): suffix = "perf samples" -output-bmk-results.py(167): if sym_type=="symbol": -output-bmk-results.py(170): item=bmk -output-bmk-results.py(172): short_diag = "{1} {2}% - {0}".format(item, locals()["prefix_" + change_kind], abs(rel_value - 100)) -output-bmk-results.py(173): long_diag = "{0} - from {1} to {2} {3}".format(short_diag, prev_value, curr_value, suffix) -output-bmk-results.py(174): return abs(rel_value - 100), short_diag, long_diag -output-bmk-results.py(239): if metric == "sample" \ -output-bmk-results.py(240): and row['symbol_md5sum_x'] == row['symbol_md5sum_y'] \ -output-bmk-results.py(241): and row['symbol_md5sum_x'] != "-1" \ -output-bmk-results.py(242): and row['symbol_md5sum_x'] != "d41d8cd98f00b204e9800998ecf8427e": -output-bmk-results.py(243): f_skip.write_csv((row["benchmark"], row["symbol"], short_diag, long_diag)) - --- modulename: output-bmk-results, funcname: write_csv -output-bmk-results.py(41): if not self.predicate or not self.csvwriter: -output-bmk-results.py(43): self.csvwriter.writerow(arr) -output-bmk-results.py(244): continue -output-bmk-results.py(224): for index, row in out_df.iterrows(): -output-bmk-results.py(226): threshold = get_threshold(sym_type, metric, mode, row["benchmark"], row["symbol"]) - --- modulename: output-bmk-results, funcname: get_threshold -output-bmk-results.py(98): if metric == "sample": -output-bmk-results.py(99): spec_thr = get_specific_thresholds(metric, mode, bmk, symb) - --- modulename: output-bmk-results, funcname: get_specific_thresholds -output-bmk-results.py(57): if specific_variability is None: -output-bmk-results.py(60): var = specific_variability[ (specific_variability['benchmark'] == bmk) & (specific_variability['symbol'].str.strip() == symb)] -output-bmk-results.py(61): if var.empty: -output-bmk-results.py(62): return np.nan -output-bmk-results.py(100): if not np.isnan(spec_thr): -output-bmk-results.py(107): if metric == "num_vect_loops" or metric == "num_sve_loops": -output-bmk-results.py(110): return default_threshold[(change_kind,metric,mode)] -output-bmk-results.py(228): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ -output-bmk-results.py(229): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold)) -output-bmk-results.py(228): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ -DEBUG: checking exe.improvement : 523.xalancbmk_r,cpuxalan_r_base.default : sample=100% (threshold=3%) -output-bmk-results.py(232): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold): - --- modulename: output-bmk-results, funcname: is_entry_improvement -output-bmk-results.py(192): if metric in metric_utils.higher_regress_metrics: -output-bmk-results.py(193): return (100 - result > threshold) -output-bmk-results.py(235): percent_change, short_diag, long_diag = get_short_long_diag(row, metric, sym_type, change_kind) - --- modulename: output-bmk-results, funcname: get_short_long_diag -output-bmk-results.py(137): bmk = row["benchmark"] -output-bmk-results.py(139): rel_value = row["rel_" + metric] -output-bmk-results.py(140): prev_value = row[metric + "_x"] -output-bmk-results.py(141): curr_value = row[metric + "_y"] -output-bmk-results.py(142): if metric == "num_vect_loops" or metric == "num_sve_loops": -output-bmk-results.py(152): suffix = "" -output-bmk-results.py(153): if metric == "sample": -output-bmk-results.py(154): prefix_regression = "slowed down by" -output-bmk-results.py(155): prefix_improvement = "sped up by" -output-bmk-results.py(156): suffix = "perf samples" -output-bmk-results.py(167): if sym_type=="symbol": -output-bmk-results.py(170): item=bmk -output-bmk-results.py(172): short_diag = "{1} {2}% - {0}".format(item, locals()["prefix_" + change_kind], abs(rel_value - 100)) -output-bmk-results.py(173): long_diag = "{0} - from {1} to {2} {3}".format(short_diag, prev_value, curr_value, suffix) -output-bmk-results.py(174): return abs(rel_value - 100), short_diag, long_diag -output-bmk-results.py(239): if metric == "sample" \ -output-bmk-results.py(240): and row['symbol_md5sum_x'] == row['symbol_md5sum_y'] \ -output-bmk-results.py(241): and row['symbol_md5sum_x'] != "-1" \ -output-bmk-results.py(242): and row['symbol_md5sum_x'] != "d41d8cd98f00b204e9800998ecf8427e": -output-bmk-results.py(243): f_skip.write_csv((row["benchmark"], row["symbol"], short_diag, long_diag)) - --- modulename: output-bmk-results, funcname: write_csv -output-bmk-results.py(41): if not self.predicate or not self.csvwriter: -output-bmk-results.py(43): self.csvwriter.writerow(arr) -output-bmk-results.py(244): continue -output-bmk-results.py(224): for index, row in out_df.iterrows(): -output-bmk-results.py(226): threshold = get_threshold(sym_type, metric, mode, row["benchmark"], row["symbol"]) - --- modulename: output-bmk-results, funcname: get_threshold -output-bmk-results.py(98): if metric == "sample": -output-bmk-results.py(99): spec_thr = get_specific_thresholds(metric, mode, bmk, symb) - --- modulename: output-bmk-results, funcname: get_specific_thresholds -output-bmk-results.py(57): if specific_variability is None: -output-bmk-results.py(60): var = specific_variability[ (specific_variability['benchmark'] == bmk) & (specific_variability['symbol'].str.strip() == symb)] -output-bmk-results.py(61): if var.empty: -output-bmk-results.py(62): return np.nan -output-bmk-results.py(100): if not np.isnan(spec_thr): -output-bmk-results.py(107): if metric == "num_vect_loops" or metric == "num_sve_loops": -output-bmk-results.py(110): return default_threshold[(change_kind,metric,mode)] -output-bmk-results.py(228): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ -output-bmk-results.py(229): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold)) -output-bmk-results.py(228): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ -DEBUG: checking exe.improvement : 531.deepsjeng_r,deepsjeng_r_base.default : sample=1% (threshold=3%) +DEBUG: checking exe.improvement : 531.deepsjeng_r,deepsjeng_r_base.default : sample=-1% (threshold=3%) output-bmk-results.py(232): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold): --- modulename: output-bmk-results, funcname: is_entry_improvement output-bmk-results.py(192): if metric in metric_utils.higher_regress_metrics: @@ -964,7 +767,7 @@ output-bmk-results.py(110): return default_threshold[(change_kind,metric,mod output-bmk-results.py(228): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ output-bmk-results.py(229): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold)) output-bmk-results.py(228): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\ -DEBUG: checking exe.improvement : 557.xz_r,xz_r_base.default : sample=0% (threshold=3%) +DEBUG: checking exe.improvement : 557.xz_r,xz_r_base.default : sample=-1% (threshold=3%) output-bmk-results.py(232): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold): --- modulename: output-bmk-results, funcname: is_entry_improvement output-bmk-results.py(192): if metric in metric_utils.higher_regress_metrics: @@ -1022,11 +825,13 @@ output-bmk-results.py(305): f_skip.close() output-bmk-results.py(29): if not self.outf: output-bmk-results.py(31): self.outf.close() output-bmk-results.py(32): if os.stat(self.filename).st_size == 0: +output-bmk-results.py(33): os.remove(self.filename) output-bmk-results.py(306): f_regr.close() --- modulename: output-bmk-results, funcname: close output-bmk-results.py(29): if not self.outf: output-bmk-results.py(31): self.outf.close() output-bmk-results.py(32): if os.stat(self.filename).st_size == 0: +output-bmk-results.py(33): os.remove(self.filename) output-bmk-results.py(307): f_ebp.close() --- modulename: output-bmk-results, funcname: close output-bmk-results.py(29): if not self.outf: diff --git a/notify/results.regressions b/notify/results.regressions deleted file mode 100644 index 4486cbe..0000000 --- a/notify/results.regressions +++ /dev/null @@ -1,5 +0,0 @@ -# namd_r_base.default,508.namd_r failed to run -# parest_r_base.default,510.parest_r failed to run -# povray_r_base.default,511.povray_r failed to run -# lbm_r_base.default,519.lbm_r failed to run -# omnetpp_r_base.default,520.omnetpp_r failed to run diff --git a/notify/status.regression b/notify/status.regression deleted file mode 100644 index 375d661..0000000 --- a/notify/status.regression +++ /dev/null @@ -1,5 +0,0 @@ -100,508.namd_r,namd_r_base.default,508.namd_r failed to run,508.namd_r failed to run
-100,510.parest_r,parest_r_base.default,510.parest_r failed to run,510.parest_r failed to run
-100,511.povray_r,povray_r_base.default,511.povray_r failed to run,511.povray_r failed to run
-100,519.lbm_r,lbm_r_base.default,519.lbm_r failed to run,519.lbm_r failed to run
-100,520.omnetpp_r,omnetpp_r_base.default,520.omnetpp_r failed to run,520.omnetpp_r failed to run
|