summaryrefslogtreecommitdiff
path: root/notify
diff options
context:
space:
mode:
authorTCWG BuildSlave <tcwg-buildslave@linaro.org>2023-11-19 06:21:34 +0000
committerTCWG BuildSlave <tcwg-buildslave@linaro.org>2023-11-19 06:21:34 +0000
commitd7d94fa5a32db4a2faf4c49e22a1dfccb2848a21 (patch)
treec87620f64899fa6e2d52b543cdee3ac2c0ddd4fc /notify
parent4e23bd6f4f14ca3b4b32473ec37abecaf5549458 (diff)
onsuccess: #84: 1: [TCWG CI] https://ci.linaro.org/job/tcwg_bmk-code_size-coremark--gnu_eabi-arm_eabi-master-O2-build/84/
Results : | # reset_artifacts: | -10 | # build_abe binutils: | -9 | # build_abe stage1 -- --set gcc_override_configure=--disable-libsanitizer --set gcc_override_configure=--disable-multilib --set gcc_override_configure=--with-cpu=cortex-m4 --set gcc_override_configure=--with-mode=thumb --set gcc_override_configure=--with-float=hard: | -8 | # build_abe newlib: | -6 | # build_abe stage2 -- --set gcc_override_configure=--disable-libsanitizer --set gcc_override_configure=--disable-multilib --set gcc_override_configure=--with-cpu=cortex-m4 --set gcc_override_configure=--with-mode=thumb --set gcc_override_configure=--with-float=hard: | -5 | # benchmark -- -O2_mthumb: | 1 check_regression status : 0
Diffstat (limited to 'notify')
-rw-r--r--notify/jira/comment-template.txt2
-rw-r--r--notify/mail-body.txt36
-rw-r--r--notify/mail-subject.txt2
-rw-r--r--notify/output-bmk-results.log252
4 files changed, 171 insertions, 121 deletions
diff --git a/notify/jira/comment-template.txt b/notify/jira/comment-template.txt
index b0a71a9..33b556a 100644
--- a/notify/jira/comment-template.txt
+++ b/notify/jira/comment-template.txt
@@ -1,3 +1,3 @@
[GNU-686]
No change
-Details: https://ci.linaro.org/job/tcwg_bmk-code_size-coremark--gnu_eabi-arm_eabi-master-O2-build/82/artifact/artifacts/notify/mail-body.txt/*view*/
+Details: https://ci.linaro.org/job/tcwg_bmk-code_size-coremark--gnu_eabi-arm_eabi-master-O2-build/84/artifact/artifacts/notify/mail-body.txt/*view*/
diff --git a/notify/mail-body.txt b/notify/mail-body.txt
index d8fe565..8be2d41 100644
--- a/notify/mail-body.txt
+++ b/notify/mail-body.txt
@@ -2,19 +2,25 @@ Dear contributor, our automatic CI has detected problems related to your patch(e
In CI config tcwg_bmk-code_size-coremark/gnu_eabi-arm_eabi-master-O2 after:
- | 77 commits in binutils,gcc
- | e5da53e26f4 Remove declaration of re_comp
- | df3926bb636 Automatic date update in version.in
- | 328e0159543 Automatic date update in version.in
- | 943e09db605 Automatic date update in version.in
- | a7a0cb6c92e bfd, binutils: add gfx11 amdgpu architectures
- | ... and 9 more commits in binutils
- | 401dc18184a RISC-V: Fix RVV dynamic frm tests failure
- | 5dfa501d2f2 RISC-V: Support FP l/ll round and rint HF mode autovec
- | 93e92b2e5d6 Revert "[PATCH v2] In the pipeline, USE or CLOBBER should delay execution if it starts a new live range."
- | 4b754091bef i386: Remove *stack_protect_set_4s_<mode>_di alternative that will never match
- | 3a5a30792f6 LibF7: Use paper-pencil method for sqrt instead of Newton-Raphson iteration.
- | ... and 58 more commits in gcc
+ | 344 commits in binutils,gcc,newlib
+ | 1e62d51f295 Automatic date update in version.in
+ | 8fbb497b720 gas: bpf: do not allow referring to register names as symbols in operands
+ | 26c7a0ea380 Automatic date update in version.in
+ | e5d6f72eb3d bpf: avoid creating wrong symbols while parsing
+ | 41336620315 gdb: pass address_space to target dcache functions
+ | ... and 109 more commits in binutils
+ | af7fa3135b6 RISC-V: Fix bug of tuple move splitter
+ | f65f63c4d86 analyzer: new warning: -Wanalyzer-undefined-behavior-strtok [PR107573]
+ | 9d58d2d8ba2 Daily bump.
+ | f73808b3b4f MAINTAINERS: Update my email address.
+ | 279e407a06c libstdc++: Check string value_type in std::make_format_args [PR112607]
+ | ... and 214 more commits in gcc
+ | 643275ed1 Cygwin: Document /dev/disk/by-* subdirectories
+ | 41e13e4d9 Cygwin: Add /dev/disk/by-label and /dev/disk/by-uuid symlinks
+ | 6ccbfe6a3 __sfvwrite_r: call __sputc_r rather than putc in __SCLE case
+ | 61ccd3f94 v{fs}printf/v{fs}wprintf: create external output helpers
+ | 7e4840bc4 newlib.hin: regenerate
+ | ... and 6 more commits in newlib
No change
@@ -34,6 +40,6 @@ This benchmarking CI is work-in-progress, and we welcome feedback and suggestion
-----------------8<--------------------------8<--------------------------8<--------------------------
The information below can be used to reproduce a debug environment:
-Current build : https://ci.linaro.org/job/tcwg_bmk-code_size-coremark--gnu_eabi-arm_eabi-master-O2-build/82/artifact/artifacts
-Reference build : https://ci.linaro.org/job/tcwg_bmk-code_size-coremark--gnu_eabi-arm_eabi-master-O2-build/81/artifact/artifacts
+Current build : https://ci.linaro.org/job/tcwg_bmk-code_size-coremark--gnu_eabi-arm_eabi-master-O2-build/84/artifact/artifacts
+Reference build : https://ci.linaro.org/job/tcwg_bmk-code_size-coremark--gnu_eabi-arm_eabi-master-O2-build/82/artifact/artifacts
diff --git a/notify/mail-subject.txt b/notify/mail-subject.txt
index 48f7bbe..3da1122 100644
--- a/notify/mail-subject.txt
+++ b/notify/mail-subject.txt
@@ -1 +1 @@
-[Linaro-TCWG-CI] 77 commits in binutils,gcc: No change on arm O2
+[Linaro-TCWG-CI] 344 commits in binutils,gcc,newlib: No change on arm O2
diff --git a/notify/output-bmk-results.log b/notify/output-bmk-results.log
index ef2ab71..76a007f 100644
--- a/notify/output-bmk-results.log
+++ b/notify/output-bmk-results.log
@@ -1,214 +1,258 @@
--- modulename: output-bmk-results, funcname: <module>
<string>(1): --- modulename: output-bmk-results, funcname: main
-output-bmk-results.py(287): results_csv = sys.argv[1]
-output-bmk-results.py(288): variability_file = sys.argv[2]
-output-bmk-results.py(289): run_step_artifacts_dir = sys.argv[3]
-output-bmk-results.py(290): metric = sys.argv[4]
-output-bmk-results.py(291): mode = sys.argv[5]
-output-bmk-results.py(292): details = sys.argv[6]
-output-bmk-results.py(294): merged_df = read_results_csv(results_csv)
+output-bmk-results.py(322): results_csv = sys.argv[1]
+output-bmk-results.py(323): variability_file = sys.argv[2]
+output-bmk-results.py(324): run_step_artifacts_dir = sys.argv[3]
+output-bmk-results.py(325): metric = sys.argv[4]
+output-bmk-results.py(326): mode = sys.argv[5]
+output-bmk-results.py(327): details = sys.argv[6]
+output-bmk-results.py(329): merged_df = read_results_csv(results_csv)
--- modulename: output-bmk-results, funcname: read_results_csv
-output-bmk-results.py(277): df = pd.read_csv(results_csv)
-output-bmk-results.py(278): df = df.fillna(-1)
-output-bmk-results.py(280): for metric in get_comparable_metrics(df):
+output-bmk-results.py(312): df = pd.read_csv(results_csv)
+output-bmk-results.py(313): df = df.fillna(-1)
+output-bmk-results.py(315): for metric in get_comparable_metrics(df):
--- modulename: output-bmk-results, funcname: get_comparable_metrics
-output-bmk-results.py(198): return set(metric[len("rel_"):] for metric in df.columns[2:] if metric.startswith("rel_")) \
+output-bmk-results.py(206): return set(metric[len("rel_"):] for metric in df.columns[2:] if metric.startswith("rel_")) \
--- modulename: output-bmk-results, funcname: <genexpr>
-output-bmk-results.py(198): return set(metric[len("rel_"):] for metric in df.columns[2:] if metric.startswith("rel_")) \
+output-bmk-results.py(206): return set(metric[len("rel_"):] for metric in df.columns[2:] if metric.startswith("rel_")) \
--- modulename: output-bmk-results, funcname: <genexpr>
-output-bmk-results.py(198): return set(metric[len("rel_"):] for metric in df.columns[2:] if metric.startswith("rel_")) \
+output-bmk-results.py(206): return set(metric[len("rel_"):] for metric in df.columns[2:] if metric.startswith("rel_")) \
--- modulename: output-bmk-results, funcname: <genexpr>
-output-bmk-results.py(198): return set(metric[len("rel_"):] for metric in df.columns[2:] if metric.startswith("rel_")) \
+output-bmk-results.py(206): return set(metric[len("rel_"):] for metric in df.columns[2:] if metric.startswith("rel_")) \
--- modulename: output-bmk-results, funcname: <genexpr>
-output-bmk-results.py(198): return set(metric[len("rel_"):] for metric in df.columns[2:] if metric.startswith("rel_")) \
+output-bmk-results.py(206): return set(metric[len("rel_"):] for metric in df.columns[2:] if metric.startswith("rel_")) \
--- modulename: output-bmk-results, funcname: <genexpr>
-output-bmk-results.py(198): return set(metric[len("rel_"):] for metric in df.columns[2:] if metric.startswith("rel_")) \
+output-bmk-results.py(206): return set(metric[len("rel_"):] for metric in df.columns[2:] if metric.startswith("rel_")) \
--- modulename: output-bmk-results, funcname: <genexpr>
-output-bmk-results.py(198): return set(metric[len("rel_"):] for metric in df.columns[2:] if metric.startswith("rel_")) \
-output-bmk-results.py(198): return set(metric[len("rel_"):] for metric in df.columns[2:] if metric.startswith("rel_")) \
-output-bmk-results.py(198): return set(metric[len("rel_"):] for metric in df.columns[2:] if metric.startswith("rel_")) \
-output-bmk-results.py(198): return set(metric[len("rel_"):] for metric in df.columns[2:] if metric.startswith("rel_")) \
-output-bmk-results.py(198): return set(metric[len("rel_"):] for metric in df.columns[2:] if metric.startswith("rel_")) \
-output-bmk-results.py(198): return set(metric[len("rel_"):] for metric in df.columns[2:] if metric.startswith("rel_")) \
-output-bmk-results.py(198): return set(metric[len("rel_"):] for metric in df.columns[2:] if metric.startswith("rel_")) \
-output-bmk-results.py(198): return set(metric[len("rel_"):] for metric in df.columns[2:] if metric.startswith("rel_")) \
-output-bmk-results.py(198): return set(metric[len("rel_"):] for metric in df.columns[2:] if metric.startswith("rel_")) \
-output-bmk-results.py(198): return set(metric[len("rel_"):] for metric in df.columns[2:] if metric.startswith("rel_")) \
-output-bmk-results.py(198): return set(metric[len("rel_"):] for metric in df.columns[2:] if metric.startswith("rel_")) \
-output-bmk-results.py(199): & metric_utils.comparable_metrics
-output-bmk-results.py(198): return set(metric[len("rel_"):] for metric in df.columns[2:] if metric.startswith("rel_")) \
-output-bmk-results.py(281): df["rel_" + metric] = df["rel_" + metric].astype("int")
-output-bmk-results.py(282): df[metric + "_x"] = df[metric + "_x"].astype("int")
-output-bmk-results.py(283): df[metric + "_y"] = df[metric + "_y"].astype("int")
-output-bmk-results.py(280): for metric in get_comparable_metrics(df):
-output-bmk-results.py(281): df["rel_" + metric] = df["rel_" + metric].astype("int")
-output-bmk-results.py(282): df[metric + "_x"] = df[metric + "_x"].astype("int")
-output-bmk-results.py(283): df[metric + "_y"] = df[metric + "_y"].astype("int")
-output-bmk-results.py(280): for metric in get_comparable_metrics(df):
-output-bmk-results.py(281): df["rel_" + metric] = df["rel_" + metric].astype("int")
-output-bmk-results.py(282): df[metric + "_x"] = df[metric + "_x"].astype("int")
-output-bmk-results.py(283): df[metric + "_y"] = df[metric + "_y"].astype("int")
-output-bmk-results.py(280): for metric in get_comparable_metrics(df):
-output-bmk-results.py(281): df["rel_" + metric] = df["rel_" + metric].astype("int")
-output-bmk-results.py(282): df[metric + "_x"] = df[metric + "_x"].astype("int")
-output-bmk-results.py(283): df[metric + "_y"] = df[metric + "_y"].astype("int")
-output-bmk-results.py(280): for metric in get_comparable_metrics(df):
-output-bmk-results.py(284): return df
-output-bmk-results.py(295): read_specific_variability_file(variability_file)
+output-bmk-results.py(206): return set(metric[len("rel_"):] for metric in df.columns[2:] if metric.startswith("rel_")) \
+ --- modulename: output-bmk-results, funcname: <genexpr>
+output-bmk-results.py(206): return set(metric[len("rel_"):] for metric in df.columns[2:] if metric.startswith("rel_")) \
+output-bmk-results.py(206): return set(metric[len("rel_"):] for metric in df.columns[2:] if metric.startswith("rel_")) \
+output-bmk-results.py(206): return set(metric[len("rel_"):] for metric in df.columns[2:] if metric.startswith("rel_")) \
+output-bmk-results.py(206): return set(metric[len("rel_"):] for metric in df.columns[2:] if metric.startswith("rel_")) \
+output-bmk-results.py(206): return set(metric[len("rel_"):] for metric in df.columns[2:] if metric.startswith("rel_")) \
+output-bmk-results.py(206): return set(metric[len("rel_"):] for metric in df.columns[2:] if metric.startswith("rel_")) \
+output-bmk-results.py(206): return set(metric[len("rel_"):] for metric in df.columns[2:] if metric.startswith("rel_")) \
+output-bmk-results.py(206): return set(metric[len("rel_"):] for metric in df.columns[2:] if metric.startswith("rel_")) \
+output-bmk-results.py(206): return set(metric[len("rel_"):] for metric in df.columns[2:] if metric.startswith("rel_")) \
+output-bmk-results.py(206): return set(metric[len("rel_"):] for metric in df.columns[2:] if metric.startswith("rel_")) \
+output-bmk-results.py(206): return set(metric[len("rel_"):] for metric in df.columns[2:] if metric.startswith("rel_")) \
+output-bmk-results.py(206): return set(metric[len("rel_"):] for metric in df.columns[2:] if metric.startswith("rel_")) \
+output-bmk-results.py(206): return set(metric[len("rel_"):] for metric in df.columns[2:] if metric.startswith("rel_")) \
+output-bmk-results.py(207): & metric_utils.comparable_metrics
+output-bmk-results.py(206): return set(metric[len("rel_"):] for metric in df.columns[2:] if metric.startswith("rel_")) \
+output-bmk-results.py(316): df["rel_" + metric] = df["rel_" + metric].astype("int")
+output-bmk-results.py(317): df[metric + "_x"] = df[metric + "_x"].astype("int")
+output-bmk-results.py(318): df[metric + "_y"] = df[metric + "_y"].astype("int")
+output-bmk-results.py(315): for metric in get_comparable_metrics(df):
+output-bmk-results.py(316): df["rel_" + metric] = df["rel_" + metric].astype("int")
+output-bmk-results.py(317): df[metric + "_x"] = df[metric + "_x"].astype("int")
+output-bmk-results.py(318): df[metric + "_y"] = df[metric + "_y"].astype("int")
+output-bmk-results.py(315): for metric in get_comparable_metrics(df):
+output-bmk-results.py(316): df["rel_" + metric] = df["rel_" + metric].astype("int")
+output-bmk-results.py(317): df[metric + "_x"] = df[metric + "_x"].astype("int")
+output-bmk-results.py(318): df[metric + "_y"] = df[metric + "_y"].astype("int")
+output-bmk-results.py(315): for metric in get_comparable_metrics(df):
+output-bmk-results.py(316): df["rel_" + metric] = df["rel_" + metric].astype("int")
+output-bmk-results.py(317): df[metric + "_x"] = df[metric + "_x"].astype("int")
+output-bmk-results.py(318): df[metric + "_y"] = df[metric + "_y"].astype("int")
+output-bmk-results.py(315): for metric in get_comparable_metrics(df):
+output-bmk-results.py(319): return df
+output-bmk-results.py(330): read_specific_variability_file(variability_file)
--- modulename: output-bmk-results, funcname: read_specific_variability_file
output-bmk-results.py(51): if not os.path.exists(bmk_specific_filename):
output-bmk-results.py(53): specific_variability = pd.read_csv(bmk_specific_filename, index_col=False)
-output-bmk-results.py(296): output_bmk_results(merged_df, run_step_artifacts_dir, metric, mode, details)
+output-bmk-results.py(331): output_bmk_results(merged_df, run_step_artifacts_dir, metric, mode, details)
--- modulename: output-bmk-results, funcname: output_bmk_results
-output-bmk-results.py(248): f_regr = Outfile("{0}/results.regressions".format(run_step_artifacts), "w")
+output-bmk-results.py(278): f_regr = Outfile("{0}/results.regressions".format(run_step_artifacts), "w")
--- modulename: output-bmk-results, funcname: __init__
output-bmk-results.py(19): self.filename=filename
output-bmk-results.py(20): self.predicate=predicate
output-bmk-results.py(21): if predicate:
output-bmk-results.py(22): self.outf = open(filename, mode)
output-bmk-results.py(23): self.csvwriter = csv.writer(self.outf)
-output-bmk-results.py(249): f_ebp = Outfile("{0}/extra-bisect-params".format(run_step_artifacts), "w")
+output-bmk-results.py(279): f_ebp = Outfile("{0}/extra-bisect-params".format(run_step_artifacts), "w")
--- modulename: output-bmk-results, funcname: __init__
output-bmk-results.py(19): self.filename=filename
output-bmk-results.py(20): self.predicate=predicate
output-bmk-results.py(21): if predicate:
output-bmk-results.py(22): self.outf = open(filename, mode)
output-bmk-results.py(23): self.csvwriter = csv.writer(self.outf)
-output-bmk-results.py(250): f_skip = Outfile("{0}/any.skipped".format(run_step_artifacts), "w", predicate=(details=="verbose"))
+output-bmk-results.py(280): f_skip = Outfile("{0}/any.skipped".format(run_step_artifacts), "w", predicate=(details=="verbose"))
--- modulename: output-bmk-results, funcname: __init__
output-bmk-results.py(19): self.filename=filename
output-bmk-results.py(20): self.predicate=predicate
output-bmk-results.py(21): if predicate:
output-bmk-results.py(22): self.outf = open(filename, mode)
output-bmk-results.py(23): self.csvwriter = csv.writer(self.outf)
-output-bmk-results.py(252): f_ebp.write("extra_build_params=")
+output-bmk-results.py(282): f_ebp.write("extra_build_params=")
--- modulename: output-bmk-results, funcname: write
output-bmk-results.py(36): if not self.predicate or not self.outf:
output-bmk-results.py(38): self.outf.write(string)
-output-bmk-results.py(256): df = merged_df[merged_df["benchmark"] != "Mean"]
-output-bmk-results.py(259): exe_df = df[df["symbol"].str.endswith("_base.default")]
-output-bmk-results.py(260): sym_df = df[~df["symbol"].str.endswith("_base.default")]
-output-bmk-results.py(262): output_bmk_results_1(exe_df, "exe", "regression", f_regr, f_skip, f_ebp, run_step_artifacts, metric, mode, details)
+output-bmk-results.py(286): df = merged_df[merged_df["benchmark"] != "Mean"]
+output-bmk-results.py(289): exe_df = df[df["symbol"].str.endswith("_base.default")]
+output-bmk-results.py(290): sym_df = df[~df["symbol"].str.endswith("_base.default")]
+output-bmk-results.py(293): output_bmk_results_status(exe_df, "regression", f_regr, f_ebp, run_step_artifacts, details)
+ --- modulename: output-bmk-results, funcname: output_bmk_results_status
+output-bmk-results.py(256): f_out = Outfile("{0}/status.{1}".format(run_step_artifacts, change_kind), "w", predicate=(details=="verbose"))
+ --- modulename: output-bmk-results, funcname: __init__
+output-bmk-results.py(19): self.filename=filename
+output-bmk-results.py(20): self.predicate=predicate
+output-bmk-results.py(21): if predicate:
+output-bmk-results.py(22): self.outf = open(filename, mode)
+output-bmk-results.py(23): self.csvwriter = csv.writer(self.outf)
+output-bmk-results.py(258): print(results_df)
+Empty DataFrame
+Columns: [benchmark, symbol, rel_sample, rel_size, rel_num_vect_loops, rel_num_sve_loops, rel_symbol_md5sum, rel_status, sample_x, sample_y, size_x, size_y, num_vect_loops_x, num_vect_loops_y, num_sve_loops_x, num_sve_loops_y, symbol_md5sum_x, symbol_md5sum_y, status_x, status_y]
+Index: []
+output-bmk-results.py(261): for index, row in results_df.iterrows():
+output-bmk-results.py(275): f_out.close()
+ --- modulename: output-bmk-results, funcname: close
+output-bmk-results.py(29): if not self.outf:
+output-bmk-results.py(31): self.outf.close()
+output-bmk-results.py(32): if os.stat(self.filename).st_size == 0:
+output-bmk-results.py(33): os.remove(self.filename)
+output-bmk-results.py(294): output_bmk_results_status(exe_df, "improvement", None, None, run_step_artifacts, details)
+ --- modulename: output-bmk-results, funcname: output_bmk_results_status
+output-bmk-results.py(256): f_out = Outfile("{0}/status.{1}".format(run_step_artifacts, change_kind), "w", predicate=(details=="verbose"))
+ --- modulename: output-bmk-results, funcname: __init__
+output-bmk-results.py(19): self.filename=filename
+output-bmk-results.py(20): self.predicate=predicate
+output-bmk-results.py(21): if predicate:
+output-bmk-results.py(22): self.outf = open(filename, mode)
+output-bmk-results.py(23): self.csvwriter = csv.writer(self.outf)
+output-bmk-results.py(258): print(results_df)
+Empty DataFrame
+Columns: [benchmark, symbol, rel_sample, rel_size, rel_num_vect_loops, rel_num_sve_loops, rel_symbol_md5sum, rel_status, sample_x, sample_y, size_x, size_y, num_vect_loops_x, num_vect_loops_y, num_sve_loops_x, num_sve_loops_y, symbol_md5sum_x, symbol_md5sum_y, status_x, status_y]
+Index: []
+output-bmk-results.py(261): for index, row in results_df.iterrows():
+output-bmk-results.py(275): f_out.close()
+ --- modulename: output-bmk-results, funcname: close
+output-bmk-results.py(29): if not self.outf:
+output-bmk-results.py(31): self.outf.close()
+output-bmk-results.py(32): if os.stat(self.filename).st_size == 0:
+output-bmk-results.py(33): os.remove(self.filename)
+output-bmk-results.py(297): output_bmk_results_1(exe_df, "exe", "regression", f_regr, f_skip, f_ebp, run_step_artifacts, metric, mode, details)
--- modulename: output-bmk-results, funcname: output_bmk_results_1
-output-bmk-results.py(210): f_out = Outfile("{0}/{1}.{2}".format(run_step_artifacts, sym_type, change_kind), "w", predicate=(details=="verbose"))
+output-bmk-results.py(218): f_out = Outfile("{0}/{1}.{2}".format(run_step_artifacts, sym_type, change_kind), "w", predicate=(details=="verbose"))
--- modulename: output-bmk-results, funcname: __init__
output-bmk-results.py(19): self.filename=filename
output-bmk-results.py(20): self.predicate=predicate
output-bmk-results.py(21): if predicate:
output-bmk-results.py(22): self.outf = open(filename, mode)
output-bmk-results.py(23): self.csvwriter = csv.writer(self.outf)
-output-bmk-results.py(212): rel_metric = "rel_" + metric
-output-bmk-results.py(213): out_df = results_df[results_df[rel_metric] != -1]
-output-bmk-results.py(216): for index, row in out_df.iterrows():
-output-bmk-results.py(245): f_out.close()
+output-bmk-results.py(220): rel_metric = "rel_" + metric
+output-bmk-results.py(221): out_df = results_df[results_df[rel_metric] != -1]
+output-bmk-results.py(224): for index, row in out_df.iterrows():
+output-bmk-results.py(253): f_out.close()
--- modulename: output-bmk-results, funcname: close
output-bmk-results.py(29): if not self.outf:
output-bmk-results.py(31): self.outf.close()
output-bmk-results.py(32): if os.stat(self.filename).st_size == 0:
output-bmk-results.py(33): os.remove(self.filename)
-output-bmk-results.py(263): output_bmk_results_1(exe_df, "exe", "improvement", None, f_skip, None, run_step_artifacts, metric, mode, details)
+output-bmk-results.py(298): output_bmk_results_1(exe_df, "exe", "improvement", None, f_skip, None, run_step_artifacts, metric, mode, details)
--- modulename: output-bmk-results, funcname: output_bmk_results_1
-output-bmk-results.py(210): f_out = Outfile("{0}/{1}.{2}".format(run_step_artifacts, sym_type, change_kind), "w", predicate=(details=="verbose"))
+output-bmk-results.py(218): f_out = Outfile("{0}/{1}.{2}".format(run_step_artifacts, sym_type, change_kind), "w", predicate=(details=="verbose"))
--- modulename: output-bmk-results, funcname: __init__
output-bmk-results.py(19): self.filename=filename
output-bmk-results.py(20): self.predicate=predicate
output-bmk-results.py(21): if predicate:
output-bmk-results.py(22): self.outf = open(filename, mode)
output-bmk-results.py(23): self.csvwriter = csv.writer(self.outf)
-output-bmk-results.py(212): rel_metric = "rel_" + metric
-output-bmk-results.py(213): out_df = results_df[results_df[rel_metric] != -1]
-output-bmk-results.py(216): for index, row in out_df.iterrows():
-output-bmk-results.py(245): f_out.close()
+output-bmk-results.py(220): rel_metric = "rel_" + metric
+output-bmk-results.py(221): out_df = results_df[results_df[rel_metric] != -1]
+output-bmk-results.py(224): for index, row in out_df.iterrows():
+output-bmk-results.py(253): f_out.close()
--- modulename: output-bmk-results, funcname: close
output-bmk-results.py(29): if not self.outf:
output-bmk-results.py(31): self.outf.close()
output-bmk-results.py(32): if os.stat(self.filename).st_size == 0:
output-bmk-results.py(33): os.remove(self.filename)
-output-bmk-results.py(265): output_bmk_results_1(sym_df, "symbol", "regression", f_regr, f_skip, f_ebp, run_step_artifacts, metric, mode, details)
+output-bmk-results.py(300): output_bmk_results_1(sym_df, "symbol", "regression", f_regr, f_skip, f_ebp, run_step_artifacts, metric, mode, details)
--- modulename: output-bmk-results, funcname: output_bmk_results_1
-output-bmk-results.py(210): f_out = Outfile("{0}/{1}.{2}".format(run_step_artifacts, sym_type, change_kind), "w", predicate=(details=="verbose"))
+output-bmk-results.py(218): f_out = Outfile("{0}/{1}.{2}".format(run_step_artifacts, sym_type, change_kind), "w", predicate=(details=="verbose"))
--- modulename: output-bmk-results, funcname: __init__
output-bmk-results.py(19): self.filename=filename
output-bmk-results.py(20): self.predicate=predicate
output-bmk-results.py(21): if predicate:
output-bmk-results.py(22): self.outf = open(filename, mode)
output-bmk-results.py(23): self.csvwriter = csv.writer(self.outf)
-output-bmk-results.py(212): rel_metric = "rel_" + metric
-output-bmk-results.py(213): out_df = results_df[results_df[rel_metric] != -1]
-output-bmk-results.py(216): for index, row in out_df.iterrows():
-output-bmk-results.py(218): threshold = get_threshold(sym_type, metric, mode, row["benchmark"], row["symbol"])
+output-bmk-results.py(220): rel_metric = "rel_" + metric
+output-bmk-results.py(221): out_df = results_df[results_df[rel_metric] != -1]
+output-bmk-results.py(224): for index, row in out_df.iterrows():
+output-bmk-results.py(226): threshold = get_threshold(sym_type, metric, mode, row["benchmark"], row["symbol"])
--- modulename: output-bmk-results, funcname: get_threshold
output-bmk-results.py(98): if metric == "sample":
output-bmk-results.py(107): if metric == "num_vect_loops" or metric == "num_sve_loops":
output-bmk-results.py(110): return default_threshold[(change_kind,metric,mode)]
-output-bmk-results.py(220): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
-output-bmk-results.py(221): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold))
-output-bmk-results.py(220): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
-DEBUG: checking symbol.regression : coremark,coremark : size=0% (threshold=10%)
-output-bmk-results.py(224): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold):
+output-bmk-results.py(228): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
+output-bmk-results.py(229): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold))
+output-bmk-results.py(228): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
+DEBUG: checking symbol.regression : coremark,coremark : size=1% (threshold=10%)
+output-bmk-results.py(232): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold):
--- modulename: output-bmk-results, funcname: is_entry_regression
-output-bmk-results.py(175): if metric in metric_utils.higher_regress_metrics:
-output-bmk-results.py(176): return (result - 100 > threshold)
-output-bmk-results.py(225): continue
-output-bmk-results.py(216): for index, row in out_df.iterrows():
-output-bmk-results.py(245): f_out.close()
+output-bmk-results.py(183): if metric in metric_utils.higher_regress_metrics:
+output-bmk-results.py(184): return (result - 100 > threshold)
+output-bmk-results.py(233): continue
+output-bmk-results.py(224): for index, row in out_df.iterrows():
+output-bmk-results.py(253): f_out.close()
--- modulename: output-bmk-results, funcname: close
output-bmk-results.py(29): if not self.outf:
output-bmk-results.py(31): self.outf.close()
output-bmk-results.py(32): if os.stat(self.filename).st_size == 0:
output-bmk-results.py(33): os.remove(self.filename)
-output-bmk-results.py(266): output_bmk_results_1(sym_df, "symbol", "improvement", None, f_skip, None, run_step_artifacts, metric, mode, details)
+output-bmk-results.py(301): output_bmk_results_1(sym_df, "symbol", "improvement", None, f_skip, None, run_step_artifacts, metric, mode, details)
--- modulename: output-bmk-results, funcname: output_bmk_results_1
-output-bmk-results.py(210): f_out = Outfile("{0}/{1}.{2}".format(run_step_artifacts, sym_type, change_kind), "w", predicate=(details=="verbose"))
+output-bmk-results.py(218): f_out = Outfile("{0}/{1}.{2}".format(run_step_artifacts, sym_type, change_kind), "w", predicate=(details=="verbose"))
--- modulename: output-bmk-results, funcname: __init__
output-bmk-results.py(19): self.filename=filename
output-bmk-results.py(20): self.predicate=predicate
output-bmk-results.py(21): if predicate:
output-bmk-results.py(22): self.outf = open(filename, mode)
output-bmk-results.py(23): self.csvwriter = csv.writer(self.outf)
-output-bmk-results.py(212): rel_metric = "rel_" + metric
-output-bmk-results.py(213): out_df = results_df[results_df[rel_metric] != -1]
-output-bmk-results.py(216): for index, row in out_df.iterrows():
-output-bmk-results.py(218): threshold = get_threshold(sym_type, metric, mode, row["benchmark"], row["symbol"])
+output-bmk-results.py(220): rel_metric = "rel_" + metric
+output-bmk-results.py(221): out_df = results_df[results_df[rel_metric] != -1]
+output-bmk-results.py(224): for index, row in out_df.iterrows():
+output-bmk-results.py(226): threshold = get_threshold(sym_type, metric, mode, row["benchmark"], row["symbol"])
--- modulename: output-bmk-results, funcname: get_threshold
output-bmk-results.py(98): if metric == "sample":
output-bmk-results.py(107): if metric == "num_vect_loops" or metric == "num_sve_loops":
output-bmk-results.py(110): return default_threshold[(change_kind,metric,mode)]
-output-bmk-results.py(220): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
-output-bmk-results.py(221): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold))
-output-bmk-results.py(220): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
-DEBUG: checking symbol.improvement : coremark,coremark : size=0% (threshold=10%)
-output-bmk-results.py(224): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold):
+output-bmk-results.py(228): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
+output-bmk-results.py(229): .format(sym_type, change_kind, row["benchmark"], row["symbol"], metric, 100-row["rel_" + metric], threshold))
+output-bmk-results.py(228): print("DEBUG: checking {0}.{1} : {2},{3} : {4}={5}% (threshold={6}%)"\
+DEBUG: checking symbol.improvement : coremark,coremark : size=1% (threshold=10%)
+output-bmk-results.py(232): if not is_entry_xxx[change_kind](metric, row["rel_" + metric], threshold):
--- modulename: output-bmk-results, funcname: is_entry_improvement
-output-bmk-results.py(184): if metric in metric_utils.higher_regress_metrics:
-output-bmk-results.py(185): return (100 - result > threshold)
-output-bmk-results.py(225): continue
-output-bmk-results.py(216): for index, row in out_df.iterrows():
-output-bmk-results.py(245): f_out.close()
+output-bmk-results.py(192): if metric in metric_utils.higher_regress_metrics:
+output-bmk-results.py(193): return (100 - result > threshold)
+output-bmk-results.py(233): continue
+output-bmk-results.py(224): for index, row in out_df.iterrows():
+output-bmk-results.py(253): f_out.close()
--- modulename: output-bmk-results, funcname: close
output-bmk-results.py(29): if not self.outf:
output-bmk-results.py(31): self.outf.close()
output-bmk-results.py(32): if os.stat(self.filename).st_size == 0:
output-bmk-results.py(33): os.remove(self.filename)
-output-bmk-results.py(268): f_ebp.write("\n")
+output-bmk-results.py(303): f_ebp.write("\n")
--- modulename: output-bmk-results, funcname: write
output-bmk-results.py(36): if not self.predicate or not self.outf:
output-bmk-results.py(38): self.outf.write(string)
-output-bmk-results.py(270): f_skip.close()
+output-bmk-results.py(305): f_skip.close()
--- modulename: output-bmk-results, funcname: close
output-bmk-results.py(29): if not self.outf:
output-bmk-results.py(31): self.outf.close()
output-bmk-results.py(32): if os.stat(self.filename).st_size == 0:
output-bmk-results.py(33): os.remove(self.filename)
-output-bmk-results.py(271): f_regr.close()
+output-bmk-results.py(306): f_regr.close()
--- modulename: output-bmk-results, funcname: close
output-bmk-results.py(29): if not self.outf:
output-bmk-results.py(31): self.outf.close()
output-bmk-results.py(32): if os.stat(self.filename).st_size == 0:
output-bmk-results.py(33): os.remove(self.filename)
-output-bmk-results.py(272): f_ebp.close()
+output-bmk-results.py(307): f_ebp.close()
--- modulename: output-bmk-results, funcname: close
output-bmk-results.py(29): if not self.outf:
output-bmk-results.py(31): self.outf.close()
output-bmk-results.py(32): if os.stat(self.filename).st_size == 0:
-output-bmk-results.py(297): return 0
+output-bmk-results.py(332): return 0