From efc19603256ab7a76a7df79c0c7c5a0822e4a219 Mon Sep 17 00:00:00 2001 From: Florin Strugariu Date: Tue, 2 Feb 2021 14:18:57 +0000 Subject: [PATCH] Bug 1672719 - Stop reporting confidence/replayed/not-replayed to Perfherder and remove all asociated code r=perftest-reviewers,AlexandruIonescu Differential Revision: https://phabricator.services.mozilla.com/D103619 --- .../mozproxy/mozproxy/backends/base.py | 4 -- .../mozproxy/mozproxy/backends/mitm/mitm.py | 59 ------------------- testing/raptor/raptor/output.py | 7 +-- testing/raptor/raptor/results.py | 4 -- testing/raptor/raptor/webextension/base.py | 16 ----- 5 files changed, 1 insertion(+), 89 deletions(-) diff --git a/testing/mozbase/mozproxy/mozproxy/backends/base.py b/testing/mozbase/mozproxy/mozproxy/backends/base.py index dad6383e0df0..f2c2ba906406 100644 --- a/testing/mozbase/mozproxy/mozproxy/backends/base.py +++ b/testing/mozbase/mozproxy/mozproxy/backends/base.py @@ -31,7 +31,3 @@ class Playback(object): @abstractmethod def stop(self): pass - - @abstractmethod - def confidence(self): - pass diff --git a/testing/mozbase/mozproxy/mozproxy/backends/mitm/mitm.py b/testing/mozbase/mozproxy/mozproxy/backends/mitm/mitm.py index 991b4ebced75..15d206eb7e31 100644 --- a/testing/mozbase/mozproxy/mozproxy/backends/mitm/mitm.py +++ b/testing/mozbase/mozproxy/mozproxy/backends/mitm/mitm.py @@ -429,62 +429,3 @@ class Mitmproxy(Playback): return True except socket.error: return False - - def confidence(self): - """Extract confidence metrics from the netlocs file - and convert them to perftest results - """ - if len(self.playback_files) == 0: - LOG.warning( - "Proxy service did not load a recording file. " - "Confidence metrics will nt be generated" - ) - return - - file_name = ( - "mitm_netlocs_%s.json" - % os.path.splitext(os.path.basename(self.playback_files[0].recording_path))[ - 0 - ] - ) - path = os.path.normpath(os.path.join(self.upload_dir, file_name)) - if os.path.exists(path): - try: - LOG.info("Reading confidence values from: %s" % path) - with open(path, "r") as f: - data = json.load(f) - return { - "replay-confidence": { - "values": data["replay-confidence"], - "subtest-prefix-type": False, - "unit": "%", - "shouldAlert": False, - "lowerIsBetter": False, - }, - "recording-proportion-used": { - "values": data["recording-proportion-used"], - "subtest-prefix-type": False, - "unit": "%", - "shouldAlert": False, - "lowerIsBetter": False, - }, - "not-replayed": { - "values": data["not-replayed"], - "subtest-prefix-type": False, - "shouldAlert": False, - "unit": "a.u.", - }, - "replayed": { - "values": data["replayed"], - "subtest-prefix-type": False, - "unit": "a.u.", - "shouldAlert": False, - "lowerIsBetter": False, - }, - } - except Exception: - LOG.info("Can't read netlocs file!", exc_info=True) - return None - else: - LOG.info("Netlocs file is not available! Cant find %s" % path) - return None diff --git a/testing/raptor/raptor/output.py b/testing/raptor/raptor/output.py index 760ec7660b2a..0a245538d7fc 100644 --- a/testing/raptor/raptor/output.py +++ b/testing/raptor/raptor/output.py @@ -34,7 +34,6 @@ class PerftestOutput(object): self.summarized_supporting_data = [] self.summarized_screenshots = [] self.subtest_alert_on = subtest_alert_on - self.mozproxy_data = False self.browser_name = None self.browser_version = None @@ -97,10 +96,6 @@ class PerftestOutput(object): data_type = data_set["type"] LOG.info("summarizing %s data" % data_type) - if "mozproxy" in data_type: - self.mozproxy_data = True - LOG.info("data: {}".format(self.supporting_data)) - if data_type not in support_data_by_type: support_data_by_type[data_type] = { "framework": {"name": "raptor"}, @@ -253,7 +248,7 @@ class PerftestOutput(object): # dumped out. TODO: Bug 1515406 - Add option to output both supplementary # data (i.e. power) and the regular Raptor test result # Both are already available as separate PERFHERDER_DATA json blobs - if len(self.summarized_supporting_data) == 0 or self.mozproxy_data: + if len(self.summarized_supporting_data) == 0: LOG.info("PERFHERDER_DATA: %s" % json.dumps(self.summarized_results)) total_perfdata = 1 else: diff --git a/testing/raptor/raptor/results.py b/testing/raptor/raptor/results.py index 125bac28e1d2..84d61bef3a9c 100644 --- a/testing/raptor/raptor/results.py +++ b/testing/raptor/raptor/results.py @@ -191,10 +191,6 @@ class PerftestResultsHandler(object): expected_perfherder = 1 - if output.mozproxy_data: - # Check if we have mozproxy data available. - expected_perfherder += 1 - if is_resource_test(): # when resource tests are run, no perfherder data is output # for the regular raptor tests (i.e. speedometer) so we diff --git a/testing/raptor/raptor/webextension/base.py b/testing/raptor/raptor/webextension/base.py index 045082faa17f..8f60dfdec08e 100644 --- a/testing/raptor/raptor/webextension/base.py +++ b/testing/raptor/raptor/webextension/base.py @@ -143,22 +143,6 @@ class WebExtension(Perftest): if self.playback is not None: self.playback.stop() - - confidence_values = self.playback.confidence() - if confidence_values: - mozproxy_replay = { - u"summarize-values": False, - u"suite-suffix-type": False, - u"type": u"mozproxy", - u"test": test["name"], - u"unit": u"a.u.", - u"values": confidence_values, - u"shouldAlert": False, # Bug 1655841 temporary disable confidence metrics - } - self.control_server.submit_supporting_data(mozproxy_replay) - else: - LOG.info("Mozproxy replay confidence data not available!") - self.playback = None self.remove_raptor_webext()