Commit ae5512e5 authored by Nghia Nguyen's avatar Nghia Nguyen Committed by Commit Bot

Move oiplan benchmarks & related code to tools/perf/contrib/ directory

This makes oilpan benchmarks no longer run on perf waterfall. However,
these benchmarks are still runnable locally & on try bot.

Bug: 717320
Change-Id: I4eb52a6a8306a03879c808821f25238dfc5bb49e
Reviewed-on: https://chromium-review.googlesource.com/492806Reviewed-by: default avatarHitoshi Yoshida <peria@chromium.org>
Reviewed-by: default avatarKentaro Hara <haraken@chromium.org>
Commit-Queue: Ned Nguyen <nednguyen@google.com>
Cr-Commit-Position: refs/heads/master@{#468594}
parent 579b7d52
This source diff could not be displayed because it is too large. You can view the blob instead.
This diff is collapsed.
......@@ -57,10 +57,6 @@ memory.long_running_idle_gmail_tbmv2,ulan@chromium.org,
memory.top_10_mobile,perezju@chromium.org,
memory.top_10_mobile_stress,perezju@chromium.org,
octane,"bmeurer@chromium.org, mvstanton@chromium.org",
oilpan_gc_times.blink_perf_stress,,
oilpan_gc_times.key_silk_cases,,
oilpan_gc_times.sync_scroll.key_mobile_sites_smooth,,
oilpan_gc_times.tough_animation_cases,peria@chromium.org,
oortonline,,
oortonline_tbmv2,,
page_cycler_v2.basic_oopif,nasko@chromium.org,
......
haraken@chromium.org
peria@chromium.org
......@@ -8,10 +8,12 @@ from core import perf_benchmark
from benchmarks import blink_perf
from benchmarks import silk_flags
from measurements import oilpan_gc_times
import page_sets
from telemetry import benchmark
import page_sets
from contrib.oilpan import oilpan_gc_times
@benchmark.Enabled('content-shell')
class OilpanGCTimesBlinkPerfStress(perf_benchmark.PerfBenchmark):
......@@ -27,7 +29,6 @@ class OilpanGCTimesBlinkPerfStress(perf_benchmark.PerfBenchmark):
return blink_perf.CreateStorySetFromPath(path, blink_perf.SKIPPED_FILE)
@benchmark.Disabled('android') # crbug.com/589567
@benchmark.Owner(emails=['peria@chromium.org'])
class OilpanGCTimesSmoothnessAnimation(perf_benchmark.PerfBenchmark):
test = oilpan_gc_times.OilpanGCTimesForSmoothness
......@@ -60,8 +61,3 @@ class OilpanGCTimesSyncScrollKeyMobileSites(perf_benchmark.PerfBenchmark):
@classmethod
def Name(cls):
return 'oilpan_gc_times.sync_scroll.key_mobile_sites_smooth'
@classmethod
def ShouldDisable(cls, possible_browser): # http://crbug.com/597656
return (possible_browser.browser_type == 'reference' and
possible_browser.platform.GetDeviceTypeName() == 'Nexus 5X')
......@@ -179,7 +179,7 @@ class OilpanGCTimesForBlinkPerf(_OilpanGCTimesBase):
def __init__(self):
super(OilpanGCTimesForBlinkPerf, self).__init__()
with open(os.path.join(os.path.dirname(__file__), '..', 'benchmarks',
with open(os.path.join(os.path.dirname(__file__), '..', '..', 'benchmarks',
'blink_perf.js'), 'r') as f:
self._blink_perf_js = f.read()
......
......@@ -2,7 +2,7 @@
# Use of this source code is governed by a BSD-style license that can be
# found in the LICENSE file.
from measurements import oilpan_gc_times
from contrib.oilpan import oilpan_gc_times
from telemetry.internal.results import page_test_results
from telemetry.page import page as page_module
......@@ -100,7 +100,7 @@ class OilpanGCTimesTest(page_test_test_case.PageTestTestCase):
tab = mock.MagicMock()
with mock.patch(
'measurements.oilpan_gc_times.TimelineModel') as MockTimelineModel:
'contrib.oilpan.oilpan_gc_times.TimelineModel') as MockTimelineModel:
MockTimelineModel.return_value = data._model
measurement.ValidateAndMeasurePage(None, tab, data.results)
......@@ -134,7 +134,7 @@ class OilpanGCTimesTest(page_test_test_case.PageTestTestCase):
measurement._timeline_model = data._model
tab = mock.MagicMock()
with mock.patch(
'measurements.oilpan_gc_times.TimelineModel') as MockTimelineModel:
'contrib.oilpan.oilpan_gc_times.TimelineModel') as MockTimelineModel:
MockTimelineModel.return_value = data._model
measurement.ValidateAndMeasurePage(None, tab, data.results)
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment