blob: 4edc38c09d236954a37edaf165231d59e9eeb7c5 [file] [log] [blame]
#!/usr/bin/env python
# Copyright (c) 2014 The Chromium Authors. All rights reserved.
# Use of this source code is governed by a BSD-style license that can be
# found in the LICENSE file.
""" Generate bench_expectations file from a given set of bench data files. """
import argparse
import bench_util
import json
import os
import re
import sys
import urllib2
# Parameters for calculating bench ranges.
RANGE_RATIO_UPPER = 1.5 # Ratio of range for upper bounds.
RANGE_RATIO_LOWER = 2.0 # Ratio of range for lower bounds.
ERR_RATIO = 0.08 # Further widens the range by the ratio of average value.
ERR_UB = 1.0 # Adds an absolute upper error to cope with small benches.
ERR_LB = 1.5
# List of bench configs to monitor. Ignore all other configs.
CONFIGS_TO_INCLUDE = ['simple_viewport_1000x1000',
# List of flaky entries that should be excluded. Each entry is defined by a list
# of 3 strings, corresponding to the substrings of [bench, config, builder] to
# search for. A bench expectations line is excluded when each of the 3 strings
# in the list is a substring of the corresponding element of the given line. For
# instance, ['desk_yahooanswers', 'gpu', 'Ubuntu'] will skip expectation entries
# of SKP benchs whose name contains 'desk_yahooanswers' on all gpu-related
# configs of all Ubuntu builders.
def compute_ranges(benches, more_benches=None):
"""Given a list of bench numbers, calculate the alert range.
benches: a list of float bench values.
more_benches: a tuple of lists of additional bench values.
The first value of each tuple is the number of commits before the current
one that set of values is at, and the second value is a list of
bench results.
a list of float [lower_bound, upper_bound].
avg = sum(benches)/len(benches)
minimum = min(benches)
maximum = max(benches)
diff = maximum - minimum
return [minimum - diff*RANGE_RATIO_LOWER - avg*ERR_RATIO - ERR_LB,
maximum + diff*RANGE_RATIO_UPPER + avg*ERR_RATIO + ERR_UB]
def create_expectations_dict(revision_data_points, builder, extra_data=None):
"""Convert list of bench data points into a dictionary of expectations data.
revision_data_points: a list of BenchDataPoint objects.
builder: string of the corresponding buildbot builder name.
a dictionary of this form:
keys = tuple of (config, bench) strings.
values = list of float [expected, lower_bound, upper_bound] for the key.
bench_dict = {}
for point in revision_data_points:
if (point.time_type or # Not walltime which has time_type ''
not point.config in CONFIGS_TO_INCLUDE):
to_skip = False
for bench_substr, config_substr, builder_substr in ENTRIES_TO_EXCLUDE:
if (bench_substr in point.bench and config_substr in point.config and
builder_substr in builder):
to_skip = True
if to_skip:
key = (point.config, point.bench)
extras = []
for idx, dataset in extra_data:
for data in dataset:
if (data.bench == point.bench and data.config == point.config and
data.time_type == point.time_type and data.per_iter_time):
extras.append((idx, data.per_iter_time))
if key in bench_dict:
raise Exception('Duplicate bench entry: ' + str(key))
bench_dict[key] = [point.time] + compute_ranges(point.per_iter_time, extras)
return bench_dict
def get_parent_commits(start_hash, num_back):
"""Returns a list of commits that are the parent of the commit passed in."""
list_commits = urllib2.urlopen(
'' %
(start_hash, num_back))
# NOTE: Very brittle. Removes the four extraneous characters
# so json can be read successfully
trunc_list =[4:]
json_data = json.loads(trunc_list)
return [revision['commit'] for revision in json_data['log']]
def get_file_suffixes(commit_hash, directory):
"""Gets all the suffixes available in the directory"""
possible_files = os.listdir(directory)
prefix = 'bench_' + commit_hash + '_data_'
return [name[len(prefix):] for name in possible_files
if name.startswith(prefix)]
def download_bench_data(builder, commit_hash, suffixes, directory):
"""Downloads data, returns the number successfully downloaded"""
cur_files = os.listdir(directory)
count = 0
for suffix in suffixes:
file_name = 'bench_'+commit_hash+'_data_'+suffix
if file_name in cur_files:
src = urllib2.urlopen(_GS_CLOUD_FORMAT % (builder, file_name))
with open(os.path.join(directory, file_name), 'w') as dest:
count += 1
except urllib2.HTTPError:
return count
def main():
"""Reads bench data points, then calculate and export expectations.
parser = argparse.ArgumentParser()
'-a', '--representation_alg', default='25th',
help='bench representation algorithm to use, see')
'-b', '--builder', required=True,
help='name of the builder whose bench ranges we are computing.')
'-d', '--input_dir', required=True,
help='a directory containing bench data files.')
'-o', '--output_file', required=True,
help='file path and name for storing the output bench expectations.')
'-r', '--git_revision', required=True,
help='the git hash to indicate the revision of input data to use.')
'-t', '--back_track', required=False, default=10,
help='the number of commit hashes backwards to look to include' +
'in the calculations.')
'-m', '--max_commits', required=False, default=1,
help='the number of commit hashes to include in the calculations.')
args = parser.parse_args()
builder = args.builder
data_points = bench_util.parse_skp_bench_data(
args.input_dir, args.git_revision, args.representation_alg)
parent_commits = get_parent_commits(args.git_revision, args.back_track)
print "Using commits: {}".format(parent_commits)
suffixes = get_file_suffixes(args.git_revision, args.input_dir)
print "Using suffixes: {}".format(suffixes)
# TODO(kelvinly): Find a better approach to than directly copying from
# the GS server?
downloaded_commits = []
for idx, commit in enumerate(parent_commits):
num_downloaded = download_bench_data(
builder, commit, suffixes, args.input_dir)
if num_downloaded > 0:
downloaded_commits.append((num_downloaded, idx, commit))
if len(downloaded_commits) < args.max_commits:
print ('Less than desired number of commits found. Please increase'
'--back_track in later runs')
trunc_commits = sorted(downloaded_commits, reverse=True)[:args.max_commits]
extra_data = []
for _, idx, commit in trunc_commits:
extra_data.append((idx, bench_util.parse_skp_bench_data(
args.input_dir, commit, args.representation_alg)))
expectations_dict = create_expectations_dict(data_points, builder,
out_lines = []
keys = expectations_dict.keys()
for (config, bench) in keys:
(expected, lower_bound, upper_bound) = expectations_dict[(config, bench)]
'%(expected)s,%(lower_bound)s,%(upper_bound)s' % {
'bench': bench,
'config': config,
'builder': builder,
'representation': args.representation_alg,
'expected': expected,
'lower_bound': lower_bound,
'upper_bound': upper_bound})
with open(args.output_file, 'w') as file_handle:
if __name__ == "__main__":