You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
451 lines
14 KiB
451 lines
14 KiB
4 months ago
|
# Copyright (c) 2013 The Chromium OS Authors. All rights reserved.
|
||
|
# Use of this source code is governed by a BSD-style license that can be
|
||
|
# found in the LICENSE file.
|
||
|
"""A reproducing entity.
|
||
|
|
||
|
Part of the Chrome build flags optimization.
|
||
|
|
||
|
The Task class is used by different modules. Each module fills in the
|
||
|
corresponding information into a Task instance. Class Task contains the bit set
|
||
|
representing the flags selection. The builder module is responsible for filling
|
||
|
the image and the checksum field of a Task. The executor module will put the
|
||
|
execution output to the execution field.
|
||
|
"""
|
||
|
|
||
|
__author__ = 'yuhenglong@google.com (Yuheng Long)'
|
||
|
|
||
|
import os
|
||
|
import subprocess
|
||
|
import sys
|
||
|
from uuid import uuid4
|
||
|
|
||
|
BUILD_STAGE = 1
|
||
|
TEST_STAGE = 2
|
||
|
|
||
|
# Message indicating that the build or test failed.
|
||
|
ERROR_STRING = 'error'
|
||
|
|
||
|
# The maximum number of tries a build can have. Some compilations may fail due
|
||
|
# to unexpected environment circumstance. This variable defines how many tries
|
||
|
# the build should attempt before giving up.
|
||
|
BUILD_TRIES = 3
|
||
|
|
||
|
# The maximum number of tries a test can have. Some tests may fail due to
|
||
|
# unexpected environment circumstance. This variable defines how many tries the
|
||
|
# test should attempt before giving up.
|
||
|
TEST_TRIES = 3
|
||
|
|
||
|
|
||
|
# Create the file/directory if it does not already exist.
|
||
|
def _CreateDirectory(file_name):
|
||
|
directory = os.path.dirname(file_name)
|
||
|
if not os.path.exists(directory):
|
||
|
os.makedirs(directory)
|
||
|
|
||
|
|
||
|
class Task(object):
|
||
|
"""A single reproducing entity.
|
||
|
|
||
|
A single test of performance with a particular set of flags. It records the
|
||
|
flag set, the image, the check sum of the image and the cost.
|
||
|
"""
|
||
|
|
||
|
# The command that will be used in the build stage to compile the tasks.
|
||
|
BUILD_COMMAND = None
|
||
|
# The command that will be used in the test stage to test the tasks.
|
||
|
TEST_COMMAND = None
|
||
|
# The directory to log the compilation and test results.
|
||
|
LOG_DIRECTORY = None
|
||
|
|
||
|
@staticmethod
|
||
|
def InitLogCommand(build_command, test_command, log_directory):
|
||
|
"""Set up the build and test command for the task and the log directory.
|
||
|
|
||
|
This framework is generic. It lets the client specify application specific
|
||
|
compile and test methods by passing different build_command and
|
||
|
test_command.
|
||
|
|
||
|
Args:
|
||
|
build_command: The command that will be used in the build stage to compile
|
||
|
this task.
|
||
|
test_command: The command that will be used in the test stage to test this
|
||
|
task.
|
||
|
log_directory: The directory to log the compilation and test results.
|
||
|
"""
|
||
|
|
||
|
Task.BUILD_COMMAND = build_command
|
||
|
Task.TEST_COMMAND = test_command
|
||
|
Task.LOG_DIRECTORY = log_directory
|
||
|
|
||
|
def __init__(self, flag_set):
|
||
|
"""Set up the optimization flag selection for this task.
|
||
|
|
||
|
Args:
|
||
|
flag_set: The optimization flag set that is encapsulated by this task.
|
||
|
"""
|
||
|
|
||
|
self._flag_set = flag_set
|
||
|
|
||
|
# A unique identifier that distinguishes this task from other tasks.
|
||
|
self._task_identifier = uuid4()
|
||
|
|
||
|
self._log_path = (Task.LOG_DIRECTORY, self._task_identifier)
|
||
|
|
||
|
# Initiate the hash value. The hash value is used so as not to recompute it
|
||
|
# every time the hash method is called.
|
||
|
self._hash_value = None
|
||
|
|
||
|
# Indicate that the task has not been compiled/tested.
|
||
|
self._build_cost = None
|
||
|
self._exe_cost = None
|
||
|
self._checksum = None
|
||
|
self._image = None
|
||
|
self._file_length = None
|
||
|
self._text_length = None
|
||
|
|
||
|
def __eq__(self, other):
|
||
|
"""Test whether two tasks are equal.
|
||
|
|
||
|
Two tasks are equal if their flag_set are equal.
|
||
|
|
||
|
Args:
|
||
|
other: The other task with which this task is tested equality.
|
||
|
Returns:
|
||
|
True if the encapsulated flag sets are equal.
|
||
|
"""
|
||
|
if isinstance(other, Task):
|
||
|
return self.GetFlags() == other.GetFlags()
|
||
|
return False
|
||
|
|
||
|
def __hash__(self):
|
||
|
if self._hash_value is None:
|
||
|
# Cache the hash value of the flags, so as not to recompute them.
|
||
|
self._hash_value = hash(self._flag_set)
|
||
|
return self._hash_value
|
||
|
|
||
|
def GetIdentifier(self, stage):
|
||
|
"""Get the identifier of the task in the stage.
|
||
|
|
||
|
The flag set uniquely identifies a task in the build stage. The checksum of
|
||
|
the image of the task uniquely identifies the task in the test stage.
|
||
|
|
||
|
Args:
|
||
|
stage: The stage (build/test) in which this method is called.
|
||
|
Returns:
|
||
|
Return the flag set in build stage and return the checksum in test stage.
|
||
|
"""
|
||
|
|
||
|
# Define the dictionary for different stage function lookup.
|
||
|
get_identifier_functions = {BUILD_STAGE: self.FormattedFlags,
|
||
|
TEST_STAGE: self.__GetCheckSum}
|
||
|
|
||
|
assert stage in get_identifier_functions
|
||
|
return get_identifier_functions[stage]()
|
||
|
|
||
|
def GetResult(self, stage):
|
||
|
"""Get the performance results of the task in the stage.
|
||
|
|
||
|
Args:
|
||
|
stage: The stage (build/test) in which this method is called.
|
||
|
Returns:
|
||
|
Performance results.
|
||
|
"""
|
||
|
|
||
|
# Define the dictionary for different stage function lookup.
|
||
|
get_result_functions = {BUILD_STAGE: self.__GetBuildResult,
|
||
|
TEST_STAGE: self.GetTestResult}
|
||
|
|
||
|
assert stage in get_result_functions
|
||
|
|
||
|
return get_result_functions[stage]()
|
||
|
|
||
|
def SetResult(self, stage, result):
|
||
|
"""Set the performance results of the task in the stage.
|
||
|
|
||
|
This method is called by the pipeling_worker to set the results for
|
||
|
duplicated tasks.
|
||
|
|
||
|
Args:
|
||
|
stage: The stage (build/test) in which this method is called.
|
||
|
result: The performance results of the stage.
|
||
|
"""
|
||
|
|
||
|
# Define the dictionary for different stage function lookup.
|
||
|
set_result_functions = {BUILD_STAGE: self.__SetBuildResult,
|
||
|
TEST_STAGE: self.__SetTestResult}
|
||
|
|
||
|
assert stage in set_result_functions
|
||
|
|
||
|
set_result_functions[stage](result)
|
||
|
|
||
|
def Done(self, stage):
|
||
|
"""Check whether the stage is done.
|
||
|
|
||
|
Args:
|
||
|
stage: The stage to be checked, build or test.
|
||
|
Returns:
|
||
|
True if the stage is done.
|
||
|
"""
|
||
|
|
||
|
# Define the dictionary for different result string lookup.
|
||
|
done_string = {BUILD_STAGE: self._build_cost, TEST_STAGE: self._exe_cost}
|
||
|
|
||
|
assert stage in done_string
|
||
|
|
||
|
return done_string[stage] is not None
|
||
|
|
||
|
def Work(self, stage):
|
||
|
"""Perform the task.
|
||
|
|
||
|
Args:
|
||
|
stage: The stage in which the task is performed, compile or test.
|
||
|
"""
|
||
|
|
||
|
# Define the dictionary for different stage function lookup.
|
||
|
work_functions = {BUILD_STAGE: self.__Compile, TEST_STAGE: self.__Test}
|
||
|
|
||
|
assert stage in work_functions
|
||
|
|
||
|
work_functions[stage]()
|
||
|
|
||
|
def FormattedFlags(self):
|
||
|
"""Format the optimization flag set of this task.
|
||
|
|
||
|
Returns:
|
||
|
The formatted optimization flag set that is encapsulated by this task.
|
||
|
"""
|
||
|
return str(self._flag_set.FormattedForUse())
|
||
|
|
||
|
def GetFlags(self):
|
||
|
"""Get the optimization flag set of this task.
|
||
|
|
||
|
Returns:
|
||
|
The optimization flag set that is encapsulated by this task.
|
||
|
"""
|
||
|
|
||
|
return self._flag_set
|
||
|
|
||
|
def __GetCheckSum(self):
|
||
|
"""Get the compilation image checksum of this task.
|
||
|
|
||
|
Returns:
|
||
|
The compilation image checksum of this task.
|
||
|
"""
|
||
|
|
||
|
# The checksum should be computed before this method is called.
|
||
|
assert self._checksum is not None
|
||
|
return self._checksum
|
||
|
|
||
|
def __Compile(self):
|
||
|
"""Run a compile.
|
||
|
|
||
|
This method compile an image using the present flags, get the image,
|
||
|
test the existent of the image and gathers monitoring information, and sets
|
||
|
the internal cost (fitness) for this set of flags.
|
||
|
"""
|
||
|
|
||
|
# Format the flags as a string as input to compile command. The unique
|
||
|
# identifier is passed to the compile command. If concurrent processes are
|
||
|
# used to compile different tasks, these processes can use the identifier to
|
||
|
# write to different file.
|
||
|
flags = self._flag_set.FormattedForUse()
|
||
|
command = '%s %s %s' % (Task.BUILD_COMMAND, ' '.join(flags),
|
||
|
self._task_identifier)
|
||
|
|
||
|
# Try BUILD_TRIES number of times before confirming that the build fails.
|
||
|
for _ in range(BUILD_TRIES):
|
||
|
try:
|
||
|
# Execute the command and get the execution status/results.
|
||
|
p = subprocess.Popen(command.split(),
|
||
|
stdout=subprocess.PIPE,
|
||
|
stderr=subprocess.PIPE)
|
||
|
(out, err) = p.communicate()
|
||
|
|
||
|
if out:
|
||
|
out = out.strip()
|
||
|
if out != ERROR_STRING:
|
||
|
# Each build results contains the checksum of the result image, the
|
||
|
# performance cost of the build, the compilation image, the length
|
||
|
# of the build, and the length of the text section of the build.
|
||
|
(checksum, cost, image, file_length, text_length) = out.split()
|
||
|
# Build successfully.
|
||
|
break
|
||
|
|
||
|
# Build failed.
|
||
|
cost = ERROR_STRING
|
||
|
except _:
|
||
|
# If there is exception getting the cost information of the build, the
|
||
|
# build failed.
|
||
|
cost = ERROR_STRING
|
||
|
|
||
|
# Convert the build cost from String to integer. The build cost is used to
|
||
|
# compare a task with another task. Set the build cost of the failing task
|
||
|
# to the max integer. The for loop will keep trying until either there is a
|
||
|
# success or BUILD_TRIES number of tries have been conducted.
|
||
|
self._build_cost = sys.maxint if cost == ERROR_STRING else float(cost)
|
||
|
|
||
|
self._checksum = checksum
|
||
|
self._file_length = file_length
|
||
|
self._text_length = text_length
|
||
|
self._image = image
|
||
|
|
||
|
self.__LogBuildCost(err)
|
||
|
|
||
|
def __Test(self):
|
||
|
"""__Test the task against benchmark(s) using the input test command."""
|
||
|
|
||
|
# Ensure that the task is compiled before being tested.
|
||
|
assert self._image is not None
|
||
|
|
||
|
# If the task does not compile, no need to test.
|
||
|
if self._image == ERROR_STRING:
|
||
|
self._exe_cost = ERROR_STRING
|
||
|
return
|
||
|
|
||
|
# The unique identifier is passed to the test command. If concurrent
|
||
|
# processes are used to compile different tasks, these processes can use the
|
||
|
# identifier to write to different file.
|
||
|
command = '%s %s %s' % (Task.TEST_COMMAND, self._image,
|
||
|
self._task_identifier)
|
||
|
|
||
|
# Try TEST_TRIES number of times before confirming that the build fails.
|
||
|
for _ in range(TEST_TRIES):
|
||
|
try:
|
||
|
p = subprocess.Popen(command.split(),
|
||
|
stdout=subprocess.PIPE,
|
||
|
stderr=subprocess.PIPE)
|
||
|
(out, err) = p.communicate()
|
||
|
|
||
|
if out:
|
||
|
out = out.strip()
|
||
|
if out != ERROR_STRING:
|
||
|
# The test results contains the performance cost of the test.
|
||
|
cost = out
|
||
|
# Test successfully.
|
||
|
break
|
||
|
|
||
|
# Test failed.
|
||
|
cost = ERROR_STRING
|
||
|
except _:
|
||
|
# If there is exception getting the cost information of the test, the
|
||
|
# test failed. The for loop will keep trying until either there is a
|
||
|
# success or TEST_TRIES number of tries have been conducted.
|
||
|
cost = ERROR_STRING
|
||
|
|
||
|
self._exe_cost = sys.maxint if (cost == ERROR_STRING) else float(cost)
|
||
|
|
||
|
self.__LogTestCost(err)
|
||
|
|
||
|
def __SetBuildResult(self, (checksum, build_cost, image, file_length,
|
||
|
text_length)):
|
||
|
self._checksum = checksum
|
||
|
self._build_cost = build_cost
|
||
|
self._image = image
|
||
|
self._file_length = file_length
|
||
|
self._text_length = text_length
|
||
|
|
||
|
def __GetBuildResult(self):
|
||
|
return (self._checksum, self._build_cost, self._image, self._file_length,
|
||
|
self._text_length)
|
||
|
|
||
|
def GetTestResult(self):
|
||
|
return self._exe_cost
|
||
|
|
||
|
def __SetTestResult(self, exe_cost):
|
||
|
self._exe_cost = exe_cost
|
||
|
|
||
|
def LogSteeringCost(self):
|
||
|
"""Log the performance results for the task.
|
||
|
|
||
|
This method is called by the steering stage and this method writes the
|
||
|
results out to a file. The results include the build and the test results.
|
||
|
"""
|
||
|
|
||
|
steering_log = '%s/%s/steering.txt' % self._log_path
|
||
|
|
||
|
_CreateDirectory(steering_log)
|
||
|
|
||
|
with open(steering_log, 'w') as out_file:
|
||
|
# Include the build and the test results.
|
||
|
steering_result = (self._flag_set, self._checksum, self._build_cost,
|
||
|
self._image, self._file_length, self._text_length,
|
||
|
self._exe_cost)
|
||
|
|
||
|
# Write out the result in the comma-separated format (CSV).
|
||
|
out_file.write('%s,%s,%s,%s,%s,%s,%s\n' % steering_result)
|
||
|
|
||
|
def __LogBuildCost(self, log):
|
||
|
"""Log the build results for the task.
|
||
|
|
||
|
The build results include the compilation time of the build, the result
|
||
|
image, the checksum, the file length and the text length of the image.
|
||
|
The file length of the image includes the length of the file of the image.
|
||
|
The text length only includes the length of the text section of the image.
|
||
|
|
||
|
Args:
|
||
|
log: The build log of this task.
|
||
|
"""
|
||
|
|
||
|
build_result_log = '%s/%s/build.txt' % self._log_path
|
||
|
|
||
|
_CreateDirectory(build_result_log)
|
||
|
|
||
|
with open(build_result_log, 'w') as out_file:
|
||
|
build_result = (self._flag_set, self._build_cost, self._image,
|
||
|
self._checksum, self._file_length, self._text_length)
|
||
|
|
||
|
# Write out the result in the comma-separated format (CSV).
|
||
|
out_file.write('%s,%s,%s,%s,%s,%s\n' % build_result)
|
||
|
|
||
|
# The build information about running the build.
|
||
|
build_run_log = '%s/%s/build_log.txt' % self._log_path
|
||
|
_CreateDirectory(build_run_log)
|
||
|
|
||
|
with open(build_run_log, 'w') as out_log_file:
|
||
|
# Write out the execution information.
|
||
|
out_log_file.write('%s' % log)
|
||
|
|
||
|
def __LogTestCost(self, log):
|
||
|
"""Log the test results for the task.
|
||
|
|
||
|
The test results include the runtime execution time of the test.
|
||
|
|
||
|
Args:
|
||
|
log: The test log of this task.
|
||
|
"""
|
||
|
|
||
|
test_log = '%s/%s/test.txt' % self._log_path
|
||
|
|
||
|
_CreateDirectory(test_log)
|
||
|
|
||
|
with open(test_log, 'w') as out_file:
|
||
|
test_result = (self._flag_set, self._checksum, self._exe_cost)
|
||
|
|
||
|
# Write out the result in the comma-separated format (CSV).
|
||
|
out_file.write('%s,%s,%s\n' % test_result)
|
||
|
|
||
|
# The execution information about running the test.
|
||
|
test_run_log = '%s/%s/test_log.txt' % self._log_path
|
||
|
|
||
|
_CreateDirectory(test_run_log)
|
||
|
|
||
|
with open(test_run_log, 'w') as out_log_file:
|
||
|
# Append the test log information.
|
||
|
out_log_file.write('%s' % log)
|
||
|
|
||
|
def IsImproved(self, other):
|
||
|
"""Compare the current task with another task.
|
||
|
|
||
|
Args:
|
||
|
other: The other task against which the current task is compared.
|
||
|
|
||
|
Returns:
|
||
|
True if this task has improvement upon the other task.
|
||
|
"""
|
||
|
|
||
|
# The execution costs must have been initiated.
|
||
|
assert self._exe_cost is not None
|
||
|
assert other.GetTestResult() is not None
|
||
|
|
||
|
return self._exe_cost < other.GetTestResult()
|