You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.

440 lines
17 KiB

# Lint as: python2, python3
# Copyright (c) 2013 The Chromium OS Authors. All rights reserved.
# Use of this source code is governed by a BSD-style license that can be
# found in the LICENSE file.
import collections
import six.moves.http_client
import logging
import os
import re
import socket
import time
import common
from autotest_lib.client.common_lib import error
from autotest_lib.client.common_lib.cros import retry
from autotest_lib.server import utils
GOOFY_JSONRPC_SERVER_PORT = 0x0FAC
GOOFY_RUNNING = 'RUNNING'
class GoofyProxyException(Exception):
"""Exception raised when a goofy rpc fails."""
pass
class GoofyRuntimeException(Exception):
"""Exception raised when something goes wrong while a test is running."""
pass
def retry_goofy_rpc(exception_tuple, timeout_min=30):
"""A decorator to use with goofy rpcs.
This decorator tries to recreate the goofy client proxy on
socket error. It will continue trying to do so until it
executes the method without any socket errors or till the
retry.retry decorator hits it's timeout.
Usage:
If you just want to recreate the proxy:
1. @retry_goofy_rpc(exception_tuple=(<exception>, socket.error),
timeout_min=<timeout>)
2. @retry_goofy_rpc(socket.error, timeout_min=<timeout>)
Note: you need to specify the socket.error exception because we
want to retry the call after recreating the proxy.
@param exception_tuple: A tuple of exceptions to pass to
the retry decorator. Any of these exceptions will result
in retries for the duration of timeout_min.
@param timeout_min: The timeout, in minutes, for which we should
retry the method ignoring any exception in exception_tuple.
"""
def inner_decorator(method):
"""Inner retry decorator applied to the method.
@param method: The method that needs to be wrapped in the decorator.
@return A wrapper function that implements the retry.
"""
@retry.retry(exception_tuple, timeout_min=timeout_min)
def wrapper(*args, **kwargs):
"""This wrapper handles socket errors.
If the method in question:
1. Throws an exception in exception_tuple and it is not a
socket.error, retry for timeout_min through retry.retry.
2. Throws a socket.error, recreate the client proxy, and
retry for timeout_min through retry.retry.
3. Throws an exception not in exception_tuple, fail.
"""
try:
return method(*args, **kwargs)
except socket.error as e:
goofy_proxy = args[0]
if type(goofy_proxy) is GoofyProxy:
logging.warning('Socket error while running factory tests '
'%s, recreating goofy proxy.', e)
goofy_proxy._create_client_proxy(timeout_min=timeout_min)
else:
logging.warning('Connectivity was lost and the retry '
'decorator was unable to recreate a goofy '
'client proxy, args: %s.', args)
raise
return wrapper
return inner_decorator
class GoofyProxy(object):
"""Client capable of making rpc calls to goofy.
Methods of this class that can cause goofy to change state
usually need a retry decorator. Methods that have a retry decorator
need to be 'pure', i.e return the same results when called multiple
times with the same argument.
There are 2 known exceptions this class can deal with, a socket.error
which happens when we try to execute an rpc when the DUT is, say, suspended
and a BadStatusLine, which we get when we try to execute an rpc while the
DUT is going through a factory_restart. Ideally we would like to handle
socket timeouts different from BadStatusLines as we can get connection
errors even when a device reboots and BadStatusLines ususally only when
factory restarts. crbug.com/281714.
"""
# This timeout was arbitrarily chosen as many tests in the factory test
# suite run for days. Ideally we would like to split this into at least 2
# timeouts, one which we use for rpcs that run while no other test is,
# running and is smaller than the second that is designed for use with rpcs
# that might execute simultaneously with a test. The latter needs a longer
# timeout since tests could suspend,resume for a long time, and a call like
# GetGoofyStatus should be tolerant to these suspend/resumes. In designing
# the base timeout one needs to allocate time to component methods of this
# class (such as _set_test_list) as a multiple of the number of rpcs it
# executes.
BASE_RPC_TIMEOUT = 1440
POLLING_INTERVAL = 5
FACTORY_BUG_RE = r'.*(/tmp/factory_bug.*tar.bz2).*'
UNTAR_COMMAND = 'tar jxf %s -C %s'
def __init__(self, host):
"""
@param host: The host object representing the DUT running goofy.
"""
self._host = host
self._raw_stop_running_tests()
self._create_client_proxy(timeout_min=self.BASE_RPC_TIMEOUT)
def _create_client_proxy(self, timeout_min=30):
"""Create a goofy client proxy.
Ping the host till it's up, then proceed to create a goofy proxy. We
don't wrap this method with a retry because it's used in the retry
decorator itself.
"""
# We don't ssh ping here as there is a potential dealy in O(minutes)
# with our ssh command against a sleeping DUT, once it wakes up, and
# that will lead to significant overhead incurred over many reboots.
self._host.ping_wait_up(timeout_min)
logging.info('Host is pingable, creating goofy client proxy')
self._client = self._host.rpc_server_tracker.jsonrpc_connect(
GOOFY_JSONRPC_SERVER_PORT)
@retry.retry((six.moves.http_client.BadStatusLine, socket.error),
timeout_min=BASE_RPC_TIMEOUT)
def _raw_stop_running_tests(self):
"""Stop running tests by shelling out to the DUT.
Use this method only before we have actually created the client
proxy, as shelling out has several pitfalls. We need to stop all
tests in a retry loop because tests will start executing as soon
as we have reimaged a DUT and trying to create the proxy while
the DUT is rebooting will lead to a spurious failure.
Note that we use the plain retry decorator for this method since
we don't need to recreate the client proxy on failure.
"""
logging.info('Stopping all tests and clearing factory state')
self._host.run('factory clear')
@retry_goofy_rpc((six.moves.http_client.BadStatusLine, socket.error),
timeout_min=BASE_RPC_TIMEOUT)
def _get_goofy_status(self):
"""Return status of goofy, ignoring socket timeouts and http exceptions.
"""
status = self._client.GetGoofyStatus().get('status')
return status
def _wait_for_goofy(self, timeout_min=BASE_RPC_TIMEOUT*2):
"""Wait till goofy is running or a timeout occurs.
@param timeout_min: Minutes to wait before timing this call out.
"""
current_time = time.time()
timeout_secs = timeout_min * 60
logging.info('Waiting on goofy')
while self._get_goofy_status() != GOOFY_RUNNING:
if time.time() - current_time > timeout_secs:
break
return
@retry_goofy_rpc(socket.error, timeout_min=BASE_RPC_TIMEOUT*2)
def _set_test_list(self, next_list):
"""Set the given test list for execution and turn on test automation.
Confirm that the given test list is a test that has been baked into
the image, then run it. Some test lists are configured to start
execution automatically when we call SetTestList, while others wait
for a corresponding RunTest.
@param next_list: The name of the test list.
@raise jsonrpclib.ProtocolError: If the test list we're trying to switch
to isn't on the DUT.
"""
# We can get a BadStatus line on 2 occassions:
# 1. As part of SwitchTestList goofy performs a factory restart, which
# will throw a BadStatusLine because the rpc can't exit cleanly. We
# don't want to retry on this exception, since we've already set the
# right test list.
# 2. If we try to set a test list while goofy is already down
# (from a previous factory restart). In this case we wouldn't have
# set the new test list, because we coulnd't connect to goofy.
# To properly set a new test list it's important to wait till goofy is
# up before attempting to set the test list, while being aware that the
# preceding httplib error is from the rpc we just executed leading to
# a factory restart. Also note that if the test list is not already on
# the DUT this method will fail, emitting the possible test lists one
# can switch to.
self._wait_for_goofy()
logging.info('Switching to test list %s', next_list)
try:
# Enable full factory test automation. Full test automation mode
# skips all manual tests and test barriers, which is what we want in
# the test lab. There are other automation modes: partial and none.
# In partial automation mode manual tests and barrier are enabled
# and user intervention is required; none disables automation.
self._client.SwitchTestList(next_list, 'full')
except six.moves.http_client.BadStatusLine:
logging.info('Switched to list %s, goofy restarting', next_list)
@retry_goofy_rpc((six.moves.http_client.BadStatusLine, socket.error),
timeout_min=BASE_RPC_TIMEOUT*2)
def _stop_running_tests(self):
"""Stop all running tests.
Wrap the StopTest rpc so we can attempt to stop tests even while a DUT
is suspended or rebooting.
"""
logging.info('Stopping tests.')
self._client.StopTest()
def _get_test_map(self):
"""Get a mapping of test suites -> tests.
Ignore entries for tests that don't have a path.
@return: A dictionary of the form
{'suite_name': ['suite_name.path_to_test', ...]}.
"""
test_all = set([test['path'] for test in self._client.GetTests()
if test.get('path')])
test_map = collections.defaultdict(list)
for names in test_all:
test_map[names.split('.')[0]].append(names)
return test_map
def _log_test_results(self, test_status, current_suite):
"""Format test status results and write them to status.log.
@param test_status: The status dictionary of a single test.
@param current_suite: The current suite name.
"""
try:
self._host.job.record('INFO', None, None,
'suite %s, test %s, status: %s' %
(current_suite, test_status.get('path'),
test_status.get('status')))
except AttributeError as e:
logging.error('Could not gather results for current test: %s', e)
@retry_goofy_rpc((six.moves.http_client.BadStatusLine, socket.error),
timeout_min=BASE_RPC_TIMEOUT*2)
def _get_test_info(self, test_name):
"""Get the status of one test.
@param test_name: The name of the test we need the status of.
@return: The entry for the test in the status dictionary.
"""
for test in self._client.GetTests():
if test['path'] == test_name:
return test
raise ValueError('Could not find test_name %s in _get_test_info.' %
test_name)
@retry_goofy_rpc((six.moves.http_client.BadStatusLine, socket.error),
timeout_min=BASE_RPC_TIMEOUT*2)
def _get_test_run_info(self, run_id):
"""Get the information about the given test run.
@param run_id: The ID of the test run.
@return: A dict of test run status.
"""
return self._client.GetTestRunStatus(run_id)
def _wait_on_run(self, run_id):
"""Wait until the given test run to end.
@param run_id: The ID of the test run.
@raises GoofyRuntimeException: If the test run does not finish
gracefully.
"""
finished_tests = set()
run_info = self._get_test_run_info(run_id)
while run_info['status'] == 'RUNNING':
finished = [(t['path'], t['status']) for t in
run_info['scheduled_tests']
if t['status'] in ('PASSED', 'FAILED')]
for t in finished:
if t not in finished_tests:
logging.info('[%s] %s', t[1], t[0])
finished_tests.add(t)
time.sleep(self.POLLING_INTERVAL)
run_info = self._get_test_run_info(run_id)
if run_info['status'] != 'FINISHED':
raise GoofyRuntimeException(
'The requested test run was interrupted.')
def _synchronous_run_suite(self, suite_name):
"""Run one suite and wait for it to finish.
Will start a test run for the specified suite_name and wait until it
ends.
@param suite_name: The name of the suite to wait for.
@raises GoofyProxyException: If the status of the suite
doesn't switch to active after we call RunTest.
@return: The result of the suite.
"""
logging.info('Starting suite: %s', suite_name)
run_id = self._client.RunTest(suite_name)
self._wait_on_run(run_id)
return self._get_test_run_info(run_id)
def monitor_tests(self, test_list):
"""Run a test list.
Will run each suite in the given list in sequence, starting each one
by name and waiting on its results. This method makes the following
assumptions:
- A test list is made up of self contained suites.
- These suites trigger several things in parallel.
- After a suite finishes it leaves goofy in an idle state.
It is not safe to pull results for individual tests during the suite
as the device could be rebooting, or goofy could be under stress.
Instead, this method synchronously waits on an entire suite, then
asks goofy for the status of each test in the suite. Since certain
test lists automatically start and others don't, this method stops
test list execution regardless, and sequentially triggers each suite.
@param test_list: The test list to run.
"""
self._set_test_list(test_list)
self._wait_for_goofy()
self._stop_running_tests()
test_map = self._get_test_map()
if test_map:
logging.info('About to execute tests: %s', test_map)
else:
raise GoofyRuntimeException('Test map is empty, you might have an '
'error in your test_list.')
for current_suite in test_map.keys():
logging.info('Processing suite %s', current_suite)
result = self._synchronous_run_suite(current_suite)
logging.info(result)
for test_names in test_map.get(current_suite):
self._log_test_results(self._get_test_info(test_names),
current_suite)
@retry_goofy_rpc((six.moves.http_client.BadStatusLine, socket.error),
timeout_min=BASE_RPC_TIMEOUT*2)
def get_results(self, resultsdir):
"""Copies results from the DUT to a local results directory.
Copy the tarball over to the results folder, untar, and delete the
tarball if everything was successful. This will effectively place
all the logs relevant to factory testing in the job's results folder.
@param resultsdir: The directory in which to untar the contents of the
tarball factory_bug generates.
"""
logging.info('Getting results logs for test_list.')
try:
factory_bug_log = self._host.run('factory_bug').stderr
except error.CmdError as e:
logging.error('Could not execute factory_bug: %s', e)
return
try:
factory_bug_tar = re.match(self.FACTORY_BUG_RE,
factory_bug_log).groups(1)[0]
except (IndexError, AttributeError):
logging.error('could not collect logs for factory results, '
'factory bug returned %s', factory_bug_log)
return
factory_bug_tar_file = os.path.basename(factory_bug_tar)
local_factory_bug_tar = os.path.join(resultsdir, factory_bug_tar_file)
try:
self._host.get_file(factory_bug_tar, local_factory_bug_tar)
except error.AutoservRunError as e:
logging.error('Failed to pull back the results tarball: %s', e)
return
try:
utils.run(self.UNTAR_COMMAND % (local_factory_bug_tar, resultsdir))
except error.CmdError as e:
logging.error('Failed to untar the results tarball: %s', e)
return
finally:
if os.path.exists(local_factory_bug_tar):
os.remove(local_factory_bug_tar)