You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.

2130 lines
88 KiB

#!/usr/bin/env python3
# -*- coding: utf-8 -*-
# Copyright (c) 2011 The Chromium OS Authors. All rights reserved.
# Use of this source code is governed by a BSD-style license that can be
# found in the LICENSE file.
"""Module of result cache unittest."""
from __future__ import print_function
import os
import shutil
import tempfile
import unittest
import unittest.mock as mock
import image_checksummer
import machine_manager
import test_flag
from label import MockLabel
from results_cache import CacheConditions
from results_cache import PerfDataReadError
from results_cache import PidVerificationError
from results_cache import Result
from results_cache import ResultsCache
from results_cache import TelemetryResult
from cros_utils import command_executer
from cros_utils import logger
from cros_utils import misc
# pylint: disable=line-too-long
OUTPUT = """CMD (True): ./test_that.sh\
--remote=172.17.128.241 --board=lumpy LibCBench
CMD (None): cd /usr/local/google/home/yunlian/gd/src/build/images/lumpy/latest/../../../../..; cros_sdk -- ./in_chroot_cmd6X7Cxu.sh
Identity added: /tmp/test_that.PO1234567/autotest_key (/tmp/test_that.PO1234567/autotest_key)
INFO : Using emerged autotests already installed at /build/lumpy/usr/local/autotest.
INFO : Running the following control files 1 times:
INFO : * 'client/site_tests/platform_LibCBench/control'
INFO : Running client test client/site_tests/platform_LibCBench/control
./server/autoserv -m 172.17.128.241 --ssh-port 22 -c client/site_tests/platform_LibCBench/control -r /tmp/test_that.PO1234567/platform_LibCBench --test-retry=0 --args
ERROR:root:import statsd failed, no stats will be reported.
14:20:22 INFO | Results placed in /tmp/test_that.PO1234567/platform_LibCBench
14:20:22 INFO | Processing control file
14:20:23 INFO | Starting master ssh connection '/usr/bin/ssh -a -x -N -o ControlMaster=yes -o ControlPath=/tmp/_autotmp_VIIP67ssh-master/socket -o StrictHostKeyChecking=no -o UserKnownHostsFile=/dev/null -o BatchMode=yes -o ConnectTimeout=30 -o ServerAliveInterval=180 -o ServerAliveCountMax=3 -o ConnectionAttempts=4 -o Protocol=2 -l root -p 22 172.17.128.241'
14:20:23 ERROR| [stderr] Warning: Permanently added '172.17.128.241' (RSA) to the list of known hosts.
14:20:23 INFO | INFO ---- ---- kernel=3.8.11 localtime=May 22 14:20:23 timestamp=1369257623
14:20:23 INFO | Installing autotest on 172.17.128.241
14:20:23 INFO | Using installation dir /usr/local/autotest
14:20:23 WARNI| No job_repo_url for <remote host: 172.17.128.241>
14:20:23 INFO | Could not install autotest using the packaging system: No repos to install an autotest client from. Trying other methods
14:20:23 INFO | Installation of autotest completed
14:20:24 WARNI| No job_repo_url for <remote host: 172.17.128.241>
14:20:24 INFO | Executing /usr/local/autotest/bin/autotest /usr/local/autotest/control phase 0
14:20:24 INFO | Entered autotestd_monitor.
14:20:24 INFO | Finished launching tail subprocesses.
14:20:24 INFO | Finished waiting on autotestd to start.
14:20:26 INFO | START ---- ---- timestamp=1369257625 localtime=May 22 14:20:25
14:20:26 INFO | START platform_LibCBench platform_LibCBench timestamp=1369257625 localtime=May 22 14:20:25
14:20:30 INFO | GOOD platform_LibCBench platform_LibCBench timestamp=1369257630 localtime=May 22 14:20:30 completed successfully
14:20:30 INFO | END GOOD platform_LibCBench platform_LibCBench timestamp=1369257630 localtime=May 22 14:20:30
14:20:31 INFO | END GOOD ---- ---- timestamp=1369257630 localtime=May 22 14:20:30
14:20:31 INFO | Got lock of exit_code_file.
14:20:31 INFO | Released lock of exit_code_file and closed it.
OUTPUT: ==============================
OUTPUT: Current time: 2013-05-22 14:20:32.818831 Elapsed: 0:01:30 ETA: Unknown
Done: 0% [ ]
OUTPUT: Thread Status:
RUNNING: 1 ('ttt: LibCBench (1)' 0:01:21)
Machine Status:
Machine Thread Lock Status Checksum
172.17.128.241 ttt: LibCBench (1) True RUNNING 3ba9f2ecbb222f20887daea5583d86ba
OUTPUT: ==============================
14:20:33 INFO | Killing child processes.
14:20:33 INFO | Client complete
14:20:33 INFO | Finished processing control file
14:20:33 INFO | Starting master ssh connection '/usr/bin/ssh -a -x -N -o ControlMaster=yes -o ControlPath=/tmp/_autotmp_aVJUgmssh-master/socket -o StrictHostKeyChecking=no -o UserKnownHostsFile=/dev/null -o BatchMode=yes -o ConnectTimeout=30 -o ServerAliveInterval=180 -o ServerAliveCountMax=3 -o ConnectionAttempts=4 -o Protocol=2 -l root -p 22 172.17.128.241'
14:20:33 ERROR| [stderr] Warning: Permanently added '172.17.128.241' (RSA) to the list of known hosts.
INFO : Test results:
-------------------------------------------------------------------
platform_LibCBench [ PASSED ]
platform_LibCBench/platform_LibCBench [ PASSED ]
platform_LibCBench/platform_LibCBench b_malloc_big1__0_ 0.00375231466667
platform_LibCBench/platform_LibCBench b_malloc_big2__0_ 0.002951359
platform_LibCBench/platform_LibCBench b_malloc_bubble__0_ 0.015066374
platform_LibCBench/platform_LibCBench b_malloc_sparse__0_ 0.015053784
platform_LibCBench/platform_LibCBench b_malloc_thread_local__0_ 0.01138439
platform_LibCBench/platform_LibCBench b_malloc_thread_stress__0_ 0.0367894733333
platform_LibCBench/platform_LibCBench b_malloc_tiny1__0_ 0.000768474333333
platform_LibCBench/platform_LibCBench b_malloc_tiny2__0_ 0.000581407333333
platform_LibCBench/platform_LibCBench b_pthread_create_serial1__0_ 0.0291785246667
platform_LibCBench/platform_LibCBench b_pthread_createjoin_serial1__0_ 0.031907936
platform_LibCBench/platform_LibCBench b_pthread_createjoin_serial2__0_ 0.043485347
platform_LibCBench/platform_LibCBench b_pthread_uselesslock__0_ 0.0294113346667
platform_LibCBench/platform_LibCBench b_regex_compile____a_b_c__d_b__ 0.00529833933333
platform_LibCBench/platform_LibCBench b_regex_search____a_b_c__d_b__ 0.00165455066667
platform_LibCBench/platform_LibCBench b_regex_search___a_25_b__ 0.0496191923333
platform_LibCBench/platform_LibCBench b_stdio_putcgetc__0_ 0.100005711667
platform_LibCBench/platform_LibCBench b_stdio_putcgetc_unlocked__0_ 0.0371443833333
platform_LibCBench/platform_LibCBench b_string_memset__0_ 0.00275405066667
platform_LibCBench/platform_LibCBench b_string_strchr__0_ 0.00456903
platform_LibCBench/platform_LibCBench b_string_strlen__0_ 0.044893587
platform_LibCBench/platform_LibCBench b_string_strstr___aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaac__ 0.118360778
platform_LibCBench/platform_LibCBench b_string_strstr___aaaaaaaaaaaaaaaaaaaaaaaaac__ 0.068957325
platform_LibCBench/platform_LibCBench b_string_strstr___aaaaaaaaaaaaaacccccccccccc__ 0.0135694476667
platform_LibCBench/platform_LibCBench b_string_strstr___abcdefghijklmnopqrstuvwxyz__ 0.0134553343333
platform_LibCBench/platform_LibCBench b_string_strstr___azbycxdwevfugthsirjqkplomn__ 0.0133123556667
platform_LibCBench/platform_LibCBench b_utf8_bigbuf__0_ 0.0473772253333
platform_LibCBench/platform_LibCBench b_utf8_onebyone__0_ 0.130938538333
-------------------------------------------------------------------
Total PASS: 2/2 (100%)
INFO : Elapsed time: 0m16s
"""
error = """
ERROR: Identity added: /tmp/test_that.Z4Ld/autotest_key (/tmp/test_that.Z4Ld/autotest_key)
INFO : Using emerged autotests already installed at /build/lumpy/usr/local/autotest.
INFO : Running the following control files 1 times:
INFO : * 'client/site_tests/platform_LibCBench/control'
INFO : Running client test client/site_tests/platform_LibCBench/control
INFO : Test results:
INFO : Elapsed time: 0m18s
"""
keyvals = {
'': 'PASS',
'b_stdio_putcgetc__0_': '0.100005711667',
'b_string_strstr___azbycxdwevfugthsirjqkplomn__': '0.0133123556667',
'b_malloc_thread_local__0_': '0.01138439',
'b_string_strlen__0_': '0.044893587',
'b_malloc_sparse__0_': '0.015053784',
'b_string_memset__0_': '0.00275405066667',
'platform_LibCBench': 'PASS',
'b_pthread_uselesslock__0_': '0.0294113346667',
'b_string_strchr__0_': '0.00456903',
'b_pthread_create_serial1__0_': '0.0291785246667',
'b_string_strstr___aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaac__': '0.118360778',
'b_string_strstr___aaaaaaaaaaaaaacccccccccccc__': '0.0135694476667',
'b_pthread_createjoin_serial1__0_': '0.031907936',
'b_malloc_thread_stress__0_': '0.0367894733333',
'b_regex_search____a_b_c__d_b__': '0.00165455066667',
'b_malloc_bubble__0_': '0.015066374',
'b_malloc_big2__0_': '0.002951359',
'b_stdio_putcgetc_unlocked__0_': '0.0371443833333',
'b_pthread_createjoin_serial2__0_': '0.043485347',
'b_regex_search___a_25_b__': '0.0496191923333',
'b_utf8_bigbuf__0_': '0.0473772253333',
'b_malloc_big1__0_': '0.00375231466667',
'b_regex_compile____a_b_c__d_b__': '0.00529833933333',
'b_string_strstr___aaaaaaaaaaaaaaaaaaaaaaaaac__': '0.068957325',
'b_malloc_tiny2__0_': '0.000581407333333',
'b_utf8_onebyone__0_': '0.130938538333',
'b_malloc_tiny1__0_': '0.000768474333333',
'b_string_strstr___abcdefghijklmnopqrstuvwxyz__': '0.0134553343333'
}
PERF_DATA_HEADER = """
# ========
# captured on : Thu Jan 01 00:00:00 1980
# header version : 1
# data offset : 536
# data size : 737678672
# feat offset : 737679208
# hostname : localhost
# os release : 5.4.61
# perf version :
# arch : aarch64
# nrcpus online : 8
# nrcpus avail : 8
# total memory : 5911496 kB
# cmdline : /usr/bin/perf record -e instructions -p {pid}
# event : name = instructions, , id = ( 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193 ), type = 8, size = 112
# event : name = dummy:u, , id = ( 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204 ), type = 1, size = 112, config = 0x9
# CPU_TOPOLOGY info available, use -I to display
# pmu mappings: software = 1, uprobe = 6, cs_etm = 8, breakpoint = 5, tracepoint = 2, armv8_pmuv3 = 7
# contains AUX area data (e.g. instruction trace)
# time of first sample : 0.000000
# time of last sample : 0.000000
# sample duration : 0.000 ms
# missing features: TRACING_DATA CPUDESC CPUID NUMA_TOPOLOGY BRANCH_STACK GROUP_DESC STAT CACHE MEM_TOPOLOGY CLOCKID DIR_FORMAT
# ========
#
"""
TURBOSTAT_LOG_OUTPUT = \
"""CPU Avg_MHz Busy% Bzy_MHz TSC_MHz IRQ CoreTmp
- 329 12.13 2723 2393 10975 77
0 336 12.41 2715 2393 6328 77
2 323 11.86 2731 2393 4647 69
CPU Avg_MHz Busy% Bzy_MHz TSC_MHz IRQ CoreTmp
- 1940 67.46 2884 2393 39920 83
0 1827 63.70 2877 2393 21184 83
2 2053 71.22 2891 2393 18736 67
CPU Avg_MHz Busy% Bzy_MHz TSC_MHz IRQ CoreTmp
- 1927 66.02 2927 2393 48946 84
0 1880 64.47 2925 2393 24457 84
2 1973 67.57 2928 2393 24489 69
CPU Avg_MHz Busy% Bzy_MHz TSC_MHz IRQ CoreTmp
- 1899 64.84 2937 2393 42540 72
0 2135 72.82 2940 2393 23615 65
2 1663 56.85 2934 2393 18925 72
CPU Avg_MHz Busy% Bzy_MHz TSC_MHz IRQ CoreTmp
- 1908 65.24 2932 2393 43172 75
0 1876 64.25 2928 2393 20743 75
2 1939 66.24 2936 2393 22429 69
CPU Avg_MHz Busy% Bzy_MHz TSC_MHz IRQ CoreTmp
- 1553 53.12 2933 2393 35488 46
0 1484 50.80 2929 2393 18246 46
2 1623 55.44 2936 2393 17242 45
CPU Avg_MHz Busy% Bzy_MHz TSC_MHz IRQ CoreTmp
- 843 29.83 2832 2393 28161 47
0 827 29.35 2826 2393 16093 47
2 858 30.31 2838 2393 12068 46
"""
TURBOSTAT_DATA = {
'cpufreq': {
'all': [2723, 2884, 2927, 2937, 2932, 2933, 2832]
},
'cputemp': {
'all': [77, 83, 84, 72, 75, 46, 47]
},
}
TOP_LOG = \
"""
PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND
4102 chronos 12 -8 3454472 238300 118188 R 41.8 6.1 0:08.37 chrome
4204 chronos 12 -8 2492716 205728 179016 S 11.8 5.3 0:03.89 chrome
4890 root 20 0 3396 2064 1596 R 11.8 0.1 0:00.03 top
375 root 0 -20 0 0 0 S 5.9 0.0 0:00.17 kworker/u13
617 syslog 20 0 25332 8372 7888 S 5.9 0.2 0:00.77 sys-journal
PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND
5745 chronos 20 0 5438580 139328 67988 R 122.8 3.6 0:04.26 chrome
912 root -51 0 0 0 0 S 2.0 0.0 0:01.04 irq/cros-ec
121 root 20 0 0 0 0 S 1.0 0.0 0:00.45 spi5
4811 root 20 0 6808 4084 3492 S 1.0 0.1 0:00.02 sshd
4890 root 20 0 3364 2148 1596 R 1.0 0.1 0:00.36 top
5205 chronos 12 -8 3673780 240928 130864 S 1.0 6.2 0:07.30 chrome
PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND
5745 chronos 20 0 5434484 139432 63892 R 107.9 3.6 0:05.35 chrome
5713 chronos 20 0 5178652 103120 50372 S 17.8 2.6 0:01.13 chrome
7 root 20 0 0 0 0 S 1.0 0.0 0:00.73 rcu_preempt
855 root 20 0 0 0 0 S 1.0 0.0 0:00.01 kworker/4:2
"""
TOP_DATA = [
{
'cmd': 'chrome-5745',
'cpu_use_avg': 115.35,
'count': 2,
'top5_cpu_use': [122.8, 107.9],
},
{
'cmd': 'chrome-5713',
'cpu_use_avg': 8.9,
'count': 1,
'top5_cpu_use': [17.8]
},
{
'cmd': 'irq/cros-ec-912',
'cpu_use_avg': 1.0,
'count': 1,
'top5_cpu_use': [2.0],
},
{
'cmd': 'chrome-5205',
'cpu_use_avg': 0.5,
'count': 1,
'top5_cpu_use': [1.0]
},
{
'cmd': 'spi5-121',
'cpu_use_avg': 0.5,
'count': 1,
'top5_cpu_use': [1.0],
},
{
'cmd': 'sshd-4811',
'cpu_use_avg': 0.5,
'count': 1,
'top5_cpu_use': [1.0],
},
{
'cmd': 'rcu_preempt-7',
'cpu_use_avg': 0.5,
'count': 1,
'top5_cpu_use': [1.0],
},
{
'cmd': 'kworker/4:2-855',
'cpu_use_avg': 0.5,
'count': 1,
'top5_cpu_use': [1.0],
},
]
TOP_OUTPUT = \
""" COMMAND AVG CPU% SEEN HIGHEST 5
chrome 128.250000 6 [122.8, 107.9, 17.8, 5.0, 2.0]
irq/230-cros-ec 1.000000 1 [2.0]
sshd 0.500000 1 [1.0]
irq/231-cros-ec 0.500000 1 [1.0]
spi5 0.500000 1 [1.0]
rcu_preempt 0.500000 1 [1.0]
kworker/4:2 0.500000 1 [1.0]
"""
CPUSTATS_UNIQ_OUTPUT = \
"""
/sys/devices/system/cpu/cpu0/cpufreq/cpuinfo_cur_freq 1512000
/sys/devices/system/cpu/cpu1/cpufreq/cpuinfo_cur_freq 1512000
/sys/devices/system/cpu/cpu3/cpufreq/cpuinfo_cur_freq 2016000
soc-thermal 44444
little-cpu 41234
big-cpu 51234
/sys/devices/system/cpu/cpu0/cpufreq/cpuinfo_cur_freq 1500000
/sys/devices/system/cpu/cpu1/cpufreq/cpuinfo_cur_freq 1600000
/sys/devices/system/cpu/cpu3/cpufreq/cpuinfo_cur_freq 2012000
soc-thermal 45456
little-cpu 42555
big-cpu 61724
"""
CPUSTATS_UNIQ_DATA = {
'cpufreq': {
'cpu0': [1512, 1500],
'cpu1': [1512, 1600],
'cpu3': [2016, 2012]
},
'cputemp': {
'soc-thermal': [44.4, 45.5],
'little-cpu': [41.2, 42.6],
'big-cpu': [51.2, 61.7]
}
}
CPUSTATS_DUPL_OUTPUT = \
"""
/sys/devices/system/cpu/cpu0/cpufreq/cpuinfo_cur_freq 1512000
/sys/devices/system/cpu/cpu1/cpufreq/cpuinfo_cur_freq 1512000
/sys/devices/system/cpu/cpu2/cpufreq/cpuinfo_cur_freq 1512000
/sys/devices/system/cpu/cpu3/cpufreq/cpuinfo_cur_freq 2016000
/sys/devices/system/cpu/cpu0/cpufreq/cpuinfo_cur_freq 1500000
/sys/devices/system/cpu/cpu1/cpufreq/cpuinfo_cur_freq 1500000
/sys/devices/system/cpu/cpu2/cpufreq/cpuinfo_cur_freq 1500000
/sys/devices/system/cpu/cpu3/cpufreq/cpuinfo_cur_freq 2016000
/sys/devices/system/cpu/cpu0/cpufreq/cpuinfo_cur_freq 1614000
/sys/devices/system/cpu/cpu1/cpufreq/cpuinfo_cur_freq 1614000
/sys/devices/system/cpu/cpu2/cpufreq/cpuinfo_cur_freq 1614000
/sys/devices/system/cpu/cpu3/cpufreq/cpuinfo_cur_freq 1982000
"""
CPUSTATS_DUPL_DATA = {
'cpufreq': {
'cpu0': [1512, 1500, 1614],
'cpu3': [2016, 2016, 1982]
},
}
TMP_DIR1 = '/tmp/tmpAbcXyz'
HISTOGRAMSET = \
"""
[
{
"values": [
"cache_temperature_cold",
"typical",
"cache_temperature:cold"
],
"guid": "db6d463b-7c07-4873-b839-db0652ccb97e",
"type": "GenericSet"
},
{
"values": [
"cache_temperature_warm",
"typical",
"cache_temperature:warm"
],
"guid": "a270eb9d-3bb0-472a-951d-74ac3398b718",
"type": "GenericSet"
},
{
"sampleValues": [
1111.672
],
"name": "timeToFirstContentfulPaint",
"diagnostics": {
"storyTags": "a270eb9d-3bb0-472a-951d-74ac3398b718"
},
"unit": "ms_smallerIsBetter"
},
{
"sampleValues": [
1146.459
],
"name": "timeToFirstContentfulPaint",
"diagnostics": {
"storyTags": "db6d463b-7c07-4873-b839-db0652ccb97e"
},
"unit": "ms_smallerIsBetter"
},
{
"sampleValues": [
888.328
],
"name": "timeToFirstContentfulPaint",
"diagnostics": {
"storyTags": "a270eb9d-3bb0-472a-951d-74ac3398b718"
},
"unit": "ms_smallerIsBetter"
},
{
"sampleValues": [
853.541
],
"name": "timeToFirstContentfulPaint",
"diagnostics": {
"storyTags": "db6d463b-7c07-4873-b839-db0652ccb97e"
},
"unit": "ms_smallerIsBetter"
},
{
"sampleValues": [
400.000
],
"name": "timeToFirstContentfulPaint",
"diagnostics": {
"storyTags": "a270eb9d-3bb0-472a-951d-74ac3398b718"
},
"unit": "ms_smallerIsBetter"
}
]
"""
# pylint: enable=line-too-long
class MockResult(Result):
"""Mock result class."""
def __init__(self, mylogger, label, logging_level, machine):
super(MockResult, self).__init__(mylogger, label, logging_level, machine)
def FindFilesInResultsDir(self, find_args):
return ''
# pylint: disable=arguments-differ
def GetKeyvals(self, temp=False):
if temp:
pass
return keyvals
class ResultTest(unittest.TestCase):
"""Result test class."""
def __init__(self, *args, **kwargs):
super(ResultTest, self).__init__(*args, **kwargs)
self.callFakeProcessResults = False
self.fakeCacheReturnResult = None
self.callGetResultsDir = False
self.callProcessResults = False
self.callGetPerfReportFiles = False
self.kv_dict = None
self.tmpdir = ''
self.callGetNewKeyvals = False
self.callGetResultsFile = False
self.callGetPerfDataFiles = False
self.callGetTurbostatFile = False
self.callGetCpustatsFile = False
self.callGetTopFile = False
self.callGetCpuinfoFile = False
self.callGetWaitTimeFile = False
self.args = None
self.callGatherPerfResults = False
self.mock_logger = mock.Mock(spec=logger.Logger)
self.mock_cmd_exec = mock.Mock(spec=command_executer.CommandExecuter)
self.mock_label = MockLabel('mock_label', 'build', 'chromeos_image',
'autotest_dir', 'debug_dir', '/tmp', 'lumpy',
'remote', 'image_args', 'cache_dir', 'average',
'gcc', False, None)
def testCreateFromRun(self):
result = MockResult.CreateFromRun(logger.GetLogger(), 'average',
self.mock_label, 'remote1', OUTPUT, error,
0, True)
self.assertEqual(result.keyvals, keyvals)
self.assertEqual(result.chroot_results_dir,
'/tmp/test_that.PO1234567/platform_LibCBench')
self.assertEqual(result.results_dir,
'/tmp/chroot/tmp/test_that.PO1234567/platform_LibCBench')
self.assertEqual(result.retval, 0)
def setUp(self):
self.result = Result(self.mock_logger, self.mock_label, 'average',
self.mock_cmd_exec)
self.result.chromeos_root = '/tmp/chromeos'
@mock.patch.object(os.path, 'isdir')
@mock.patch.object(command_executer.CommandExecuter, 'RunCommand')
@mock.patch.object(command_executer.CommandExecuter, 'CopyFiles')
def test_copy_files_to(self, mock_copyfiles, mock_runcmd, mock_isdir):
files = ['src_file_1', 'src_file_2', 'src_file_3']
dest_dir = '/tmp/test'
self.mock_cmd_exec.RunCommand = mock_runcmd
self.mock_cmd_exec.CopyFiles = mock_copyfiles
mock_copyfiles.return_value = 0
# test 1. dest_dir exists; CopyFiles returns 0.
mock_isdir.return_value = True
self.result.CopyFilesTo(dest_dir, files)
self.assertEqual(mock_runcmd.call_count, 0)
self.assertEqual(mock_copyfiles.call_count, 3)
first_args = mock_copyfiles.call_args_list[0][0]
second_args = mock_copyfiles.call_args_list[1][0]
third_args = mock_copyfiles.call_args_list[2][0]
self.assertEqual(first_args, ('src_file_1', '/tmp/test/src_file_1.0'))
self.assertEqual(second_args, ('src_file_2', '/tmp/test/src_file_2.0'))
self.assertEqual(third_args, ('src_file_3', '/tmp/test/src_file_3.0'))
mock_runcmd.reset_mock()
mock_copyfiles.reset_mock()
# test 2. dest_dir does not exist; CopyFiles returns 0.
mock_isdir.return_value = False
self.result.CopyFilesTo(dest_dir, files)
self.assertEqual(mock_runcmd.call_count, 3)
self.assertEqual(mock_copyfiles.call_count, 3)
self.assertEqual(mock_runcmd.call_args_list[0],
mock_runcmd.call_args_list[1])
self.assertEqual(mock_runcmd.call_args_list[0],
mock_runcmd.call_args_list[2])
self.assertEqual(mock_runcmd.call_args_list[0][0], ('mkdir -p /tmp/test',))
# test 3. CopyFiles returns 1 (fails).
mock_copyfiles.return_value = 1
self.assertRaises(Exception, self.result.CopyFilesTo, dest_dir, files)
@mock.patch.object(Result, 'CopyFilesTo')
def test_copy_results_to(self, mockCopyFilesTo):
results_file = [
'/tmp/result.json.0', '/tmp/result.json.1', '/tmp/result.json.2'
]
perf_data_files = [
'/tmp/perf.data.0', '/tmp/perf.data.1', '/tmp/perf.data.2'
]
perf_report_files = [
'/tmp/perf.report.0', '/tmp/perf.report.1', '/tmp/perf.report.2'
]
self.result.results_file = results_file
self.result.perf_data_files = perf_data_files
self.result.perf_report_files = perf_report_files
self.result.CopyFilesTo = mockCopyFilesTo
self.result.CopyResultsTo('/tmp/results/')
self.assertEqual(mockCopyFilesTo.call_count, 3)
self.assertEqual(len(mockCopyFilesTo.call_args_list), 3)
self.assertEqual(mockCopyFilesTo.call_args_list[0][0],
('/tmp/results/', results_file))
self.assertEqual(mockCopyFilesTo.call_args_list[1][0],
('/tmp/results/', perf_data_files))
self.assertEqual(mockCopyFilesTo.call_args_list[2][0],
('/tmp/results/', perf_report_files))
def test_get_new_keyvals(self):
kv_dict = {}
def FakeGetDataMeasurementsFiles():
filename = os.path.join(os.getcwd(), 'unittest_keyval_file.txt')
return [filename]
self.result.GetDataMeasurementsFiles = FakeGetDataMeasurementsFiles
kv_dict2, udict = self.result.GetNewKeyvals(kv_dict)
self.assertEqual(
kv_dict2, {
u'Box2D__Box2D': 4775,
u'Mandreel__Mandreel': 6620,
u'Gameboy__Gameboy': 9901,
u'Crypto__Crypto': 8737,
u'telemetry_page_measurement_results__num_errored': 0,
u'telemetry_page_measurement_results__num_failed': 0,
u'PdfJS__PdfJS': 6455,
u'Total__Score': 7918,
u'EarleyBoyer__EarleyBoyer': 14340,
u'MandreelLatency__MandreelLatency': 5188,
u'CodeLoad__CodeLoad': 6271,
u'DeltaBlue__DeltaBlue': 14401,
u'Typescript__Typescript': 9815,
u'SplayLatency__SplayLatency': 7653,
u'zlib__zlib': 16094,
u'Richards__Richards': 10358,
u'RegExp__RegExp': 1765,
u'NavierStokes__NavierStokes': 9815,
u'Splay__Splay': 4425,
u'RayTrace__RayTrace': 16600
})
self.assertEqual(
udict, {
u'Box2D__Box2D': u'score',
u'Mandreel__Mandreel': u'score',
u'Gameboy__Gameboy': u'score',
u'Crypto__Crypto': u'score',
u'telemetry_page_measurement_results__num_errored': u'count',
u'telemetry_page_measurement_results__num_failed': u'count',
u'PdfJS__PdfJS': u'score',
u'Total__Score': u'score',
u'EarleyBoyer__EarleyBoyer': u'score',
u'MandreelLatency__MandreelLatency': u'score',
u'CodeLoad__CodeLoad': u'score',
u'DeltaBlue__DeltaBlue': u'score',
u'Typescript__Typescript': u'score',
u'SplayLatency__SplayLatency': u'score',
u'zlib__zlib': u'score',
u'Richards__Richards': u'score',
u'RegExp__RegExp': u'score',
u'NavierStokes__NavierStokes': u'score',
u'Splay__Splay': u'score',
u'RayTrace__RayTrace': u'score'
})
def test_append_telemetry_units(self):
kv_dict = {
u'Box2D__Box2D': 4775,
u'Mandreel__Mandreel': 6620,
u'Gameboy__Gameboy': 9901,
u'Crypto__Crypto': 8737,
u'PdfJS__PdfJS': 6455,
u'Total__Score': 7918,
u'EarleyBoyer__EarleyBoyer': 14340,
u'MandreelLatency__MandreelLatency': 5188,
u'CodeLoad__CodeLoad': 6271,
u'DeltaBlue__DeltaBlue': 14401,
u'Typescript__Typescript': 9815,
u'SplayLatency__SplayLatency': 7653,
u'zlib__zlib': 16094,
u'Richards__Richards': 10358,
u'RegExp__RegExp': 1765,
u'NavierStokes__NavierStokes': 9815,
u'Splay__Splay': 4425,
u'RayTrace__RayTrace': 16600
}
units_dict = {
u'Box2D__Box2D': u'score',
u'Mandreel__Mandreel': u'score',
u'Gameboy__Gameboy': u'score',
u'Crypto__Crypto': u'score',
u'PdfJS__PdfJS': u'score',
u'Total__Score': u'score',
u'EarleyBoyer__EarleyBoyer': u'score',
u'MandreelLatency__MandreelLatency': u'score',
u'CodeLoad__CodeLoad': u'score',
u'DeltaBlue__DeltaBlue': u'score',
u'Typescript__Typescript': u'score',
u'SplayLatency__SplayLatency': u'score',
u'zlib__zlib': u'score',
u'Richards__Richards': u'score',
u'RegExp__RegExp': u'score',
u'NavierStokes__NavierStokes': u'score',
u'Splay__Splay': u'score',
u'RayTrace__RayTrace': u'score'
}
results_dict = self.result.AppendTelemetryUnits(kv_dict, units_dict)
self.assertEqual(
results_dict, {
u'Box2D__Box2D': [4775, u'score'],
u'Splay__Splay': [4425, u'score'],
u'Gameboy__Gameboy': [9901, u'score'],
u'Crypto__Crypto': [8737, u'score'],
u'PdfJS__PdfJS': [6455, u'score'],
u'Total__Score': [7918, u'score'],
u'EarleyBoyer__EarleyBoyer': [14340, u'score'],
u'MandreelLatency__MandreelLatency': [5188, u'score'],
u'DeltaBlue__DeltaBlue': [14401, u'score'],
u'SplayLatency__SplayLatency': [7653, u'score'],
u'Mandreel__Mandreel': [6620, u'score'],
u'Richards__Richards': [10358, u'score'],
u'zlib__zlib': [16094, u'score'],
u'CodeLoad__CodeLoad': [6271, u'score'],
u'Typescript__Typescript': [9815, u'score'],
u'RegExp__RegExp': [1765, u'score'],
u'RayTrace__RayTrace': [16600, u'score'],
u'NavierStokes__NavierStokes': [9815, u'score']
})
@mock.patch.object(misc, 'GetInsideChrootPath')
@mock.patch.object(tempfile, 'mkdtemp')
@mock.patch.object(command_executer.CommandExecuter, 'RunCommand')
@mock.patch.object(command_executer.CommandExecuter,
'ChrootRunCommandWOutput')
def test_get_keyvals(self, mock_chrootruncmd, mock_runcmd, mock_mkdtemp,
mock_getpath):
self.kv_dict = {}
self.callGetNewKeyvals = False
def reset():
self.kv_dict = {}
self.callGetNewKeyvals = False
mock_chrootruncmd.reset_mock()
mock_runcmd.reset_mock()
mock_mkdtemp.reset_mock()
mock_getpath.reset_mock()
def FakeGetNewKeyvals(kv_dict):
self.kv_dict = kv_dict
self.callGetNewKeyvals = True
return_kvdict = {'first_time': 680, 'Total': 10}
return_udict = {'first_time': 'ms', 'Total': 'score'}
return return_kvdict, return_udict
mock_mkdtemp.return_value = TMP_DIR1
mock_chrootruncmd.return_value = [
'', ('%s,PASS\n%s/telemetry_Crosperf,PASS\n') % (TMP_DIR1, TMP_DIR1), ''
]
mock_getpath.return_value = TMP_DIR1
self.result.ce.ChrootRunCommandWOutput = mock_chrootruncmd
self.result.ce.RunCommand = mock_runcmd
self.result.GetNewKeyvals = FakeGetNewKeyvals
self.result.suite = 'telemetry_Crosperf'
self.result.results_dir = '/tmp/test_that_resultsNmq'
# Test 1. no self.temp_dir.
res = self.result.GetKeyvals()
self.assertTrue(self.callGetNewKeyvals)
self.assertEqual(self.kv_dict, {'': 'PASS', 'telemetry_Crosperf': 'PASS'})
self.assertEqual(mock_runcmd.call_count, 1)
self.assertEqual(mock_runcmd.call_args_list[0][0],
('cp -r /tmp/test_that_resultsNmq/* %s' % TMP_DIR1,))
self.assertEqual(mock_chrootruncmd.call_count, 1)
self.assertEqual(
mock_chrootruncmd.call_args_list[0][0],
(self.result.chromeos_root,
('./generate_test_report --no-color --csv %s') % TMP_DIR1))
self.assertEqual(mock_getpath.call_count, 1)
self.assertEqual(mock_mkdtemp.call_count, 1)
self.assertEqual(res, {'Total': [10, 'score'], 'first_time': [680, 'ms']})
# Test 2. self.temp_dir
reset()
mock_chrootruncmd.return_value = [
'', ('/tmp/tmpJCajRG,PASS\n/tmp/tmpJCajRG/'
'telemetry_Crosperf,PASS\n'), ''
]
mock_getpath.return_value = '/tmp/tmpJCajRG'
self.result.temp_dir = '/tmp/tmpJCajRG'
res = self.result.GetKeyvals()
self.assertEqual(mock_runcmd.call_count, 0)
self.assertEqual(mock_mkdtemp.call_count, 0)
self.assertEqual(mock_chrootruncmd.call_count, 1)
self.assertTrue(self.callGetNewKeyvals)
self.assertEqual(self.kv_dict, {'': 'PASS', 'telemetry_Crosperf': 'PASS'})
self.assertEqual(res, {'Total': [10, 'score'], 'first_time': [680, 'ms']})
# Test 3. suite != telemetry_Crosperf. Normally this would be for
# running non-Telemetry autotests, such as BootPerfServer. In this test
# case, the keyvals we have set up were returned from a Telemetry test run;
# so this pass is basically testing that we don't append the units to the
# test results (which we do for Telemetry autotest runs).
reset()
self.result.suite = ''
res = self.result.GetKeyvals()
self.assertEqual(res, {'Total': 10, 'first_time': 680})
@mock.patch.object(misc, 'GetInsideChrootPath')
@mock.patch.object(command_executer.CommandExecuter,
'ChrootRunCommandWOutput')
def test_get_samples(self, mock_chrootruncmd, mock_getpath):
fake_file = '/usr/chromeos/chroot/tmp/results/fake_file'
self.result.perf_data_files = ['/tmp/results/perf.data']
self.result.board = 'samus'
mock_getpath.return_value = fake_file
self.result.ce.ChrootRunCommandWOutput = mock_chrootruncmd
mock_chrootruncmd.return_value = ['', '45.42% 237210 chrome ', '']
samples = self.result.GetSamples()
self.assertEqual(samples, [237210, u'samples'])
def test_get_results_dir(self):
self.result.out = ''
self.assertRaises(Exception, self.result.GetResultsDir)
self.result.out = OUTPUT
resdir = self.result.GetResultsDir()
self.assertEqual(resdir, '/tmp/test_that.PO1234567/platform_LibCBench')
@mock.patch.object(command_executer.CommandExecuter, 'RunCommandGeneric')
def test_find_files_in_results_dir(self, mock_runcmd):
self.result.results_dir = None
res = self.result.FindFilesInResultsDir('-name perf.data')
self.assertEqual(res, '')
self.result.ce.RunCommand = mock_runcmd
self.result.results_dir = '/tmp/test_results'
mock_runcmd.return_value = [0, '/tmp/test_results/perf.data', '']
res = self.result.FindFilesInResultsDir('-name perf.data')
self.assertEqual(mock_runcmd.call_count, 1)
self.assertEqual(mock_runcmd.call_args_list[0][0],
('find /tmp/test_results -name perf.data',))
self.assertEqual(res, '/tmp/test_results/perf.data')
mock_runcmd.reset_mock()
mock_runcmd.return_value = [1, '', '']
self.assertRaises(Exception, self.result.FindFilesInResultsDir,
'-name perf.data')
@mock.patch.object(Result, 'FindFilesInResultsDir')
def test_get_perf_data_files(self, mock_findfiles):
self.args = None
mock_findfiles.return_value = 'line1\nline1\n'
self.result.FindFilesInResultsDir = mock_findfiles
res = self.result.GetPerfDataFiles()
self.assertEqual(res, ['line1', 'line1'])
self.assertEqual(mock_findfiles.call_args_list[0][0], ('-name perf.data',))
def test_get_perf_report_files(self):
self.args = None
def FakeFindFiles(find_args):
self.args = find_args
return 'line1\nline1\n'
self.result.FindFilesInResultsDir = FakeFindFiles
res = self.result.GetPerfReportFiles()
self.assertEqual(res, ['line1', 'line1'])
self.assertEqual(self.args, '-name perf.data.report')
def test_get_data_measurement_files(self):
self.args = None
def FakeFindFiles(find_args):
self.args = find_args
return 'line1\nline1\n'
self.result.FindFilesInResultsDir = FakeFindFiles
res = self.result.GetDataMeasurementsFiles()
self.assertEqual(res, ['line1', 'line1'])
self.assertEqual(self.args, '-name perf_measurements')
@mock.patch.object(command_executer.CommandExecuter, 'RunCommandWOutput')
def test_get_turbostat_file_finds_single_log(self, mock_runcmd):
"""Expected behavior when a single log file found."""
self.result.results_dir = '/tmp/test_results'
self.result.ce.RunCommandWOutput = mock_runcmd
mock_runcmd.return_value = (0, 'some/long/path/turbostat.log', '')
found_single_log = self.result.GetTurbostatFile()
self.assertEqual(found_single_log, 'some/long/path/turbostat.log')
@mock.patch.object(command_executer.CommandExecuter, 'RunCommandWOutput')
def test_get_turbostat_file_finds_multiple_logs(self, mock_runcmd):
"""Error case when multiple files found."""
self.result.results_dir = '/tmp/test_results'
self.result.ce.RunCommandWOutput = mock_runcmd
mock_runcmd.return_value = (0,
'some/long/path/turbostat.log\nturbostat.log',
'')
found_first_logs = self.result.GetTurbostatFile()
self.assertEqual(found_first_logs, 'some/long/path/turbostat.log')
@mock.patch.object(command_executer.CommandExecuter, 'RunCommandWOutput')
def test_get_turbostat_file_finds_no_logs(self, mock_runcmd):
"""Error case when no log file found."""
self.result.results_dir = '/tmp/test_results'
self.result.ce.RunCommandWOutput = mock_runcmd
mock_runcmd.return_value = (0, '', '')
found_no_logs = self.result.GetTurbostatFile()
self.assertEqual(found_no_logs, '')
@mock.patch.object(command_executer.CommandExecuter, 'RunCommandWOutput')
def test_get_turbostat_file_with_failing_find(self, mock_runcmd):
"""Error case when file search returns an error."""
self.result.results_dir = '/tmp/test_results'
mock_runcmd.return_value = (-1, '', 'error')
with self.assertRaises(RuntimeError):
self.result.GetTurbostatFile()
@mock.patch.object(command_executer.CommandExecuter, 'RunCommandWOutput')
def test_get_top_file_finds_single_log(self, mock_runcmd):
"""Expected behavior when a single top log file found."""
self.result.results_dir = '/tmp/test_results'
self.result.ce.RunCommandWOutput = mock_runcmd
mock_runcmd.return_value = (0, 'some/long/path/top.log', '')
found_single_log = self.result.GetTopFile()
self.assertEqual(found_single_log, 'some/long/path/top.log')
@mock.patch.object(command_executer.CommandExecuter, 'RunCommandWOutput')
def test_get_top_file_finds_multiple_logs(self, mock_runcmd):
"""The case when multiple top files found."""
self.result.results_dir = '/tmp/test_results'
self.result.ce.RunCommandWOutput = mock_runcmd
mock_runcmd.return_value = (0, 'some/long/path/top.log\ntop.log', '')
found_first_logs = self.result.GetTopFile()
self.assertEqual(found_first_logs, 'some/long/path/top.log')
@mock.patch.object(command_executer.CommandExecuter, 'RunCommandWOutput')
def test_get_top_file_finds_no_logs(self, mock_runcmd):
"""Error case when no log file found."""
self.result.results_dir = '/tmp/test_results'
self.result.ce.RunCommandWOutput = mock_runcmd
mock_runcmd.return_value = (0, '', '')
found_no_logs = self.result.GetTopFile()
self.assertEqual(found_no_logs, '')
@mock.patch.object(command_executer.CommandExecuter, 'RunCommandWOutput')
def test_get_cpuinfo_file_finds_single_log(self, mock_runcmd):
"""Expected behavior when a single cpuinfo file found."""
self.result.results_dir = '/tmp/test_results'
self.result.ce.RunCommandWOutput = mock_runcmd
mock_runcmd.return_value = (0, 'some/long/path/cpuinfo.log', '')
found_single_log = self.result.GetCpuinfoFile()
self.assertEqual(found_single_log, 'some/long/path/cpuinfo.log')
@mock.patch.object(command_executer.CommandExecuter, 'RunCommandWOutput')
def test_get_cpustats_file_finds_single_log(self, mock_runcmd):
"""Expected behavior when a single log file found."""
self.result.results_dir = '/tmp/test_results'
self.result.ce.RunCommandWOutput = mock_runcmd
mock_runcmd.return_value = (0, 'some/long/path/cpustats.log', '')
found_single_log = self.result.GetCpustatsFile()
self.assertEqual(found_single_log, 'some/long/path/cpustats.log')
@mock.patch.object(command_executer.CommandExecuter, 'RunCommandWOutput')
def test_get_cpustats_file_finds_multiple_logs(self, mock_runcmd):
"""The case when multiple files found."""
self.result.results_dir = '/tmp/test_results'
self.result.ce.RunCommandWOutput = mock_runcmd
mock_runcmd.return_value = (0, 'some/long/path/cpustats.log\ncpustats.log',
'')
found_first_logs = self.result.GetCpustatsFile()
self.assertEqual(found_first_logs, 'some/long/path/cpustats.log')
@mock.patch.object(command_executer.CommandExecuter, 'RunCommandWOutput')
def test_get_cpustats_file_finds_no_logs(self, mock_runcmd):
"""Error case when no log file found."""
self.result.results_dir = '/tmp/test_results'
self.result.ce.RunCommandWOutput = mock_runcmd
mock_runcmd.return_value = (0, '', '')
found_no_logs = self.result.GetCpustatsFile()
self.assertEqual(found_no_logs, '')
def test_verify_perf_data_pid_ok(self):
"""Verify perf PID which is present in TOP_DATA."""
self.result.top_cmds = TOP_DATA
# pid is present in TOP_DATA.
with mock.patch.object(
Result, 'ReadPidFromPerfData', return_value=['5713']):
self.result.VerifyPerfDataPID()
def test_verify_perf_data_pid_fail(self):
"""Test perf PID missing in top raises the error."""
self.result.top_cmds = TOP_DATA
# pid is not in the list of top processes.
with mock.patch.object(
Result, 'ReadPidFromPerfData', return_value=['9999']):
with self.assertRaises(PidVerificationError):
self.result.VerifyPerfDataPID()
@mock.patch.object(command_executer.CommandExecuter,
'ChrootRunCommandWOutput')
def test_read_pid_from_perf_data_ok(self, mock_runcmd):
"""Test perf header parser, normal flow."""
self.result.ce.ChrootRunCommandWOutput = mock_runcmd
self.result.perf_data_files = ['/tmp/chromeos/chroot/tmp/results/perf.data']
exp_pid = '12345'
mock_runcmd.return_value = (0, PERF_DATA_HEADER.format(pid=exp_pid), '')
pids = self.result.ReadPidFromPerfData()
self.assertEqual(pids, [exp_pid])
@mock.patch.object(command_executer.CommandExecuter,
'ChrootRunCommandWOutput')
def test_read_pid_from_perf_data_mult_profiles(self, mock_runcmd):
"""Test multiple perf.data files with PID."""
self.result.ce.ChrootRunCommandWOutput = mock_runcmd
# self.result.chromeos_root = '/tmp/chromeos'
self.result.perf_data_files = [
'/tmp/chromeos/chroot/tmp/results/perf.data.0',
'/tmp/chromeos/chroot/tmp/results/perf.data.1',
]
# There is '-p <pid>' in command line but it's still system-wide: '-a'.
cmd_line = '# cmdline : /usr/bin/perf record -e instructions -p {pid}'
exp_perf_pids = ['1111', '2222']
mock_runcmd.side_effect = [
(0, cmd_line.format(pid=exp_perf_pids[0]), ''),
(0, cmd_line.format(pid=exp_perf_pids[1]), ''),
]
pids = self.result.ReadPidFromPerfData()
self.assertEqual(pids, exp_perf_pids)
@mock.patch.object(command_executer.CommandExecuter,
'ChrootRunCommandWOutput')
def test_read_pid_from_perf_data_no_pid(self, mock_runcmd):
"""Test perf.data without PID."""
self.result.ce.ChrootRunCommandWOutput = mock_runcmd
self.result.perf_data_files = ['/tmp/chromeos/chroot/tmp/results/perf.data']
cmd_line = '# cmdline : /usr/bin/perf record -e instructions'
mock_runcmd.return_value = (0, cmd_line, '')
pids = self.result.ReadPidFromPerfData()
# pids is empty.
self.assertEqual(pids, [])
@mock.patch.object(command_executer.CommandExecuter,
'ChrootRunCommandWOutput')
def test_read_pid_from_perf_data_system_wide(self, mock_runcmd):
"""Test reading from system-wide profile with PID."""
self.result.ce.ChrootRunCommandWOutput = mock_runcmd
self.result.perf_data_files = ['/tmp/chromeos/chroot/tmp/results/perf.data']
# There is '-p <pid>' in command line but it's still system-wide: '-a'.
cmd_line = '# cmdline : /usr/bin/perf record -e instructions -a -p 1234'
mock_runcmd.return_value = (0, cmd_line, '')
pids = self.result.ReadPidFromPerfData()
# pids should be empty since it's not a per-process profiling.
self.assertEqual(pids, [])
@mock.patch.object(command_executer.CommandExecuter,
'ChrootRunCommandWOutput')
def test_read_pid_from_perf_data_read_fail(self, mock_runcmd):
"""Failure to read perf.data raises the error."""
self.result.ce.ChrootRunCommandWOutput = mock_runcmd
self.result.perf_data_files = ['/tmp/chromeos/chroot/tmp/results/perf.data']
# Error status of the profile read.
mock_runcmd.return_value = (1, '', '')
with self.assertRaises(PerfDataReadError):
self.result.ReadPidFromPerfData()
@mock.patch.object(command_executer.CommandExecuter,
'ChrootRunCommandWOutput')
def test_read_pid_from_perf_data_fail(self, mock_runcmd):
"""Failure to find cmdline in perf.data header raises the error."""
self.result.ce.ChrootRunCommandWOutput = mock_runcmd
self.result.perf_data_files = ['/tmp/chromeos/chroot/tmp/results/perf.data']
# Empty output.
mock_runcmd.return_value = (0, '', '')
with self.assertRaises(PerfDataReadError):
self.result.ReadPidFromPerfData()
def test_process_turbostat_results_with_valid_data(self):
"""Normal case when log exists and contains valid data."""
self.result.turbostat_log_file = '/tmp/somelogfile.log'
with mock.patch('builtins.open',
mock.mock_open(read_data=TURBOSTAT_LOG_OUTPUT)) as mo:
cpustats = self.result.ProcessTurbostatResults()
# Check that the log got opened and data were read/parsed.
calls = [mock.call('/tmp/somelogfile.log')]
mo.assert_has_calls(calls)
self.assertEqual(cpustats, TURBOSTAT_DATA)
def test_process_turbostat_results_from_empty_file(self):
"""Error case when log exists but file is empty."""
self.result.turbostat_log_file = '/tmp/emptylogfile.log'
with mock.patch('builtins.open', mock.mock_open(read_data='')) as mo:
cpustats = self.result.ProcessTurbostatResults()
# Check that the log got opened and parsed successfully and empty data
# returned.
calls = [mock.call('/tmp/emptylogfile.log')]
mo.assert_has_calls(calls)
self.assertEqual(cpustats, {})
def test_process_turbostat_results_when_file_doesnt_exist(self):
"""Error case when file does not exist."""
nonexistinglog = '/tmp/1'
while os.path.exists(nonexistinglog):
# Extend file path if it happens to exist.
nonexistinglog = os.path.join(nonexistinglog, '1')
self.result.turbostat_log_file = nonexistinglog
# Allow the tested function to call a 'real' open and hopefully crash.
with self.assertRaises(IOError):
self.result.ProcessTurbostatResults()
def test_process_cpustats_results_with_uniq_data(self):
"""Process cpustats log which has freq unique to each core.
Testing normal case when frequency data vary between
different cores.
Expecting that data for all cores will be present in
returned cpustats.
"""
self.result.cpustats_log_file = '/tmp/somelogfile.log'
with mock.patch('builtins.open',
mock.mock_open(read_data=CPUSTATS_UNIQ_OUTPUT)) as mo:
cpustats = self.result.ProcessCpustatsResults()
# Check that the log got opened and data were read/parsed.
calls = [mock.call('/tmp/somelogfile.log')]
mo.assert_has_calls(calls)
self.assertEqual(cpustats, CPUSTATS_UNIQ_DATA)
def test_process_cpustats_results_with_dupl_data(self):
"""Process cpustats log where cores have duplicate freq.
Testing normal case when frequency data on some cores
are duplicated.
Expecting that duplicated data is discarded in
returned cpustats.
"""
self.result.cpustats_log_file = '/tmp/somelogfile.log'
with mock.patch('builtins.open',
mock.mock_open(read_data=CPUSTATS_DUPL_OUTPUT)) as mo:
cpustats = self.result.ProcessCpustatsResults()
# Check that the log got opened and data were read/parsed.
calls = [mock.call('/tmp/somelogfile.log')]
mo.assert_has_calls(calls)
self.assertEqual(cpustats, CPUSTATS_DUPL_DATA)
def test_process_cpustats_results_from_empty_file(self):
"""Error case when log exists but file is empty."""
self.result.cpustats_log_file = '/tmp/emptylogfile.log'
with mock.patch('builtins.open', mock.mock_open(read_data='')) as mo:
cpustats = self.result.ProcessCpustatsResults()
# Check that the log got opened and parsed successfully and empty data
# returned.
calls = [mock.call('/tmp/emptylogfile.log')]
mo.assert_has_calls(calls)
self.assertEqual(cpustats, {})
def test_process_top_results_with_valid_data(self):
"""Process top log with valid data."""
self.result.top_log_file = '/tmp/fakelogfile.log'
with mock.patch('builtins.open', mock.mock_open(read_data=TOP_LOG)) as mo:
topproc = self.result.ProcessTopResults()
# Check that the log got opened and data were read/parsed.
calls = [mock.call('/tmp/fakelogfile.log')]
mo.assert_has_calls(calls)
self.assertEqual(topproc, TOP_DATA)
def test_process_top_results_from_empty_file(self):
"""Error case when log exists but file is empty."""
self.result.top_log_file = '/tmp/emptylogfile.log'
with mock.patch('builtins.open', mock.mock_open(read_data='')) as mo:
topcalls = self.result.ProcessTopResults()
# Check that the log got opened and parsed successfully and empty data
# returned.
calls = [mock.call('/tmp/emptylogfile.log')]
mo.assert_has_calls(calls)
self.assertEqual(topcalls, [])
def test_format_string_top_cmds(self):
"""Test formatted string with top commands."""
self.result.top_cmds = [
{
'cmd': 'chrome-111',
'cpu_use_avg': 119.753453465,
'count': 44444,
'top5_cpu_use': [222.8, 217.9, 217.8, 191.0, 189.9],
},
{
'cmd': 'chrome-222',
'cpu_use_avg': 100,
'count': 33333,
'top5_cpu_use': [200.0, 195.0, 190.0, 185.0, 180.0],
},
{
'cmd': 'irq/230-cros-ec',
'cpu_use_avg': 10.000000000000001,
'count': 1000,
'top5_cpu_use': [11.5, 11.4, 11.3, 11.2, 11.1],
},
{
'cmd': 'powerd',
'cpu_use_avg': 2.0,
'count': 2,
'top5_cpu_use': [3.0, 1.0]
},
{
'cmd': 'cmd3',
'cpu_use_avg': 1.0,
'count': 1,
'top5_cpu_use': [1.0],
},
{
'cmd': 'cmd4',
'cpu_use_avg': 1.0,
'count': 1,
'top5_cpu_use': [1.0],
},
{
'cmd': 'cmd5',
'cpu_use_avg': 1.0,
'count': 1,
'top5_cpu_use': [1.0],
},
{
'cmd': 'cmd6_not_for_print',
'cpu_avg': 1.0,
'count': 1,
'top5': [1.0],
},
]
form_str = self.result.FormatStringTopCommands()
self.assertEqual(
form_str, '\n'.join([
'Top commands with highest CPU usage:',
' COMMAND AVG CPU% COUNT HIGHEST 5',
'-' * 50,
' chrome-111 119.75 44444 '
'[222.8, 217.9, 217.8, 191.0, 189.9]',
' chrome-222 100.00 33333 '
'[200.0, 195.0, 190.0, 185.0, 180.0]',
' irq/230-cros-ec 10.00 1000 '
'[11.5, 11.4, 11.3, 11.2, 11.1]',
' powerd 2.00 2 [3.0, 1.0]',
' cmd3 1.00 1 [1.0]',
' cmd4 1.00 1 [1.0]',
' cmd5 1.00 1 [1.0]',
'-' * 50,
]))
def test_format_string_top_calls_no_data(self):
"""Test formatted string of top with no data."""
self.result.top_cmds = []
form_str = self.result.FormatStringTopCommands()
self.assertEqual(
form_str, '\n'.join([
'Top commands with highest CPU usage:',
' COMMAND AVG CPU% COUNT HIGHEST 5',
'-' * 50,
'[NO DATA FROM THE TOP LOG]',
'-' * 50,
]))
@mock.patch.object(misc, 'GetInsideChrootPath')
@mock.patch.object(command_executer.CommandExecuter, 'ChrootRunCommand')
def test_generate_perf_report_files(self, mock_chrootruncmd, mock_getpath):
fake_file = '/usr/chromeos/chroot/tmp/results/fake_file'
self.result.perf_data_files = ['/tmp/results/perf.data']
self.result.board = 'lumpy'
mock_getpath.return_value = fake_file
self.result.ce.ChrootRunCommand = mock_chrootruncmd
mock_chrootruncmd.return_value = 0
# Debug path not found
self.result.label.debug_path = ''
tmp = self.result.GeneratePerfReportFiles()
self.assertEqual(tmp, ['/tmp/chromeos/chroot%s' % fake_file])
self.assertEqual(mock_chrootruncmd.call_args_list[0][0],
(self.result.chromeos_root,
('/usr/sbin/perf report -n '
'-i %s --stdio > %s') % (fake_file, fake_file)))
@mock.patch.object(misc, 'GetInsideChrootPath')
@mock.patch.object(command_executer.CommandExecuter, 'ChrootRunCommand')
def test_generate_perf_report_files_debug(self, mock_chrootruncmd,
mock_getpath):
fake_file = '/usr/chromeos/chroot/tmp/results/fake_file'
self.result.perf_data_files = ['/tmp/results/perf.data']
self.result.board = 'lumpy'
mock_getpath.return_value = fake_file
self.result.ce.ChrootRunCommand = mock_chrootruncmd
mock_chrootruncmd.return_value = 0
# Debug path found
self.result.label.debug_path = '/tmp/debug'
tmp = self.result.GeneratePerfReportFiles()
self.assertEqual(tmp, ['/tmp/chromeos/chroot%s' % fake_file])
self.assertEqual(mock_chrootruncmd.call_args_list[0][0],
(self.result.chromeos_root,
('/usr/sbin/perf report -n --symfs /tmp/debug '
'--vmlinux /tmp/debug/boot/vmlinux '
'-i %s --stdio > %s') % (fake_file, fake_file)))
@mock.patch.object(misc, 'GetOutsideChrootPath')
def test_populate_from_run(self, mock_getpath):
def FakeGetResultsDir():
self.callGetResultsDir = True
return '/tmp/results_dir'
def FakeGetResultsFile():
self.callGetResultsFile = True
return []
def FakeGetPerfDataFiles():
self.callGetPerfDataFiles = True
return []
def FakeGetPerfReportFiles():
self.callGetPerfReportFiles = True
return []
def FakeGetTurbostatFile():
self.callGetTurbostatFile = True
return []
def FakeGetCpustatsFile():
self.callGetCpustatsFile = True
return []
def FakeGetTopFile():
self.callGetTopFile = True
return []
def FakeGetCpuinfoFile():
self.callGetCpuinfoFile = True
return []
def FakeGetWaitTimeFile():
self.callGetWaitTimeFile = True
return []
def FakeProcessResults(show_results=False):
if show_results:
pass
self.callProcessResults = True
if mock_getpath:
pass
mock.get_path = '/tmp/chromeos/tmp/results_dir'
self.callGetResultsDir = False
self.callGetResultsFile = False
self.callGetPerfDataFiles = False
self.callGetPerfReportFiles = False
self.callGetTurbostatFile = False
self.callGetCpustatsFile = False
self.callGetTopFile = False
self.callGetCpuinfoFile = False
self.callGetWaitTimeFile = False
self.callProcessResults = False
self.result.GetResultsDir = FakeGetResultsDir
self.result.GetResultsFile = FakeGetResultsFile
self.result.GetPerfDataFiles = FakeGetPerfDataFiles
self.result.GeneratePerfReportFiles = FakeGetPerfReportFiles
self.result.GetTurbostatFile = FakeGetTurbostatFile
self.result.GetCpustatsFile = FakeGetCpustatsFile
self.result.GetTopFile = FakeGetTopFile
self.result.GetCpuinfoFile = FakeGetCpuinfoFile
self.result.GetWaitTimeFile = FakeGetWaitTimeFile
self.result.ProcessResults = FakeProcessResults
self.result.PopulateFromRun(OUTPUT, '', 0, 'test', 'telemetry_Crosperf',
'chrome')
self.assertTrue(self.callGetResultsDir)
self.assertTrue(self.callGetResultsFile)
self.assertTrue(self.callGetPerfDataFiles)
self.assertTrue(self.callGetPerfReportFiles)
self.assertTrue(self.callGetTurbostatFile)
self.assertTrue(self.callGetCpustatsFile)
self.assertTrue(self.callGetTopFile)
self.assertTrue(self.callGetCpuinfoFile)
self.assertTrue(self.callGetWaitTimeFile)
self.assertTrue(self.callProcessResults)
def FakeGetKeyvals(self, show_all=False):
if show_all:
return {'first_time': 680, 'Total': 10}
else:
return {'Total': 10}
def test_process_results(self):
def FakeGatherPerfResults():
self.callGatherPerfResults = True
def FakeGetSamples():
return (1, 'samples')
# Test 1
self.callGatherPerfResults = False
self.result.GetKeyvals = self.FakeGetKeyvals
self.result.GatherPerfResults = FakeGatherPerfResults
self.result.retval = 0
self.result.ProcessResults()
self.assertTrue(self.callGatherPerfResults)
self.assertEqual(len(self.result.keyvals), 2)
self.assertEqual(self.result.keyvals, {'Total': 10, 'retval': 0})
# Test 2
self.result.retval = 1
self.result.ProcessResults()
self.assertEqual(len(self.result.keyvals), 2)
self.assertEqual(self.result.keyvals, {'Total': 10, 'retval': 1})
# Test 3
self.result.cwp_dso = 'chrome'
self.result.retval = 0
self.result.GetSamples = FakeGetSamples
self.result.ProcessResults()
self.assertEqual(len(self.result.keyvals), 3)
self.assertEqual(self.result.keyvals, {
'Total': 10,
'samples': (1, 'samples'),
'retval': 0
})
# Test 4. Parse output of benchmarks with multiple sotries in histogram
# format
self.result.suite = 'telemetry_Crosperf'
self.result.results_file = [tempfile.mkdtemp() + '/histograms.json']
with open(self.result.results_file[0], 'w') as f:
f.write(HISTOGRAMSET)
self.result.ProcessResults()
shutil.rmtree(os.path.dirname(self.result.results_file[0]))
# Verify the summary for the story is correct
self.assertEqual(self.result.keyvals['timeToFirstContentfulPaint__typical'],
[880.000, u'ms_smallerIsBetter'])
# Veirfy the summary for a certain stroy tag is correct
self.assertEqual(
self.result
.keyvals['timeToFirstContentfulPaint__cache_temperature:cold'],
[1000.000, u'ms_smallerIsBetter'])
self.assertEqual(
self.result
.keyvals['timeToFirstContentfulPaint__cache_temperature:warm'],
[800.000, u'ms_smallerIsBetter'])
@mock.patch.object(Result, 'ProcessCpustatsResults')
@mock.patch.object(Result, 'ProcessTurbostatResults')
def test_process_results_with_turbostat_log(self, mock_proc_turbo,
mock_proc_cpustats):
self.result.GetKeyvals = self.FakeGetKeyvals
self.result.retval = 0
self.result.turbostat_log_file = '/tmp/turbostat.log'
mock_proc_turbo.return_value = {
'cpufreq': {
'all': [1, 2, 3]
},
'cputemp': {
'all': [5.0, 6.0, 7.0]
}
}
self.result.ProcessResults()
mock_proc_turbo.assert_has_calls([mock.call()])
mock_proc_cpustats.assert_not_called()
self.assertEqual(len(self.result.keyvals), 8)
self.assertEqual(
self.result.keyvals, {
'Total': 10,
'cpufreq_all_avg': 2,
'cpufreq_all_max': 3,
'cpufreq_all_min': 1,
'cputemp_all_avg': 6.0,
'cputemp_all_min': 5.0,
'cputemp_all_max': 7.0,
'retval': 0
})
@mock.patch.object(Result, 'ProcessCpustatsResults')
@mock.patch.object(Result, 'ProcessTurbostatResults')
def test_process_results_with_cpustats_log(self, mock_proc_turbo,
mock_proc_cpustats):
self.result.GetKeyvals = self.FakeGetKeyvals
self.result.retval = 0
self.result.cpustats_log_file = '/tmp/cpustats.log'
mock_proc_cpustats.return_value = {
'cpufreq': {
'cpu0': [100, 100, 100],
'cpu1': [4, 5, 6]
},
'cputemp': {
'little': [20.2, 20.2, 20.2],
'big': [55.2, 66.1, 77.3]
}
}
self.result.ProcessResults()
mock_proc_turbo.assert_not_called()
mock_proc_cpustats.assert_has_calls([mock.call()])
self.assertEqual(len(self.result.keyvals), 10)
self.assertEqual(
self.result.keyvals, {
'Total': 10,
'cpufreq_cpu0_avg': 100,
'cpufreq_cpu1_avg': 5,
'cpufreq_cpu1_max': 6,
'cpufreq_cpu1_min': 4,
'cputemp_big_avg': 66.2,
'cputemp_big_max': 77.3,
'cputemp_big_min': 55.2,
'cputemp_little_avg': 20.2,
'retval': 0
})
@mock.patch.object(Result, 'ProcessCpustatsResults')
@mock.patch.object(Result, 'ProcessTurbostatResults')
def test_process_results_with_turbostat_and_cpustats_logs(
self, mock_proc_turbo, mock_proc_cpustats):
self.result.GetKeyvals = self.FakeGetKeyvals
self.result.retval = 0
self.result.turbostat_log_file = '/tmp/turbostat.log'
self.result.cpustats_log_file = '/tmp/cpustats.log'
mock_proc_turbo.return_value = {
'cpufreq': {
'all': [1, 2, 3]
},
'cputemp': {
'all': [5.0, 6.0, 7.0]
}
}
self.result.ProcessResults()
mock_proc_turbo.assert_has_calls([mock.call()])
mock_proc_cpustats.assert_not_called()
self.assertEqual(len(self.result.keyvals), 8)
self.assertEqual(
self.result.keyvals, {
'Total': 10,
'cpufreq_all_avg': 2,
'cpufreq_all_max': 3,
'cpufreq_all_min': 1,
'cputemp_all_avg': 6.0,
'cputemp_all_min': 5.0,
'cputemp_all_max': 7.0,
'retval': 0
})
@mock.patch.object(Result, 'ProcessCpustatsResults')
@mock.patch.object(Result, 'ProcessTurbostatResults')
def test_process_results_without_cpu_data(self, mock_proc_turbo,
mock_proc_cpustats):
self.result.GetKeyvals = self.FakeGetKeyvals
self.result.retval = 0
self.result.turbostat_log_file = ''
self.result.cpustats_log_file = ''
self.result.ProcessResults()
mock_proc_turbo.assert_not_called()
mock_proc_cpustats.assert_not_called()
self.assertEqual(len(self.result.keyvals), 2)
self.assertEqual(self.result.keyvals, {'Total': 10, 'retval': 0})
@mock.patch.object(misc, 'GetInsideChrootPath')
@mock.patch.object(command_executer.CommandExecuter,
'ChrootRunCommandWOutput')
def test_populate_from_cache_dir(self, mock_runchrootcmd, mock_getpath):
# pylint: disable=redefined-builtin
def FakeMkdtemp(dir=None):
if dir:
pass
return self.tmpdir
def FakeGetSamples():
return [1, u'samples']
current_path = os.getcwd()
cache_dir = os.path.join(current_path, 'test_cache/test_input')
self.result.ce = command_executer.GetCommandExecuter(log_level='average')
self.result.ce.ChrootRunCommandWOutput = mock_runchrootcmd
mock_runchrootcmd.return_value = [
'', ('%s,PASS\n%s/\telemetry_Crosperf,PASS\n') % (TMP_DIR1, TMP_DIR1),
''
]
mock_getpath.return_value = TMP_DIR1
self.tmpdir = tempfile.mkdtemp()
save_real_mkdtemp = tempfile.mkdtemp
tempfile.mkdtemp = FakeMkdtemp
self.result.PopulateFromCacheDir(cache_dir, 'sunspider',
'telemetry_Crosperf', '')
self.assertEqual(
self.result.keyvals, {
u'Total__Total': [444.0, u'ms'],
u'regexp-dna__regexp-dna': [16.2, u'ms'],
u'telemetry_page_measurement_results__num_failed': [0, u'count'],
u'telemetry_page_measurement_results__num_errored': [0, u'count'],
u'string-fasta__string-fasta': [23.2, u'ms'],
u'crypto-sha1__crypto-sha1': [11.6, u'ms'],
u'bitops-3bit-bits-in-byte__bitops-3bit-bits-in-byte': [3.2, u'ms'],
u'access-nsieve__access-nsieve': [7.9, u'ms'],
u'bitops-nsieve-bits__bitops-nsieve-bits': [9.4, u'ms'],
u'string-validate-input__string-validate-input': [19.3, u'ms'],
u'3d-raytrace__3d-raytrace': [24.7, u'ms'],
u'3d-cube__3d-cube': [28.0, u'ms'],
u'string-unpack-code__string-unpack-code': [46.7, u'ms'],
u'date-format-tofte__date-format-tofte': [26.3, u'ms'],
u'math-partial-sums__math-partial-sums': [22.0, u'ms'],
'\telemetry_Crosperf': ['PASS', ''],
u'crypto-aes__crypto-aes': [15.2, u'ms'],
u'bitops-bitwise-and__bitops-bitwise-and': [8.4, u'ms'],
u'crypto-md5__crypto-md5': [10.5, u'ms'],
u'string-tagcloud__string-tagcloud': [52.8, u'ms'],
u'access-nbody__access-nbody': [8.5, u'ms'],
'retval': 0,
u'math-spectral-norm__math-spectral-norm': [6.6, u'ms'],
u'math-cordic__math-cordic': [8.7, u'ms'],
u'access-binary-trees__access-binary-trees': [4.5, u'ms'],
u'controlflow-recursive__controlflow-recursive': [4.4, u'ms'],
u'access-fannkuch__access-fannkuch': [17.8, u'ms'],
u'string-base64__string-base64': [16.0, u'ms'],
u'date-format-xparb__date-format-xparb': [20.9, u'ms'],
u'3d-morph__3d-morph': [22.1, u'ms'],
u'bitops-bits-in-byte__bitops-bits-in-byte': [9.1, u'ms']
})
self.result.GetSamples = FakeGetSamples
self.result.PopulateFromCacheDir(cache_dir, 'sunspider',
'telemetry_Crosperf', 'chrome')
self.assertEqual(
self.result.keyvals, {
u'Total__Total': [444.0, u'ms'],
u'regexp-dna__regexp-dna': [16.2, u'ms'],
u'telemetry_page_measurement_results__num_failed': [0, u'count'],
u'telemetry_page_measurement_results__num_errored': [0, u'count'],
u'string-fasta__string-fasta': [23.2, u'ms'],
u'crypto-sha1__crypto-sha1': [11.6, u'ms'],
u'bitops-3bit-bits-in-byte__bitops-3bit-bits-in-byte': [3.2, u'ms'],
u'access-nsieve__access-nsieve': [7.9, u'ms'],
u'bitops-nsieve-bits__bitops-nsieve-bits': [9.4, u'ms'],
u'string-validate-input__string-validate-input': [19.3, u'ms'],
u'3d-raytrace__3d-raytrace': [24.7, u'ms'],
u'3d-cube__3d-cube': [28.0, u'ms'],
u'string-unpack-code__string-unpack-code': [46.7, u'ms'],
u'date-format-tofte__date-format-tofte': [26.3, u'ms'],
u'math-partial-sums__math-partial-sums': [22.0, u'ms'],
'\telemetry_Crosperf': ['PASS', ''],
u'crypto-aes__crypto-aes': [15.2, u'ms'],
u'bitops-bitwise-and__bitops-bitwise-and': [8.4, u'ms'],
u'crypto-md5__crypto-md5': [10.5, u'ms'],
u'string-tagcloud__string-tagcloud': [52.8, u'ms'],
u'access-nbody__access-nbody': [8.5, u'ms'],
'retval': 0,
u'math-spectral-norm__math-spectral-norm': [6.6, u'ms'],
u'math-cordic__math-cordic': [8.7, u'ms'],
u'access-binary-trees__access-binary-trees': [4.5, u'ms'],
u'controlflow-recursive__controlflow-recursive': [4.4, u'ms'],
u'access-fannkuch__access-fannkuch': [17.8, u'ms'],
u'string-base64__string-base64': [16.0, u'ms'],
u'date-format-xparb__date-format-xparb': [20.9, u'ms'],
u'3d-morph__3d-morph': [22.1, u'ms'],
u'bitops-bits-in-byte__bitops-bits-in-byte': [9.1, u'ms'],
u'samples': [1, u'samples']
})
# Clean up after test.
tempfile.mkdtemp = save_real_mkdtemp
command = 'rm -Rf %s' % self.tmpdir
self.result.ce.RunCommand(command)
@mock.patch.object(misc, 'GetRoot')
@mock.patch.object(command_executer.CommandExecuter, 'RunCommand')
def test_cleanup(self, mock_runcmd, mock_getroot):
# Test 1. 'rm_chroot_tmp' is True; self.results_dir exists;
# self.temp_dir exists; results_dir name contains 'test_that_results_'.
mock_getroot.return_value = ['/tmp/tmp_AbcXyz', 'test_that_results_fake']
self.result.ce.RunCommand = mock_runcmd
self.result.results_dir = 'test_results_dir'
self.result.temp_dir = 'testtemp_dir'
self.result.CleanUp(True)
self.assertEqual(mock_getroot.call_count, 1)
self.assertEqual(mock_runcmd.call_count, 2)
self.assertEqual(mock_runcmd.call_args_list[0][0],
('rm -rf test_results_dir',))
self.assertEqual(mock_runcmd.call_args_list[1][0], ('rm -rf testtemp_dir',))
# Test 2. Same, except ath results_dir name does not contain
# 'test_that_results_'
mock_getroot.reset_mock()
mock_runcmd.reset_mock()
mock_getroot.return_value = ['/tmp/tmp_AbcXyz', 'other_results_fake']
self.result.ce.RunCommand = mock_runcmd
self.result.results_dir = 'test_results_dir'
self.result.temp_dir = 'testtemp_dir'
self.result.CleanUp(True)
self.assertEqual(mock_getroot.call_count, 1)
self.assertEqual(mock_runcmd.call_count, 2)
self.assertEqual(mock_runcmd.call_args_list[0][0],
('rm -rf /tmp/tmp_AbcXyz',))
self.assertEqual(mock_runcmd.call_args_list[1][0], ('rm -rf testtemp_dir',))
# Test 3. mock_getroot returns nothing; 'rm_chroot_tmp' is False.
mock_getroot.reset_mock()
mock_runcmd.reset_mock()
self.result.CleanUp(False)
self.assertEqual(mock_getroot.call_count, 0)
self.assertEqual(mock_runcmd.call_count, 1)
self.assertEqual(mock_runcmd.call_args_list[0][0], ('rm -rf testtemp_dir',))
# Test 4. 'rm_chroot_tmp' is True, but result_dir & temp_dir are None.
mock_getroot.reset_mock()
mock_runcmd.reset_mock()
self.result.results_dir = None
self.result.temp_dir = None
self.result.CleanUp(True)
self.assertEqual(mock_getroot.call_count, 0)
self.assertEqual(mock_runcmd.call_count, 0)
@mock.patch.object(misc, 'GetInsideChrootPath')
@mock.patch.object(command_executer.CommandExecuter, 'ChrootRunCommand')
def test_store_to_cache_dir(self, mock_chrootruncmd, mock_getpath):
def FakeMkdtemp(directory=''):
if directory:
pass
return self.tmpdir
if mock_chrootruncmd or mock_getpath:
pass
current_path = os.getcwd()
cache_dir = os.path.join(current_path, 'test_cache/test_output')
self.result.ce = command_executer.GetCommandExecuter(log_level='average')
self.result.out = OUTPUT
self.result.err = error
self.result.retval = 0
self.tmpdir = tempfile.mkdtemp()
if not os.path.exists(self.tmpdir):
os.makedirs(self.tmpdir)
self.result.results_dir = os.path.join(os.getcwd(), 'test_cache')
save_real_mkdtemp = tempfile.mkdtemp
tempfile.mkdtemp = FakeMkdtemp
mock_mm = machine_manager.MockMachineManager('/tmp/chromeos_root', 0,
'average', '')
mock_mm.machine_checksum_string['mock_label'] = 'fake_machine_checksum123'
mock_keylist = ['key1', 'key2', 'key3']
test_flag.SetTestMode(True)
self.result.StoreToCacheDir(cache_dir, mock_mm, mock_keylist)
# Check that the correct things were written to the 'cache'.
test_dir = os.path.join(os.getcwd(), 'test_cache/test_output')
base_dir = os.path.join(os.getcwd(), 'test_cache/compare_output')
self.assertTrue(os.path.exists(os.path.join(test_dir, 'autotest.tbz2')))
self.assertTrue(os.path.exists(os.path.join(test_dir, 'machine.txt')))
self.assertTrue(os.path.exists(os.path.join(test_dir, 'results.txt')))
f1 = os.path.join(test_dir, 'machine.txt')
f2 = os.path.join(base_dir, 'machine.txt')
cmd = 'diff %s %s' % (f1, f2)
[_, out, _] = self.result.ce.RunCommandWOutput(cmd)
self.assertEqual(len(out), 0)
f1 = os.path.join(test_dir, 'results.txt')
f2 = os.path.join(base_dir, 'results.txt')
cmd = 'diff %s %s' % (f1, f2)
[_, out, _] = self.result.ce.RunCommandWOutput(cmd)
self.assertEqual(len(out), 0)
# Clean up after test.
tempfile.mkdtemp = save_real_mkdtemp
command = 'rm %s/*' % test_dir
self.result.ce.RunCommand(command)
TELEMETRY_RESULT_KEYVALS = {
'http://www.webkit.org/perf/sunspider-1.0.2/sunspider-1.0.2/driver.html '
'math-cordic (ms)':
'11.4',
'http://www.webkit.org/perf/sunspider-1.0.2/sunspider-1.0.2/driver.html '
'access-nbody (ms)':
'6.9',
'http://www.webkit.org/perf/sunspider-1.0.2/sunspider-1.0.2/driver.html '
'access-fannkuch (ms)':
'26.3',
'http://www.webkit.org/perf/sunspider-1.0.2/sunspider-1.0.2/driver.html '
'math-spectral-norm (ms)':
'6.3',
'http://www.webkit.org/perf/sunspider-1.0.2/sunspider-1.0.2/driver.html '
'bitops-nsieve-bits (ms)':
'9.3',
'http://www.webkit.org/perf/sunspider-1.0.2/sunspider-1.0.2/driver.html '
'math-partial-sums (ms)':
'32.8',
'http://www.webkit.org/perf/sunspider-1.0.2/sunspider-1.0.2/driver.html '
'regexp-dna (ms)':
'16.1',
'http://www.webkit.org/perf/sunspider-1.0.2/sunspider-1.0.2/driver.html '
'3d-cube (ms)':
'42.7',
'http://www.webkit.org/perf/sunspider-1.0.2/sunspider-1.0.2/driver.html '
'crypto-md5 (ms)':
'10.8',
'http://www.webkit.org/perf/sunspider-1.0.2/sunspider-1.0.2/driver.html '
'crypto-sha1 (ms)':
'12.4',
'http://www.webkit.org/perf/sunspider-1.0.2/sunspider-1.0.2/driver.html '
'string-tagcloud (ms)':
'47.2',
'http://www.webkit.org/perf/sunspider-1.0.2/sunspider-1.0.2/driver.html '
'string-fasta (ms)':
'36.3',
'http://www.webkit.org/perf/sunspider-1.0.2/sunspider-1.0.2/driver.html '
'access-binary-trees (ms)':
'7.3',
'http://www.webkit.org/perf/sunspider-1.0.2/sunspider-1.0.2/driver.html '
'date-format-xparb (ms)':
'138.1',
'http://www.webkit.org/perf/sunspider-1.0.2/sunspider-1.0.2/driver.html '
'crypto-aes (ms)':
'19.2',
'http://www.webkit.org/perf/sunspider-1.0.2/sunspider-1.0.2/driver.html '
'Total (ms)':
'656.5',
'http://www.webkit.org/perf/sunspider-1.0.2/sunspider-1.0.2/driver.html '
'string-base64 (ms)':
'17.5',
'http://www.webkit.org/perf/sunspider-1.0.2/sunspider-1.0.2/driver.html '
'string-validate-input (ms)':
'24.8',
'http://www.webkit.org/perf/sunspider-1.0.2/sunspider-1.0.2/driver.html '
'3d-raytrace (ms)':
'28.7',
'http://www.webkit.org/perf/sunspider-1.0.2/sunspider-1.0.2/driver.html '
'controlflow-recursive (ms)':
'5.3',
'http://www.webkit.org/perf/sunspider-1.0.2/sunspider-1.0.2/driver.html '
'bitops-bits-in-byte (ms)':
'9.8',
'http://www.webkit.org/perf/sunspider-1.0.2/sunspider-1.0.2/driver.html '
'3d-morph (ms)':
'50.2',
'http://www.webkit.org/perf/sunspider-1.0.2/sunspider-1.0.2/driver.html '
'bitops-bitwise-and (ms)':
'8.8',
'http://www.webkit.org/perf/sunspider-1.0.2/sunspider-1.0.2/driver.html '
'access-nsieve (ms)':
'8.6',
'http://www.webkit.org/perf/sunspider-1.0.2/sunspider-1.0.2/driver.html '
'date-format-tofte (ms)':
'31.2',
'http://www.webkit.org/perf/sunspider-1.0.2/sunspider-1.0.2/driver.html '
'bitops-3bit-bits-in-byte (ms)':
'3.5',
'retval':
0,
'http://www.webkit.org/perf/sunspider-1.0.2/sunspider-1.0.2/driver.html '
'string-unpack-code (ms)':
'45.0'
}
PURE_TELEMETRY_OUTPUT = """
page_name,3d-cube (ms),3d-morph (ms),3d-raytrace (ms),Total (ms),access-binary-trees (ms),access-fannkuch (ms),access-nbody (ms),access-nsieve (ms),bitops-3bit-bits-in-byte (ms),bitops-bits-in-byte (ms),bitops-bitwise-and (ms),bitops-nsieve-bits (ms),controlflow-recursive (ms),crypto-aes (ms),crypto-md5 (ms),crypto-sha1 (ms),date-format-tofte (ms),date-format-xparb (ms),math-cordic (ms),math-partial-sums (ms),math-spectral-norm (ms),regexp-dna (ms),string-base64 (ms),string-fasta (ms),string-tagcloud (ms),string-unpack-code (ms),string-validate-input (ms)\r\nhttp://www.webkit.org/perf/sunspider-1.0.2/sunspider-1.0.2/driver.html,42.7,50.2,28.7,656.5,7.3,26.3,6.9,8.6,3.5,9.8,8.8,9.3,5.3,19.2,10.8,12.4,31.2,138.1,11.4,32.8,6.3,16.1,17.5,36.3,47.2,45.0,24.8\r
"""
class TelemetryResultTest(unittest.TestCase):
"""Telemetry result test."""
def __init__(self, *args, **kwargs):
super(TelemetryResultTest, self).__init__(*args, **kwargs)
self.callFakeProcessResults = False
self.result = None
self.mock_logger = mock.Mock(spec=logger.Logger)
self.mock_cmd_exec = mock.Mock(spec=command_executer.CommandExecuter)
self.mock_label = MockLabel('mock_label', 'build', 'chromeos_image',
'autotest_dir', 'debug_dir', '/tmp', 'lumpy',
'remote', 'image_args', 'cache_dir', 'average',
'gcc', False, None)
self.mock_machine = machine_manager.MockCrosMachine('falco.cros',
'/tmp/chromeos',
'average')
def test_populate_from_run(self):
def FakeProcessResults():
self.callFakeProcessResults = True
self.callFakeProcessResults = False
self.result = TelemetryResult(self.mock_logger, self.mock_label, 'average',
self.mock_cmd_exec)
self.result.ProcessResults = FakeProcessResults
self.result.PopulateFromRun(OUTPUT, error, 3, 'fake_test',
'telemetry_Crosperf', '')
self.assertTrue(self.callFakeProcessResults)
self.assertEqual(self.result.out, OUTPUT)
self.assertEqual(self.result.err, error)
self.assertEqual(self.result.retval, 3)
def test_populate_from_cache_dir_and_process_results(self):
self.result = TelemetryResult(self.mock_logger, self.mock_label, 'average',
self.mock_machine)
current_path = os.getcwd()
cache_dir = os.path.join(current_path,
'test_cache/test_puretelemetry_input')
self.result.PopulateFromCacheDir(cache_dir, '', '', '')
self.assertEqual(self.result.out.strip(), PURE_TELEMETRY_OUTPUT.strip())
self.assertEqual(self.result.err, '')
self.assertEqual(self.result.retval, 0)
self.assertEqual(self.result.keyvals, TELEMETRY_RESULT_KEYVALS)
class ResultsCacheTest(unittest.TestCase):
"""Resultcache test class."""
def __init__(self, *args, **kwargs):
super(ResultsCacheTest, self).__init__(*args, **kwargs)
self.fakeCacheReturnResult = None
self.mock_logger = mock.Mock(spec=logger.Logger)
self.mock_label = MockLabel('mock_label', 'build', 'chromeos_image',
'autotest_dir', 'debug_dir', '/tmp', 'lumpy',
'remote', 'image_args', 'cache_dir', 'average',
'gcc', False, None)
def setUp(self):
self.results_cache = ResultsCache()
mock_machine = machine_manager.MockCrosMachine('falco.cros',
'/tmp/chromeos', 'average')
mock_mm = machine_manager.MockMachineManager('/tmp/chromeos_root', 0,
'average', '')
mock_mm.machine_checksum_string['mock_label'] = 'fake_machine_checksum123'
self.results_cache.Init(
self.mock_label.chromeos_image,
self.mock_label.chromeos_root,
'sunspider',
1, # benchmark_run.iteration,
'', # benchmark_run.test_args,
'', # benchmark_run.profiler_args,
mock_mm,
mock_machine,
self.mock_label.board,
[CacheConditions.CACHE_FILE_EXISTS, CacheConditions.CHECKSUMS_MATCH],
self.mock_logger,
'average',
self.mock_label,
'', # benchmark_run.share_cache
'telemetry_Crosperf',
True, # benchmark_run.show_all_results
False, # benchmark_run.run_local
'') # benchmark_run.cwp_dso
@mock.patch.object(image_checksummer.ImageChecksummer, 'Checksum')
def test_get_cache_dir_for_write(self, mock_checksum):
def FakeGetMachines(label):
if label:
pass
m1 = machine_manager.MockCrosMachine('lumpy1.cros',
self.results_cache.chromeos_root,
'average')
m2 = machine_manager.MockCrosMachine('lumpy2.cros',
self.results_cache.chromeos_root,
'average')
return [m1, m2]
mock_checksum.return_value = 'FakeImageChecksumabc123'
self.results_cache.machine_manager.GetMachines = FakeGetMachines
self.results_cache.machine_manager.machine_checksum['mock_label'] = \
'FakeMachineChecksumabc987'
# Based on the label, benchmark and machines, get the directory in which
# to store the cache information for this test run.
result_path = self.results_cache.GetCacheDirForWrite()
# Verify that the returned directory is correct (since the label
# contained a cache_dir, named 'cache_dir', that's what is expected in
# the result, rather than '~/cros_scratch').
comp_path = os.path.join(
os.getcwd(), 'cache_dir/54524606abaae4fdf7b02f49f7ae7127_'
'sunspider_1_fda29412ceccb72977516c4785d08e2c_'
'FakeImageChecksumabc123_FakeMachineChecksum'
'abc987__6')
self.assertEqual(result_path, comp_path)
def test_form_cache_dir(self):
# This is very similar to the previous test (FormCacheDir is called
# from GetCacheDirForWrite).
cache_key_list = ('54524606abaae4fdf7b02f49f7ae7127', 'sunspider', '1',
'7215ee9c7d9dc229d2921a40e899ec5f',
'FakeImageChecksumabc123', '*', '*', '6')
path = self.results_cache.FormCacheDir(cache_key_list)
self.assertEqual(len(path), 1)
path1 = path[0]
test_dirname = ('54524606abaae4fdf7b02f49f7ae7127_sunspider_1_7215ee9'
'c7d9dc229d2921a40e899ec5f_FakeImageChecksumabc123_*_*_6')
comp_path = os.path.join(os.getcwd(), 'cache_dir', test_dirname)
self.assertEqual(path1, comp_path)
@mock.patch.object(image_checksummer.ImageChecksummer, 'Checksum')
def test_get_cache_key_list(self, mock_checksum):
# This tests the mechanism that generates the various pieces of the
# cache directory name, based on various conditions.
def FakeGetMachines(label):
if label:
pass
m1 = machine_manager.MockCrosMachine('lumpy1.cros',
self.results_cache.chromeos_root,
'average')
m2 = machine_manager.MockCrosMachine('lumpy2.cros',
self.results_cache.chromeos_root,
'average')
return [m1, m2]
mock_checksum.return_value = 'FakeImageChecksumabc123'
self.results_cache.machine_manager.GetMachines = FakeGetMachines
self.results_cache.machine_manager.machine_checksum['mock_label'] = \
'FakeMachineChecksumabc987'
# Test 1. Generating cache name for reading (not writing).
key_list = self.results_cache.GetCacheKeyList(True)
self.assertEqual(key_list[0], '*') # Machine checksum value, for read.
self.assertEqual(key_list[1], 'sunspider')
self.assertEqual(key_list[2], '1')
self.assertEqual(key_list[3], 'fda29412ceccb72977516c4785d08e2c')
self.assertEqual(key_list[4], 'FakeImageChecksumabc123')
self.assertEqual(key_list[5], '*')
self.assertEqual(key_list[6], '*')
self.assertEqual(key_list[7], '6')
# Test 2. Generating cache name for writing, with local image type.
key_list = self.results_cache.GetCacheKeyList(False)
self.assertEqual(key_list[0], '54524606abaae4fdf7b02f49f7ae7127')
self.assertEqual(key_list[1], 'sunspider')
self.assertEqual(key_list[2], '1')
self.assertEqual(key_list[3], 'fda29412ceccb72977516c4785d08e2c')
self.assertEqual(key_list[4], 'FakeImageChecksumabc123')
self.assertEqual(key_list[5], 'FakeMachineChecksumabc987')
self.assertEqual(key_list[6], '')
self.assertEqual(key_list[7], '6')
# Test 3. Generating cache name for writing, with trybot image type.
self.results_cache.label.image_type = 'trybot'
key_list = self.results_cache.GetCacheKeyList(False)
self.assertEqual(key_list[0], '54524606abaae4fdf7b02f49f7ae7127')
self.assertEqual(key_list[3], 'fda29412ceccb72977516c4785d08e2c')
self.assertEqual(key_list[4], '54524606abaae4fdf7b02f49f7ae7127')
self.assertEqual(key_list[5], 'FakeMachineChecksumabc987')
# Test 4. Generating cache name for writing, with official image type.
self.results_cache.label.image_type = 'official'
key_list = self.results_cache.GetCacheKeyList(False)
self.assertEqual(key_list[0], '54524606abaae4fdf7b02f49f7ae7127')
self.assertEqual(key_list[1], 'sunspider')
self.assertEqual(key_list[2], '1')
self.assertEqual(key_list[3], 'fda29412ceccb72977516c4785d08e2c')
self.assertEqual(key_list[4], '*')
self.assertEqual(key_list[5], 'FakeMachineChecksumabc987')
self.assertEqual(key_list[6], '')
self.assertEqual(key_list[7], '6')
# Test 5. Generating cache name for writing, with local image type, and
# specifying that the image path must match the cached image path.
self.results_cache.label.image_type = 'local'
self.results_cache.cache_conditions.append(CacheConditions.IMAGE_PATH_MATCH)
key_list = self.results_cache.GetCacheKeyList(False)
self.assertEqual(key_list[0], '54524606abaae4fdf7b02f49f7ae7127')
self.assertEqual(key_list[3], 'fda29412ceccb72977516c4785d08e2c')
self.assertEqual(key_list[4], 'FakeImageChecksumabc123')
self.assertEqual(key_list[5], 'FakeMachineChecksumabc987')
@mock.patch.object(command_executer.CommandExecuter, 'RunCommand')
@mock.patch.object(os.path, 'isdir')
@mock.patch.object(Result, 'CreateFromCacheHit')
def test_read_result(self, mock_create, mock_isdir, mock_runcmd):
self.fakeCacheReturnResult = None
def FakeGetCacheDirForRead():
return self.fakeCacheReturnResult
def FakeGetCacheDirForWrite():
return self.fakeCacheReturnResult
mock_cmd_exec = mock.Mock(spec=command_executer.CommandExecuter)
fake_result = Result(self.mock_logger, self.mock_label, 'average',
mock_cmd_exec)
fake_result.retval = 0
# Set up results_cache _GetCacheDirFor{Read,Write} to return
# self.fakeCacheReturnResult, which is initially None (see above).
# So initially, no cache dir is returned.
self.results_cache.GetCacheDirForRead = FakeGetCacheDirForRead
self.results_cache.GetCacheDirForWrite = FakeGetCacheDirForWrite
mock_isdir.return_value = True
save_cc = [
CacheConditions.CACHE_FILE_EXISTS, CacheConditions.CHECKSUMS_MATCH
]
self.results_cache.cache_conditions.append(CacheConditions.FALSE)
# Test 1. CacheCondition.FALSE, which means do not read from the cache.
# (force re-running of test). Result should be None.
res = self.results_cache.ReadResult()
self.assertIsNone(res)
self.assertEqual(mock_runcmd.call_count, 1)
# Test 2. Remove CacheCondition.FALSE. Result should still be None,
# because GetCacheDirForRead is returning None at the moment.
mock_runcmd.reset_mock()
self.results_cache.cache_conditions = save_cc
res = self.results_cache.ReadResult()
self.assertIsNone(res)
self.assertEqual(mock_runcmd.call_count, 0)
# Test 3. Now set up cache dir to be returned by GetCacheDirForRead.
# Since cache_dir is found, will call Result.CreateFromCacheHit, which
# which will actually all our mock_create and should return fake_result.
self.fakeCacheReturnResult = 'fake/cache/dir'
mock_create.return_value = fake_result
res = self.results_cache.ReadResult()
self.assertEqual(mock_runcmd.call_count, 0)
self.assertEqual(res, fake_result)
# Test 4. os.path.isdir(cache_dir) will now return false, so result
# should be None again (no cache found).
mock_isdir.return_value = False
res = self.results_cache.ReadResult()
self.assertEqual(mock_runcmd.call_count, 0)
self.assertIsNone(res)
# Test 5. os.path.isdir returns true, but mock_create now returns None
# (the call to CreateFromCacheHit returns None), so overal result is None.
mock_isdir.return_value = True
mock_create.return_value = None
res = self.results_cache.ReadResult()
self.assertEqual(mock_runcmd.call_count, 0)
self.assertIsNone(res)
# Test 6. Everything works 'as expected', result should be fake_result.
mock_create.return_value = fake_result
res = self.results_cache.ReadResult()
self.assertEqual(mock_runcmd.call_count, 0)
self.assertEqual(res, fake_result)
# Test 7. The run failed; result should be None.
mock_create.return_value = fake_result
fake_result.retval = 1
self.results_cache.cache_conditions.append(CacheConditions.RUN_SUCCEEDED)
res = self.results_cache.ReadResult()
self.assertEqual(mock_runcmd.call_count, 0)
self.assertIsNone(res)
if __name__ == '__main__':
unittest.main()