# Copyright 2012 The Chromium Authors
# Use of this source code is governed by a BSD-style license that can be
# found in the LICENSE file.
"""Generic utilities for all python scripts."""
import atexit
import base64
import http.client
import json
import os
import platform
import random
import signal
import socket
import stat
import subprocess
import sys
import tempfile
import time
import zipfile
import requests
import chrome_paths
def GetPlatformName():
"""Return a string to be used in paths for the platform."""
if IsWindows():
return 'win'
if IsMac():
return 'mac'
if IsLinux():
return 'linux'
raise NotImplementedError('Unknown platform "%s".' % sys.platform)
def IsWindows():
return sys.platform == 'cygwin' or sys.platform.startswith('win')
def IsLinux():
return sys.platform.startswith('linux')
def IsMac():
return sys.platform.startswith('darwin')
def Is64Bit():
# This function checks whether the target (not host) word size is 64-bits.
# Since 64-bit hosts can cross-compile 32-bit binaries, check the GN args to
# see what CPU we're targetting.
try:
args_gn = os.path.join(chrome_paths.GetBuildDir(['args.gn']), 'args.gn')
with open(args_gn, encoding='utf-8') as build_args:
for build_arg in build_args:
decommented = build_arg.split('#')[0]
key_and_value = decommented.split('=')
if len(key_and_value) != 2:
continue
key = key_and_value[0].strip()
value = key_and_value[1].strip()
if key == 'target_cpu':
return value.endswith('64')
except:
pass
# If we don't find anything, or if there is no GN args file, default to the
# host architecture.
return platform.architecture()[0] == '64bit'
def GetAbsolutePathOfUserPath(user_path):
"""Expand the given |user_path| (like "~/file") and return its absolute path.
"""
if user_path is None:
return None
return os.path.abspath(os.path.expanduser(user_path))
def _DeleteDir(path):
"""Deletes a directory recursively, which must exist.
Note that this function can fail and raise OSError if another process is
making changes to the directory at the same time.
"""
# Don't use shutil.rmtree because it can't delete read-only files on Win.
for root, dirs, files in os.walk(path, topdown=False):
for name in files:
filename = os.path.join(root, name)
try:
os.chmod(filename, stat.S_IWRITE)
except OSError:
pass
os.remove(filename)
for name in dirs:
os.rmdir(os.path.join(root, name))
os.rmdir(path)
def Delete(path):
"""Deletes the given file or directory (recursively), which must exist."""
if os.path.isdir(path):
_DeleteDir(path)
else:
os.remove(path)
def MaybeDelete(path):
"""Deletes the given file or directory (recurisvely), if it exists."""
if os.path.exists(path):
Delete(path)
def MakeTempDir(parent_dir=None):
"""Creates a temporary directory and returns an absolute path to it.
The temporary directory is automatically deleted when the python interpreter
exits normally.
Args:
parent_dir: the directory to create the temp dir in. If None, the system
temp dir is used.
Returns:
The absolute path to the temporary directory.
"""
path = tempfile.mkdtemp(prefix='chromedriver_', dir=parent_dir)
atexit.register(MaybeDelete, path)
return path
def Zip(path):
"""Zips the given path and returns the zipped file."""
zip_path = os.path.join(MakeTempDir(), 'build.zip')
f = zipfile.ZipFile(zip_path, 'w', zipfile.ZIP_DEFLATED)
f.write(path, os.path.basename(path))
f.close()
return zip_path
def Unzip(zip_path, output_dir):
"""Unzips the given zip file using a system installed unzip tool.
Args:
zip_path: zip file to unzip.
output_dir: directory to unzip the contents of the zip file. The directory
must exist.
Raises:
RuntimeError if the unzip operation fails.
"""
if IsWindows():
unzip_cmd = ['C:\\Program Files\\7-Zip\\7z.exe', 'x', '-y']
else:
unzip_cmd = ['unzip', '-o']
unzip_cmd += [zip_path]
if RunCommand(unzip_cmd, output_dir) != 0:
raise RuntimeError('Unable to unzip %s to %s' % (zip_path, output_dir))
def Kill(pid):
"""Terminate the given pid."""
if IsWindows():
subprocess.call(['taskkill.exe', '/T', '/F', '/PID', str(pid)])
else:
os.kill(pid, signal.SIGTERM)
def RunCommand(cmd, cwd=None, fileName=None):
"""Runs the given command and returns the exit code.
Args:
fileName: file name to redirect output
cmd: list of command arguments.
cwd: working directory to execute the command, or None if the current
working directory should be used.
Returns:
The exit code of the command.
"""
sys.stdout.flush()
if fileName is not None:
with open(fileName, "wb", encoding='utf-8') as out:
process = subprocess.Popen(cmd, cwd=cwd,stdout=out,stderr=out)
else:
process = subprocess.Popen(cmd, cwd=cwd)
process.wait()
sys.stdout.flush()
return process.returncode
def MarkBuildStepStart(name):
print('@@@BUILD_STEP %s@@@' % name)
sys.stdout.flush()
def MarkBuildStepError():
print('@@@STEP_FAILURE@@@')
sys.stdout.flush()
def AddBuildStepText(text):
print('@@@STEP_TEXT@%s@@@' % text)
sys.stdout.flush()
def PrintAndFlush(text):
print(text)
sys.stdout.flush()
def AddLink(label, url):
"""Adds a link with name |label| linking to |url| to current buildbot step.
Args:
label: A string with the name of the label.
url: A string of the URL.
"""
print('@@@STEP_LINK@%s@%s@@@' % (label, url))
def FindProbableFreePorts():
"""Get an generator returning random free ports on the system.
Note that this function has an inherent race condition: some other process
may bind to the port after we return it, so it may no longer be free by then.
The workaround is to do this inside a retry loop. Do not use this function
if there is any alternative.
"""
# This is the range of dynamic ports. See RFC6335 page 10.
dynamic_ports = list(range(49152, 65535))
random.shuffle(dynamic_ports)
for port in dynamic_ports:
try:
socket.create_connection(('127.0.0.1', port), 0.2).close()
except socket.error:
# If we can't connect to the port, then clearly nothing is listening on
# it.
yield port
raise RuntimeError('Cannot find open port')
def WriteResultToJSONFile(test_suites, results, json_path):
"""Aggregate a list of unittest result object and write to a file as a JSON.
This takes a list of result object from one or more runs (for retry purpose)
of Python unittest tests; aggregates the list by appending each test result
from each run and writes to a file in the correct format
for the --isolated-script-test-output argument passed to test isolates.
Args:
test_suites: a list of unittest.TestSuite that were run to get
the list of result object; each test_suite contains
the tests run and is iterated to get all test cases ran.
results: a list of unittest.TextTestResult object returned
from running unittest tests.
json_path: desired path to JSON file of result.
"""
output = {
'interrupted': False,
'num_failures_by_type': {},
'path_delimiter': '.',
'seconds_since_epoch': time.time(),
'tests': {},
'version': 3,
}
def initialize(test_suite):
for test_name in test_suite:
if test_name not in output['tests']:
output['tests'][test_name] = {
'expected': 'PASS',
'actual': []
}
for test_suite in test_suites:
initialize(test_suite)
def get_pass_fail(test_suite, result):
success = []
fail = []
for failure in result.failures + result.errors:
fail.append(failure[0].id())
for test_name in test_suite:
if test_name not in fail:
success.append(test_name)
return {
'success': success,
'fail': fail,
}
for test_suite, result in zip(test_suites, results):
pass_fail = get_pass_fail(test_suite, result)
for s in pass_fail['success']:
output['tests'][s]['actual'].append('PASS')
for f in pass_fail['fail']:
output['tests'][f]['actual'].append('FAIL')
num_fails = 0
for test_result in output['tests'].values():
if test_result['actual'][-1] == 'FAIL':
num_fails += 1
test_result['is_unexpected'] = True
test_result['actual'] = ' '.join(test_result['actual'])
output['num_failures_by_type']['FAIL'] = num_fails
output['num_failures_by_type']['PASS'] = len(output['tests']) - num_fails
with open(json_path, 'w', encoding='utf-8') as script_out_file:
json.dump(output, script_out_file)
script_out_file.write('\n')
def TryUploadingResultToResultSink(results):
def parse(result):
test_results = []
for test_case in result.successes:
test_results.append({
'testId': test_case.id(),
'expected': True,
'status': 'PASS',
})
for (test_case, stack_trace) in result.failures + result.errors:
test_results.append({
'testId': test_case.id(),
'expected': False,
'status': 'FAIL',
# Uses <text-artifact> tag to embed the artifact content
# in summaryHtml.
'summaryHtml': '<p><text-artifact artifact-id="stack_trace"></p>',
# A map of artifacts. The keys are artifact ids which uniquely
# identify an artifact within the test result.
'artifacts': {
'stack_trace': {
'contents': base64.b64encode(stack_trace.encode()).decode(),
},
},
})
return test_results
def getResultSinkTestResults(results):
test_results = []
for r in results:
test_results.extend(parse(r))
return test_results
try:
with open(os.environ['LUCI_CONTEXT'], encoding='utf-8') as f:
sink = json.load(f)['result_sink']
except KeyError:
return
test_results = getResultSinkTestResults(results)
# Uploads all test results at once.
res = requests.post(
url='http://%s/prpc/luci.resultsink.v1.Sink/ReportTestResults' % sink['address'],
headers={
'Content-Type': 'application/json',
'Accept': 'application/json',
'Authorization': 'ResultSink %s' % sink['auth_token'],
},
data=json.dumps({'testResults': test_results})
)
res.raise_for_status()