2018-12-21 06:08:57 +00:00
|
|
|
"""Show information about the test environment."""
|
2019-07-12 06:46:20 +00:00
|
|
|
from __future__ import (absolute_import, division, print_function)
|
|
|
|
__metaclass__ = type
|
2018-12-21 06:08:57 +00:00
|
|
|
|
|
|
|
import datetime
|
|
|
|
import json
|
2019-03-05 19:58:13 +00:00
|
|
|
import functools
|
2018-12-21 06:08:57 +00:00
|
|
|
import os
|
|
|
|
import platform
|
|
|
|
import re
|
2019-03-05 19:58:13 +00:00
|
|
|
import signal
|
2018-12-21 06:08:57 +00:00
|
|
|
import sys
|
2019-03-05 19:58:13 +00:00
|
|
|
import time
|
2018-12-21 06:08:57 +00:00
|
|
|
|
2019-08-06 21:43:29 +00:00
|
|
|
from .config import (
|
2018-12-21 06:08:57 +00:00
|
|
|
CommonConfig,
|
2019-03-05 19:58:13 +00:00
|
|
|
TestConfig,
|
2018-12-21 06:08:57 +00:00
|
|
|
)
|
|
|
|
|
2019-08-06 21:43:29 +00:00
|
|
|
from .util import (
|
2018-12-21 06:08:57 +00:00
|
|
|
display,
|
|
|
|
find_executable,
|
|
|
|
SubprocessError,
|
|
|
|
ApplicationError,
|
2019-08-26 21:02:55 +00:00
|
|
|
get_ansible_version,
|
|
|
|
get_available_python_versions,
|
2018-12-21 06:08:57 +00:00
|
|
|
)
|
|
|
|
|
2019-08-28 06:40:06 +00:00
|
|
|
from .util_common import (
|
|
|
|
write_json_test_results,
|
|
|
|
write_json_file,
|
|
|
|
ResultType,
|
|
|
|
)
|
|
|
|
|
2019-08-06 21:43:29 +00:00
|
|
|
from .git import (
|
2018-12-21 06:08:57 +00:00
|
|
|
Git,
|
|
|
|
)
|
|
|
|
|
2019-08-06 21:43:29 +00:00
|
|
|
from .docker_util import (
|
2018-12-21 06:08:57 +00:00
|
|
|
docker_info,
|
|
|
|
docker_version
|
|
|
|
)
|
|
|
|
|
2019-08-06 21:43:29 +00:00
|
|
|
from .thread import (
|
2019-03-05 19:58:13 +00:00
|
|
|
WrappedThread,
|
|
|
|
)
|
|
|
|
|
2019-08-06 21:43:29 +00:00
|
|
|
from .constants import (
|
2019-03-05 19:58:13 +00:00
|
|
|
TIMEOUT_PATH,
|
|
|
|
)
|
|
|
|
|
2019-08-06 21:43:29 +00:00
|
|
|
from .test import (
|
2019-03-05 19:58:13 +00:00
|
|
|
TestTimeout,
|
|
|
|
)
|
|
|
|
|
2019-08-26 21:02:55 +00:00
|
|
|
from .executor import (
|
|
|
|
SUPPORTED_PYTHON_VERSIONS,
|
|
|
|
)
|
|
|
|
|
2018-12-21 06:08:57 +00:00
|
|
|
|
|
|
|
class EnvConfig(CommonConfig):
|
|
|
|
"""Configuration for the tools command."""
|
|
|
|
def __init__(self, args):
|
|
|
|
"""
|
|
|
|
:type args: any
|
|
|
|
"""
|
|
|
|
super(EnvConfig, self).__init__(args, 'env')
|
|
|
|
|
2019-03-05 19:58:13 +00:00
|
|
|
self.show = args.show
|
2018-12-21 06:08:57 +00:00
|
|
|
self.dump = args.dump
|
2019-03-05 19:58:13 +00:00
|
|
|
self.timeout = args.timeout
|
2018-12-21 06:08:57 +00:00
|
|
|
|
2019-08-26 21:02:55 +00:00
|
|
|
if not self.show and not self.dump and self.timeout is None:
|
|
|
|
# default to --show if no options were given
|
|
|
|
self.show = True
|
|
|
|
|
2018-12-21 06:08:57 +00:00
|
|
|
|
|
|
|
def command_env(args):
|
|
|
|
"""
|
|
|
|
:type args: EnvConfig
|
|
|
|
"""
|
2019-03-05 19:58:13 +00:00
|
|
|
show_dump_env(args)
|
|
|
|
set_timeout(args)
|
|
|
|
|
|
|
|
|
|
|
|
def show_dump_env(args):
|
|
|
|
"""
|
|
|
|
:type args: EnvConfig
|
|
|
|
"""
|
|
|
|
if not args.show and not args.dump:
|
|
|
|
return
|
|
|
|
|
2018-12-21 06:08:57 +00:00
|
|
|
data = dict(
|
|
|
|
ansible=dict(
|
2019-08-08 23:14:19 +00:00
|
|
|
version=get_ansible_version(),
|
2018-12-21 06:08:57 +00:00
|
|
|
),
|
|
|
|
docker=get_docker_details(args),
|
|
|
|
environ=os.environ.copy(),
|
2019-08-26 21:02:55 +00:00
|
|
|
location=dict(
|
|
|
|
pwd=os.environ.get('PWD', None),
|
|
|
|
cwd=os.getcwd(),
|
|
|
|
),
|
2018-12-21 06:08:57 +00:00
|
|
|
git=get_git_details(args),
|
|
|
|
platform=dict(
|
|
|
|
datetime=datetime.datetime.utcnow().strftime('%Y-%m-%dT%H:%M:%SZ'),
|
|
|
|
platform=platform.platform(),
|
|
|
|
uname=platform.uname(),
|
|
|
|
),
|
|
|
|
python=dict(
|
|
|
|
executable=sys.executable,
|
|
|
|
version=platform.python_version(),
|
|
|
|
),
|
2019-08-26 21:02:55 +00:00
|
|
|
interpreters=get_available_python_versions(SUPPORTED_PYTHON_VERSIONS),
|
2018-12-21 06:08:57 +00:00
|
|
|
)
|
|
|
|
|
|
|
|
if args.show:
|
|
|
|
verbose = {
|
|
|
|
'docker': 3,
|
|
|
|
'docker.executable': 0,
|
|
|
|
'environ': 2,
|
|
|
|
'platform.uname': 1,
|
|
|
|
}
|
|
|
|
|
|
|
|
show_dict(data, verbose)
|
|
|
|
|
|
|
|
if args.dump and not args.explain:
|
2019-08-28 06:40:06 +00:00
|
|
|
write_json_test_results(ResultType.BOT, 'data-environment.json', data)
|
2018-12-21 06:08:57 +00:00
|
|
|
|
|
|
|
|
2019-03-05 19:58:13 +00:00
|
|
|
def set_timeout(args):
|
|
|
|
"""
|
|
|
|
:type args: EnvConfig
|
|
|
|
"""
|
|
|
|
if args.timeout is None:
|
|
|
|
return
|
|
|
|
|
|
|
|
if args.timeout:
|
|
|
|
deadline = (datetime.datetime.utcnow() + datetime.timedelta(minutes=args.timeout)).strftime('%Y-%m-%dT%H:%M:%SZ')
|
|
|
|
|
|
|
|
display.info('Setting a %d minute test timeout which will end at: %s' % (args.timeout, deadline), verbosity=1)
|
|
|
|
else:
|
|
|
|
deadline = None
|
|
|
|
|
|
|
|
display.info('Clearing existing test timeout.', verbosity=1)
|
|
|
|
|
|
|
|
if args.explain:
|
|
|
|
return
|
|
|
|
|
|
|
|
if deadline:
|
|
|
|
data = dict(
|
|
|
|
duration=args.timeout,
|
|
|
|
deadline=deadline,
|
|
|
|
)
|
|
|
|
|
2019-08-28 06:40:06 +00:00
|
|
|
write_json_file(TIMEOUT_PATH, data)
|
2019-03-05 19:58:13 +00:00
|
|
|
elif os.path.exists(TIMEOUT_PATH):
|
|
|
|
os.remove(TIMEOUT_PATH)
|
|
|
|
|
|
|
|
|
|
|
|
def get_timeout():
|
|
|
|
"""
|
|
|
|
:rtype: dict[str, any] | None
|
|
|
|
"""
|
|
|
|
if not os.path.exists(TIMEOUT_PATH):
|
|
|
|
return None
|
|
|
|
|
|
|
|
with open(TIMEOUT_PATH, 'r') as timeout_fd:
|
|
|
|
data = json.load(timeout_fd)
|
|
|
|
|
|
|
|
data['deadline'] = datetime.datetime.strptime(data['deadline'], '%Y-%m-%dT%H:%M:%SZ')
|
|
|
|
|
|
|
|
return data
|
|
|
|
|
|
|
|
|
|
|
|
def configure_timeout(args):
|
|
|
|
"""
|
|
|
|
:type args: CommonConfig
|
|
|
|
"""
|
|
|
|
if isinstance(args, TestConfig):
|
|
|
|
configure_test_timeout(args) # only tests are subject to the timeout
|
|
|
|
|
|
|
|
|
|
|
|
def configure_test_timeout(args):
|
|
|
|
"""
|
|
|
|
:type args: TestConfig
|
|
|
|
"""
|
|
|
|
timeout = get_timeout()
|
|
|
|
|
|
|
|
if not timeout:
|
|
|
|
return
|
|
|
|
|
|
|
|
timeout_start = datetime.datetime.utcnow()
|
|
|
|
timeout_duration = timeout['duration']
|
|
|
|
timeout_deadline = timeout['deadline']
|
|
|
|
timeout_remaining = timeout_deadline - timeout_start
|
|
|
|
|
|
|
|
test_timeout = TestTimeout(timeout_duration)
|
|
|
|
|
|
|
|
if timeout_remaining <= datetime.timedelta():
|
|
|
|
test_timeout.write(args)
|
|
|
|
|
|
|
|
raise ApplicationError('The %d minute test timeout expired %s ago at %s.' % (
|
|
|
|
timeout_duration, timeout_remaining * -1, timeout_deadline))
|
|
|
|
|
|
|
|
display.info('The %d minute test timeout expires in %s at %s.' % (
|
|
|
|
timeout_duration, timeout_remaining, timeout_deadline), verbosity=1)
|
|
|
|
|
|
|
|
def timeout_handler(_dummy1, _dummy2):
|
|
|
|
"""Runs when SIGUSR1 is received."""
|
|
|
|
test_timeout.write(args)
|
|
|
|
|
|
|
|
raise ApplicationError('Tests aborted after exceeding the %d minute time limit.' % timeout_duration)
|
|
|
|
|
|
|
|
def timeout_waiter(timeout_seconds):
|
|
|
|
"""
|
|
|
|
:type timeout_seconds: int
|
|
|
|
"""
|
|
|
|
time.sleep(timeout_seconds)
|
|
|
|
os.kill(os.getpid(), signal.SIGUSR1)
|
|
|
|
|
|
|
|
signal.signal(signal.SIGUSR1, timeout_handler)
|
|
|
|
|
|
|
|
instance = WrappedThread(functools.partial(timeout_waiter, timeout_remaining.seconds))
|
|
|
|
instance.daemon = True
|
|
|
|
instance.start()
|
|
|
|
|
|
|
|
|
2018-12-21 06:08:57 +00:00
|
|
|
def show_dict(data, verbose, root_verbosity=0, path=None):
|
|
|
|
"""
|
|
|
|
:type data: dict[str, any]
|
|
|
|
:type verbose: dict[str, int]
|
|
|
|
:type root_verbosity: int
|
|
|
|
:type path: list[str] | None
|
|
|
|
"""
|
|
|
|
path = path if path else []
|
|
|
|
|
|
|
|
for key, value in sorted(data.items()):
|
|
|
|
indent = ' ' * len(path)
|
|
|
|
key_path = path + [key]
|
|
|
|
key_name = '.'.join(key_path)
|
|
|
|
verbosity = verbose.get(key_name, root_verbosity)
|
|
|
|
|
|
|
|
if isinstance(value, (tuple, list)):
|
|
|
|
display.info(indent + '%s:' % key, verbosity=verbosity)
|
|
|
|
for item in value:
|
|
|
|
display.info(indent + ' - %s' % item, verbosity=verbosity)
|
|
|
|
elif isinstance(value, dict):
|
|
|
|
min_verbosity = min([verbosity] + [v for k, v in verbose.items() if k.startswith('%s.' % key)])
|
|
|
|
display.info(indent + '%s:' % key, verbosity=min_verbosity)
|
|
|
|
show_dict(value, verbose, verbosity, key_path)
|
|
|
|
else:
|
|
|
|
display.info(indent + '%s: %s' % (key, value), verbosity=verbosity)
|
|
|
|
|
|
|
|
|
|
|
|
def get_docker_details(args):
|
|
|
|
"""
|
|
|
|
:type args: CommonConfig
|
|
|
|
:rtype: dict[str, any]
|
|
|
|
"""
|
|
|
|
docker = find_executable('docker', required=False)
|
|
|
|
info = None
|
|
|
|
version = None
|
|
|
|
|
|
|
|
if docker:
|
|
|
|
try:
|
|
|
|
info = docker_info(args)
|
|
|
|
except SubprocessError as ex:
|
|
|
|
display.warning('Failed to collect docker info:\n%s' % ex)
|
|
|
|
|
|
|
|
try:
|
|
|
|
version = docker_version(args)
|
|
|
|
except SubprocessError as ex:
|
|
|
|
display.warning('Failed to collect docker version:\n%s' % ex)
|
|
|
|
|
|
|
|
docker_details = dict(
|
|
|
|
executable=docker,
|
|
|
|
info=info,
|
|
|
|
version=version,
|
|
|
|
)
|
|
|
|
|
|
|
|
return docker_details
|
|
|
|
|
|
|
|
|
|
|
|
def get_git_details(args):
|
|
|
|
"""
|
|
|
|
:type args: CommonConfig
|
|
|
|
:rtype: dict[str, any]
|
|
|
|
"""
|
|
|
|
commit = os.environ.get('COMMIT')
|
|
|
|
base_commit = os.environ.get('BASE_COMMIT')
|
|
|
|
|
|
|
|
git_details = dict(
|
|
|
|
base_commit=base_commit,
|
|
|
|
commit=commit,
|
|
|
|
merged_commit=get_merged_commit(args, commit),
|
|
|
|
)
|
|
|
|
|
|
|
|
return git_details
|
|
|
|
|
|
|
|
|
2019-07-11 20:03:49 +00:00
|
|
|
# noinspection PyUnusedLocal
|
2019-06-27 19:00:26 +00:00
|
|
|
def get_merged_commit(args, commit): # pylint: disable=unused-argument
|
2018-12-21 06:08:57 +00:00
|
|
|
"""
|
|
|
|
:type args: CommonConfig
|
|
|
|
:type commit: str
|
|
|
|
:rtype: str | None
|
|
|
|
"""
|
|
|
|
if not commit:
|
|
|
|
return None
|
|
|
|
|
2019-06-27 19:00:26 +00:00
|
|
|
git = Git()
|
2018-12-21 06:08:57 +00:00
|
|
|
|
|
|
|
try:
|
|
|
|
show_commit = git.run_git(['show', '--no-patch', '--no-abbrev', commit])
|
|
|
|
except SubprocessError as ex:
|
|
|
|
# This should only fail for pull requests where the commit does not exist.
|
|
|
|
# Merge runs would fail much earlier when attempting to checkout the commit.
|
|
|
|
raise ApplicationError('Commit %s was not found:\n\n%s\n\n'
|
2019-04-12 18:52:44 +00:00
|
|
|
'GitHub may not have fully replicated the commit across their infrastructure.\n'
|
|
|
|
'It is also possible the commit was removed by a force push between job creation and execution.\n'
|
2018-12-21 06:08:57 +00:00
|
|
|
'Find the latest run for the pull request and restart failed jobs as needed.'
|
|
|
|
% (commit, ex.stderr.strip()))
|
|
|
|
|
|
|
|
head_commit = git.run_git(['show', '--no-patch', '--no-abbrev', 'HEAD'])
|
|
|
|
|
|
|
|
if show_commit == head_commit:
|
|
|
|
# Commit is HEAD, so this is not a pull request or the base branch for the pull request is up-to-date.
|
|
|
|
return None
|
|
|
|
|
|
|
|
match_merge = re.search(r'^Merge: (?P<parents>[0-9a-f]{40} [0-9a-f]{40})$', head_commit, flags=re.MULTILINE)
|
|
|
|
|
|
|
|
if not match_merge:
|
|
|
|
# The most likely scenarios resulting in a failure here are:
|
|
|
|
# A new run should or does supersede this job, but it wasn't cancelled in time.
|
|
|
|
# A job was superseded and then later restarted.
|
|
|
|
raise ApplicationError('HEAD is not commit %s or a merge commit:\n\n%s\n\n'
|
|
|
|
'This job has likely been superseded by another run due to additional commits being pushed.\n'
|
|
|
|
'Find the latest run for the pull request and restart failed jobs as needed.'
|
|
|
|
% (commit, head_commit.strip()))
|
|
|
|
|
|
|
|
parents = set(match_merge.group('parents').split(' '))
|
|
|
|
|
|
|
|
if len(parents) != 2:
|
|
|
|
raise ApplicationError('HEAD is a %d-way octopus merge.' % len(parents))
|
|
|
|
|
|
|
|
if commit not in parents:
|
|
|
|
raise ApplicationError('Commit %s is not a parent of HEAD.' % commit)
|
|
|
|
|
|
|
|
parents.remove(commit)
|
|
|
|
|
|
|
|
last_commit = parents.pop()
|
|
|
|
|
|
|
|
return last_commit
|