78ced5318f
Copy module was walking over files in subdirectories repeatedly (a directory tree a few levels deep could bring the time spent into the tens of minutes) This was traced to the fix for this bug report: https://github.com/ansible/ansible/issues/13013 Fixed #13013 a different way and added an integration test to check for regressions of #13013 as we optimize this code. Fixes #21513 |
||
---|---|---|
.. | ||
group_vars | ||
host_vars | ||
roles | ||
targets | ||
vars | ||
amazon.yml | ||
asa.yaml | ||
azure.yml | ||
cleanup_azure.py | ||
cleanup_ec2.py | ||
cleanup_gce.py | ||
cleanup_rax.py | ||
cloudflare.yml | ||
cloudscale.yml | ||
cloudstack.yml | ||
cnos.yaml | ||
consul.yml | ||
consul_inventory.yml | ||
consul_running.py | ||
credentials.template | ||
dellos6.yaml | ||
dellos9.yaml | ||
dellos10.yaml | ||
destructive.yml | ||
eos.yaml | ||
exoscale.yml | ||
galaxy_playbook.yml | ||
galaxy_playbook_git.yml | ||
galaxy_roles.yml | ||
galaxy_rolesfile | ||
gce.yml | ||
gce_credentials.py | ||
integration_config.yml | ||
inventory | ||
inventory.network | ||
inventory.remote.template | ||
inventory.winrm.template | ||
ios.yaml | ||
iosxr.yaml | ||
jenkins.yml | ||
junos.yaml | ||
Makefile | ||
net_command.yml | ||
network-all.yaml | ||
non_destructive.yml | ||
nxos.yaml | ||
ops.yaml | ||
rackspace.yml | ||
README.md | ||
setup_gce.py | ||
target-prefixes.network | ||
test_win_group1.yml | ||
test_win_group2.yml | ||
test_win_group3.yml | ||
vyos.yaml |
Integration tests
The ansible integration system.
Tests for playbooks, by playbooks.
Some tests may require credentials. Credentials may be specified with credentials.yml
.
Some tests may require root.
Quick Start
It is highly recommended that you install and activate the argcomplete
python package.
It provides tab completion in bash
for the ansible-test
test runner.
To get started quickly using Docker containers for testing, see Tests in Docker containers.
Configuration
Making your own version of integration_config.yml
can allow for setting some
tunable parameters to help run the tests better in your environment. Some
tests (e.g. cloud) will only run when access credentials are provided. For
more information about supported credentials, refer to credentials.template
.
Prerequisites
The tests will assume things like hg, svn, and git are installed and in path.
(Complete list pending)
Non-destructive Tests
These tests will modify files in subdirectories, but will not do things that install or remove packages or things outside of those test subdirectories. They will also not reconfigure or bounce system services.
Run as follows for all POSIX platform tests executed by our CI system:
test/runner/ansible-test integration -v posix/ci/
You can select specific tests as well, such as for individual modules:
test/runner/ansible-test integration -v ping
Destructive Tests
These tests are allowed to install and remove some trivial packages. You will likely want to devote these to a virtual environment. They won't reformat your filesystem, however :)
test/runner/ansible-test integration -v destructive/
Cloud Tests
Cloud tests exercise capabilities of cloud modules (e.g. ec2_key). These are not 'tests run in the cloud' so much as tests that leverage the cloud modules and are organized by cloud provider.
In order to run cloud tests, you must provide access credentials in a file
named credentials.yml
. A sample credentials file named
credentials.template
is available for syntax help.
Provide cloud credentials:
cp credentials.template credentials.yml
${EDITOR:-vi} credentials.yml
Run the tests: make cloud
WARNING running cloud integration tests will create and destroy cloud resources. Running these tests may result in additional fees associated with your cloud account. Care is taken to ensure that created resources are removed. However, it is advisable to inspect your AWS console to ensure no unexpected resources are running.
Windows Tests
These tests exercise the winrm connection plugin and Windows modules. You'll need to define an inventory with a remote Windows 2008 or 2012 Server to use for testing, and enable PowerShell Remoting to continue.
Running these tests may result in changes to your Windows host, so don't run them against a production/critical Windows environment.
Enable PowerShell Remoting (run on the Windows host via Remote Desktop): Enable-PSRemoting -Force
Define Windows inventory:
cp inventory.winrm.template inventory.winrm
${EDITOR:-vi} inventory.winrm
Run the Windows tests executed by our CI system:
test/runner/ansible-test windows-integration -v windows/ci/
Tests in Docker containers
If you have a Linux system with Docker installed, running integration tests using the same Docker containers used by the Ansible continuous integration (CI) system is recommended.
Using Docker Engine to run Docker on a non-Linux host is not recommended. Some tests may fail, depending on the image used for testing. Using the
--docker-privileged
option may resolve the issue.
Running Integration Tests
To run all CI integration test targets for POSIX platforms in a Ubuntu 16.04 container:
test/runner/ansible-test integration -v posix/ci/ --docker
You can also run specific tests or select a different Linux distribution.
For example, to run tests for the ping
module on a Ubuntu 14.04 container:
test/runner/ansible-test integration -v ping --docker ubuntu1404
Container Images
Python 2
Most container images are for testing with Python 2:
- centos6
- centos7
- fedora24
- fedora25
- opensuse42.1
- opensuse42.2
- ubuntu1204
- ubuntu1404
- ubuntu1604
Python 3
To test with Python 3 use the following images:
- ubuntu1604py3
Network Tests
Note: From Ansible 2.3, for any new Network Module to be accepted it must be accompanied by a corresponding test.
For further help with this please contact gundalow
in #ansible-devel
on FreeNode IRC.
$ ANSIBLE_ROLES_PATH=targets ansible-playbook network-all.yaml
NOTE To run the network tests you will need a number of test machines and sutabily configured inventory file, a sample is included in test/integration/inventory.network
NOTE As with the rest of the integration tests, they can be found grouped by module in test/integration/targets/MODULENAME/
To filter a set of test cases set limit_to
to the name of the group, generally this is the name of the module:
$ ANSIBLE_ROLES_PATH=targets ansible-playbook -i inventory.network network-all.yaml -e "limit_to=eos_command"
To filter a singular test case set the tags options to eapi or cli, set limit_to to the test group, and test_cases to the name of the test:
$ ANSIBLE_ROLES_PATH=targets ansible-playbook -i inventory.network network-all.yaml --tags="cli" -e "limit_to=eos_command test_case=notequal"
Contributing Test Cases
Test cases are added to roles based on the module being testing. Test cases
should include both cli
and eapi
test cases. Cli test cases should be
added to test/integration/targets/modulename/tests/cli
and eapi tests should be added to
test/integration/targets/modulename/tests/eapi
.
In addition to positive testing, negative tests are required to ensure user friendly warnings & errors are generated, rather than backtraces, for example:
- name: test invalid subset (foobar)
eos_facts:
provider: "{{ cli }}"
gather_subset:
- "foobar"
register: result
ignore_errors: true
- assert:
that:
# Failures shouldn't return changes
- "result.changed == false"
# It's a failure
- "result.failed == true"
# Sensible Failure message
- "'Subset must be one of' in result.msg"
Conventions
-
Each test case should generally follow the pattern:
setup —> test —> assert —> test again (idempotent) —> assert —> -teardown (if needed) -> done
This keeps test playbooks from becoming monolithic and difficult to troubleshoot.
-
Include a name for each task that is not an assertion. (It's OK to add names to assertions too. But to make it easy to identify the broken task within a failed test, at least provide a helpful name for each task.)
-
Files containing test cases must end in
.yaml
Adding a new Network Platform
A top level playbook is required such as ansible/test/integration/eos.yaml
which needs to be references by ansible/test/integration/network-all.yaml