diff --git a/bin/ansible-connection b/bin/ansible-connection index a3e3c7cf41..e29c681815 100755 --- a/bin/ansible-connection +++ b/bin/ansible-connection @@ -45,7 +45,7 @@ from ansible.module_utils.six import PY3 from ansible.module_utils.six.moves import cPickle from ansible.module_utils.connection import send_data, recv_data from ansible.playbook.play_context import PlayContext -from ansible.plugins import connection_loader +from ansible.plugins.loader import connection_loader from ansible.utils.path import unfrackpath, makedirs_safe from ansible.errors import AnsibleConnectionFailure from ansible.utils.display import Display diff --git a/lib/ansible/cli/adhoc.py b/lib/ansible/cli/adhoc.py index 4891323946..8f1d706e96 100644 --- a/lib/ansible/cli/adhoc.py +++ b/lib/ansible/cli/adhoc.py @@ -29,7 +29,7 @@ from ansible.executor.task_queue_manager import TaskQueueManager from ansible.module_utils._text import to_text from ansible.parsing.splitter import parse_kv from ansible.playbook.play import Play -from ansible.plugins import get_all_plugin_loaders +from ansible.plugins.loader import get_all_plugin_loaders try: from __main__ import display @@ -105,6 +105,9 @@ class AdHocCLI(CLI): (sshpass, becomepass) = self.ask_passwords() passwords = {'conn_pass': sshpass, 'become_pass': becomepass} + # dynamically load any plugins + get_all_plugin_loaders() + loader, inventory, variable_manager = self._play_prereqs(self.options) no_hosts = False @@ -138,13 +141,6 @@ class AdHocCLI(CLI): if self.options.module_name in ('include', 'include_role'): raise AnsibleOptionsError("'%s' is not a valid action for ad-hoc commands" % self.options.module_name) - # dynamically load any plugins from the playbook directory - for name, obj in get_all_plugin_loaders(): - if obj.subdir: - plugin_path = os.path.join('.', obj.subdir) - if os.path.isdir(plugin_path): - obj.add_directory(plugin_path) - play_ds = self._play_ds(pattern, self.options.seconds, self.options.poll_interval) play = Play().load(play_ds, variable_manager=variable_manager, loader=loader) diff --git a/lib/ansible/cli/config.py b/lib/ansible/cli/config.py index bf36550735..89913f16ee 100644 --- a/lib/ansible/cli/config.py +++ b/lib/ansible/cli/config.py @@ -26,8 +26,7 @@ import sys import yaml from ansible.cli import CLI -from ansible.config.data import Setting -from ansible.config.manager import ConfigManager +from ansible.config.manager import ConfigManager, Setting from ansible.errors import AnsibleError, AnsibleOptionsError from ansible.module_utils._text import to_native, to_text from ansible.parsing.yaml.dumper import AnsibleDumper @@ -68,6 +67,8 @@ class ConfigCLI(CLI): if self.action == "list": self.parser.set_usage("usage: %prog list [options] ") if self.action == "dump": + self.parser.add_option('--only-changed', dest='only_changed', action='store_true', + help="Only show configurations that have changed from the default") self.parser.set_usage("usage: %prog dump [options] [-c ansible.cfg]") elif self.action == "view": self.parser.set_usage("usage: %prog view [options] [-c ansible.cfg] ") @@ -154,14 +155,15 @@ class ConfigCLI(CLI): ''' list all current configs reading lib/constants.py and shows env and config file setting names ''' - self.pager(to_text(yaml.dump(self.config.initial_defs, Dumper=AnsibleDumper), errors='surrogate_or_strict')) + self.pager(to_text(yaml.dump(self.config.get_configuration_definitions(), Dumper=AnsibleDumper), errors='surrogate_or_strict')) def execute_dump(self): ''' Shows the current settings, merges ansible.cfg if specified ''' + # FIXME: deal with plugins, not just base config text = [] - defaults = self.config.initial_defs.copy() + defaults = self.config.get_configuration_definitions().copy() for setting in self.config.data.get_settings(): if setting.name in defaults: defaults[setting.name] = setting @@ -176,6 +178,7 @@ class ConfigCLI(CLI): else: color = 'green' msg = "%s(%s) = %s" % (setting, 'default', defaults[setting].get('default')) - text.append(stringc(msg, color)) + if not self.options.only_changed or color == 'yellow': + text.append(stringc(msg, color)) self.pager(to_text('\n'.join(text), errors='surrogate_or_strict')) diff --git a/lib/ansible/cli/console.py b/lib/ansible/cli/console.py index 94d8de583e..9e9d93ee53 100644 --- a/lib/ansible/cli/console.py +++ b/lib/ansible/cli/console.py @@ -44,7 +44,7 @@ from ansible.module_utils._text import to_native, to_text from ansible.module_utils.parsing.convert_bool import boolean from ansible.parsing.splitter import parse_kv from ansible.playbook.play import Play -from ansible.plugins import module_loader +from ansible.plugins.loader import module_loader from ansible.utils import plugin_docs from ansible.utils.color import stringc diff --git a/lib/ansible/cli/doc.py b/lib/ansible/cli/doc.py index dab8677d14..664d2b7285 100644 --- a/lib/ansible/cli/doc.py +++ b/lib/ansible/cli/doc.py @@ -30,7 +30,7 @@ from ansible.cli import CLI from ansible.errors import AnsibleError, AnsibleOptionsError from ansible.module_utils.six import string_types from ansible.parsing.yaml.dumper import AnsibleDumper -from ansible.plugins import module_loader, action_loader, lookup_loader, callback_loader, cache_loader, connection_loader, strategy_loader, PluginLoader +from ansible.plugins.loader import module_loader, action_loader, lookup_loader, callback_loader, cache_loader, connection_loader, strategy_loader, PluginLoader from ansible.utils import plugin_docs try: from __main__ import display @@ -66,7 +66,8 @@ class DocCLI(CLI): self.parser.add_option("-a", "--all", action="store_true", default=False, dest='all_plugins', help='Show documentation for all plugins') self.parser.add_option("-t", "--type", action="store", default='module', dest='type', type='choice', - help='Choose which plugin type', choices=['cache', 'callback', 'connection', 'inventory', 'lookup', 'module', 'strategy']) + help='Choose which plugin type (defaults to "module")', + choices=['cache', 'callback', 'connection', 'inventory', 'lookup', 'module', 'strategy']) super(DocCLI, self).parse() @@ -99,6 +100,10 @@ class DocCLI(CLI): for i in self.options.module_path.split(os.pathsep): loader.add_directory(i) + # save only top level paths for errors + search_paths = DocCLI.print_paths(loader) + loader._paths = None # reset so we can use subdirs below + # list plugins for type if self.options.list_dir: paths = loader._get_paths() @@ -125,7 +130,7 @@ class DocCLI(CLI): # if the plugin lives in a non-python file (eg, win_X.ps1), require the corresponding python file for docs filename = loader.find_plugin(plugin, mod_type='.py', ignore_deprecated=True) if filename is None: - display.warning("%s %s not found in %s\n" % (plugin_type, plugin, DocCLI.print_paths(loader))) + display.warning("%s %s not found in:\n%s\n" % (plugin_type, plugin, search_paths)) continue if any(filename.endswith(x) for x in C.BLACKLIST_EXTS): @@ -255,7 +260,7 @@ class DocCLI(CLI): # Uses a list to get the order right ret = [] - for i in finder._get_paths(): + for i in finder._get_paths(subdirs=False): if i not in ret: ret.append(i) return os.pathsep.join(ret) @@ -288,6 +293,9 @@ class DocCLI(CLI): return "\n".join(text) + def _dump_yaml(self, struct, indent): + return CLI.tty_ify('\n'.join([indent + line for line in yaml.dump(struct, default_flow_style=False, Dumper=AnsibleDumper).split('\n')])) + def add_fields(self, text, fields, limit, opt_indent): for o in sorted(fields): @@ -322,123 +330,109 @@ class DocCLI(CLI): del opt['choices'] default = '' if 'default' in opt or not required: - default = "[Default: " + str(opt.pop('default', '(null)')) + "]" + default = "[Default: %s" % str(opt.pop('default', '(null)')) + "]" + text.append(textwrap.fill(CLI.tty_ify(aliases + choices + default), limit, initial_indent=opt_indent, subsequent_indent=opt_indent)) if 'options' in opt: - text.append(opt_indent + "options:\n") - self.add_fields(text, opt['options'], limit, opt_indent + opt_indent) - text.append('') - del opt['options'] + text.append("%soptions:\n" % opt_indent) + self.add_fields(text, opt.pop('options'), limit, opt_indent + opt_indent) if 'spec' in opt: - text.append(opt_indent + "spec:\n") - self.add_fields(text, opt['spec'], limit, opt_indent + opt_indent) - text.append('') - del opt['spec'] + text.append("%sspec:\n" % opt_indent) + self.add_fields(text, opt.pop('spec'), limit, opt_indent + opt_indent) - for conf in ('config', 'env_vars', 'host_vars'): - if conf in opt: - text.append(textwrap.fill(CLI.tty_ify("%s: " % conf), limit, initial_indent=opt_indent, subsequent_indent=opt_indent)) - for entry in opt[conf]: - if isinstance(entry, dict): - pre = " -" - for key in entry: - text.append(textwrap.fill(CLI.tty_ify("%s %s: %s" % (pre, key, entry[key])), - limit, initial_indent=opt_indent, subsequent_indent=opt_indent)) - pre = " " - else: - text.append(textwrap.fill(CLI.tty_ify(" - %s" % entry), limit, initial_indent=opt_indent, subsequent_indent=opt_indent)) - del opt[conf] + conf = {} + for config in ('env', 'ini', 'yaml', 'vars'): + if config in opt and opt[config]: + conf[config] = opt.pop(config) - # unspecified keys - for k in opt: + if conf: + text.append(self._dump_yaml({'set_via': conf}, opt_indent)) + + for k in sorted(opt): if k.startswith('_'): continue if isinstance(opt[k], string_types): - text.append(textwrap.fill(CLI.tty_ify("%s: %s" % (k, opt[k])), limit, initial_indent=opt_indent, subsequent_indent=opt_indent)) - elif isinstance(opt[k], (list, dict)): - text.append(textwrap.fill(CLI.tty_ify("%s: %s" % (k, yaml.dump(opt[k], Dumper=AnsibleDumper, default_flow_style=False))), - limit, initial_indent=opt_indent, subsequent_indent=opt_indent)) + text.append('%s%s: %s' % (opt_indent, k, textwrap.fill(CLI.tty_ify(opt[k]), limit - (len(k) + 2), subsequent_indent=opt_indent))) + elif isinstance(opt[k], (list, tuple)): + text.append(CLI.tty_ify('%s%s: %s' % (opt_indent, k, ', '.join(opt[k])))) else: - display.vv("Skipping %s key cuase we don't know how to handle eet" % k) + text.append(self._dump_yaml({k: opt[k]}, opt_indent)) + text.append('') def get_man_text(self, doc): + + IGNORE = frozenset(['module', 'docuri', 'version_added', 'short_description', 'now_date']) opt_indent = " " text = [] - text.append("> %s (%s)\n" % (doc[self.options.type].upper(), doc['filename'])) + + text.append("> %s (%s)\n" % (doc[self.options.type].upper(), doc.pop('filename'))) pad = display.columns * 0.20 limit = max(display.columns - int(pad), 70) if isinstance(doc['description'], list): - desc = " ".join(doc['description']) + desc = " ".join(doc.pop('description')) else: - desc = doc['description'] + desc = doc.pop('description') - text.append("%s\n" % textwrap.fill(CLI.tty_ify(desc), limit, initial_indent=" ", subsequent_indent=" ")) + text.append("%s\n" % textwrap.fill(CLI.tty_ify(desc), limit, initial_indent=opt_indent, subsequent_indent=opt_indent)) if 'deprecated' in doc and doc['deprecated'] is not None and len(doc['deprecated']) > 0: - text.append("DEPRECATED: \n%s\n" % doc['deprecated']) + text.append("DEPRECATED: \n%s\n" % doc.pop('deprecated')) - if 'action' in doc and doc['action']: + if doc.pop('action', False): text.append(" * note: %s\n" % "This module has a corresponding action plugin.") if 'options' in doc and doc['options']: - text.append("Options (= is mandatory):\n") - self.add_fields(text, doc['options'], limit, opt_indent) + text.append("OPTIONS (= is mandatory):\n") + self.add_fields(text, doc.pop('options'), limit, opt_indent) text.append('') if 'notes' in doc and doc['notes'] and len(doc['notes']) > 0: - text.append("Notes:") + text.append("NOTES:") for note in doc['notes']: - text.append(textwrap.fill(CLI.tty_ify(note), limit - 6, initial_indent=" * ", subsequent_indent=opt_indent)) + text.append(textwrap.fill(CLI.tty_ify(note), limit - 6, initial_indent=opt_indent[:-2] + "* ", subsequent_indent=opt_indent)) + text.append('') + del doc['notes'] if 'requirements' in doc and doc['requirements'] is not None and len(doc['requirements']) > 0: - req = ", ".join(doc['requirements']) - text.append("Requirements:%s\n" % textwrap.fill(CLI.tty_ify(req), limit - 16, initial_indent=" ", subsequent_indent=opt_indent)) - - if 'examples' in doc and len(doc['examples']) > 0: - text.append("Example%s:\n" % ('' if len(doc['examples']) < 2 else 's')) - for ex in doc['examples']: - text.append("%s\n" % (ex['code'])) + req = ", ".join(doc.pop('requirements')) + text.append("REQUIREMENTS:%s\n" % textwrap.fill(CLI.tty_ify(req), limit - 16, initial_indent=" ", subsequent_indent=opt_indent)) if 'plainexamples' in doc and doc['plainexamples'] is not None: - text.append("EXAMPLES:\n") + text.append("EXAMPLES:") if isinstance(doc['plainexamples'], string_types): - text.append(doc['plainexamples']) + text.append(doc.pop('plainexamples').strip()) else: - text.append(yaml.dump(doc['plainexamples'], indent=2, default_flow_style=False)) + text.append(yaml.dump(doc.pop('plainexamples'), indent=2, default_flow_style=False)) + text.append('') if 'returndocs' in doc and doc['returndocs'] is not None: text.append("RETURN VALUES:\n") if isinstance(doc['returndocs'], string_types): - text.append(doc['returndocs']) + text.append(doc.pop('returndocs')) else: - text.append(yaml.dump(doc['returndocs'], indent=2, default_flow_style=False)) + text.append(yaml.dump(doc.pop('returndocs'), indent=2, default_flow_style=False)) text.append('') - maintainers = set() - if 'author' in doc: - if isinstance(doc['author'], string_types): - maintainers.add(doc['author']) + # Control rest of keys on verbosity (3 == full, 0 only adds small list) + rest = [] + if self.options.verbosity >= 3: + rest = doc + elif 'author' in doc: + rest = ['author'] + + # Generic handler + for k in sorted(rest): + if k in IGNORE or not doc[k]: + continue + if isinstance(doc[k], string_types): + text.append('%s: %s' % (k.upper(), textwrap.fill(CLI.tty_ify(doc[k]), limit - (len(k) + 2), subsequent_indent=opt_indent))) + elif isinstance(doc[k], (list, tuple)): + text.append('%s: %s' % (k.upper(), ', '.join(doc[k]))) else: - maintainers.update(doc['author']) - - if 'maintainers' in doc: - if isinstance(doc['maintainers'], string_types): - maintainers.add(doc['author']) - else: - maintainers.update(doc['author']) - - text.append('MAINTAINERS: ' + ', '.join(maintainers)) - text.append('') - - if 'metadata' in doc and doc['metadata']: - text.append("METADATA:") - for k in doc['metadata']: - if isinstance(k, list): - text.append("\t%s: %s" % (k.capitalize(), ", ".join(doc['metadata'][k]))) - else: - text.append("\t%s: %s" % (k.capitalize(), doc['metadata'][k])) + text.append(self._dump_yaml({k.upper(): doc[k]}, opt_indent)) text.append('') + return "\n".join(text) diff --git a/lib/ansible/cli/pull.py b/lib/ansible/cli/pull.py index 94103edaa0..38130fab7f 100644 --- a/lib/ansible/cli/pull.py +++ b/lib/ansible/cli/pull.py @@ -31,7 +31,7 @@ import time from ansible.cli import CLI from ansible.errors import AnsibleOptionsError from ansible.module_utils._text import to_native -from ansible.plugins import module_loader +from ansible.plugins.loader import module_loader from ansible.utils.cmd_functions import run_cmd try: diff --git a/lib/ansible/config/data/config.yml b/lib/ansible/config/base.yml similarity index 65% rename from lib/ansible/config/data/config.yml rename to lib/ansible/config/base.yml index 82f9c478a0..89282c799f 100644 --- a/lib/ansible/config/data/config.yml +++ b/lib/ansible/config/base.yml @@ -1,6 +1,9 @@ +# Copyright (c) 2017 Ansible Project +# GNU General Public License v3.0+ (see COPYING or https://www.gnu.org/licenses/gpl-3.0.txt) +--- ACCELERATE_CONNECT_TIMEOUT: default: 1.0 - desc: + description: - This setting controls the timeout for the socket connect call, and should be kept relatively low. The connection to the accelerate_port will be attempted 3 times before Ansible will fall back to ssh or paramiko (depending on your default connection setting) to try and start the accelerate daemon remotely. @@ -9,152 +12,167 @@ ACCELERATE_CONNECT_TIMEOUT: env: [{name: ACCELERATE_CONNECT_TIMEOUT }] ini: - {key: accelerate_connect_timeout, section: accelerate} - value_type: float - deprecated: True - vars: [] - yaml: {key: accelerate.accelerate_connect_timeout} + type: float + deprecated: + why: Removing accelerate as a connection method, settings not needed either. + version: "2.5" + alternatives: ssh and paramiko version_added: "1.4" ACCELERATE_DAEMON_TIMEOUT: default: 30 - desc: + description: - This setting controls the timeout for the accelerated daemon, as measured in minutes. The default daemon timeout is 30 minutes. - Prior to 1.6, the timeout was hard-coded from the time of the daemon’s launch. - For version 1.6+, the timeout is now based on the last activity to the daemon and is configurable via this option. env: [{name: ACCELERATE_DAEMON_TIMEOUT}] ini: - {key: accelerate_daemon_timeout, section: accelerate} - value_type: integer - deprecated: True - vars: [] - yaml: {key: accelerate.accelerate_daemon_timeout} + type: integer + deprecated: + why: Removing accelerate as a connection method, settings not needed either. + version: "2.5" + alternatives: ssh and paramiko version_added: "1.6" ACCELERATE_KEYS_DIR: default: ~/.fireball.keys - desc: '' - deprecated: True + description: '' + deprecated: + why: Removing accelerate as a connection method, settings not needed either. + version: "2.5" + alternatives: ssh and paramiko env: [{name: ACCELERATE_KEYS_DIR}] ini: - {key: accelerate_keys_dir, section: accelerate} - vars: [] - yaml: {key: accelerate.accelerate_keys_dir} ACCELERATE_KEYS_DIR_PERMS: default: '700' - desc: 'TODO: write it' + description: 'TODO: write it' env: [{name: ACCELERATE_KEYS_DIR_PERMS}] ini: - {key: accelerate_keys_dir_perms, section: accelerate} - deprecated: True - vars: [] - yaml: {key: accelerate.accelerate_keys_dir_perms} + deprecated: + why: Removing accelerate as a connection method, settings not needed either. + version: "2.5" + alternatives: ssh and paramiko ACCELERATE_KEYS_FILE_PERMS: default: '600' - desc: 'TODO: write it' + description: 'TODO: write it' env: [{name: ACCELERATE_KEYS_FILE_PERMS}] ini: - {key: accelerate_keys_file_perms, section: accelerate} - deprecated: True - vars: [] - yaml: {key: accelerate.accelerate_keys_file_perms} + deprecated: + why: Removing accelerate as a connection method, settings not needed either. + version: "2.5" + alternatives: ssh and paramiko ACCELERATE_MULTI_KEY: default: False - desc: 'TODO: write it' + description: 'TODO: write it' env: [{name: ACCELERATE_MULTI_KEY}] ini: - {key: accelerate_multi_key, section: accelerate} - value_type: boolean - deprecated: True - vars: [] - yaml: {key: accelerate.accelerate_multi_key} + type: boolean + deprecated: + why: Removing accelerate as a connection method, settings not needed either. + version: "2.5" + alternatives: ssh and paramiko ACCELERATE_PORT: default: 5099 - desc: 'TODO: write it' + description: 'TODO: write it' env: [{name: ACCELERATE_PORT}] ini: - {key: accelerate_port, section: accelerate} - value_type: integer - deprecated: True - vars: [] - yaml: {key: accelerate.accelerate_port} + type: integer + deprecated: + why: Removing accelerate as a connection method, settings not needed either. + version: "2.5" + alternatives: ssh and paramiko ACCELERATE_TIMEOUT: default: 30 - desc: 'TODO: write it' + description: 'TODO: write it' env: [{name: ACCELERATE_TIMEOUT}] ini: - {key: accelerate_timeout, section: accelerate} - value_type: integer - deprecated: True - vars: [] - yaml: {key: accelerate.accelerate_timeout} + type: integer + deprecated: + why: Removing accelerate as a connection method, settings not needed either. + version: "2.5" + alternatives: ssh and paramiko ALLOW_WORLD_READABLE_TMPFILES: default: False - desc: + description: - This makes the temporary files created on the machine to be world readable and will issue a warning instead of failing the task. - It is useful when becoming an unprivileged user. env: [] ini: - {key: allow_world_readable_tmpfiles, section: defaults} - value_type: boolean - vars: [] + type: boolean yaml: {key: defaults.allow_world_readable_tmpfiles} version_added: "2.1" ANSIBLE_COW_SELECTION: default: default - desc: This allows you to chose a specific cowsay stencil for the banners or use 'random' to cycle through them. + description: This allows you to chose a specific cowsay stencil for the banners or use 'random' to cycle through them. env: [{name: ANSIBLE_COW_SELECTION}] ini: - {key: cow_selection, section: defaults} - vars: [] yaml: {key: defaults.cow_selection} ANSIBLE_COW_WHITELIST: default: ['bud-frogs', 'bunny', 'cheese', 'daemon', 'default', 'dragon', 'elephant-in-snake', 'elephant', 'eyes', 'hellokitty', 'kitty', 'luke-koala', 'meow', 'milk', 'moofasa', 'moose', 'ren', 'sheep', 'small', 'stegosaurus', 'stimpy', 'supermilker', 'three-eyes', 'turkey', 'turtle', 'tux', 'udder', 'vader-koala', 'vader', 'www'] - desc: White list of cowsay templates that are 'safe' to use, set to empty list if you want to enable all installed templates. + description: White list of cowsay templates that are 'safe' to use, set to empty list if you want to enable all installed templates. env: [{name: ANSIBLE_COW_WHITELIST}] ini: - {key: cow_whitelist, section: defaults} - value_type: list - vars: [] + type: list yaml: {key: defaults.cow_whitelist} ANSIBLE_FORCE_COLOR: default: False - desc: This options forces color mode even when running without a TTY + description: This options forces color mode even when running without a TTY env: [{name: ANSIBLE_FORCE_COLOR}] ini: - {key: force_color, section: defaults} - value_type: boolean - vars: [] + type: boolean yaml: {key: defaults.force_color} ANSIBLE_NOCOLOR: default: False - desc: This setting allows suppressing colorizing output, which is used to give a better indication of failure and status information. + description: This setting allows suppressing colorizing output, which is used to give a better indication of failure and status information. env: [{name: ANSIBLE_NOCOLOR}] ini: - {key: nocolor, section: defaults} - value_type: boolean - vars: [] + type: boolean yaml: {key: defaults.nocolor} ANSIBLE_NOCOWS: default: False - desc: If you have cowsay installed but want to avoid the 'cows' (why????), use this. + description: If you have cowsay installed but want to avoid the 'cows' (why????), use this. env: [{name: ANSIBLE_NOCOWS}] ini: - {key: nocows, section: defaults} - value_type: boolean - vars: [] + type: boolean yaml: {key: defaults.nocows} +ANSIBLE_PIPELINING: + default: False + description: + - Pipelining, if supported by the connection plugin, reduces the number of network operations required to execute a module on the remote server, + by executing many Ansible modules without actual file transfer. + - This can result in a very significant performance improvement when enabled. + - However this conflicts with privilege escalation (become). For example, when using “sudo:” operations you must first + disable ‘requiretty’ in /etc/sudoers on all managed hosts, which is why it is disabled by default. + env: [{name: ANSIBLE_PIPELINING}] + ini: + - {key: pipelining, section: connection} + type: boolean + yaml: {key: plugins.connection.pipelining} ANSIBLE_SSH_ARGS: default: -C -o ControlMaster=auto -o ControlPersist=60s - desc: + description: - If set, this will override the Ansible default ssh arguments. - In particular, users may wish to raise the ControlPersist time to encourage performance. A value of 30 minutes may be appropriate. - Be aware that if `-o ControlPath` is set in ssh_args, the control path setting is not used. env: [{name: ANSIBLE_SSH_ARGS}] ini: - {key: ssh_args, section: ssh_connection} - vars: [] yaml: {key: ssh_connection.ssh_args} ANSIBLE_SSH_CONTROL_PATH: + # TODO: move to ssh plugin default: null - desc: + description: - This is the location to save ssh's ControlPath sockets, it uses ssh's variable substitution. - Since 2.3, if null, ansible will generate a unique hash. Use `%(directory)s` to indicate where to use the control dir path setting. - Before 2.3 it defaulted to `control_path=%(directory)s/ansible-ssh-%%h-%%p-%%r`. @@ -162,33 +180,31 @@ ANSIBLE_SSH_CONTROL_PATH: env: [{name: ANSIBLE_SSH_CONTROL_PATH}] ini: - {key: control_path, section: ssh_connection} - vars: [] yaml: {key: ssh_connection.control_path} ANSIBLE_SSH_CONTROL_PATH_DIR: + # TODO: move to ssh plugin default: ~/.ansible/cp - desc: + description: - This sets the directory to use for ssh control path if the control path setting is null. - Also, provides the `%(directory)s` variable for the control path setting. env: [{name: ANSIBLE_SSH_CONTROL_PATH_DIR}] ini: - {key: control_path_dir, section: ssh_connection} - vars: [] yaml: {key: ssh_connection.control_path_dir} ANSIBLE_SSH_EXECUTABLE: default: ssh - desc: + description: - This defines the location of the ssh binary. It defaults to `ssh` which will use the first ssh binary available in $PATH. - This option is usually not required, it might be useful when access to system ssh is restricted, or when using ssh wrappers to connect to remote hosts. env: [{name: ANSIBLE_SSH_EXECUTABLE}] ini: - {key: ssh_executable, section: ssh_connection} - vars: [] yaml: {key: ssh_connection.ssh_executable} version_added: "2.2" ANSIBLE_SSH_PIPELINING: default: False - desc: + description: - Pipelining reduces the number of SSH operations required to execute a module on the remote server, by executing many Ansible modules without actual file transfer. - This can result in a very significant performance improvement when enabled. @@ -196,654 +212,592 @@ ANSIBLE_SSH_PIPELINING: env: [{name: ANSIBLE_SSH_PIPELINING}] ini: - {key: pipelining, section: ssh_connection} - value_type: boolean - vars: [] + type: boolean yaml: {key: ssh_connection.pipelining} ANSIBLE_SSH_RETRIES: default: 0 - desc: 'TODO: write it' + description: 'TODO: write it' env: [{name: ANSIBLE_SSH_RETRIES}] ini: - {key: retries, section: ssh_connection} - value_type: integer - vars: [] + type: integer yaml: {key: ssh_connection.retries} ANY_ERRORS_FATAL: default: False - desc: Sets the default value for the any_errors_fatal keyword + description: Sets the default value for the any_errors_fatal keyword env: - name: ANSIBLE_ANY_ERRORS_FATAL ini: - section: defaults key: any_errors_fatal - value_type: boolean - vars: [] + type: boolean yaml: {key: errors.anyerrors_fatal} version_added: "2.4" BECOME_ALLOW_SAME_USER: default: False - desc: This setting controls if become is skipped when remote user and become user are the same. + description: This setting controls if become is skipped when remote user and become user are the same. env: [{name: ANSIBLE_BECOME_ALLOW_SAME_USER}] ini: - {key: become_allow_same_user, section: privilege_escalation} - value_type: boolean - vars: [] + type: boolean yaml: {key: privilege_escalation.become_allow_same_user} CACHE_PLUGIN: default: memory - desc: Chooses which cache plugin to use + description: Chooses which cache plugin to use env: [{name: ANSIBLE_CACHE_PLUGIN}] ini: - {key: fact_caching, section: defaults} - vars: [] yaml: {key: defaults.fact_caching} CACHE_PLUGIN_CONNECTION: default: - desc: Defines connection or path information for the cache plugin + description: Defines connection or path information for the cache plugin env: [{name: ANSIBLE_CACHE_PLUGIN_CONNECTION}] ini: - {key: fact_caching_connection, section: defaults} - vars: [] yaml: {key: defaults.fact_caching_connection} CACHE_PLUGIN_PREFIX: default: ansible_facts - desc: Prefix to use for cache plugin files/tables + description: Prefix to use for cache plugin files/tables env: [{name: ANSIBLE_CACHE_PLUGIN_PREFIX}] ini: - {key: fact_caching_prefix, section: defaults} - vars: [] yaml: {key: defaults.fact_caching_prefix} CACHE_PLUGIN_TIMEOUT: default: 86400 - desc: Expiration timeout for the cache plugin data + description: Expiration timeout for the cache plugin data env: [{name: ANSIBLE_CACHE_PLUGIN_TIMEOUT}] ini: - {key: fact_caching_timeout, section: defaults} - value_type: integer - vars: [] + type: integer yaml: {key: defaults.fact_caching_timeout} COLOR_CHANGED: default: yellow - desc: Defines the color to use on 'Changed' status + description: Defines the color to use on 'Changed' status env: [{name: ANSIBLE_COLOR_CHANGED}] ini: - {key: changed, section: colors} - vars: [] yaml: {key: display.colors.changed} COLOR_DEBUG: default: dark gray - desc: Defines the color to use when emitting debug messages + description: Defines the color to use when emitting debug messages env: [{name: ANSIBLE_COLOR_DEBUG}] ini: - {key: debug, section: colors} - vars: [] yaml: {key: colors.debug} COLOR_DEPRECATE: default: purple - desc: Defines the color to use when emitting deprecation messages + description: Defines the color to use when emitting deprecation messages env: [{name: ANSIBLE_COLOR_DEPRECATE}] ini: - {key: deprecate, section: colors} - vars: [] yaml: {key: colors.deprecate} COLOR_DIFF_ADD: default: green - desc: Defines the color to use when showing added lines in diffs + description: Defines the color to use when showing added lines in diffs env: [{name: ANSIBLE_COLOR_DIFF_ADD}] ini: - {key: diff_add, section: colors} - vars: [] yaml: {key: colors.diff_add} COLOR_DIFF_LINES: default: cyan - desc: Defines the color to use when showing diffs + description: Defines the color to use when showing diffs env: [{name: ANSIBLE_COLOR_DIFF_LINES}] ini: - {key: diff_lines, section: colors} - vars: [] yaml: {key: colors.diff_lines} COLOR_DIFF_REMOVE: default: red - desc: Defines the color to use when showing removed lines in diffs + description: Defines the color to use when showing removed lines in diffs env: [{name: ANSIBLE_COLOR_DIFF_REMOVE}] ini: - {key: diff_remove, section: colors} - vars: [] yaml: {key: colors.diff_remove} COLOR_ERROR: default: red - desc: Defines the color to use when emitting error messages + description: Defines the color to use when emitting error messages env: [{name: ANSIBLE_COLOR_ERROR}] ini: - {key: error, section: colors} - vars: [] yaml: {key: colors.error} COLOR_HIGHLIGHT: default: white - desc: 'TODO: write it' + description: 'TODO: write it' env: [{name: ANSIBLE_COLOR_HIGHLIGHT}] ini: - {key: highlight, section: colors} - vars: [] yaml: {key: colors.highlight} COLOR_OK: default: green - desc: Defines the color to use when showing 'OK' status + description: Defines the color to use when showing 'OK' status env: [{name: ANSIBLE_COLOR_OK}] ini: - {key: ok, section: colors} - vars: [] yaml: {key: colors.ok} COLOR_SKIP: default: cyan - desc: Defines the color to use when showing 'Skipped' status + description: Defines the color to use when showing 'Skipped' status env: [{name: ANSIBLE_COLOR_SKIP}] ini: - {key: skip, section: colors} - vars: [] yaml: {key: colors.skip} COLOR_UNREACHABLE: default: bright red - desc: Defines the color to use on 'Unreachable' status + description: Defines the color to use on 'Unreachable' status env: [{name: ANSIBLE_COLOR_UNREACHABLE}] ini: - {key: unreachable, section: colors} - vars: [] yaml: {key: colors.unreachable} COLOR_VERBOSE: default: blue - desc: Defines the color to use when emitting verbose messages + description: Defines the color to use when emitting verbose messages env: [{name: ANSIBLE_COLOR_VERBOSE}] ini: - {key: verbose, section: colors} - vars: [] yaml: {key: colors.verbose} COLOR_WARN: default: bright purple - desc: Defines the color to use when emitting warning messages + description: Defines the color to use when emitting warning messages env: [{name: ANSIBLE_COLOR_WARN}] ini: - {key: warn, section: colors} - vars: [] yaml: {key: colors.warn} COMMAND_WARNINGS: default: True - desc: 'TODO: write it' + description: 'TODO: write it' env: [{name: ANSIBLE_COMMAND_WARNINGS}] ini: - {key: command_warnings, section: defaults} - value_type: boolean - vars: [] + type: boolean yaml: {key: defaults.command_warnings} DEFAULT_ACTION_PLUGIN_PATH: default: ~/.ansible/plugins/action:/usr/share/ansible/plugins/action - desc: 'TODO: write it' + description: Colon separated paths in which Ansible will search for Action Plugins. env: [{name: ANSIBLE_ACTION_PLUGINS}] ini: - {key: action_plugins, section: defaults} - value_type: pathlist - vars: [] + type: pathlist yaml: {key: defaults.action_plugins} DEFAULT_ALLOW_UNSAFE_LOOKUPS: default: False - desc: 'TODO: write it' + description: 'TODO: write it' env: [] ini: - {key: allow_unsafe_lookups, section: defaults} - value_type: boolean - vars: [] + type: boolean yaml: {key: defaults.allow_unsafe_lookups} DEFAULT_ASK_PASS: default: False - desc: 'TODO: write it' + description: 'TODO: write it' env: [{name: ANSIBLE_ASK_PASS}] ini: - {key: ask_pass, section: defaults} - value_type: boolean - vars: [] + type: boolean yaml: {key: defaults.ask_pass} DEFAULT_ASK_SUDO_PASS: default: False - desc: 'TODO: write it' + deprecated: + why: In favor of become which is a generic framework + version: "2.8" + alternatives: become + description: 'TODO: write it' env: [{name: ANSIBLE_ASK_SUDO_PASS}] ini: - {key: ask_sudo_pass, section: defaults} - value_type: boolean - vars: [] - yaml: {key: defaults.ask_sudo_pass} + type: boolean DEFAULT_ASK_SU_PASS: default: False - desc: 'TODO: write it' + deprecated: + why: In favor of become which is a generic framework + version: "2.8" + alternatives: become + description: 'TODO: write it' env: [{name: ANSIBLE_ASK_SU_PASS}] ini: - {key: ask_su_pass, section: defaults} - value_type: boolean - vars: [] - yaml: {key: defaults.ask_su_pass} + type: boolean DEFAULT_ASK_VAULT_PASS: default: False - desc: 'TODO: write it' + description: 'TODO: write it' env: [{name: ANSIBLE_ASK_VAULT_PASS}] ini: - {key: ask_vault_pass, section: defaults} - value_type: boolean - vars: [] + type: boolean yaml: {key: defaults.ask_vault_pass} DEFAULT_BECOME: default: False - desc: 'TODO: write it' + description: Toggles the use of privilege escalation, allowing you to 'become' another user after login. env: [{name: ANSIBLE_BECOME}] ini: - {key: become, section: privilege_escalation} - value_type: boolean - vars: [] + type: boolean yaml: {key: privilege_escalation.become} DEFAULT_BECOME_ASK_PASS: default: False - desc: 'TODO: write it' + description: Toggle to prompt for privilege escalation password. env: [{name: ANSIBLE_BECOME_ASK_PASS}] ini: - {key: become_ask_pass, section: privilege_escalation} - value_type: boolean - vars: [] + type: boolean yaml: {key: privilege_escalation.become_ask_pass} -DEFAULT_BECOME_EXE: - default: - desc: 'TODO: write it' - env: [{name: ANSIBLE_BECOME_EXE}] - ini: - - {key: become_exe, section: privilege_escalation} - vars: [] - yaml: {key: privilege_escalation.become_exe} -DEFAULT_BECOME_FLAGS: - default: - desc: 'TODO: write it' - env: [{name: ANSIBLE_BECOME_FLAGS}] - ini: - - {key: become_flags, section: privilege_escalation} - vars: [] - yaml: {key: privilege_escalation.become_flags} DEFAULT_BECOME_METHOD: default: 'sudo' - desc: + description: Privilege escalation method to use when `become` is enabled. env: [{name: ANSIBLE_BECOME_METHOD}] ini: - {section: privilege_escalation, key: become_method} - vars: [] yaml: {key: privilege_escalation.become_method} +DEFAULT_BECOME_EXE: + default: ~ + description: 'executable to use for privilege escalation, otherwise Ansible will depend on PATh' + env: [{name: ANSIBLE_BECOME_EXE}] + ini: + - {key: become_exe, section: privilege_escalation} + yaml: {key: privilege_escalation.become_exe} +DEFAULT_BECOME_FLAGS: + default: ~ + description: Flags to pass to the privilege escalation executable. + env: [{name: ANSIBLE_BECOME_FLAGS}] + ini: + - {key: become_flags, section: privilege_escalation} + yaml: {key: privilege_escalation.become_flags} DEFAULT_BECOME_USER: default: root - desc: 'TODO: write it' + description: User your become when using privilege escalation, most systems will use 'root' when no user is specified. env: [{name: ANSIBLE_BECOME_USER}] ini: - {key: become_user, section: privilege_escalation} - vars: [] yaml: {key: privilege_escalation.become_user} DEFAULT_CACHE_PLUGIN_PATH: default: ~/.ansible/plugins/cache:/usr/share/ansible/plugins/cache - desc: 'TODO: write it' + description: 'TODO: write it' env: [{name: ANSIBLE_CACHE_PLUGINS}] ini: - {key: cache_plugins, section: defaults} - value_type: pathlist - vars: [] + type: pathlist yaml: {key: defaults.cache_plugins} DEFAULT_CALLABLE_WHITELIST: default: [] - desc: 'TODO: write it' + description: 'TODO: write it' env: [{name: ANSIBLE_CALLABLE_WHITELIST}] ini: - {key: callable_whitelist, section: defaults} - value_type: list - vars: [] + type: list yaml: {key: defaults.callable_whitelist} DEFAULT_CALLBACK_PLUGIN_PATH: default: ~/.ansible/plugins/callback:/usr/share/ansible/plugins/callback - desc: 'TODO: write it' + description: 'TODO: write it' env: [{name: ANSIBLE_CALLBACK_PLUGINS}] ini: - {key: callback_plugins, section: defaults} - value_type: pathlist - vars: [] + type: pathlist yaml: {key: defaults.callback_plugins} DEFAULT_CALLBACK_WHITELIST: default: [] - desc: 'TODO: write it' + description: 'TODO: write it' env: [{name: ANSIBLE_CALLBACK_WHITELIST}] ini: - {key: callback_whitelist, section: defaults} - value_type: list - vars: [] + type: list yaml: {key: defaults.callback_whitelist} DEFAULT_CONNECTION_PLUGIN_PATH: default: ~/.ansible/plugins/connection:/usr/share/ansible/plugins/connection - desc: 'TODO: write it' + description: 'TODO: write it' env: [{name: ANSIBLE_CONNECTION_PLUGINS}] ini: - {key: connection_plugins, section: defaults} - value_type: pathlist - vars: [] + type: pathlist yaml: {key: defaults.connection_plugins} DEFAULT_DEBUG: default: False - desc: 'TODO: write it' + description: Toggles debug output in Ansible, VERY verbose and can hinder multiprocessing. env: [{name: ANSIBLE_DEBUG}] ini: - {key: debug, section: defaults} - value_type: boolean - vars: [] + type: boolean yaml: {key: defaults.debug} DEFAULT_EXECUTABLE: default: /bin/sh - desc: 'TODO: write it' + description: 'TODO: write it' env: [{name: ANSIBLE_EXECUTABLE}] ini: - {key: executable, section: defaults} - vars: [] yaml: {key: defaults.executable} DEFAULT_FACT_PATH: default: - desc: 'TODO: write it' + description: 'TODO: write it' env: [{name: ANSIBLE_FACT_PATH}] ini: - {key: fact_path, section: defaults} - value_type: path - vars: [] + type: path yaml: {key: defaults.fact_path} DEFAULT_FILTER_PLUGIN_PATH: default: ~/.ansible/plugins/filter:/usr/share/ansible/plugins/filter - desc: 'TODO: write it' + description: 'TODO: write it' env: [{name: ANSIBLE_FILTER_PLUGINS}] ini: - {key: filter_plugins, section: defaults} - value_type: pathlist - vars: [] + type: pathlist yaml: {key: defaults.filter_plugins} DEFAULT_FORCE_HANDLERS: default: False - desc: 'TODO: write it' + description: 'TODO: write it' env: [{name: ANSIBLE_FORCE_HANDLERS}] ini: - {key: force_handlers, section: defaults} - value_type: boolean - vars: [] + type: boolean yaml: {key: defaults.force_handlers} DEFAULT_FORKS: default: 5 - desc: 'TODO: write it' + description: Maximum number of forks Ansible will use to execute tasks on target hosts. env: [{name: ANSIBLE_FORKS}] ini: - {key: forks, section: defaults} - value_type: integer - vars: [] + type: integer yaml: {key: defaults.forks} DEFAULT_GATHERING: default: 'implicit' - desc: 'TODO: write it' + description: 'TODO: write it' env: [{name: ANSIBLE_GATHERING}] ini: - key: gathering section: defaults - vars: [] yaml: {key: defaults.gathering} DEFAULT_GATHER_SUBSET: default: 'all' - desc: 'TODO: write it' + description: 'TODO: write it' env: [{name: ANSIBLE_GATHER_SUBSET}] ini: - key: gather_subset section: defaults - vars: [] yaml: {key: defaults.gather_subset} DEFAULT_GATHER_TIMEOUT: default: 10 - desc: 'TODO: write it' + description: 'TODO: write it' env: [{name: ANSIBLE_GATHER_TIMEOUT}] ini: - {key: gather_timeout, section: defaults} - value_type: integer - vars: [] + type: integer yaml: {key: defaults.gather_timeout} DEFAULT_HANDLER_INCLUDES_STATIC: default: False - desc: 'TODO: write it' + description: 'TODO: write it' env: [{name: ANSIBLE_HANDLER_INCLUDES_STATIC}] ini: - {key: handler_includes_static, section: defaults} - value_type: boolean - vars: [] + type: boolean yaml: {key: defaults.handler_includes_static} + deprecated: + why: include itself is deprecated and this setting will not matter in the future + version: "2.8" + alternatives: none as its already built into the decision between include_tasks and import_tasks DEFAULT_HASH_BEHAVIOUR: default: replace - desc: 'TODO: write it' + description: 'TODO: write it' env: [{name: ANSIBLE_HASH_BEHAVIOUR}] ini: - {key: hash_behaviour, section: defaults} - vars: [] yaml: {key: defaults.hash_behaviour} DEFAULT_HOST_LIST: default: /etc/ansible/hosts - desc: 'TODO: write it' + description: Location of the Ansible inventory source. env: [{name: ANSIBLE_INVENTORY}] expand_relative_paths: True ini: - {key: inventory, section: defaults} - value_type: path - vars: [] + type: path yaml: {key: defaults.inventory} DEFAULT_INTERNAL_POLL_INTERVAL: default: 0.001 - desc: 'TODO: write it' + description: 'TODO: write it' env: [] ini: - {key: internal_poll_interval, section: defaults} - value_type: float - vars: [] + type: float yaml: {key: defaults.internal_poll_interval} DEFAULT_INVENTORY_PLUGIN_PATH: default: ~/.ansible/plugins/inventory:/usr/share/ansible/plugins/inventory - desc: 'TODO: write it' + description: 'TODO: write it' env: [{name: ANSIBLE_INVENTORY_PLUGINS}] ini: - {key: inventory_plugins, section: defaults} - value_type: pathlist - vars: [] + type: pathlist yaml: {key: defaults.inventory_plugins} DEFAULT_JINJA2_EXTENSIONS: default: - desc: 'TODO: write it' + description: 'TODO: write it' env: [{name: ANSIBLE_JINJA2_EXTENSIONS}] ini: - {key: jinja2_extensions, section: defaults} - vars: [] yaml: {key: defaults.jinja2_extensions} DEFAULT_KEEP_REMOTE_FILES: default: False - desc: 'TODO: write it' + description: Enables/disables the cleaning up of the temporary files Ansible used to execute the tasks on the remote. env: [{name: ANSIBLE_KEEP_REMOTE_FILES}] ini: - {key: keep_remote_files, section: defaults} - value_type: boolean - vars: [] + type: boolean yaml: {key: defaults.keep_remote_files} DEFAULT_LIBVIRT_LXC_NOSECLABEL: default: False - desc: 'TODO: write it' + description: 'TODO: write it' env: [{name: LIBVIRT_LXC_NOSECLABEL}] ini: - {key: libvirt_lxc_noseclabel, section: selinux} - value_type: boolean - vars: [] + type: boolean yaml: {key: selinux.libvirt_lxc_noseclabel} DEFAULT_LOAD_CALLBACK_PLUGINS: default: False - desc: 'TODO: write it' + description: 'TODO: write it' env: [{name: ANSIBLE_LOAD_CALLBACK_PLUGINS}] ini: - {key: bin_ansible_callbacks, section: defaults} - value_type: boolean - vars: [] + type: boolean yaml: {key: defaults.bin_ansible_callbacks} DEFAULT_LOCAL_TMP: default: ~/.ansible/tmp - desc: 'TODO: write it' + description: Temporary directory for Ansible to use on the controller. env: [{name: ANSIBLE_LOCAL_TEMP}] ini: - {key: local_tmp, section: defaults} - value_type: tmppath - vars: [] + type: tmppath yaml: {key: defaults.local_tmp} DEFAULT_LOG_PATH: default: '' - desc: 'TODO: write it' + description: File to which Ansible will log on the controller. When empty logging is disabled. env: [{name: ANSIBLE_LOG_PATH}] ini: - {key: log_path, section: defaults} - value_type: path - vars: [] + type: path yaml: {key: defaults.log_path} DEFAULT_LOOKUP_PLUGIN_PATH: default: ~/.ansible/plugins/lookup:/usr/share/ansible/plugins/lookup - desc: 'TODO: write it' + description: 'TODO: write it' env: [{name: ANSIBLE_LOOKUP_PLUGINS}] ini: - {key: lookup_plugins, section: defaults} - value_type: pathlist - vars: [] + type: pathlist yaml: {key: defaults.lookup_plugins} DEFAULT_MANAGED_STR: default: Ansible managed - desc: 'TODO: write it' + description: Sets the macro for the 'ansible_managed' variable available for 'tempalte' tasks. env: [] ini: - {key: ansible_managed, section: defaults} - vars: [] yaml: {key: defaults.ansible_managed} DEFAULT_MODULE_ARGS: default: '' - desc: 'TODO: write it' + description: 'TODO: write it' env: [{name: ANSIBLE_MODULE_ARGS}] ini: - {key: module_args, section: defaults} - vars: [] yaml: {key: defaults.module_args} DEFAULT_MODULE_COMPRESSION: default: ZIP_DEFLATED - desc: 'TODO: write it' + description: 'TODO: write it' env: [] ini: - {key: module_compression, section: defaults} - vars: [] yaml: {key: defaults.module_compression} DEFAULT_MODULE_LANG: - default: os.getenv('LANG', 'en_US.UTF-8') - desc: 'TODO: write it' + # TODO: allow setting to function: os.getenv('LANG', 'en_US.UTF-8') + default: eval(os.getenv('LANG', 'en_US.UTF-8')) + description: "Language locale setting to use for modules when they execute on the target, if empty it defaults to 'en_US.UTF-8'" env: [{name: ANSIBLE_MODULE_LANG}] ini: - {key: module_lang, section: defaults} - vars: [] yaml: {key: defaults.module_lang} DEFAULT_MODULE_NAME: default: command - desc: 'TODO: write it' + description: Module to use with the `ansible` AdHoc command, if none is specified. env: [] ini: - {key: module_name, section: defaults} - vars: [] yaml: {key: defaults.module_name} DEFAULT_MODULE_PATH: default: ~/.ansible/plugins/modules:/usr/share/ansible/plugins/modules - desc: 'TODO: write it' + description: 'TODO: write it' env: [{name: ANSIBLE_LIBRARY}] ini: - {key: library, section: defaults} - value_type: pathlist - vars: [] + type: pathlist yaml: {key: defaults.library} DEFAULT_MODULE_SET_LOCALE: default: False - desc: 'TODO: write it' + description: Controls if we set locale for modules when executing on the target. env: [{name: ANSIBLE_MODULE_SET_LOCALE}] ini: - {key: module_set_locale, section: defaults} - value_type: boolean - vars: [] + type: boolean yaml: {key: defaults.module_set_locale} DEFAULT_MODULE_UTILS_PATH: default: ~/.ansible/plugins/module_utils:/usr/share/ansible/plugins/module_utils - desc: 'TODO: write it' + description: 'TODO: write it' env: [{name: ANSIBLE_MODULE_UTILS}] ini: - {key: module_utils, section: defaults} - value_type: pathlist - vars: [] + type: pathlist yaml: {key: defaults.module_utils} DEFAULT_NO_LOG: default: False - desc: 'TODO: write it' + description: Toggle Ansible's display and logging of task details, mainly used to avoid security disclosures. env: [{name: ANSIBLE_NO_LOG}] ini: - {key: no_log, section: defaults} - value_type: boolean - vars: [] + type: boolean yaml: {key: defaults.no_log} DEFAULT_NO_TARGET_SYSLOG: default: False - desc: 'TODO: write it' + description: Toggle Ansbile logging to syslog on the target when it executes tasks. env: [{name: ANSIBLE_NO_TARGET_SYSLOG}] ini: - {key: no_target_syslog, section: defaults} - value_type: boolean - vars: [] + type: boolean yaml: {key: defaults.no_target_syslog} DEFAULT_NULL_REPRESENTATION: default: - desc: 'TODO: write it' + description: 'TODO: write it' env: [{name: ANSIBLE_NULL_REPRESENTATION}] ini: - {key: null_representation, section: defaults} - value_type: none - vars: [] + type: none yaml: {key: defaults.null_representation} DEFAULT_POLL_INTERVAL: default: 15 - desc: 'TODO: write it' + description: 'TODO: write it' env: [{name: ANSIBLE_POLL_INTERVAL}] ini: - {key: poll_interval, section: defaults} - value_type: integer - vars: [] + type: integer yaml: {key: defaults.poll_interval} DEFAULT_PRIVATE_KEY_FILE: default: - desc: 'TODO: write it' + description: 'TODO: write it' env: [{name: ANSIBLE_PRIVATE_KEY_FILE}] ini: - {key: private_key_file, section: defaults} - value_type: path - vars: [] + type: path yaml: {key: defaults.private_key_file} DEFAULT_PRIVATE_ROLE_VARS: default: False - desc: 'TODO: write it' + description: 'TODO: write it' env: [{name: ANSIBLE_PRIVATE_ROLE_VARS}] ini: - {key: private_role_vars, section: defaults} - value_type: boolean - vars: [] + type: boolean yaml: {key: defaults.private_role_vars} DEFAULT_REMOTE_PORT: default: - desc: 'TODO: write it' + description: Port to use in remote connections, when blank it will use the connection plugin default. env: [{name: ANSIBLE_REMOTE_PORT}] ini: - {key: remote_port, section: defaults} - value_type: integer - vars: [] + type: integer yaml: {key: defaults.remote_port} DEFAULT_REMOTE_TMP: default: ~/.ansible/tmp - desc: 'TODO: write it' + description: + - Temporary directory to use on targets when executing tasks. + - In some cases Ansible may still choose to use a system temporary dir to avoid permission issues. env: [{name: ANSIBLE_REMOTE_TEMP}] ini: - {key: remote_tmp, section: defaults} @@ -852,225 +806,244 @@ DEFAULT_REMOTE_TMP: yaml: {key: defaults.remote_tmp} DEFAULT_REMOTE_USER: default: - desc: 'TODO: write it' + description: + - Sets the login user for the target machines + - When blank it uses the connection plugin's default, normally the user currently executing Ansible. env: [{name: ANSIBLE_REMOTE_USER}] ini: - {key: remote_user, section: defaults} - vars: [] yaml: {key: defaults.remote_user} DEFAULT_ROLES_PATH: default: ~/.ansible/roles:/usr/share/ansible/roles:/etc/ansible/roles - desc: 'TODO: write it' + description: 'TODO: write it' env: [{name: ANSIBLE_ROLES_PATH}] expand_relative_paths: True ini: - {key: roles_path, section: defaults} - value_type: pathlist - vars: [] + type: pathlist yaml: {key: defaults.roles_path} DEFAULT_SCP_IF_SSH: default: smart - desc: 'TODO: write it' + description: + - "Prefered method to use when transfering files over ssh" + - When set to smart, Ansible will try them until one succeeds or they all fail + - If set to True, it will force 'scp', if False it will use 'sftp' env: [{name: ANSIBLE_SCP_IF_SSH}] ini: - {key: scp_if_ssh, section: ssh_connection} - vars: [] yaml: {key: ssh_connection.scp_if_ssh} DEFAULT_SELINUX_SPECIAL_FS: default: fuse, nfs, vboxsf, ramfs, 9p - desc: 'TODO: write it' + description: + - "Some filesystems do not support safe operations and/or return inconsistent errors, + this setting makes Ansible 'tolerate' those in the list w/o causing fatal errors." + - Data corruption may occur and writes are not always verified when a filesystem is in the list. + env: [] ini: - {key: special_context_filesystems, section: selinux} - value_type: list - vars: [] + type: list yaml: {key: selinux.special_context_filesystems} DEFAULT_SFTP_BATCH_MODE: default: True - desc: 'TODO: write it' + description: 'TODO: write it' env: [{name: ANSIBLE_SFTP_BATCH_MODE}] ini: - {key: sftp_batch_mode, section: ssh_connection} - value_type: boolean - vars: [] + type: boolean yaml: {key: ssh_connection.sftp_batch_mode} DEFAULT_SQUASH_ACTIONS: default: apk, apt, dnf, homebrew, openbsd_pkg, pacman, pkgng, yum, zypper - desc: 'TODO: write it' + description: + - Ansible can optimise actions that call modules that support list parameters when using with_ looping. + Instead of calling the module once for each item, the module is called once with the full list. + - The default value for this setting is only for certain package managers, but it can be used for any module + - Currently, this is only supported for modules that have a name or pkg parameter, and only when the item is the only thing being passed to the parameter. env: [{name: ANSIBLE_SQUASH_ACTIONS}] ini: - {key: squash_actions, section: defaults} - value_type: list - vars: [] + type: list yaml: {key: defaults.squash_actions} + version_added: "2.0" DEFAULT_SSH_TRANSFER_METHOD: default: - desc: 'TODO: write it' + description: 'unused?' + # - "Prefered method to use when transfering files over ssh" + # - Setting to smart will try them until one succeeds or they all fail + #choices: ['sftp', 'scp', 'dd', 'smart'] env: [{name: ANSIBLE_SSH_TRANSFER_METHOD}] ini: - {key: transfer_method, section: ssh_connection} - vars: [] yaml: {key: ssh_connection.transfer_method} DEFAULT_STDOUT_CALLBACK: default: default - desc: 'TODO: write it' + description: + - "Set the main callback used to display Ansible output, you can only have one at a time." + - You can have many other callbacks, but just one can be in charge of stdout. env: [{name: ANSIBLE_STDOUT_CALLBACK}] ini: - {key: stdout_callback, section: defaults} - vars: [] yaml: {key: defaults.stdout_callback} DEFAULT_STRATEGY: - default: linear - desc: 'TODO: write it' + default: 'linear' + description: Set the default strategy used for plays. env: [{name: ANSIBLE_STRATEGY}] ini: - {key: strategy, section: defaults} - vars: [] yaml: {key: defaults.strategy} + version_added: "2.3" DEFAULT_STRATEGY_PLUGIN_PATH: default: ~/.ansible/plugins/strategy:/usr/share/ansible/plugins/strategy - desc: 'TODO: write it' + description: 'TODO: write it' env: [{name: ANSIBLE_STRATEGY_PLUGINS}] ini: - {key: strategy_plugins, section: defaults} - value_type: pathlist - vars: [] + type: pathlist yaml: {key: defaults.strategy_plugins} DEFAULT_SU: default: False - desc: 'TODO: write it' + description: 'Toggle the use of "su" for tasks.' env: [{name: ANSIBLE_SU}] ini: - {key: su, section: defaults} - value_type: boolean - vars: [] + type: boolean yaml: {key: defaults.su} DEFAULT_SUDO: default: False - desc: 'TODO: write it' + deprecated: + why: In favor of become which is a generic framework + version: "2.8" + alternatives: become + description: 'Toggle the use of "sudo" for tasks.' env: [{name: ANSIBLE_SUDO}] ini: - {key: sudo, section: defaults} - value_type: boolean - vars: [] - yaml: {key: defaults.sudo} + type: boolean DEFAULT_SUDO_EXE: - default: - desc: 'TODO: write it' + default: sudo + deprecated: + why: In favor of become which is a generic framework + version: "2.8" + alternatives: become + description: 'specify an "sudo" executable, otherwise it relies on PATH.' env: [{name: ANSIBLE_SUDO_EXE}] ini: - {key: sudo_exe, section: defaults} - vars: [] - yaml: {key: defaults.sudo_exe} DEFAULT_SUDO_FLAGS: - default: -H -S -n - desc: 'TODO: write it' + default: '-H -S -n' + deprecated: + why: In favor of become which is a generic framework + version: "2.8" + alternatives: become + description: 'Flags to pass to "sudo"' env: [{name: ANSIBLE_SUDO_FLAGS}] ini: - {key: sudo_flags, section: defaults} - vars: [] - yaml: {key: defaults.sudo_flags} DEFAULT_SUDO_USER: - default: root - desc: 'TODO: write it' + default: + deprecated: + why: In favor of become which is a generic framework + version: "2.8" + alternatives: become + description: 'User you become when using "sudo", leaving it blank will use the default configured on the target (normally root)' env: [{name: ANSIBLE_SUDO_USER}] ini: - {key: sudo_user, section: defaults} - vars: [] - yaml: {key: defaults.sudo_user} DEFAULT_SU_EXE: - default: - desc: 'TODO: write it' + default: su + deprecated: + why: In favor of become which is a generic framework + version: "2.8" + alternatives: become + description: 'specify an "su" executable, otherwise it relies on PATH.' env: [{name: ANSIBLE_SU_EXE}] ini: - {key: su_exe, section: defaults} - vars: [] - yaml: {key: defaults.su_exe} DEFAULT_SU_FLAGS: - default: - desc: 'TODO: write it' + default: ~ + deprecated: + why: In favor of become which is a generic framework + version: "2.8" + alternatives: become + description: 'Flags to pass to su' env: [{name: ANSIBLE_SU_FLAGS}] ini: - {key: su_flags, section: defaults} - vars: [] - yaml: {key: defaults.su_flags} DEFAULT_SU_USER: - default: root - desc: 'TODO: write it' + default: + description: 'User you become when using "su", leaving it blank will use the default configured on the target (normally root)' env: [{name: ANSIBLE_SU_USER}] ini: - {key: su_user, section: defaults} - vars: [] - yaml: {key: defaults.su_user} + deprecated: + why: In favor of become which is a generic framework + version: "2.8" + alternatives: become DEFAULT_SYSLOG_FACILITY: default: LOG_USER - desc: 'TODO: write it' + description: 'TODO: write it' env: [{name: ANSIBLE_SYSLOG_FACILITY}] ini: - {key: syslog_facility, section: defaults} - vars: [] yaml: {key: defaults.syslog_facility} DEFAULT_TASK_INCLUDES_STATIC: default: False - desc: 'TODO: write it' + description: 'TODO: write it' env: [{name: ANSIBLE_TASK_INCLUDES_STATIC}] ini: - {key: task_includes_static, section: defaults} - value_type: boolean - vars: [] + type: boolean yaml: {key: defaults.task_includes_static} + deprecated: + why: include itself is deprecated and this setting will not matter in the future + version: "2.8" + alternatives: None, as its already built into the decision between include_tasks and import_tasks DEFAULT_TEST_PLUGIN_PATH: default: ~/.ansible/plugins/test:/usr/share/ansible/plugins/test - desc: 'TODO: write it' + description: 'TODO: write it' env: [{name: ANSIBLE_TEST_PLUGINS}] ini: - {key: test_plugins, section: defaults} - value_type: pathlist - vars: [] + type: pathlist yaml: {key: defaults.test_plugins} DEFAULT_TIMEOUT: default: 10 - desc: 'TODO: write it' + description: 'TODO: write it' env: [{name: ANSIBLE_TIMEOUT}] ini: - {key: timeout, section: defaults} - value_type: integer - vars: [] + type: integer yaml: {key: defaults.timeout} DEFAULT_TRANSPORT: default: smart - desc: 'TODO: write it' + description: "Default connection plugin to use, the 'smart' option will toggle between 'ssh' and 'paramiko' depending on controller OS and ssh versions" env: [{name: ANSIBLE_TRANSPORT}] ini: - {key: transport, section: defaults} - vars: [] yaml: {key: defaults.transport} DEFAULT_UNDEFINED_VAR_BEHAVIOR: default: True - desc: 'TODO: write it' + description: 'TODO: write it' env: [{name: ANSIBLE_ERROR_ON_UNDEFINED_VARS}] ini: - {key: error_on_undefined_vars, section: defaults} - value_type: boolean - vars: [] + type: boolean yaml: {key: defaults.error_on_undefined_vars} DEFAULT_VARS_PLUGIN_PATH: default: ~/.ansible/plugins/vars:/usr/share/ansible/plugins/vars - desc: 'TODO: write it' + description: 'TODO: write it' env: [{name: ANSIBLE_VARS_PLUGINS}] ini: - {key: vars_plugins, section: defaults} - value_type: pathlist - vars: [] + type: pathlist yaml: {key: defaults.vars_plugins} DEFAULT_VAR_COMPRESSION_LEVEL: default: 0 - desc: 'TODO: write it' + description: 'TODO: write it' env: [{name: ANSIBLE_VAR_COMPRESSION_LEVEL}] ini: - {key: var_compression_level, section: defaults} - value_type: integer - vars: [] + type: integer yaml: {key: defaults.var_compression_level} DEFAULT_VAULT_ID_MATCH: default: False @@ -1098,256 +1071,244 @@ DEFAULT_VAULT_IDENTITY_LIST: vars: [] yaml: {key: defaults.vault_identity_list} DEFAULT_VAULT_PASSWORD_FILE: - default: - desc: 'TODO: write it' + default: ~ + description: 'TODO: write it' env: [{name: ANSIBLE_VAULT_PASSWORD_FILE}] ini: - {key: vault_password_file, section: defaults} - value_type: path - vars: [] + type: path yaml: {key: defaults.vault_password_file} DEFAULT_VERBOSITY: default: 0 - desc: 'TODO: write it' + description: 'TODO: write it' env: [{name: ANSIBLE_VERBOSITY}] ini: - {key: verbosity, section: defaults} - value_type: integer - vars: [] + type: integer yaml: {key: defaults.verbosity} DEPRECATION_WARNINGS: default: True - desc: 'TODO: write it' + description: "Toggle to control the showing of deprecation warnings" env: [{name: ANSIBLE_DEPRECATION_WARNINGS}] ini: - {key: deprecation_warnings, section: defaults} - value_type: boolean - vars: [] + type: boolean yaml: {key: defaults.deprecation_warnings} DIFF_ALWAYS: default: False - desc: 'TODO: write it' + description: 'TODO: write it' env: [{name: ANSIBLE_DIFF_ALWAYS}] ini: - {key: always, section: diff} - value_type: bool - vars: [] + type: bool yaml: {key: diff.always} DIFF_CONTEXT: default: 3 - desc: 'TODO: write it' + description: 'TODO: write it' env: [{name: ANSIBLE_DIFF_CONTEXT}] ini: - {key: context, section: diff} - value_type: integer - vars: [] + type: integer yaml: {key: diff.context} DISPLAY_ARGS_TO_STDOUT: default: False - desc: 'TODO: write it' + description: 'TODO: write it' env: [{name: ANSIBLE_DISPLAY_ARGS_TO_STDOUT}] ini: - {key: display_args_to_stdout, section: defaults} - value_type: boolean - vars: [] + type: boolean yaml: {key: defaults.display_args_to_stdout} DISPLAY_SKIPPED_HOSTS: default: True - desc: 'TODO: write it' + description: "Toggle to control displaying skipped host entries in a task in the default callback" env: [{name: DISPLAY_SKIPPED_HOSTS}] ini: - {key: display_skipped_hosts, section: defaults} - value_type: boolean - vars: [] + type: boolean yaml: {key: defaults.display_skipped_hosts} ERROR_ON_MISSING_HANDLER: default: True - desc: 'TODO: write it' + description: "Toggle to allow missing handlers to become a warning instead of an error when notifying." env: [{name: ANSIBLE_ERROR_ON_MISSING_HANDLER}] ini: - {key: error_on_missing_handler, section: defaults} - value_type: boolean - vars: [] + type: boolean yaml: {key: defaults.error_on_missing_handler} GALAXY_IGNORE_CERTS: default: False - desc: 'TODO: write it' + description: 'TODO: write it' env: [{name: ANSIBLE_GALAXY_IGNORE}] ini: - {key: ignore_certs, section: galaxy} - value_type: boolean - vars: [] + type: boolean yaml: {key: galaxy.ignore_certs} GALAXY_ROLE_SKELETON: default: - desc: 'TODO: write it' + description: 'TODO: write it' env: [{name: ANSIBLE_GALAXY_ROLE_SKELETON}] ini: - {key: role_skeleton, section: galaxy} - value_type: path - vars: [] + type: path yaml: {key: galaxy.role_skeleton} GALAXY_ROLE_SKELETON_IGNORE: default: [^.git$, ^.*/.git_keep$] - desc: 'TODO: write it' + description: 'TODO: write it' env: [{name: ANSIBLE_GALAXY_ROLE_SKELETON_IGNORE}] ini: - {key: role_skeleton_ignore, section: galaxy} - value_type: list - vars: [] + type: list yaml: {key: galaxy.role_skeleton_ignore} GALAXY_SCMS: default: git, hg - desc: 'TODO: write it' + description: 'TODO: write it' env: [{name: ANSIBLE_GALAXY_SCMS}] ini: - {key: scms, section: galaxy} - value_type: list - vars: [] + type: list yaml: {key: galaxy.scms} GALAXY_SERVER: default: https://galaxy.ansible.com - desc: 'TODO: write it' + description: "URL to prepend when roles don't specify the full URI, assume they are referencing this server as the source." env: [{name: ANSIBLE_GALAXY_SERVER}] ini: - {key: server, section: galaxy} - vars: [] yaml: {key: galaxy.server} HOST_KEY_CHECKING: default: True - desc: 'TODO: write it' + description: 'Set this to "False" if you want to avoid host key checking by the underlying tools Ansible uses to connect to the host' env: [{name: ANSIBLE_HOST_KEY_CHECKING}] ini: - {key: host_key_checking, section: defaults} - value_type: boolean - vars: [] + type: boolean yaml: {key: defaults.host_key_checking} INVENTORY_ENABLED: default: [host_list, script, yaml, ini] - desc: List of enabled inventory plugins, it also determines the order in which they are used. + description: List of enabled inventory plugins, it also determines the order in which they are used. env: [{name: ANSIBLE_INVENTORY_ENABLED}] ini: - {key: inventory_enabled, section: defaults} - value_type: list - vars: [] + type: list yaml: {key: inventory.enabled_plugins} INVENTORY_IGNORE_EXTS: - default: BLACKLIST_EXTS + [ '~', '.orig', '.ini', '.cfg', '.retry'] - desc: List of extensions to ignore when using a directory as an inventory source + default: eval(BLACKLIST_EXTS + ( '~', '.orig', '.ini', '.cfg', '.retry')) + description: List of extensions to ignore when using a directory as an inventory source env: [{name: ANSIBLE_INVENTORY_IGNORE}] ini: - {key: inventory_ignore_extensions, section: defaults} - value_type: list - vars: [] + type: list yaml: {key: inventory.ignore_extensions} INVENTORY_IGNORE_PATTERNS: default: [] - desc: List of patterns to ignore when using a directory as an inventory source + description: List of patterns to ignore when using a directory as an inventory source env: [{name: ANSIBLE_INVENTORY_IGNORE_REGEX}] ini: - {key: inventory_ignore_patterns, section: defaults} - value_type: list - vars: [] + type: list yaml: {key: inventory.ignore_patterns} MAX_FILE_SIZE_FOR_DIFF: default: 104448 - desc: Maximum size of files to be considered for diff display + description: Maximum size of files to be considered for diff display env: [{name: ANSIBLE_MAX_DIFF_SIZE}] ini: - {key: max_diff_size, section: defaults} - value_type: integer - vars: [] + type: integer yaml: {key: defaults.max_diff_size} MERGE_MULTIPLE_CLI_TAGS: default: True - desc: 'TODO: write it' + description: 'TODO: write it' env: [{name: ANSIBLE_MERGE_MULTIPLE_CLI_TAGS}] ini: - - {key: merge_multiple_cli_tags, section: defaults} - value_type: boolean - vars: [] + - {key: merge_multiple_cli_tags, section: defaults} + type: boolean yaml: {key: defaults.merge_multiple_cli_tags} NETWORK_GROUP_MODULES: default: [eos, nxos, ios, iosxr, junos, ce, vyos, sros, dellos9, dellos10, dellos6, asa, aruba, aireos] - desc: 'TODO: write it' + description: 'TODO: write it' env: [{name: NETWORK_GROUP_MODULES}] ini: - {key: network_group_modules, section: defaults} - value_type: list - vars: [] + type: list yaml: {key: defaults.network_group_modules} ONLY_NAMESPACE_FACTS: default: False - desc: + description: - Facts normally get injected as top level variables, this setting prevents that. - Facts are still available in the `ansible_facts` variable w/o the `ansible_` prefix. env: [{name: ANSIBLE_RESTRICT_FACTS}] ini: - {key: restrict_facts_namespace, section: defaults} - value_type: boolean - vars: [] + type: boolean yaml: {key: defaults.restrict_facts_namespace} version_added: "2.4" PARAMIKO_HOST_KEY_AUTO_ADD: default: False - desc: 'TODO: write it' + description: 'TODO: write it' env: [{name: ANSIBLE_PARAMIKO_HOST_KEY_AUTO_ADD}] ini: - {key: host_key_auto_add, section: paramiko_connection} - value_type: boolean - vars: [] + type: boolean yaml: {key: paramiko_connection.host_key_auto_add} PARAMIKO_LOOK_FOR_KEYS: default: True - desc: 'TODO: write it' + description: 'TODO: write it' env: [{name: ANSIBLE_PARAMIKO_LOOK_FOR_KEYS}] ini: - {key: look_for_keys, section: paramiko_connection} - value_type: boolean - vars: [] + type: boolean yaml: {key: paramiko_connection.look_for_keys} PARAMIKO_PROXY_COMMAND: default: - desc: 'TODO: write it' + description: 'TODO: write it' env: [{name: ANSIBLE_PARAMIKO_PROXY_COMMAND}] ini: - {key: proxy_command, section: paramiko_connection} - vars: [] yaml: {key: paramiko_connection.proxy_command} PARAMIKO_PTY: default: True - desc: 'TODO: write it' + description: 'TODO: write it' env: [{name: ANSIBLE_PARAMIKO_PTY}] ini: - {key: pty, section: paramiko_connection} - value_type: boolean - vars: [] + type: boolean yaml: {key: paramiko_connection.pty} PARAMIKO_RECORD_HOST_KEYS: default: True - desc: 'TODO: write it' + description: 'TODO: write it' env: [{name: ANSIBLE_PARAMIKO_RECORD_HOST_KEYS}] ini: - {key: record_host_keys, section: paramiko_connection} - value_type: boolean - vars: [] + type: boolean yaml: {key: paramiko_connection.record_host_keys} +PERSISTENT_CONNECT_INTERVAL: + default: 1 + description: 'TODO: write it' + env: [{name: ANSIBLE_PERSISTENT_CONNECT_INTERVAL}] + ini: + - {key: connect_interval, section: persistent_connection} + type: integer + yaml: {key: persistent_connection.connect_interval} PERSISTENT_CONTROL_PATH_DIR: default: ~/.ansible/pc - desc: 'TODO: write it' + description: 'TODO: write it' env: [{name: ANSIBLE_PERSISTENT_CONTROL_PATH_DIR}] ini: - {key: control_path_dir, section: persistent_connection} - vars: [] yaml: {key: persistent_connection.control_path_dir} +PERSISTENT_CONNECT_RETRIES: + default: 30 + description: 'TODO: write it' + env: [{name: ANSIBLE_PERSISTENT_CONNECT_RETRIES}] + ini: + - {key: connect_retries, section: persistent_connection} + type: integer + yaml: {key: persistent_connection.connect_retries} PERSISTENT_CONNECT_TIMEOUT: default: 30 - desc: 'TODO: write it' + description: 'TODO: write it' env: [{name: ANSIBLE_PERSISTENT_CONNECT_TIMEOUT}] ini: - {key: connect_timeout, section: persistent_connection} - value_type: integer - vars: [] + type: integer yaml: {key: persistent_connection.connect_timeout} PERSISTENT_CONNECT_RETRY_TIMEOUT: default: 15 @@ -1369,74 +1330,72 @@ PERSISTENT_COMMAND_TIMEOUT: yaml: {key: persistent_connection.command_timeout} RETRY_FILES_ENABLED: default: True - desc: This controls whether a failed Ansible playbook should create a .retry file. + description: This controls whether a failed Ansible playbook should create a .retry file. env: [{name: ANSIBLE_RETRY_FILES_ENABLED}] ini: - {key: retry_files_enabled, section: defaults} - value_type: boolean - vars: [] + type: boolean yaml: {key: errors.retry.enabled} RETRY_FILES_SAVE_PATH: default: ~ - desc: This sets the path in which Ansible will save .retry files when a playbook fails and retry files are enabled. + description: This sets the path in which Ansible will save .retry files when a playbook fails and retry files are enabled. env: [{name: ANSIBLE_RETRY_FILES_SAVE_PATH}] ini: - {key: retry_files_save_path, section: defaults} - value_type: path - vars: [] + type: path yaml: {key: errors.retry.path} SHOW_CUSTOM_STATS: default: False - desc: 'TODO: write it' + description: 'This adds the custom stats set via the set_stats plugin to the default output' env: [{name: ANSIBLE_SHOW_CUSTOM_STATS}] ini: - {key: show_custom_stats, section: defaults} - value_type: boolean - vars: [] + type: boolean yaml: {key: defaults.show_custom_stats} STRING_TYPE_FILTERS: default: [string, to_json, to_nice_json, to_yaml, ppretty, json] - desc: 'TODO: write it' + description: + - "This list of filters avoids 'type conversion' when templating variables" + - Useful when you want to avoid conversion into lists or dictionaries for JSON strings, for example. env: [{name: ANSIBLE_STRING_TYPE_FILTERS}] ini: - {key: dont_type_filters, section: jinja2} - value_type: list - vars: [] + type: list yaml: {key: jinja2.dont_type_filters} SYSTEM_WARNINGS: default: True - desc: 'TODO: write it' + description: 'TODO: write it' env: [{name: ANSIBLE_SYSTEM_WARNINGS}] ini: - {key: system_warnings, section: defaults} - value_type: boolean - vars: [] + type: boolean yaml: {key: defaults.system_warnings} USE_PERSISTENT_CONNECTIONS: default: False - desc: 'TODO: write it' + description: 'TODO: write it' env: [{name: ANSIBLE_USE_PERSISTENT_CONNECTIONS}] ini: - {key: use_persistent_connections, section: defaults} - value_type: boolean - vars: [] + type: boolean yaml: {key: defaults.use_persistent_connections} VARIABLE_PRECEDENCE: default: [all_inventory, groups_inventory, all_plugins_inventory, all_plugins_play, groups_plugins_inventory, groups_plugins_play] - desc: 'TODO: write it' + description: 'TODO: write it' env: [{name: ANSIBLE_PRECEDENCE}] ini: - {key: precedence, section: defaults} - value_type: list - vars: [] + type: list yaml: {key: defaults.precedence} YAML_FILENAME_EXTENSIONS: default: [".yml", ".yaml", ".json"] - desc: "check all of these extensions when looking for 'variable' files which should be YAML or JSON or vaulted versions of theses." + description: + - "Check all of these extensions when looking for 'variable' files which should be YAML or JSON or vaulted versions of these." + - 'This affects vars_files, include_vars, inventory and vars plugins among others.' env: - name: ANSIBLE_YAML_FILENAME_EXT ini: - - {key: defaults, section: yaml_valid_extensions} - value_type: list - vars: [] + - section: yaml_valid_extensions + key: defaults + type: list yaml: {key: defaults.yaml_valid_extensions} +... diff --git a/lib/ansible/config/data.py b/lib/ansible/config/data.py index 903b9cf00c..fb947e88c1 100644 --- a/lib/ansible/config/data.py +++ b/lib/ansible/config/data.py @@ -1,27 +1,10 @@ -# (c) 2017, Ansible by Red Hat, inc -# -# This file is part of Ansible -# -# Ansible is free software: you can redistribute it and/or modify -# it under the terms of the GNU General Public License as published by -# the Free Software Foundation, either version 3 of the License, or -# (at your option) any later version. -# -# Ansible is distributed in the hope that it will be useful, -# but WITHOUT ANY WARRANTY; without even the implied warranty of -# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the -# GNU General Public License for more details. -# -# You should have received a copy of the GNU General Public License -# along with Ansible. If not, see . +# Copyright (c) 2017 Ansible Project +# GNU General Public License v3.0+ (see COPYING or https://www.gnu.org/licenses/gpl-3.0.txt) # Make coding more python3-ish from __future__ import (absolute_import, division, print_function) __metaclass__ = type -from collections import namedtuple - -Setting = namedtuple('Setting','name value origin') class ConfigData(object): @@ -59,3 +42,4 @@ class ConfigData(object): if plugin.name not in self._plugins[plugin.type]: self._plugins[plugin.type][plugin.name] = {} self._plugins[plugin.type][plugin.name][setting.name] = setting + diff --git a/lib/ansible/config/manager.py b/lib/ansible/config/manager.py index 6bef96f7af..77b4d17a4d 100644 --- a/lib/ansible/config/manager.py +++ b/lib/ansible/config/manager.py @@ -1,19 +1,5 @@ -# (c) 2017, Ansible by Red Hat, inc -# -# This file is part of Ansible -# -# Ansible is free software: you can redistribute it and/or modify -# it under the terms of the GNU General Public License as published by -# the Free Software Foundation, either version 3 of the License, or -# (at your option) any later version. -# -# Ansible is distributed in the hope that it will be useful, -# but WITHOUT ANY WARRANTY; without even the implied warranty of -# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the -# GNU General Public License for more details. -# -# You should have received a copy of the GNU General Public License -# along with Ansible. If not, see . +# Copyright (c) 2017 Ansible Project +# GNU General Public License v3.0+ (see COPYING or https://www.gnu.org/licenses/gpl-3.0.txt) # Make coding more python3-ish from __future__ import (absolute_import, division, print_function) @@ -24,7 +10,9 @@ import sys import tempfile import yaml -from ansible.config.data import ConfigData, Setting +from collections import namedtuple + +from ansible.config.data import ConfigData from ansible.errors import AnsibleOptionsError, AnsibleError from ansible.module_utils.six import string_types from ansible.module_utils.six.moves import configparser @@ -34,131 +22,169 @@ from ansible.parsing.quoting import unquote from ansible.utils.path import unfrackpath from ansible.utils.path import makedirs_safe +Plugin = namedtuple('Plugin','name type') +Setting = namedtuple('Setting','name value origin') +# FIXME: see if we can unify in module_utils with similar function used by argspec +def ensure_type(value, value_type): + ''' return a configuration variable with casting + :arg value: The value to ensure correct typing of + :kwarg value_type: The type of the value. This can be any of the following strings: + :boolean: sets the value to a True or False value + :integer: Sets the value to an integer or raises a ValueType error + :float: Sets the value to a float or raises a ValueType error + :list: Treats the value as a comma separated list. Split the value + and return it as a python list. + :none: Sets the value to None + :path: Expands any environment variables and tilde's in the value. + :tmp_path: Create a unique temporary directory inside of the directory + specified by value and return its path. + :pathlist: Treat the value as a typical PATH string. (On POSIX, this + means colon separated strings.) Split the value and then expand + each part for environment variables and tildes. + ''' + if value_type: + value_type = value_type.lower() + + if value_type in ('boolean', 'bool'): + value = boolean(value, strict=False) + + elif value: + if value_type in ('integer', 'int'): + value = int(value) + + elif value_type == 'float': + value = float(value) + + elif value_type == 'list': + if isinstance(value, string_types): + value = [x.strip() for x in value.split(',')] + + elif value_type == 'none': + if value == "None": + value = None + + elif value_type == 'path': + value = resolve_path(value) + + elif value_type in ('tmp', 'temppath', 'tmppath'): + value = resolve_path(value) + if not os.path.exists(value): + makedirs_safe(value, 0o700) + prefix = 'ansible-local-%s' % os.getpid() + value = tempfile.mkdtemp(prefix=prefix, dir=value) + + elif value_type == 'pathlist': + if isinstance(value, string_types): + value = [resolve_path(x) for x in value.split(os.pathsep)] + + # defaults to string types + elif isinstance(value, string_types): + value = unquote(value) + + return to_text(value, errors='surrogate_or_strict', nonstring='passthru') + +# FIXME: see if this can live in utils/path def resolve_path(path): - + ''' resolve relative or 'varaible' paths ''' if '{{CWD}}' in path: # allow users to force CWD using 'magic' {{CWD}} path = path.replace('{{CWD}}', os.getcwd()) return unfrackpath(path, follow=False) +# FIXME: generic file type? +def get_config_type(cfile): -def get_ini_config(p, entries): + ftype = None + if cfile is not None: + ext = os.path.splitext(cfile)[-1] + if ext in ('.ini', '.cfg'): + ftype = 'ini' + elif ext in ('.yaml', '.yml'): + ftype = 'yaml' + else: + raise AnsibleOptionsError("Unsupported configuration file extension for %s: %s" % (cfile, to_native(ext))) + + return ftype + +# FIXME: can move to module_utils for use for ini plugins also? +def get_ini_config_value(p, entry): ''' returns the value of last ini entry found ''' value = None if p is not None: - for entry in entries: - try: - value = p.get(entry.get('section','defaults'), entry.get('key',''), raw=True) - except: - pass - + try: + value = p.get(entry.get('section','defaults'), entry.get('key',''), raw=True) + except: # FIXME: actually report issues here + pass return value class ConfigManager(object): + UNABLE = [] + DEPRECATED = [] + def __init__(self, conf_file=None): + self._base_defs = {} + self._plugins = {} + self._parser = None + + self._config_file = conf_file self.data = ConfigData() - #FIXME: make dynamic? - bconfig_def = to_bytes('%s/data/config.yml' % os.path.dirname(__file__)) + + #FIXME: make dynamic? scan for more? make it's own method? + # Create configuration definitions from source + bconfig_def = to_bytes('%s/base.yml' % os.path.dirname(__file__)) if os.path.exists(bconfig_def): with open(bconfig_def, 'rb') as config_def: - self.initial_defs = yaml.safe_load(config_def) + self._base_defs = yaml.safe_load(config_def) else: raise AnsibleError("Missing base configuration definition file (bad install?): %s" % to_native(bconfig_def)) - ftype = None - if conf_file is None: + if self._config_file is None: # set config using ini - conf_file = self.find_ini_config_file() - ftype = 'ini' - else: - ext = os.path.splitext(conf_file)[-1] - if ext in ('.ini', '.cfg'): - ftype = 'ini' - elif ext in ('.yaml', '.yml'): - ftype = 'yaml' - else: - raise AnsibleOptionsError("Unsupported configuration file extension: \n{0}".format(ext)) + self._config_file = self._find_ini_config_file() - self.parse_config(conf_file, ftype) + if self._config_file: + if os.path.exists(self._config_file): + # initialize parser and read config + self._parse_config_file() - def parse_config(self, cfile, ftype): + # update constants + self.update_config_data() + + def _parse_config_file(self, cfile=None): + ''' return flat configuration settings from file(s) ''' # TODO: take list of files with merge/nomerge - parser = None - if cfile: + if cfile is None: + cfile = self._config_file + + ftype = get_config_type(cfile) + if cfile is not None: if ftype == 'ini': - parser = configparser.ConfigParser() + self._parser = configparser.ConfigParser() try: - parser.read(cfile) + self._parser.read(cfile) except configparser.Error as e: - raise AnsibleOptionsError("Error reading config file: \n{0}".format(e)) - elif ftype == 'yaml': - with open(cfile, 'rb') as config_stream: - parser = yaml.safe_load(config_stream) + raise AnsibleOptionsError("Error reading config file (%s): %s" % (cfile, to_native(e))) + # FIXME: this should eventually handle yaml config files + #elif ftype == 'yaml': + # with open(cfile, 'rb') as config_stream: + # self._parser = yaml.safe_load(config_stream) else: - raise AnsibleOptionsError("Unsupported configuration file type: \n{0}".format(ftype)) - - self.update_config(cfile, self.initial_defs, parser, ftype) - - def update_config(self, configfile, defs, parser, ftype): - - # update the constant for config file - self.data.update_setting(Setting('CONFIG_FILE', configfile, '')) - - origin = None - # env and config defs can have several entries, ordered in list from lowest to highest precedence - for config in self.initial_defs: - - value = None - # env vars are highest precedence - if defs[config].get('env'): - try: - for env_var in defs[config]['env']: - env_value = os.environ.get(env_var.get('name'), None) - if env_value is not None: # only set if env var is defined - value = env_value - origin = 'env: %s' % env_var.get('name') - except: - sys.stderr.write("Error while loading environment configs for %s\n" % config) - - # try config file entries next - if value is None and defs[config].get(ftype): - if ftype == 'ini': - # load from ini config - try: - value = get_ini_config(parser, defs[config]['ini']) - origin = configfile - except Exception as e: - sys.stderr.write("Error while loading ini config %s: %s" % (configfile, str(e))) - elif ftype == 'yaml': - # FIXME: break down key from defs (. notation???) - key = 'name' - value = parser.get(key) - origin = configfile - - # set default if we got here w/o a value - if value is None: - value = defs[config].get('default') - origin = 'default' - - # ensure correct type - try: - value = self.ensure_type(value, defs[config].get('value_type')) - except: - sys.stderr.write("Unable to set correct type for %s, skipping" % config) - continue - - # set the constant - self.data.update_setting(Setting(config, value, origin)) + raise AnsibleOptionsError("Unsupported configuration file type: %s" % to_native(ftype)) - def find_ini_config_file(self): - ''' Load Config File order(first found is used): ENV, CWD, HOME, /etc/ansible ''' + def _find_yaml_config_files(self): + ''' Load YAML Config Files in order, check merge flags, keep origin of settings''' + pass + + def _find_ini_config_file(self): + ''' Load INI Config File order(first found is used): ENV, CWD, HOME, /etc/ansible ''' + # FIXME: eventually deprecate ini configs path0 = os.getenv("ANSIBLE_CONFIG", None) if path0 is not None: @@ -180,57 +206,163 @@ class ConfigManager(object): return path - def ensure_type(self, value, value_type): - ''' return a configuration variable with casting - :arg value: The value to ensure correct typing of - :kwarg value_type: The type of the value. This can be any of the following strings: - :boolean: sets the value to a True or False value - :integer: Sets the value to an integer or raises a ValueType error - :float: Sets the value to a float or raises a ValueType error - :list: Treats the value as a comma separated list. Split the value - and return it as a python list. - :none: Sets the value to None - :path: Expands any environment variables and tilde's in the value. - :tmp_path: Create a unique temporary directory inside of the directory - specified by value and return its path. - :pathlist: Treat the value as a typical PATH string. (On POSIX, this - means colon separated strings.) Split the value and then expand - each part for environment variables and tildes. + def get_configuration_definitions(self, plugin_type=None, name=None): + ''' just list the possible settings, either base or for specific plugins or plugin ''' + + ret = {} + if plugin_type is None: + ret = self._base_defs + elif name is None: + ret = self._plugins.get(plugin_type, {}) + else: + ret = {name: self._plugins.get(plugin_type, {}).get(name, {})} + + return ret + + def _loop_entries(self, container, entry_list): + ''' repeat code for value entry assignment ''' + + value = None + origin = None + for entry in entry_list: + name = entry.get('name') + temp_value = container.get(name, None) + if temp_value is not None: # only set if env var is defined + value = temp_value + origin = name + + # deal with deprecation of setting source, if used + #FIXME: if entry.get('deprecated'): + + return value, origin + + def get_config_value(self, config, cfile=None, plugin_type=None, plugin_name=None, variables=None): + ''' wrapper ''' + value, _drop = self.get_config_value_and_origin(config, cfile=cfile, plugin_type=plugin_type, plugin_name=plugin_name, variables=variables) + return value + + def get_config_value_and_origin(self, config, cfile=None, plugin_type=None, plugin_name=None, variables=None): + ''' Given a config key figure out the actual value and report on the origin of the settings ''' + + if cfile is None: + cfile = self._config_file + + # Note: sources that are lists listed in low to high precedence (last one wins) + value = None + defs = {} + if plugin_type is None: + defs = self._base_defs + elif plugin_name is None: + defs = self._plugins[plugin_type] + else: + defs = self._plugins[plugin_type][plugin_name] + + # Use 'variable overrides' if present, highest precedence, but only present when querying running play + if variables: + value, origin = self._loop_entries(variables, defs[config]['vars']) + origin = 'var: %s' % origin + + # env vars are next precedence + if value is None and defs[config].get('env'): + value, origin = self._loop_entries(os.environ, defs[config]['env']) + origin = 'env: %s' % origin + + # try config file entries next, if we have one + if value is None and cfile is not None: + ftype = get_config_type(cfile) + if ftype and defs[config].get(ftype): + if ftype == 'ini': + # load from ini config + try: # FIXME: generaelize _loop_entries to allow for files also + for ini_entry in defs[config]['ini']: + value = get_ini_config_value(self._parser, ini_entry) + origin = cfile + #FIXME: if ini_entry.get('deprecated'): + except Exception as e: + sys.stderr.write("Error while loading ini config %s: %s" % (cfile, to_native(e))) + elif ftype == 'yaml': + pass # FIXME: implement, also , break down key from defs (. notation???) + origin = cfile + + ''' + # for plugins, try using existing constants, this is for backwards compatiblity + if plugin_name and defs[config].get('constants'): + value, origin = self._loop_entries(self.data, defs[config]['constants']) + origin = 'constant: %s' % origin ''' - if value_type == 'boolean': - value = boolean(value, strict=False) - elif value: - if value_type == 'integer': - value = int(value) + # set default if we got here w/o a value + if value is None: + value = defs[config].get('default') + origin = 'default' + # FIXME: moved eval to constants as this does not have access to previous vars + if plugin_type is None and isinstance(value, string_types) and (value.startswith('eval(') and value.endswith(')')): + return value, origin + #default_value = defs[config].get('default') + #if plugin_type is None and isinstance(default_value, string_types) and (default_value.startswith('eval(') and default_value.endswith(')')): + # try: + # eval_string = default_value.replace('eval(', '', 1)[:-1] + # value = eval(eval_string) # FIXME: safe eval? + # except: + # value = default_value + #else: + # value = default_value - elif value_type == 'float': - value = float(value) + # ensure correct type + try: + value = ensure_type(value, defs[config].get('type')) + except Exception as e: + self.UNABLE.append(config) - elif value_type == 'list': - if isinstance(value, string_types): - value = [x.strip() for x in value.split(',')] + # deal with deprecation of the setting + if defs[config].get('deprecated') and origin != 'default': + self.DEPRECATED.append((config, defs[config].get('deprecated'))) - elif value_type == 'none': - if value == "None": - value = None + return value, origin - elif value_type == 'path': - value = resolve_path(value) + def update_plugin_config(self, plugin_type, name, defs): + ''' really: update constants ''' + # no sense? + self.initialize_plugin_configuration_definitions(plugin_type, name, defs) + self.update_config_data(defs) - elif value_type == 'tmppath': - value = resolve_path(value) - if not os.path.exists(value): - makedirs_safe(value, 0o700) - prefix = 'ansible-local-%s' % os.getpid() - value = tempfile.mkdtemp(prefix=prefix, dir=value) + def initialize_plugin_configuration_definitions(self, plugin_type, name, defs): - elif value_type == 'pathlist': - if isinstance(value, string_types): - value = [resolve_path(x) for x in value.split(os.pathsep)] + if plugin_type not in self._plugins: + self._plugins[plugin_type] = {} - elif isinstance(value, string_types): - value = unquote(value) + self._plugins[plugin_type][name] = defs - return to_text(value, errors='surrogate_or_strict', nonstring='passthru') + def update_config_data(self, defs=None, configfile=None): + ''' really: update constants ''' + if defs is None: + defs = self._base_defs + + if configfile is None: + configfile = self._config_file + + if not isinstance(defs, dict): + raise AnsibleOptionsError("Invalid configuration definition type: %s for %s" % (type(defs), defs)) + + # update the constant for config file + self.data.update_setting(Setting('CONFIG_FILE', configfile, '')) + + origin = None + # env and config defs can have several entries, ordered in list from lowest to highest precedence + for config in defs: + if not isinstance(defs[config], dict): + raise AnsibleOptionsError("Invalid configuration definition '%s': type is %s" % (to_native(config), type(defs[config]))) + + # get value and origin + value, origin = self.get_config_value_and_origin(config, configfile) + + # set the constant + self.data.update_setting(Setting(config, value, origin)) + + # FIXME: find better way to do this by passing back to where display is available + if self.UNABLE: + sys.stderr.write("Unable to set correct type for:\n\t%s\n" % '\n\t'.join(self.UNABLE)) + if self.DEPRECATED: + for k, reason in self.DEPRECATED: + sys.stderr.write("[DEPRECATED] %s: %(why)s. It will be removed in %(version)s. As alternative %(alternative)s", (k, reason)) diff --git a/lib/ansible/constants.py b/lib/ansible/constants.py index 3bf4a24ea8..25f1a00d8e 100644 --- a/lib/ansible/constants.py +++ b/lib/ansible/constants.py @@ -1,46 +1,58 @@ -# (c) 2012-2014, Michael DeHaan -# -# This file is part of Ansible -# -# Ansible is free software: you can redistribute it and/or modify -# it under the terms of the GNU General Public License as published by -# the Free Software Foundation, either version 3 of the License, or -# (at your option) any later version. -# -# Ansible is distributed in the hope that it will be useful, -# but WITHOUT ANY WARRANTY; without even the implied warranty of -# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the -# GNU General Public License for more details. -# -# You should have received a copy of the GNU General Public License -# along with Ansible. If not, see . +# Copyright (c) 2012-2014, Michael DeHaan +# Copyright (c) 2017 Ansible Project +# GNU General Public License v3.0+ (see COPYING or https://www.gnu.org/licenses/gpl-3.0.txt) # Make coding more python3-ish from __future__ import (absolute_import, division, print_function) __metaclass__ = type +import os # used to set lang + from string import ascii_letters, digits from ansible.module_utils._text import to_text from ansible.module_utils.parsing.convert_bool import boolean, BOOLEANS_TRUE +from ansible.module_utils.six import string_types from ansible.config.manager import ConfigManager -_config = ConfigManager() - -# Generate constants from config -for setting in _config.data.get_settings(): - vars()[setting.name] = setting.value - +def _deprecated(msg): + ''' display is not guaranteed here, nor it being the full class, but try anyways, fallback to sys.stderr.write ''' + try: + from __main__ import display + display.deprecated(msg, version='2.8') + except: + import sys + sys.stderr.write('[DEPRECATED] %s, to be removed in 2.8' % msg) def mk_boolean(value): ''' moved to module_utils''' - # We don't have a display here so we can't call deprecated - # display.deprecated('ansible.constants.mk_boolean() is deprecated. Use ansible.module_utils.parsing.convert_bool.boolean() instead', version='2.8') + _deprecated('ansible.constants.mk_boolean() is deprecated. Use ansible.module_utils.parsing.convert_bool.boolean() instead') return boolean(value, strict=False) +def get_config(parser, section, key, env_var, default_value, value_type=None, expand_relative_paths=False): + ''' kept for backwarsd compatibility, but deprecated ''' + _deprecated('ansible.constants.get_config() is deprecated. There is new config API, see porting docs.') -# ### CONSTANTS ### yes, actual ones + import os + value = None + # small reconstruction of the old code env/ini/default + value = os.environ.get(env_var, None) + if value is None: + try: + value = config.get_ini_config(parser, [{'key': key, 'section': section}]) + except: + pass + if value is None: + value = default_value + try: + value = config.ensure_type(value, value_type) + except: + pass + + return value + +### CONSTANTS ### yes, actual ones BLACKLIST_EXTS = ('.pyc', '.pyo', '.swp', '.bak', '~', '.rpm', '.md', '.txt') BECOME_METHODS = ['sudo', 'su', 'pbrun', 'pfexec', 'doas', 'dzdo', 'ksu', 'runas', 'pmrun'] BECOME_ERROR_STRINGS = { @@ -79,3 +91,22 @@ RESTRICTED_RESULT_KEYS = ['ansible_rsync_path', 'ansible_playbook_python'] TREE_DIR = None VAULT_VERSION_MIN = 1.0 VAULT_VERSION_MAX = 1.0 + +### POPULATE SETTINGS FROM CONFIG ### +config = ConfigManager() + +# Generate constants from config +for setting in config.data.get_settings(): + + # FIXME: find better way to do in manager class and/or ensure types + if isinstance(setting.value, string_types) and (setting.value.startswith('eval(') and setting.value.endswith(')')): + try: + eval_string = setting.value.replace('eval(', '', 1)[:-1] + vars()[setting.name] = eval(eval_string) # FIXME: safe eval? + continue + except: + pass + + vars()[setting.name] = setting.value + + diff --git a/lib/ansible/errors/__init__.py b/lib/ansible/errors/__init__.py index c3249d9679..18119d6d95 100644 --- a/lib/ansible/errors/__init__.py +++ b/lib/ansible/errors/__init__.py @@ -20,6 +20,8 @@ from __future__ import (absolute_import, division, print_function) __metaclass__ = type from collections import Sequence +import traceback +import sys from ansible.errors.yaml_strings import ( YAML_COMMON_DICT_ERROR, @@ -68,6 +70,8 @@ class AnsibleError(Exception): self.message += '\nexception type: %s' % to_native(type(orig_exc)) self.message += '\nexception: %s' % to_native(orig_exc) + self.tb = ''.join(traceback.format_tb(sys.exc_info()[2])) + def __str__(self): return self.message diff --git a/lib/ansible/executor/module_common.py b/lib/ansible/executor/module_common.py index d13d37d800..474c71dcbf 100644 --- a/lib/ansible/executor/module_common.py +++ b/lib/ansible/executor/module_common.py @@ -36,7 +36,7 @@ from ansible.release import __version__, __author__ from ansible import constants as C from ansible.errors import AnsibleError from ansible.module_utils._text import to_bytes, to_text -from ansible.plugins import module_utils_loader, ps_module_utils_loader +from ansible.plugins.loader import module_utils_loader, ps_module_utils_loader from ansible.plugins.shell.powershell import async_watchdog, async_wrapper, become_wrapper, leaf_exec, exec_wrapper # Must import strategy and use write_locks from there # If we import write_locks directly then we end up binding a @@ -579,7 +579,7 @@ def recursive_finder(name, data, py_module_names, py_module_cache, zf): zf.writestr(os.path.join("ansible/module_utils", py_module_file_name), py_module_cache[py_module_name][0]) - display.vvv("Using module_utils file %s" % py_module_cache[py_module_name][1]) + display.vvvvv("Using module_utils file %s" % py_module_cache[py_module_name][1]) # Add the names of the files we're scheduling to examine in the loop to # py_module_names so that we don't re-examine them in the next pass diff --git a/lib/ansible/executor/task_executor.py b/lib/ansible/executor/task_executor.py index f4e96e2065..ae1a9fdebf 100644 --- a/lib/ansible/executor/task_executor.py +++ b/lib/ansible/executor/task_executor.py @@ -731,6 +731,7 @@ class TaskExecutor: conn_type = self._play_context.connection connection = self._shared_loader_obj.connection_loader.get(conn_type, self._play_context, self._new_stdin) + self._play_context.set_options_from_plugin(connection) if not connection: raise AnsibleError("the connection plugin '%s' was not found" % conn_type) diff --git a/lib/ansible/executor/task_queue_manager.py b/lib/ansible/executor/task_queue_manager.py index 49b84364bb..6dbf51c716 100644 --- a/lib/ansible/executor/task_queue_manager.py +++ b/lib/ansible/executor/task_queue_manager.py @@ -31,7 +31,7 @@ from ansible.module_utils.six import string_types from ansible.module_utils._text import to_text from ansible.playbook.block import Block from ansible.playbook.play_context import PlayContext -from ansible.plugins import callback_loader, strategy_loader, module_loader +from ansible.plugins.loader import callback_loader, strategy_loader, module_loader from ansible.plugins.callback import CallbackBase from ansible.template import Templar from ansible.utils.helpers import pct_to_int diff --git a/lib/ansible/inventory/manager.py b/lib/ansible/inventory/manager.py index 50a42ccbc5..245cade032 100644 --- a/lib/ansible/inventory/manager.py +++ b/lib/ansible/inventory/manager.py @@ -30,7 +30,7 @@ from ansible.inventory.data import InventoryData from ansible.module_utils.six import string_types from ansible.module_utils._text import to_bytes, to_text from ansible.parsing.utils.addresses import parse_address -from ansible.plugins import PluginLoader +from ansible.plugins.loader import PluginLoader from ansible.utils.path import unfrackpath try: @@ -260,14 +260,15 @@ class InventoryManager(object): display.vvv(u'Parsed %s inventory source with %s plugin' % (to_text(source), plugin_name)) break except AnsibleParserError as e: - failures.append(u'\n* Failed to parse %s with %s inventory plugin: %s\n' % (to_text(source), plugin_name, to_text(e))) + failures.append({'src': source, 'plugin': plugin_name, 'exc': e}) else: display.debug(u'%s did not meet %s requirements' % (to_text(source), plugin_name)) else: if failures: # only if no plugin processed files should we show errors. for fail in failures: - display.warning(fail) + display.warning(u'\n* Failed to parse %s with %s inventory plugin: %s' % (to_text(fail['src']), fail['plugin'], to_text(fail['exc']))) + display.vvv(fail['exc'].tb) if not parsed: display.warning(u"Unable to parse %s as an inventory source" % to_text(source)) diff --git a/lib/ansible/parsing/mod_args.py b/lib/ansible/parsing/mod_args.py index 39d9ff25dc..632911b5dc 100644 --- a/lib/ansible/parsing/mod_args.py +++ b/lib/ansible/parsing/mod_args.py @@ -23,7 +23,7 @@ from ansible.errors import AnsibleParserError, AnsibleError from ansible.module_utils.six import iteritems, string_types from ansible.module_utils._text import to_text from ansible.parsing.splitter import parse_kv, split_args -from ansible.plugins import module_loader, action_loader +from ansible.plugins.loader import module_loader, action_loader from ansible.template import Templar diff --git a/lib/ansible/parsing/plugin_docs.py b/lib/ansible/parsing/plugin_docs.py new file mode 100644 index 0000000000..0f48b0080f --- /dev/null +++ b/lib/ansible/parsing/plugin_docs.py @@ -0,0 +1,83 @@ +# Copyright (c) 2017 Ansible Project +# GNU General Public License v3.0+ (see COPYING or https://www.gnu.org/licenses/gpl-3.0.txt) + +# Make coding more python3-ish +from __future__ import (absolute_import, division, print_function) +__metaclass__ = type + +import ast +import yaml + +from ansible.parsing.yaml.loader import AnsibleLoader + +try: + from __main__ import display +except ImportError: + from ansible.utils.display import Display + display = Display() + + +def read_docstring(filename, verbose=True, ignore_errors=True): + """ + Search for assignment of the DOCUMENTATION and EXAMPLES variables in the given file. + Parse DOCUMENTATION from YAML and return the YAML doc or None together with EXAMPLES, as plain text. + """ + + data = { + 'doc': None, + 'plainexamples': None, + 'returndocs': None, + 'metadata': None + } + + string_to_vars = { + 'DOCUMENTATION': 'doc', + 'EXAMPLES': 'plainexamples', + 'RETURN': 'returndocs', + 'ANSIBLE_METADATA': 'metadata' + } + + try: + M = ast.parse(''.join(open(filename))) + try: + display.debug('Attempt first docstring is yaml docs') + docstring = yaml.load(M.body[0].value.s) + for string in string_to_vars.keys(): + if string in docstring: + data[string_to_vars[string]] = docstring[string] + display.debug('assigned :%s' % string_to_vars[string]) + except Exception as e: + display.debug('failed docstring parsing: %s' % str(e)) + + if 'docs' not in data or not data['docs']: + display.debug('Fallback to vars parsing') + for child in M.body: + if isinstance(child, ast.Assign): + for t in child.targets: + try: + theid = t.id + except AttributeError: + # skip errors can happen when trying to use the normal code + display.warning("Failed to assign id for %s on %s, skipping" % (t, filename)) + continue + + if theid in string_to_vars: + varkey = string_to_vars[theid] + if isinstance(child.value, ast.Dict): + data[varkey] = ast.literal_eval(child.value) + else: + if theid in ['DOCUMENTATION', 'ANSIBLE_METADATA']: + # string should be yaml + data[varkey] = AnsibleLoader(child.value.s, file_name=filename).get_single_data() + else: + # not yaml, should be a simple string + data[varkey] = child.value.s + display.debug('assigned :%s' % varkey) + + except: + if verbose: + display.error("unable to parse %s" % filename) + if not ignore_errors: + raise + + return data diff --git a/lib/ansible/playbook/__init__.py b/lib/ansible/playbook/__init__.py index cc0451a2f8..7c20843925 100644 --- a/lib/ansible/playbook/__init__.py +++ b/lib/ansible/playbook/__init__.py @@ -26,7 +26,7 @@ from ansible.errors import AnsibleParserError from ansible.module_utils._text import to_text from ansible.playbook.play import Play from ansible.playbook.playbook_include import PlaybookInclude -from ansible.plugins import get_all_plugin_loaders +from ansible.plugins.loader import get_all_plugin_loaders try: from __main__ import display diff --git a/lib/ansible/playbook/base.py b/lib/ansible/playbook/base.py index 775fd9878e..66caf73bcf 100644 --- a/lib/ansible/playbook/base.py +++ b/lib/ansible/playbook/base.py @@ -197,9 +197,10 @@ class Base(with_metaclass(BaseMeta, object)): self.vars = dict() def dump_me(self, depth=0): + ''' this is never called from production code, it is here to be used when debugging as a 'complex print' ''' if depth == 0: - print("DUMPING OBJECT ------------------------------------------------------") - print("%s- %s (%s, id=%s)" % (" " * depth, self.__class__.__name__, self, id(self))) + display.debug("DUMPING OBJECT ------------------------------------------------------") + display.debug("%s- %s (%s, id=%s)" % (" " * depth, self.__class__.__name__, self, id(self))) if hasattr(self, '_parent') and self._parent: self._parent.dump_me(depth + 2) dep_chain = self._parent.get_dep_chain() diff --git a/lib/ansible/playbook/play_context.py b/lib/ansible/playbook/play_context.py index d3f738fffe..f86709f202 100644 --- a/lib/ansible/playbook/play_context.py +++ b/lib/ansible/playbook/play_context.py @@ -36,6 +36,7 @@ from ansible.module_utils._text import to_bytes from ansible.module_utils.parsing.convert_bool import boolean from ansible.playbook.attribute import FieldAttribute from ansible.playbook.base import Base +from ansible.plugins import get_plugin_class from ansible.utils.ssh_functions import check_for_controlpersist @@ -54,31 +55,47 @@ __all__ = ['PlayContext'] # in variable names. MAGIC_VARIABLE_MAPPING = dict( + accelerate_port=('ansible_accelerate_port', ), + + # base connection=('ansible_connection', ), + module_compression=('ansible_module_compression', ), + shell=('ansible_shell_type', ), + executable=('ansible_shell_executable', ), + remote_tmp_dir=('ansible_remote_tmp', ), + + # connection common remote_addr=('ansible_ssh_host', 'ansible_host'), remote_user=('ansible_ssh_user', 'ansible_user'), - remote_tmp_dir=('ansible_remote_tmp', ), - port=('ansible_ssh_port', 'ansible_port'), - timeout=('ansible_ssh_timeout', 'ansible_timeout'), - ssh_executable=('ansible_ssh_executable', ), - accelerate_port=('ansible_accelerate_port', ), password=('ansible_ssh_pass', 'ansible_password'), - private_key_file=('ansible_ssh_private_key_file', 'ansible_private_key_file'), + port=('ansible_ssh_port', 'ansible_port'), pipelining=('ansible_ssh_pipelining', 'ansible_pipelining'), - shell=('ansible_shell_type', ), + timeout=('ansible_ssh_timeout', 'ansible_timeout'), + private_key_file=('ansible_ssh_private_key_file', 'ansible_private_key_file'), + + # networking modules network_os=('ansible_network_os', ), + + # ssh TODO: remove + ssh_executable=('ansible_ssh_executable', ), + ssh_common_args=('ansible_ssh_common_args', ), + sftp_extra_args=('ansible_sftp_extra_args', ), + scp_extra_args=('ansible_scp_extra_args', ), + ssh_extra_args=('ansible_ssh_extra_args', ), + ssh_transfer_method=('ansible_ssh_transfer_method', ), + + # docker TODO: remove + docker_extra_args=('ansible_docker_extra_args', ), + + # become become=('ansible_become', ), become_method=('ansible_become_method', ), become_user=('ansible_become_user', ), become_pass=('ansible_become_password', 'ansible_become_pass'), become_exe=('ansible_become_exe', ), become_flags=('ansible_become_flags', ), - ssh_common_args=('ansible_ssh_common_args', ), - docker_extra_args=('ansible_docker_extra_args', ), - sftp_extra_args=('ansible_sftp_extra_args', ), - scp_extra_args=('ansible_scp_extra_args', ), - ssh_extra_args=('ansible_ssh_extra_args', ), - ssh_transfer_method=('ansible_ssh_transfer_method', ), + + # deprecated sudo=('ansible_sudo', ), sudo_user=('ansible_sudo_user', ), sudo_pass=('ansible_sudo_password', 'ansible_sudo_pass'), @@ -89,10 +106,9 @@ MAGIC_VARIABLE_MAPPING = dict( su_pass=('ansible_su_password', 'ansible_su_pass'), su_exe=('ansible_su_exe', ), su_flags=('ansible_su_flags', ), - executable=('ansible_shell_executable', ), - module_compression=('ansible_module_compression', ), ) +# TODO: needs to be configurable b_SU_PROMPT_LOCALIZATIONS = [ to_bytes('Password'), to_bytes('암호'), @@ -135,7 +151,7 @@ TASK_ATTRIBUTE_OVERRIDES = ( 'become_method', 'become_flags', 'connection', - 'docker_extra_args', + 'docker_extra_args', # TODO: remove 'delegate_to', 'no_log', 'remote_user', @@ -143,6 +159,11 @@ TASK_ATTRIBUTE_OVERRIDES = ( RESET_VARS = ( 'ansible_connection', + 'ansible_user', + 'ansible_host', + 'ansible_port', + + # TODO: ??? 'ansible_docker_extra_args', 'ansible_ssh_host', 'ansible_ssh_pass', @@ -151,9 +172,6 @@ RESET_VARS = ( 'ansible_ssh_private_key_file', 'ansible_ssh_pipelining', 'ansible_ssh_executable', - 'ansible_user', - 'ansible_host', - 'ansible_port', ) @@ -165,47 +183,59 @@ class PlayContext(Base): connection/authentication information. ''' + # base + _module_compression = FieldAttribute(isa='string', default=C.DEFAULT_MODULE_COMPRESSION) + _shell = FieldAttribute(isa='string') + _executable = FieldAttribute(isa='string', default=C.DEFAULT_EXECUTABLE) + # connection fields, some are inherited from Base: # (connection, port, remote_user, environment, no_log) - _docker_extra_args = FieldAttribute(isa='string') _remote_addr = FieldAttribute(isa='string') _remote_tmp_dir = FieldAttribute(isa='string', default=C.DEFAULT_REMOTE_TMP) _password = FieldAttribute(isa='string') - _private_key_file = FieldAttribute(isa='string', default=C.DEFAULT_PRIVATE_KEY_FILE) _timeout = FieldAttribute(isa='int', default=C.DEFAULT_TIMEOUT) - _shell = FieldAttribute(isa='string') - _network_os = FieldAttribute(isa='string') _connection_user = FieldAttribute(isa='string') + _private_key_file = FieldAttribute(isa='string', default=C.DEFAULT_PRIVATE_KEY_FILE) + _pipelining = FieldAttribute(isa='bool', default=C.ANSIBLE_PIPELINING) + + # networking modules + _network_os = FieldAttribute(isa='string') + + # docker FIXME: remove these + _docker_extra_args = FieldAttribute(isa='string') + + # ssh # FIXME: remove these + _ssh_executable = FieldAttribute(isa='string', default=C.ANSIBLE_SSH_EXECUTABLE) _ssh_args = FieldAttribute(isa='string', default=C.ANSIBLE_SSH_ARGS) _ssh_common_args = FieldAttribute(isa='string') _sftp_extra_args = FieldAttribute(isa='string') _scp_extra_args = FieldAttribute(isa='string') _ssh_extra_args = FieldAttribute(isa='string') - _ssh_executable = FieldAttribute(isa='string', default=C.ANSIBLE_SSH_EXECUTABLE) _ssh_transfer_method = FieldAttribute(isa='string', default=C.DEFAULT_SSH_TRANSFER_METHOD) + + # ??? _connection_lockfd = FieldAttribute(isa='int') - _pipelining = FieldAttribute(isa='bool', default=C.ANSIBLE_SSH_PIPELINING) + + # accelerate FIXME: remove as soon as deprecation period expires _accelerate = FieldAttribute(isa='bool', default=False) _accelerate_ipv6 = FieldAttribute(isa='bool', default=False, always_post_validate=True) _accelerate_port = FieldAttribute(isa='int', default=C.ACCELERATE_PORT, always_post_validate=True) - _executable = FieldAttribute(isa='string', default=C.DEFAULT_EXECUTABLE) - _module_compression = FieldAttribute(isa='string', default=C.DEFAULT_MODULE_COMPRESSION) # privilege escalation fields _become = FieldAttribute(isa='bool') _become_method = FieldAttribute(isa='string') _become_user = FieldAttribute(isa='string') _become_pass = FieldAttribute(isa='string') - _become_exe = FieldAttribute(isa='string') - _become_flags = FieldAttribute(isa='string') + _become_exe = FieldAttribute(isa='string', default=C.DEFAULT_BECOME_EXE) + _become_flags = FieldAttribute(isa='string', default=C.DEFAULT_BECOME_FLAGS) _prompt = FieldAttribute(isa='string') - # backwards compatibility fields for sudo/su - _sudo_exe = FieldAttribute(isa='string') - _sudo_flags = FieldAttribute(isa='string') + # DEPRECATED: backwards compatibility fields for sudo/su + _sudo_exe = FieldAttribute(isa='string', default=C.DEFAULT_SUDO_EXE) + _sudo_flags = FieldAttribute(isa='string', default=C.DEFAULT_SUDO_FLAGS) _sudo_pass = FieldAttribute(isa='string') - _su_exe = FieldAttribute(isa='string') - _su_flags = FieldAttribute(isa='string') + _su_exe = FieldAttribute(isa='string', default=C.DEFAULT_SU_EXE) + _su_flags = FieldAttribute(isa='string', default=C.DEFAULT_SU_FLAGS) _su_pass = FieldAttribute(isa='string') # general flags @@ -277,6 +307,22 @@ class PlayContext(Base): if play.force_handlers is not None: self.force_handlers = play.force_handlers + def set_options_from_plugin(self, plugin): + # generic derived from connection plugin + + # get options for plugins + options = C.config.get_configuration_definitions(get_plugin_class(plugin), plugin._load_name) + for option in options: + if option: + flag = options[option].get('name') + if flag: + setattr(self, flag, self.connection.get_option(flag)) + + # TODO: made irrelavent by above + # get ssh options FIXME: make these common to all connections + # for flag in ('ssh_common_args', 'docker_extra_args', 'sftp_extra_args', 'scp_extra_args', 'ssh_extra_args'): + # setattr(self, flag, getattr(options, flag, '')) + def set_options(self, options): ''' Configures this connection information instance with data from @@ -291,12 +337,10 @@ class PlayContext(Base): self.check_mode = boolean(options.check, strict=False) - # get ssh options FIXME: make these common to all connections - for flag in ['ssh_common_args', 'docker_extra_args', 'sftp_extra_args', 'scp_extra_args', 'ssh_extra_args']: - setattr(self, flag, getattr(options, flag, '')) - - # general flags (should we move out?) - for flag in ['connection', 'remote_user', 'private_key_file', 'verbosity', 'force_handlers', 'step', 'start_at_task', 'diff']: + # general flags (should we move out?) + # for flag in ('connection', 'remote_user', 'private_key_file', 'verbosity', 'force_handlers', 'step', 'start_at_task', 'diff'): + # should only be 'non plugin' flags + for flag in ('connection', 'private_key_file', 'verbosity', 'force_handlers', 'step', 'start_at_task', 'diff'): attribute = getattr(options, flag, False) if attribute: setattr(self, flag, attribute) @@ -492,22 +536,18 @@ class PlayContext(Base): command = success_cmd # set executable to use for the privilege escalation method, with various overrides - exe = ( - self.become_exe or - getattr(self, '%s_exe' % self.become_method, None) or - C.DEFAULT_BECOME_EXE or - getattr(C, 'DEFAULT_%s_EXE' % self.become_method.upper(), None) or - self.become_method - ) + exe = self.become_method + for myexe in (getattr(self, '%s_exe' % self.become_method, None), self.become_exe): + if myexe: + exe = myexe + break # set flags to use for the privilege escalation method, with various overrides - flags = ( - self.become_flags or - getattr(self, '%s_flags' % self.become_method, None) or - C.DEFAULT_BECOME_FLAGS or - getattr(C, 'DEFAULT_%s_FLAGS' % self.become_method.upper(), None) or - '' - ) + flags = '' + for myflag in (getattr(self, '%s_flags' % self.become_method, None), self.become_flags): + if myflag is not None: + flags = myflag + break if self.become_method == 'sudo': # If we have a password, we run sudo with a randomly-generated diff --git a/lib/ansible/playbook/role/__init__.py b/lib/ansible/playbook/role/__init__.py index d9805b00fa..b0101d700f 100644 --- a/lib/ansible/playbook/role/__init__.py +++ b/lib/ansible/playbook/role/__init__.py @@ -31,7 +31,7 @@ from ansible.playbook.conditional import Conditional from ansible.playbook.helpers import load_list_of_blocks from ansible.playbook.role.metadata import RoleMetadata from ansible.playbook.taggable import Taggable -from ansible.plugins import get_all_plugin_loaders +from ansible.plugins.loader import get_all_plugin_loaders from ansible.utils.vars import combine_vars diff --git a/lib/ansible/playbook/task.py b/lib/ansible/playbook/task.py index 35979b68a3..a1db1d1beb 100644 --- a/lib/ansible/playbook/task.py +++ b/lib/ansible/playbook/task.py @@ -26,7 +26,7 @@ from ansible.module_utils.six import iteritems, string_types from ansible.module_utils._text import to_native from ansible.parsing.mod_args import ModuleArgsParser from ansible.parsing.yaml.objects import AnsibleBaseYAMLObject, AnsibleMapping, AnsibleUnicode -from ansible.plugins import lookup_loader +from ansible.plugins.loader import lookup_loader from ansible.playbook.attribute import FieldAttribute from ansible.playbook.base import Base from ansible.playbook.become import Become diff --git a/lib/ansible/plugins/__init__.py b/lib/ansible/plugins/__init__.py index 29d54519c2..b9e7a299e0 100644 --- a/lib/ansible/plugins/__init__.py +++ b/lib/ansible/plugins/__init__.py @@ -21,18 +21,10 @@ from __future__ import (absolute_import, division, print_function) __metaclass__ = type -import glob -import imp -import os -import os.path -import sys -import warnings - -from collections import defaultdict +from abc import ABCMeta from ansible import constants as C -from ansible.module_utils._text import to_text - +from ansible.module_utils.six import with_metaclass try: from __main__ import display @@ -46,537 +38,11 @@ PATH_CACHE = {} PLUGIN_PATH_CACHE = {} -def get_all_plugin_loaders(): - return [(name, obj) for (name, obj) in globals().items() if isinstance(obj, PluginLoader)] +def get_plugin_class(obj): + return obj.__class__.__name__.lower().replace('module', '') -class PluginLoader: +class AnsiblePlugin(with_metaclass(ABCMeta, object)): - ''' - PluginLoader loads plugins from the configured plugin directories. - - It searches for plugins by iterating through the combined list of - play basedirs, configured paths, and the python path. - The first match is used. - ''' - - def __init__(self, class_name, package, config, subdir, aliases={}, required_base_class=None): - - self.class_name = class_name - self.base_class = required_base_class - self.package = package - self.subdir = subdir - self.aliases = aliases - - if config and not isinstance(config, list): - config = [config] - elif not config: - config = [] - - self.config = config - - if class_name not in MODULE_CACHE: - MODULE_CACHE[class_name] = {} - if class_name not in PATH_CACHE: - PATH_CACHE[class_name] = None - if class_name not in PLUGIN_PATH_CACHE: - PLUGIN_PATH_CACHE[class_name] = defaultdict(dict) - - self._module_cache = MODULE_CACHE[class_name] - self._paths = PATH_CACHE[class_name] - self._plugin_path_cache = PLUGIN_PATH_CACHE[class_name] - - self._extra_dirs = [] - self._searched_paths = set() - - def __setstate__(self, data): - ''' - Deserializer. - ''' - - class_name = data.get('class_name') - package = data.get('package') - config = data.get('config') - subdir = data.get('subdir') - aliases = data.get('aliases') - base_class = data.get('base_class') - - PATH_CACHE[class_name] = data.get('PATH_CACHE') - PLUGIN_PATH_CACHE[class_name] = data.get('PLUGIN_PATH_CACHE') - - self.__init__(class_name, package, config, subdir, aliases, base_class) - self._extra_dirs = data.get('_extra_dirs', []) - self._searched_paths = data.get('_searched_paths', set()) - - def __getstate__(self): - ''' - Serializer. - ''' - - return dict( - class_name=self.class_name, - base_class=self.base_class, - package=self.package, - config=self.config, - subdir=self.subdir, - aliases=self.aliases, - _extra_dirs=self._extra_dirs, - _searched_paths=self._searched_paths, - PATH_CACHE=PATH_CACHE[self.class_name], - PLUGIN_PATH_CACHE=PLUGIN_PATH_CACHE[self.class_name], - ) - - def format_paths(self, paths): - ''' Returns a string suitable for printing of the search path ''' - - # Uses a list to get the order right - ret = [] - for i in paths: - if i not in ret: - ret.append(i) - return os.pathsep.join(ret) - - def print_paths(self): - return self.format_paths(self._get_paths()) - - def _all_directories(self, dir): - results = [] - results.append(dir) - for root, subdirs, files in os.walk(dir, followlinks=True): - if '__init__.py' in files: - for x in subdirs: - results.append(os.path.join(root, x)) - return results - - def _get_package_paths(self, subdirs=True): - ''' Gets the path of a Python package ''' - - if not self.package: - return [] - if not hasattr(self, 'package_path'): - m = __import__(self.package) - parts = self.package.split('.')[1:] - for parent_mod in parts: - m = getattr(m, parent_mod) - self.package_path = os.path.dirname(m.__file__) - if subdirs: - return self._all_directories(self.package_path) - return [self.package_path] - - def _get_paths(self, subdirs=True): - ''' Return a list of paths to search for plugins in ''' - - # FIXME: This is potentially buggy if subdirs is sometimes True and - # sometimes False. In current usage, everything calls this with - # subdirs=True except for module_utils_loader which always calls it - # with subdirs=False. So there currently isn't a problem with this - # caching. - if self._paths is not None: - return self._paths - - ret = self._extra_dirs[:] - - # look in any configured plugin paths, allow one level deep for subcategories - if self.config is not None: - for path in self.config: - path = os.path.realpath(os.path.expanduser(path)) - if subdirs: - contents = glob.glob("%s/*" % path) + glob.glob("%s/*/*" % path) - for c in contents: - if os.path.isdir(c) and c not in ret: - ret.append(c) - if path not in ret: - ret.append(path) - - # look for any plugins installed in the package subtree - # Note package path always gets added last so that every other type of - # path is searched before it. - ret.extend(self._get_package_paths(subdirs=subdirs)) - - # HACK: because powershell modules are in the same directory - # hierarchy as other modules we have to process them last. This is - # because powershell only works on windows but the other modules work - # anywhere (possibly including windows if the correct language - # interpreter is installed). the non-powershell modules can have any - # file extension and thus powershell modules are picked up in that. - # The non-hack way to fix this is to have powershell modules be - # a different PluginLoader/ModuleLoader. But that requires changing - # other things too (known thing to change would be PATHS_CACHE, - # PLUGIN_PATHS_CACHE, and MODULE_CACHE. Since those three dicts key - # on the class_name and neither regular modules nor powershell modules - # would have class_names, they would not work as written. - reordered_paths = [] - win_dirs = [] - - for path in ret: - if path.endswith('windows'): - win_dirs.append(path) - else: - reordered_paths.append(path) - reordered_paths.extend(win_dirs) - - # cache and return the result - self._paths = reordered_paths - return reordered_paths - - def add_directory(self, directory, with_subdir=False): - ''' Adds an additional directory to the search path ''' - - directory = os.path.realpath(directory) - - if directory is not None: - if with_subdir: - directory = os.path.join(directory, self.subdir) - if directory not in self._extra_dirs: - # append the directory and invalidate the path cache - self._extra_dirs.append(directory) - self._paths = None - - def find_plugin(self, name, mod_type='', ignore_deprecated=False): - ''' Find a plugin named name ''' - - if mod_type: - suffix = mod_type - elif self.class_name: - # Ansible plugins that run in the controller process (most plugins) - suffix = '.py' - else: - # Only Ansible Modules. Ansible modules can be any executable so - # they can have any suffix - suffix = '' - - # The particular cache to look for modules within. This matches the - # requested mod_type - pull_cache = self._plugin_path_cache[suffix] - try: - return pull_cache[name] - except KeyError: - # Cache miss. Now let's find the plugin - pass - - # TODO: Instead of using the self._paths cache (PATH_CACHE) and - # self._searched_paths we could use an iterator. Before enabling that - # we need to make sure we don't want to add additional directories - # (add_directory()) once we start using the iterator. Currently, it - # looks like _get_paths() never forces a cache refresh so if we expect - # additional directories to be added later, it is buggy. - for path in (p for p in self._get_paths() if p not in self._searched_paths and os.path.isdir(p)): - try: - full_paths = (os.path.join(path, f) for f in os.listdir(path)) - except OSError as e: - display.warning("Error accessing plugin paths: %s" % to_text(e)) - - for full_path in (f for f in full_paths if os.path.isfile(f) and not f.endswith('__init__.py')): - full_name = os.path.basename(full_path) - - # HACK: We have no way of executing python byte - # compiled files as ansible modules so specifically exclude them - # FIXME: I believe this is only correct for modules and - # module_utils. For all other plugins we want .pyc and .pyo should - # bew valid - if full_path.endswith(('.pyc', '.pyo')): - continue - - splitname = os.path.splitext(full_name) - base_name = splitname[0] - try: - extension = splitname[1] - except IndexError: - extension = '' - - # Module found, now enter it into the caches that match - # this file - if base_name not in self._plugin_path_cache['']: - self._plugin_path_cache[''][base_name] = full_path - - if full_name not in self._plugin_path_cache['']: - self._plugin_path_cache[''][full_name] = full_path - - if base_name not in self._plugin_path_cache[extension]: - self._plugin_path_cache[extension][base_name] = full_path - - if full_name not in self._plugin_path_cache[extension]: - self._plugin_path_cache[extension][full_name] = full_path - - self._searched_paths.add(path) - try: - return pull_cache[name] - except KeyError: - # Didn't find the plugin in this directory. Load modules from - # the next one - pass - - # if nothing is found, try finding alias/deprecated - if not name.startswith('_'): - alias_name = '_' + name - # We've already cached all the paths at this point - if alias_name in pull_cache: - if not ignore_deprecated and not os.path.islink(pull_cache[alias_name]): - display.deprecated('%s is kept for backwards compatibility ' - 'but usage is discouraged. The module ' - 'documentation details page may explain ' - 'more about this rationale.' % - name.lstrip('_')) - return pull_cache[alias_name] - - return None - - def has_plugin(self, name): - ''' Checks if a plugin named name exists ''' - - return self.find_plugin(name) is not None - - __contains__ = has_plugin - - def _load_module_source(self, name, path): - - # avoid collisions across plugins - full_name = '.'.join([self.package, name]) - - if full_name in sys.modules: - # Avoids double loading, See https://github.com/ansible/ansible/issues/13110 - return sys.modules[full_name] - - with warnings.catch_warnings(): - warnings.simplefilter("ignore", RuntimeWarning) - with open(path, 'rb') as module_file: - module = imp.load_source(full_name, path, module_file) - return module - - def get(self, name, *args, **kwargs): - ''' instantiates a plugin of the given name using arguments ''' - - found_in_cache = True - class_only = kwargs.pop('class_only', False) - if name in self.aliases: - name = self.aliases[name] - path = self.find_plugin(name) - if path is None: - return None - - if path not in self._module_cache: - self._module_cache[path] = self._load_module_source(name, path) - found_in_cache = False - - obj = getattr(self._module_cache[path], self.class_name) - if self.base_class: - # The import path is hardcoded and should be the right place, - # so we are not expecting an ImportError. - module = __import__(self.package, fromlist=[self.base_class]) - # Check whether this obj has the required base class. - try: - plugin_class = getattr(module, self.base_class) - except AttributeError: - return None - if not issubclass(obj, plugin_class): - return None - - self._display_plugin_load(self.class_name, name, self._searched_paths, path, - found_in_cache=found_in_cache, class_only=class_only) - if not class_only: - try: - obj = obj(*args, **kwargs) - except TypeError as e: - if "abstract" in e.args[0]: - # Abstract Base Class. The found plugin file does not - # fully implement the defined interface. - return None - raise - - # set extra info on the module, in case we want it later - setattr(obj, '_original_path', path) - setattr(obj, '_load_name', name) - return obj - - def _display_plugin_load(self, class_name, name, searched_paths, path, found_in_cache=None, class_only=None): - msg = 'Loading %s \'%s\' from %s' % (class_name, os.path.basename(name), path) - - if len(searched_paths) > 1: - msg = '%s (searched paths: %s)' % (msg, self.format_paths(searched_paths)) - - if found_in_cache or class_only: - msg = '%s (found_in_cache=%s, class_only=%s)' % (msg, found_in_cache, class_only) - - display.debug(msg) - - def all(self, *args, **kwargs): - ''' instantiates all plugins with the same arguments ''' - - path_only = kwargs.pop('path_only', False) - class_only = kwargs.pop('class_only', False) - all_matches = [] - found_in_cache = True - - for i in self._get_paths(): - all_matches.extend(glob.glob(os.path.join(i, "*.py"))) - - for path in sorted(all_matches, key=lambda match: os.path.basename(match)): - name, _ = os.path.splitext(path) - if '__init__' in name: - continue - - if path_only: - yield path - continue - - if path not in self._module_cache: - self._module_cache[path] = self._load_module_source(name, path) - found_in_cache = False - - try: - obj = getattr(self._module_cache[path], self.class_name) - except AttributeError as e: - display.warning("Skipping plugin (%s) as it seems to be invalid: %s" % (path, to_text(e))) - continue - - if self.base_class: - # The import path is hardcoded and should be the right place, - # so we are not expecting an ImportError. - module = __import__(self.package, fromlist=[self.base_class]) - # Check whether this obj has the required base class. - try: - plugin_class = getattr(module, self.base_class) - except AttributeError: - continue - if not issubclass(obj, plugin_class): - continue - - self._display_plugin_load(self.class_name, name, self._searched_paths, path, found_in_cache=found_in_cache, class_only=class_only) - if not class_only: - try: - obj = obj(*args, **kwargs) - except TypeError as e: - display.warning("Skipping plugin (%s) as it seems to be incomplete: %s" % (path, to_text(e))) - - # set extra info on the module, in case we want it later - setattr(obj, '_original_path', path) - setattr(obj, '_load_name', name) - yield obj - -action_loader = PluginLoader( - 'ActionModule', - 'ansible.plugins.action', - C.DEFAULT_ACTION_PLUGIN_PATH, - 'action_plugins', - required_base_class='ActionBase', -) - -cache_loader = PluginLoader( - 'CacheModule', - 'ansible.plugins.cache', - C.DEFAULT_CACHE_PLUGIN_PATH, - 'cache_plugins', -) - -callback_loader = PluginLoader( - 'CallbackModule', - 'ansible.plugins.callback', - C.DEFAULT_CALLBACK_PLUGIN_PATH, - 'callback_plugins', -) - -connection_loader = PluginLoader( - 'Connection', - 'ansible.plugins.connection', - C.DEFAULT_CONNECTION_PLUGIN_PATH, - 'connection_plugins', - aliases={'paramiko': 'paramiko_ssh'}, - required_base_class='ConnectionBase', -) - -shell_loader = PluginLoader( - 'ShellModule', - 'ansible.plugins.shell', - 'shell_plugins', - 'shell_plugins', -) - -module_loader = PluginLoader( - '', - 'ansible.modules', - C.DEFAULT_MODULE_PATH, - 'library', -) - -module_utils_loader = PluginLoader( - '', - 'ansible.module_utils', - C.DEFAULT_MODULE_UTILS_PATH, - 'module_utils', -) - -# NB: dedicated loader is currently necessary because PS module_utils expects "with subdir" lookup where -# regular module_utils doesn't. This can be revisited once we have more granular loaders. -ps_module_utils_loader = PluginLoader( - '', - 'ansible.module_utils', - C.DEFAULT_MODULE_UTILS_PATH, - 'module_utils', -) - -lookup_loader = PluginLoader( - 'LookupModule', - 'ansible.plugins.lookup', - C.DEFAULT_LOOKUP_PLUGIN_PATH, - 'lookup_plugins', - required_base_class='LookupBase', -) - -filter_loader = PluginLoader( - 'FilterModule', - 'ansible.plugins.filter', - C.DEFAULT_FILTER_PLUGIN_PATH, - 'filter_plugins', -) - -test_loader = PluginLoader( - 'TestModule', - 'ansible.plugins.test', - C.DEFAULT_TEST_PLUGIN_PATH, - 'test_plugins' -) - -fragment_loader = PluginLoader( - 'ModuleDocFragment', - 'ansible.utils.module_docs_fragments', - os.path.join(os.path.dirname(__file__), 'module_docs_fragments'), - '', -) - -strategy_loader = PluginLoader( - 'StrategyModule', - 'ansible.plugins.strategy', - C.DEFAULT_STRATEGY_PLUGIN_PATH, - 'strategy_plugins', - required_base_class='StrategyBase', -) - -terminal_loader = PluginLoader( - 'TerminalModule', - 'ansible.plugins.terminal', - 'terminal_plugins', - 'terminal_plugins' -) - -vars_loader = PluginLoader( - 'VarsModule', - 'ansible.plugins.vars', - C.DEFAULT_VARS_PLUGIN_PATH, - 'vars_plugins', -) - -cliconf_loader = PluginLoader( - 'Cliconf', - 'ansible.plugins.cliconf', - 'cliconf_plugins', - 'cliconf_plugins', - required_base_class='CliconfBase' -) - -netconf_loader = PluginLoader( - 'Netconf', - 'ansible.plugins.netconf', - 'netconf_plugins', - 'netconf_plugins', - required_base_class='NetconfBase' -) + def get_option(self, option): + return C.get_plugin_option(get_plugin_class(self), self.name, option) diff --git a/lib/ansible/plugins/action/__init__.py b/lib/ansible/plugins/action/__init__.py index c7ff4c1bd7..1a38d05b32 100644 --- a/lib/ansible/plugins/action/__init__.py +++ b/lib/ansible/plugins/action/__init__.py @@ -40,6 +40,7 @@ from ansible.parsing.utils.jsonify import jsonify from ansible.playbook.play_context import MAGIC_VARIABLE_MAPPING from ansible.release import __version__ from ansible.utils.unsafe_proxy import wrap_var +from ansible.vars.manager import remove_internal_keys try: @@ -743,7 +744,7 @@ class ActionBase(with_metaclass(ABCMeta, object)): tmpdir_delete = (not data.pop("_ansible_suppress_tmpdir_delete", False) and wrap_async) # remove internal keys - self._remove_internal_keys(data) + remove_internal_keys(data) # cleanup tmp? if (self._play_context.become and self._play_context.become_user != 'root') and not persist_files and delete_remote_tmp or tmpdir_delete: @@ -766,17 +767,6 @@ class ActionBase(with_metaclass(ABCMeta, object)): display.debug("done with _execute_module (%s, %s)" % (module_name, module_args)) return data - def _remove_internal_keys(self, data): - for key in list(data.keys()): - if key.startswith('_ansible_') and key != '_ansible_parsed' or key in C.INTERNAL_RESULT_KEYS: - display.warning("Removed unexpected internal key in module return: %s = %s" % (key, data[key])) - del data[key] - - # remove bad/empty internal keys - for key in ['warnings', 'deprecations']: - if key in data and not data[key]: - del data[key] - def _clean_returned_data(self, data): remove_keys = set() fact_keys = set(data.keys()) @@ -817,7 +807,7 @@ class ActionBase(with_metaclass(ABCMeta, object)): display.warning("Removed restricted key from module data: %s = %s" % (r_key, r_val)) del data[r_key] - self._remove_internal_keys(data) + remove_internal_keys(data) def _parse_returned_data(self, res): try: diff --git a/lib/ansible/plugins/action/junos.py b/lib/ansible/plugins/action/junos.py index d55908e40a..15d993d77d 100644 --- a/lib/ansible/plugins/action/junos.py +++ b/lib/ansible/plugins/action/junos.py @@ -26,7 +26,7 @@ from ansible import constants as C from ansible.module_utils.basic import AnsibleFallbackNotFound from ansible.module_utils.junos import junos_argument_spec from ansible.module_utils.six import iteritems -from ansible.plugins import connection_loader, module_loader +from ansible.plugins.loader import connection_loader, module_loader from ansible.plugins.action.normal import ActionModule as _ActionModule from ansible.module_utils.connection import Connection diff --git a/lib/ansible/plugins/action/synchronize.py b/lib/ansible/plugins/action/synchronize.py index be142fe770..d4d3589cc1 100644 --- a/lib/ansible/plugins/action/synchronize.py +++ b/lib/ansible/plugins/action/synchronize.py @@ -26,7 +26,7 @@ from ansible.module_utils._text import to_text from ansible.module_utils.parsing.convert_bool import boolean from ansible.playbook.play_context import MAGIC_VARIABLE_MAPPING from ansible.plugins.action import ActionBase -from ansible.plugins import connection_loader +from ansible.plugins.loader import connection_loader class ActionModule(ActionBase): diff --git a/lib/ansible/plugins/cache/__init__.py b/lib/ansible/plugins/cache/__init__.py index c36f00bba2..16e3107e58 100644 --- a/lib/ansible/plugins/cache/__init__.py +++ b/lib/ansible/plugins/cache/__init__.py @@ -27,7 +27,7 @@ from ansible import constants as C from ansible.errors import AnsibleError from ansible.module_utils.six import with_metaclass from ansible.module_utils._text import to_bytes -from ansible.plugins import cache_loader +from ansible.plugins.loader import cache_loader try: from __main__ import display diff --git a/lib/ansible/plugins/callback/json.py b/lib/ansible/plugins/callback/json.py index f1221d9351..1dd886ab2c 100644 --- a/lib/ansible/plugins/callback/json.py +++ b/lib/ansible/plugins/callback/json.py @@ -15,6 +15,17 @@ # You should have received a copy of the GNU General Public License # along with Ansible. If not, see . +''' +DOCUMENTATION: + callback: json + short_description: Ansbile screen output asjson + version_added: "2.2" + description: + - This callback converts all events into JSON output + type: stdout + plugin_api_version: "2.0" +''' + # Make coding more python3-ish from __future__ import (absolute_import, division, print_function) __metaclass__ = type diff --git a/lib/ansible/plugins/connection/__init__.py b/lib/ansible/plugins/connection/__init__.py index 813cf9dabe..8a30cbbbdc 100644 --- a/lib/ansible/plugins/connection/__init__.py +++ b/lib/ansible/plugins/connection/__init__.py @@ -23,14 +23,15 @@ import fcntl import gettext import os import shlex -from abc import ABCMeta, abstractmethod, abstractproperty +from abc import abstractmethod, abstractproperty from functools import wraps from ansible import constants as C from ansible.errors import AnsibleError -from ansible.module_utils.six import string_types, with_metaclass +from ansible.module_utils.six import string_types from ansible.module_utils._text import to_bytes, to_text -from ansible.plugins import shell_loader +from ansible.plugins import AnsiblePlugin +from ansible.plugins.loader import shell_loader try: from __main__ import display @@ -53,7 +54,7 @@ def ensure_connect(func): return wrapped -class ConnectionBase(with_metaclass(ABCMeta, object)): +class ConnectionBase(AnsiblePlugin): ''' A base class for connections to contain common code. ''' diff --git a/lib/ansible/plugins/connection/buildah.py b/lib/ansible/plugins/connection/buildah.py index 1901a28af2..41fc26786d 100644 --- a/lib/ansible/plugins/connection/buildah.py +++ b/lib/ansible/plugins/connection/buildah.py @@ -29,16 +29,24 @@ DOCUMENTATION: author: Tomas Tomecek (ttomecek@redhat.com) version_added: 2.4 options: + remote_addr: + description: + - The ID of the container you want to access. + default: inventory_hostname + config: + vars: + - name: ansible_host remote_user: description: - User specified via name or ID which is used to execute commands inside the container. config: - - section: defaults - key: remote_user - env_vars: - - ANSIBLE_REMOTE_USER - host_vars: - - ansible_user + ini: + - section: defaults + key: remote_user + env: + - name: ANSIBLE_REMOTE_USER + vars: + - name: ansible_user """ from __future__ import (absolute_import, division, print_function) diff --git a/lib/ansible/plugins/connection/netconf.py b/lib/ansible/plugins/connection/netconf.py index e176282887..55d87e5d07 100644 --- a/lib/ansible/plugins/connection/netconf.py +++ b/lib/ansible/plugins/connection/netconf.py @@ -25,7 +25,7 @@ import json from ansible import constants as C from ansible.errors import AnsibleConnectionFailure, AnsibleError from ansible.module_utils._text import to_bytes, to_native, to_text -from ansible.plugins import netconf_loader +from ansible.plugins.loader import netconf_loader from ansible.plugins.connection import ConnectionBase, ensure_connect from ansible.utils.jsonrpc import Rpc diff --git a/lib/ansible/plugins/connection/network_cli.py b/lib/ansible/plugins/connection/network_cli.py index bec126210b..f506298385 100644 --- a/lib/ansible/plugins/connection/network_cli.py +++ b/lib/ansible/plugins/connection/network_cli.py @@ -31,8 +31,7 @@ from ansible import constants as C from ansible.errors import AnsibleConnectionFailure from ansible.module_utils.six import BytesIO, binary_type from ansible.module_utils._text import to_bytes, to_text -from ansible.plugins import cliconf_loader -from ansible.plugins import terminal_loader +from ansible.plugins.loader import cliconf_loader, terminal_loader from ansible.plugins.connection.paramiko_ssh import Connection as _Connection from ansible.utils.jsonrpc import Rpc diff --git a/lib/ansible/plugins/connection/ssh.py b/lib/ansible/plugins/connection/ssh.py index b526905ff3..85669ff4ce 100644 --- a/lib/ansible/plugins/connection/ssh.py +++ b/lib/ansible/plugins/connection/ssh.py @@ -26,73 +26,122 @@ DOCUMENTATION: author: ansible (@core) version_added: historical options: - _host: + host: description: Hostname/ip to connect to. default: inventory_hostname - host_vars: - - ansible_host - - ansible_ssh_host - _host_key_checking: - type: bool + vars: + - name: ansible_host + - name: ansible_ssh_host + host_key_checking: + constants: + - name: HOST_KEY_CHECKING description: Determines if ssh should check host keys - config: + type: boolean + ini: - section: defaults key: 'host_key_checking' - env_vars: - - ANSIBLE_HOST_KEY_CHECKING - _password: + env: + - name: ANSIBLE_HOST_KEY_CHECKING + password: description: Authentication password for the C(remote_user). Can be supplied as CLI option. - host_vars: - - ansible_password - - ansible_ssh_pass - _ssh_args: + vars: + - name: ansible_password + - name: ansible_ssh_pass + ssh_args: description: Arguments to pass to all ssh cli tools default: '-C -o ControlMaster=auto -o ControlPersist=60s' - config: + ini: - section: 'ssh_connection' key: 'ssh_args' - env_vars: - - ANSIBLE_SSH_ARGS - _ssh_common_args: - description: Common extra args for ssh CLI tools - host_vars: - - ansible_ssh_common_args - _scp_extra_args: + env: + - name: ANSIBLE_SSH_ARGS + ssh_common_args: + description: Common extra args for all ssh CLI tools + vars: + - name: ansible_ssh_common_args + ssh_executable: + default: ssh + description: + - This defines the location of the ssh binary. It defaults to `ssh` which will use the first ssh binary available in $PATH. + - This option is usually not required, it might be useful when access to system ssh is restricted, + or when using ssh wrappers to connect to remote hosts. + env: [{name: ANSIBLE_SSH_EXECUTABLE}] + ini: + - {key: ssh_executable, section: ssh_connection} + yaml: {key: ssh_connection.ssh_executable} + const: + - name: ANSIBLE_SSH_EXECUTABLE + version_added: "2.2" + scp_extra_args: description: Extra exclusive to the 'scp' CLI - host_vars: - - ansible_scp_extra_args - _sftp_extra_args: + vars: + - name: ansible_scp_extra_args + sftp_extra_args: description: Extra exclusive to the 'sftp' CLI - host_vars: - - ansible_sftp_extra_args - _ssh_extra_args: + vars: + - name: ansible_sftp_extra_args + ssh_extra_args: description: Extra exclusive to the 'ssh' CLI - host_vars: - - ansible_ssh_extra_args + vars: + - name: ansible_ssh_extra_args + ssh_retries: + # constant: ANSIBLE_SSH_RETRIES + description: Number of attempts to connect. + default: 3 + env: + - name: ANSIBLE_SSH_RETRIES + ini: + - section: connection + key: retries + - section: ssh_connection + key: retries port: description: Remote port to connect to. type: int - config: - - section: defaults - key: remote_port - default: 22 - env_vars: - - ANSIBLE_REMOTE_PORT - host_vars: - - ansible_port - - ansible_ssh_port + default: 22 + ini: + - section: defaults + key: remote_port + env: + - name: ANSIBLE_REMOTE_PORT + vars: + - name: ansible_port + - name: ansible_ssh_port remote_user: description: - User name with which to login to the remote server, normally set by the remote_user keyword. - If no user is supplied, Ansible will let the ssh client binary choose the user as it normally - config: - - section: defaults - key: remote_user - env_vars: - - ANSIBLE_REMOTE_USER - host_vars: - - ansible_user - - ansible_ssh_user + ini: + - section: defaults + key: remote_user + env: + - name: ANSIBLE_REMOTE_USER + vars: + - name: ansible_user + - name: ansible_ssh_user + pipelining: + default: ANSIBLE_PIPELINING + description: + - Pipelining reduces the number of SSH operations required to execute a module on the remote server, + by executing many Ansible modules without actual file transfer. + - This can result in a very significant performance improvement when enabled. + - However this conflicts with privilege escalation (become). + For example, when using sudo operations you must first disable 'requiretty' in the sudoers file for the target hosts, + which is why this feature is disabled by default. + env: [{name: ANSIBLE_SSH_PIPELINING}] + ini: + - {key: pipelining, section: ssh_connection} + type: boolean + vars: [{name: ansible_ssh_pipelining}] + +# TODO: +# ANSIBLE_SSH_RETRIES + +# self._play_context.private_key_file +# ANSIBLE_SSH_CONTROL_PATH +# ANSIBLE_SSH_CONTROL_PATH_DIR +# DEFAULT_SFTP_BATCH_MODE +# DEFAULT_SCP_IF_SSH ''' from __future__ import (absolute_import, division, print_function) diff --git a/lib/ansible/plugins/inventory/constructed_groups.py b/lib/ansible/plugins/inventory/constructed_groups.py index a7315b39f1..302b289bdb 100644 --- a/lib/ansible/plugins/inventory/constructed_groups.py +++ b/lib/ansible/plugins/inventory/constructed_groups.py @@ -25,21 +25,20 @@ DOCUMENTATION: - Uses a YAML configuration file to identify group and the Jinja2 expressions that qualify a host for membership. - Only variables already in inventory are available for expressions (no facts). - Failed expressions will be ignored (assumes vars were missing). -EXAMPLES: -# inventory.config file in YAML format -plugin: constructed_groups -groups: - # simple name matching - webservers: inventory_hostname.startswith('web') +EXAMPLES: | # inventory.config file in YAML format + plugin: constructed_groups + groups: + # simple name matching + webservers: inventory_hostname.startswith('web') - # using ec2 'tags' (assumes aws inventory) - development: "'devel' in (ec2_tags|list)" + # using ec2 'tags' (assumes aws inventory) + development: "'devel' in (ec2_tags|list)" - # using other host properties populated in inventory - private_only: not (public_dns_name is defined or ip_address is defined) + # using other host properties populated in inventory + private_only: not (public_dns_name is defined or ip_address is defined) - # complex group membership - multi_group: (group_names|intersection(['alpha', 'beta', 'omega']))|length >= 2 + # complex group membership + multi_group: (group_names|intersection(['alpha', 'beta', 'omega']))|length >= 2 ''' from __future__ import (absolute_import, division, print_function) @@ -77,7 +76,7 @@ class InventoryModule(BaseInventoryPlugin): def parse(self, inventory, loader, path, cache=False): ''' parses the inventory file ''' - super(InventoryModule, self).parse(inventory, loader, path) + super(InventoryModule, self).parse(inventory, loader, path, cache=True) try: data = self.loader.load_from_file(path) @@ -94,19 +93,19 @@ class InventoryModule(BaseInventoryPlugin): for host in inventory.hosts: # get available variables to templar - hostvars = host.get_vars() - if host.name in inventory.cache: # adds facts if cache is active - hostvars = combine_vars(hostvars, inventory.cache[host.name]) + hostvars = inventory.hosts[host].get_vars() + if host in inventory.cache: # adds facts if cache is active + hostvars = combine_vars(hostvars, inventory.cache[host]) templar.set_available_variables(hostvars) # process each 'group entry' - for group_name, expression in data.get('groups', {}): - conditional = u"{%% if %s %%} True {%% else %%} False {%% endif %%}" % expression + for group_name in data.get('groups', {}): + conditional = u"{%% if %s %%} True {%% else %%} False {%% endif %%}" % data['groups'][group_name] result = templar.template(conditional) if result and bool(result): # ensure group exists inventory.add_group(group_name) # add host to group - inventory.add_child(group_name, host.name) + inventory.add_child(group_name, host) except Exception as e: raise AnsibleParserError("failed to parse %s: %s " % (to_native(path), to_native(e))) diff --git a/lib/ansible/plugins/inventory/yaml.py b/lib/ansible/plugins/inventory/yaml.py index 520d8543a3..a7ea874d85 100644 --- a/lib/ansible/plugins/inventory/yaml.py +++ b/lib/ansible/plugins/inventory/yaml.py @@ -21,17 +21,19 @@ DOCUMENTATION: version_added: "2.4" short_description: Uses a specifically YAML file as inventory source. description: - - YAML based inventory, starts with the 'all' group and has hosts/vars/children entries. + - "YAML based inventory, starts with the 'all' group and has hosts/vars/children entries." - Host entries can have sub-entries defined, which will be treated as variables. - Vars entries are normal group vars. - - Children are 'child groups', which can also have their own vars/hosts/children and so on. - - File MUST have a valid extension: yaml, yml, json. + - "Children are 'child groups', which can also have their own vars/hosts/children and so on." + - File MUST have a valid extension, defined in configuration notes: - It takes the place of the previously hardcoded YAML inventory. - To function it requires being whitelisted in configuration. options: - _yaml_extensions: + yaml_extensions: description: list of 'valid' extensions for files containing YAML + type: list + default: ['.yaml', '.yml', '.json'] EXAMPLES: all: # keys must be unique, i.e. only one 'hosts' per group hosts: diff --git a/lib/ansible/plugins/loader.py b/lib/ansible/plugins/loader.py new file mode 100644 index 0000000000..4e4ddd33cc --- /dev/null +++ b/lib/ansible/plugins/loader.py @@ -0,0 +1,588 @@ +# (c) 2012, Daniel Hokka Zakrisson +# (c) 2012-2014, Michael DeHaan and others +# (c) 2017, Toshio Kuratomi +# +# This file is part of Ansible +# +# Ansible is free software: you can redistribute it and/or modify +# it under the terms of the GNU General Public License as published by +# the Free Software Foundation, either version 3 of the License, or +# (at your option) any later version. +# +# Ansible is distributed in the hope that it will be useful, +# but WITHOUT ANY WARRANTY; without even the implied warranty of +# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the +# GNU General Public License for more details. +# +# You should have received a copy of the GNU General Public License +# along with Ansible. If not, see . + +# Make coding more python3-ish +from __future__ import (absolute_import, division, print_function) +__metaclass__ = type + +import glob +import imp +import os +import os.path +import sys +import warnings + +from collections import defaultdict + +from ansible import constants as C +from ansible.plugins import get_plugin_class, MODULE_CACHE, PATH_CACHE, PLUGIN_PATH_CACHE +from ansible.module_utils._text import to_text +from ansible.parsing.plugin_docs import read_docstring + +try: + from __main__ import display +except ImportError: + from ansible.utils.display import Display + display = Display() + + +def get_all_plugin_loaders(): + return [(name, obj) for (name, obj) in globals().items() if isinstance(obj, PluginLoader)] + + +class PluginLoader: + + ''' + PluginLoader loads plugins from the configured plugin directories. + + It searches for plugins by iterating through the combined list of + play basedirs, configured paths, and the python path. + The first match is used. + ''' + + def __init__(self, class_name, package, config, subdir, aliases={}, required_base_class=None): + + self.class_name = class_name + self.base_class = required_base_class + self.package = package + self.subdir = subdir + self.aliases = aliases + + if config and not isinstance(config, list): + config = [config] + elif not config: + config = [] + + self.config = config + + if class_name not in MODULE_CACHE: + MODULE_CACHE[class_name] = {} + if class_name not in PATH_CACHE: + PATH_CACHE[class_name] = None + if class_name not in PLUGIN_PATH_CACHE: + PLUGIN_PATH_CACHE[class_name] = defaultdict(dict) + + self._module_cache = MODULE_CACHE[class_name] + self._paths = PATH_CACHE[class_name] + self._plugin_path_cache = PLUGIN_PATH_CACHE[class_name] + + self._extra_dirs = [] + self._searched_paths = set() + + def __setstate__(self, data): + ''' + Deserializer. + ''' + + class_name = data.get('class_name') + package = data.get('package') + config = data.get('config') + subdir = data.get('subdir') + aliases = data.get('aliases') + base_class = data.get('base_class') + + PATH_CACHE[class_name] = data.get('PATH_CACHE') + PLUGIN_PATH_CACHE[class_name] = data.get('PLUGIN_PATH_CACHE') + + self.__init__(class_name, package, config, subdir, aliases, base_class) + self._extra_dirs = data.get('_extra_dirs', []) + self._searched_paths = data.get('_searched_paths', set()) + + def __getstate__(self): + ''' + Serializer. + ''' + + return dict( + class_name=self.class_name, + base_class=self.base_class, + package=self.package, + config=self.config, + subdir=self.subdir, + aliases=self.aliases, + _extra_dirs=self._extra_dirs, + _searched_paths=self._searched_paths, + PATH_CACHE=PATH_CACHE[self.class_name], + PLUGIN_PATH_CACHE=PLUGIN_PATH_CACHE[self.class_name], + ) + + def format_paths(self, paths): + ''' Returns a string suitable for printing of the search path ''' + + # Uses a list to get the order right + ret = [] + for i in paths: + if i not in ret: + ret.append(i) + return os.pathsep.join(ret) + + def print_paths(self): + return self.format_paths(self._get_paths(subdirs=False)) + + def _all_directories(self, dir): + results = [] + results.append(dir) + for root, subdirs, files in os.walk(dir, followlinks=True): + if '__init__.py' in files: + for x in subdirs: + results.append(os.path.join(root, x)) + return results + + def _get_package_paths(self, subdirs=True): + ''' Gets the path of a Python package ''' + + if not self.package: + return [] + if not hasattr(self, 'package_path'): + m = __import__(self.package) + parts = self.package.split('.')[1:] + for parent_mod in parts: + m = getattr(m, parent_mod) + self.package_path = os.path.dirname(m.__file__) + if subdirs: + return self._all_directories(self.package_path) + return [self.package_path] + + def _get_paths(self, subdirs=True): + ''' Return a list of paths to search for plugins in ''' + + # FIXME: This is potentially buggy if subdirs is sometimes True and sometimes False. + # In current usage, everything calls this with subdirs=True except for module_utils_loader and ansible-doc + # which always calls it with subdirs=False. So there currently isn't a problem with this caching. + if self._paths is not None: + return self._paths + + ret = self._extra_dirs[:] + + # look in any configured plugin paths, allow one level deep for subcategories + if self.config is not None: + for path in self.config: + path = os.path.realpath(os.path.expanduser(path)) + if subdirs: + contents = glob.glob("%s/*" % path) + glob.glob("%s/*/*" % path) + for c in contents: + if os.path.isdir(c) and c not in ret: + ret.append(c) + if path not in ret: + ret.append(path) + + # look for any plugins installed in the package subtree + # Note package path always gets added last so that every other type of + # path is searched before it. + ret.extend(self._get_package_paths(subdirs=subdirs)) + + # HACK: because powershell modules are in the same directory + # hierarchy as other modules we have to process them last. This is + # because powershell only works on windows but the other modules work + # anywhere (possibly including windows if the correct language + # interpreter is installed). the non-powershell modules can have any + # file extension and thus powershell modules are picked up in that. + # The non-hack way to fix this is to have powershell modules be + # a different PluginLoader/ModuleLoader. But that requires changing + # other things too (known thing to change would be PATHS_CACHE, + # PLUGIN_PATHS_CACHE, and MODULE_CACHE. Since those three dicts key + # on the class_name and neither regular modules nor powershell modules + # would have class_names, they would not work as written. + reordered_paths = [] + win_dirs = [] + + for path in ret: + if path.endswith('windows'): + win_dirs.append(path) + else: + reordered_paths.append(path) + reordered_paths.extend(win_dirs) + + # cache and return the result + self._paths = reordered_paths + return reordered_paths + + def _load_config_defs(self, name, path): + ''' Reads plugin docs to find configuration setting definitions, to push to config manager for later use ''' + + # plugins w/o class name don't support config + if self.class_name and self.class_name in ('Connection'): + # FIXME: expand from just connection + type_name = get_plugin_class(self) + dstring = read_docstring(path, verbose=False, ignore_errors=False) + if dstring.get('doc', False): + if 'options' in dstring['doc'] and isinstance(dstring['doc']['options'], dict): + C.config.initialize_plugin_configuration_definitions(type_name, name, dstring['doc']['options']) + display.debug('Loaded config def from plugin (%s/%s)' % (type_name, name)) + + def add_directory(self, directory, with_subdir=False): + ''' Adds an additional directory to the search path ''' + + directory = os.path.realpath(directory) + + if directory is not None: + if with_subdir: + directory = os.path.join(directory, self.subdir) + if directory not in self._extra_dirs: + # append the directory and invalidate the path cache + self._extra_dirs.append(directory) + self._paths = None + + def find_plugin(self, name, mod_type='', ignore_deprecated=False): + ''' Find a plugin named name ''' + + if mod_type: + suffix = mod_type + elif self.class_name: + # Ansible plugins that run in the controller process (most plugins) + suffix = '.py' + else: + # Only Ansible Modules. Ansible modules can be any executable so + # they can have any suffix + suffix = '' + + # The particular cache to look for modules within. This matches the + # requested mod_type + pull_cache = self._plugin_path_cache[suffix] + try: + return pull_cache[name] + except KeyError: + # Cache miss. Now let's find the plugin + pass + + # TODO: Instead of using the self._paths cache (PATH_CACHE) and + # self._searched_paths we could use an iterator. Before enabling that + # we need to make sure we don't want to add additional directories + # (add_directory()) once we start using the iterator. Currently, it + # looks like _get_paths() never forces a cache refresh so if we expect + # additional directories to be added later, it is buggy. + for path in (p for p in self._get_paths() if p not in self._searched_paths and os.path.isdir(p)): + try: + full_paths = (os.path.join(path, f) for f in os.listdir(path)) + except OSError as e: + display.warning("Error accessing plugin paths: %s" % to_text(e)) + + for full_path in (f for f in full_paths if os.path.isfile(f) and not f.endswith('__init__.py')): + full_name = os.path.basename(full_path) + + # HACK: We have no way of executing python byte compiled files as ansible modules so specifically exclude them + # FIXME: I believe this is only correct for modules and module_utils. + # For all other plugins we want .pyc and .pyo should be valid + if full_path.endswith(('.pyc', '.pyo')): + continue + + splitname = os.path.splitext(full_name) + base_name = splitname[0] + try: + extension = splitname[1] + except IndexError: + extension = '' + + # Module found, now enter it into the caches that match this file + if base_name not in self._plugin_path_cache['']: + self._plugin_path_cache[''][base_name] = full_path + + if full_name not in self._plugin_path_cache['']: + self._plugin_path_cache[''][full_name] = full_path + + if base_name not in self._plugin_path_cache[extension]: + self._plugin_path_cache[extension][base_name] = full_path + + if full_name not in self._plugin_path_cache[extension]: + self._plugin_path_cache[extension][full_name] = full_path + + self._searched_paths.add(path) + try: + return pull_cache[name] + except KeyError: + # Didn't find the plugin in this directory. Load modules from the next one + pass + + # if nothing is found, try finding alias/deprecated + if not name.startswith('_'): + alias_name = '_' + name + # We've already cached all the paths at this point + if alias_name in pull_cache: + if not ignore_deprecated and not os.path.islink(pull_cache[alias_name]): + # FIXME: this is not always the case, some are just aliases + display.deprecated('%s is kept for backwards compatibility but usage is discouraged. ' + 'The module documentation details page may explain more about this rationale.' % name.lstrip('_')) + return pull_cache[alias_name] + + return None + + def has_plugin(self, name): + ''' Checks if a plugin named name exists ''' + + return self.find_plugin(name) is not None + + __contains__ = has_plugin + + def _load_module_source(self, name, path): + + # avoid collisions across plugins + full_name = '.'.join([self.package, name]) + + if full_name in sys.modules: + # Avoids double loading, See https://github.com/ansible/ansible/issues/13110 + return sys.modules[full_name] + + with warnings.catch_warnings(): + warnings.simplefilter("ignore", RuntimeWarning) + with open(path, 'rb') as module_file: + module = imp.load_source(full_name, path, module_file) + return module + + def _update_object(self, obj, name, path): + + # load plugin config data + self._load_config_defs(name, path) + + # set extra info on the module, in case we want it later + setattr(obj, '_original_path', path) + setattr(obj, '_load_name', name) + + def get(self, name, *args, **kwargs): + ''' instantiates a plugin of the given name using arguments ''' + + found_in_cache = True + class_only = kwargs.pop('class_only', False) + if name in self.aliases: + name = self.aliases[name] + path = self.find_plugin(name) + if path is None: + return None + + if path not in self._module_cache: + self._module_cache[path] = self._load_module_source(name, path) + found_in_cache = False + + obj = getattr(self._module_cache[path], self.class_name) + if self.base_class: + # The import path is hardcoded and should be the right place, + # so we are not expecting an ImportError. + module = __import__(self.package, fromlist=[self.base_class]) + # Check whether this obj has the required base class. + try: + plugin_class = getattr(module, self.base_class) + except AttributeError: + return None + if not issubclass(obj, plugin_class): + return None + + self._display_plugin_load(self.class_name, name, self._searched_paths, path, found_in_cache=found_in_cache, class_only=class_only) + if not class_only: + try: + obj = obj(*args, **kwargs) + except TypeError as e: + if "abstract" in e.args[0]: + # Abstract Base Class. The found plugin file does not + # fully implement the defined interface. + return None + raise + + self._update_object(obj, name, path) + return obj + + def _display_plugin_load(self, class_name, name, searched_paths, path, found_in_cache=None, class_only=None): + msg = 'Loading %s \'%s\' from %s' % (class_name, os.path.basename(name), path) + + if len(searched_paths) > 1: + msg = '%s (searched paths: %s)' % (msg, self.format_paths(searched_paths)) + + if found_in_cache or class_only: + msg = '%s (found_in_cache=%s, class_only=%s)' % (msg, found_in_cache, class_only) + + display.debug(msg) + + def all(self, *args, **kwargs): + ''' instantiates all plugins with the same arguments ''' + + path_only = kwargs.pop('path_only', False) + class_only = kwargs.pop('class_only', False) + all_matches = [] + found_in_cache = True + + for i in self._get_paths(): + all_matches.extend(glob.glob(os.path.join(i, "*.py"))) + + for path in sorted(all_matches, key=lambda match: os.path.basename(match)): + name = os.path.basename(os.path.splitext(path)[0]) + + if '__init__' in name: + continue + + if path_only: + yield path + continue + + if path not in self._module_cache: + self._module_cache[path] = self._load_module_source(name, path) + found_in_cache = False + + try: + obj = getattr(self._module_cache[path], self.class_name) + except AttributeError as e: + display.warning("Skipping plugin (%s) as it seems to be invalid: %s" % (path, to_text(e))) + continue + + if self.base_class: + # The import path is hardcoded and should be the right place, + # so we are not expecting an ImportError. + module = __import__(self.package, fromlist=[self.base_class]) + # Check whether this obj has the required base class. + try: + plugin_class = getattr(module, self.base_class) + except AttributeError: + continue + if not issubclass(obj, plugin_class): + continue + + self._display_plugin_load(self.class_name, name, self._searched_paths, path, found_in_cache=found_in_cache, class_only=class_only) + if not class_only: + try: + obj = obj(*args, **kwargs) + except TypeError as e: + display.warning("Skipping plugin (%s) as it seems to be incomplete: %s" % (path, to_text(e))) + + self._update_object(obj, name, path) + yield obj + +action_loader = PluginLoader( + 'ActionModule', + 'ansible.plugins.action', + C.DEFAULT_ACTION_PLUGIN_PATH, + 'action_plugins', + required_base_class='ActionBase', +) + +cache_loader = PluginLoader( + 'CacheModule', + 'ansible.plugins.cache', + C.DEFAULT_CACHE_PLUGIN_PATH, + 'cache_plugins', +) + +callback_loader = PluginLoader( + 'CallbackModule', + 'ansible.plugins.callback', + C.DEFAULT_CALLBACK_PLUGIN_PATH, + 'callback_plugins', +) + +connection_loader = PluginLoader( + 'Connection', + 'ansible.plugins.connection', + C.DEFAULT_CONNECTION_PLUGIN_PATH, + 'connection_plugins', + aliases={'paramiko': 'paramiko_ssh'}, + required_base_class='ConnectionBase', +) + +shell_loader = PluginLoader( + 'ShellModule', + 'ansible.plugins.shell', + 'shell_plugins', + 'shell_plugins', +) + +module_loader = PluginLoader( + '', + 'ansible.modules', + C.DEFAULT_MODULE_PATH, + 'library', +) + +module_utils_loader = PluginLoader( + '', + 'ansible.module_utils', + C.DEFAULT_MODULE_UTILS_PATH, + 'module_utils', +) + +# NB: dedicated loader is currently necessary because PS module_utils expects "with subdir" lookup where +# regular module_utils doesn't. This can be revisited once we have more granular loaders. +ps_module_utils_loader = PluginLoader( + '', + 'ansible.module_utils', + C.DEFAULT_MODULE_UTILS_PATH, + 'module_utils', +) + +lookup_loader = PluginLoader( + 'LookupModule', + 'ansible.plugins.lookup', + C.DEFAULT_LOOKUP_PLUGIN_PATH, + 'lookup_plugins', + required_base_class='LookupBase', +) + +filter_loader = PluginLoader( + 'FilterModule', + 'ansible.plugins.filter', + C.DEFAULT_FILTER_PLUGIN_PATH, + 'filter_plugins', +) + +test_loader = PluginLoader( + 'TestModule', + 'ansible.plugins.test', + C.DEFAULT_TEST_PLUGIN_PATH, + 'test_plugins' +) + +fragment_loader = PluginLoader( + 'ModuleDocFragment', + 'ansible.utils.module_docs_fragments', + os.path.join(os.path.dirname(__file__), 'module_docs_fragments'), + '', +) + +strategy_loader = PluginLoader( + 'StrategyModule', + 'ansible.plugins.strategy', + C.DEFAULT_STRATEGY_PLUGIN_PATH, + 'strategy_plugins', + required_base_class='StrategyBase', +) + +terminal_loader = PluginLoader( + 'TerminalModule', + 'ansible.plugins.terminal', + 'terminal_plugins', + 'terminal_plugins' +) + +vars_loader = PluginLoader( + 'VarsModule', + 'ansible.plugins.vars', + C.DEFAULT_VARS_PLUGIN_PATH, + 'vars_plugins', +) + +cliconf_loader = PluginLoader( + 'Cliconf', + 'ansible.plugins.cliconf', + 'cliconf_plugins', + 'cliconf_plugins', + required_base_class='CliconfBase' +) + +netconf_loader = PluginLoader( + 'Netconf', + 'ansible.plugins.netconf', + 'netconf_plugins', + 'netconf_plugins', + required_base_class='NetconfBase' +) diff --git a/lib/ansible/plugins/lookup/etcd.py b/lib/ansible/plugins/lookup/etcd.py index c0de93c3be..30b562fb30 100644 --- a/lib/ansible/plugins/lookup/etcd.py +++ b/lib/ansible/plugins/lookup/etcd.py @@ -29,20 +29,24 @@ DOCUMENTATION: description: - the list of keys to lookup on the etcd server type: list - element_type: string + elements: string required: True _etcd_url: description: - Environment variable with the url for the etcd server default: 'http://127.0.0.1:4001' - env_vars: - - name: ANSIBLE_ETCD_URL + env: + - name: ANSIBLE_ETCD_URL + yaml: + - key: etcd.url _etcd_version: description: - Environment variable with the etcd protocol version default: 'v1' - env_vars: - - name: ANSIBLE_ETCD_VERSION + env: + - name: ANSIBLE_ETCD_VERSION + yaml: + - key: etcd.version EXAMPLES: - name: "a value from a locally running etcd" debug: msg={{ lookup('etcd', 'foo/bar') }} @@ -50,10 +54,11 @@ EXAMPLES: - name: "a values from a folder on a locally running etcd" debug: msg={{ lookup('etcd', 'foo') }} RETURN: - _list: + _raw: description: - list of values associated with input keys - type: strings + type: list + elements: strings ''' from __future__ import (absolute_import, division, print_function) __metaclass__ = type diff --git a/lib/ansible/plugins/strategy/__init__.py b/lib/ansible/plugins/strategy/__init__.py index a3698aabdf..4e54fcaf2a 100644 --- a/lib/ansible/plugins/strategy/__init__.py +++ b/lib/ansible/plugins/strategy/__init__.py @@ -40,7 +40,7 @@ from ansible.playbook.helpers import load_list_of_blocks from ansible.playbook.included_file import IncludedFile from ansible.playbook.task_include import TaskInclude from ansible.playbook.role_include import IncludeRole -from ansible.plugins import action_loader, connection_loader, filter_loader, lookup_loader, module_loader, test_loader +from ansible.plugins.loader import action_loader, connection_loader, filter_loader, lookup_loader, module_loader, test_loader from ansible.template import Templar from ansible.utils.vars import combine_vars from ansible.vars.manager import strip_internal_keys @@ -899,6 +899,7 @@ class StrategyBase: msg = "ending play" elif meta_action == 'reset_connection': connection = connection_loader.get(play_context.connection, play_context, os.devnull) + play_context.set_options_from_plugin(connection) if connection: connection.reset() msg = 'reset connection' diff --git a/lib/ansible/plugins/strategy/free.py b/lib/ansible/plugins/strategy/free.py index 9b009bc8f5..43e5c05ddd 100644 --- a/lib/ansible/plugins/strategy/free.py +++ b/lib/ansible/plugins/strategy/free.py @@ -34,7 +34,7 @@ import time from ansible import constants as C from ansible.errors import AnsibleError from ansible.playbook.included_file import IncludedFile -from ansible.plugins import action_loader +from ansible.plugins.loader import action_loader from ansible.plugins.strategy import StrategyBase from ansible.template import Templar from ansible.module_utils._text import to_text diff --git a/lib/ansible/plugins/strategy/linear.py b/lib/ansible/plugins/strategy/linear.py index 76bb946ab8..250be3d6e0 100644 --- a/lib/ansible/plugins/strategy/linear.py +++ b/lib/ansible/plugins/strategy/linear.py @@ -38,7 +38,7 @@ from ansible.module_utils._text import to_text from ansible.playbook.block import Block from ansible.playbook.included_file import IncludedFile from ansible.playbook.task import Task -from ansible.plugins import action_loader +from ansible.plugins.loader import action_loader from ansible.plugins.strategy import StrategyBase from ansible.template import Templar diff --git a/lib/ansible/template/__init__.py b/lib/ansible/template/__init__.py index 19741a997a..c23bdf43ce 100644 --- a/lib/ansible/template/__init__.py +++ b/lib/ansible/template/__init__.py @@ -45,7 +45,7 @@ from ansible import constants as C from ansible.errors import AnsibleError, AnsibleFilterError, AnsibleUndefinedVariable from ansible.module_utils.six import string_types, text_type from ansible.module_utils._text import to_native, to_text, to_bytes -from ansible.plugins import filter_loader, lookup_loader, test_loader +from ansible.plugins.loader import filter_loader, lookup_loader, test_loader from ansible.template.safe_eval import safe_eval from ansible.template.template import AnsibleJ2Template from ansible.template.vars import AnsibleJ2Vars diff --git a/lib/ansible/template/safe_eval.py b/lib/ansible/template/safe_eval.py index 2f819b40b1..d2ccc09683 100644 --- a/lib/ansible/template/safe_eval.py +++ b/lib/ansible/template/safe_eval.py @@ -24,7 +24,7 @@ import sys from ansible import constants as C from ansible.module_utils.six import string_types from ansible.module_utils.six.moves import builtins -from ansible.plugins import filter_loader, test_loader +from ansible.plugins.loader import filter_loader, test_loader def safe_eval(expr, locals={}, include_exceptions=False): diff --git a/lib/ansible/utils/plugin_docs.py b/lib/ansible/utils/plugin_docs.py index f59f115401..717f023f4d 100644 --- a/lib/ansible/utils/plugin_docs.py +++ b/lib/ansible/utils/plugin_docs.py @@ -20,15 +20,12 @@ from __future__ import (absolute_import, division, print_function) __metaclass__ = type -import ast -import yaml - from collections import MutableMapping, MutableSet, MutableSequence from ansible.module_utils.six import string_types -from ansible.parsing.metadata import extract_metadata +from ansible.parsing.plugin_docs import read_docstring from ansible.parsing.yaml.loader import AnsibleLoader -from ansible.plugins import fragment_loader +from ansible.plugins.loader import fragment_loader try: from __main__ import display @@ -93,94 +90,13 @@ def add_fragments(doc, filename): def get_docstring(filename, verbose=False): """ - Search for assignment of the DOCUMENTATION and EXAMPLES variables - in the given file. - Parse DOCUMENTATION from YAML and return the YAML doc or None - together with EXAMPLES, as plain text. - - DOCUMENTATION can be extended using documentation fragments - loaded by the PluginLoader from the module_docs_fragments - directory. + DOCUMENTATION can be extended using documentation fragments loaded by the PluginLoader from the module_docs_fragments directory. """ - # FIXME: Should refactor this so that we have a docstring parsing - # function and a separate variable parsing function - # Can have a function one higher that invokes whichever is needed - # - # Should look roughly like this: - # get_plugin_doc(filename, verbose=False) - # documentation = extract_docstring(plugin_ast, identifier, verbose=False) - # if not documentation and not (filter or test): - # documentation = extract_variables(plugin_ast) - # documentation['metadata'] = extract_metadata(plugin_ast) + data = read_docstring(filename, verbose=verbose) - data = { - 'doc': None, - 'plainexamples': None, - 'returndocs': None, - 'metadata': None - } - - string_to_vars = { - 'DOCUMENTATION': 'doc', - 'EXAMPLES': 'plainexamples', - 'RETURN': 'returndocs', - } - - try: - b_module_data = open(filename, 'rb').read() - M = ast.parse(b_module_data) - try: - display.debug('Attempt first docstring is yaml docs') - docstring = yaml.load(M.body[0].value.s) - for string in string_to_vars.keys(): - if string in docstring: - data[string_to_vars[string]] = docstring[string] - display.debug('assigned :%s' % string_to_vars[string]) - except Exception as e: - display.debug('failed docstring parsing: %s' % str(e)) - - if 'docs' not in data or not data['docs']: - display.debug('Fallback to vars parsing') - for child in M.body: - if isinstance(child, ast.Assign): - for t in child.targets: - try: - theid = t.id - except AttributeError: - # skip errors can happen when trying to use the normal code - display.warning("Failed to assign id for %s on %s, skipping" % (t, filename)) - continue - - if theid in string_to_vars: - varkey = string_to_vars[theid] - if isinstance(child.value, ast.Dict): - data[varkey] = ast.literal_eval(child.value) - else: - if theid == 'DOCUMENTATION': - # string should be yaml - data[varkey] = AnsibleLoader(child.value.s, file_name=filename).get_single_data() - else: - # not yaml, should be a simple string - data[varkey] = child.value.s - display.debug('assigned :%s' % varkey) - - # Metadata is per-file rather than per-plugin/function - data['metadata'] = extract_metadata(module_ast=M)[0] - - # add fragments to documentation - if data['doc']: - add_fragments(data['doc'], filename) - - # remove version - if data['metadata']: - for x in ('version', 'metadata_version'): - if x in data['metadata']: - del data['metadata'][x] - except Exception as e: - display.error("unable to parse %s" % filename) - if verbose is True: - display.display("unable to parse %s" % filename) - raise + # add fragments to documentation + if data.get('doc', False): + add_fragments(data['doc'], filename) return data['doc'], data['plainexamples'], data['returndocs'], data['metadata'] diff --git a/lib/ansible/vars/manager.py b/lib/ansible/vars/manager.py index af0f4dcde6..0659b825e0 100644 --- a/lib/ansible/vars/manager.py +++ b/lib/ansible/vars/manager.py @@ -37,7 +37,7 @@ from ansible.inventory.host import Host from ansible.inventory.helpers import sort_groups, get_group_vars from ansible.module_utils._text import to_native from ansible.module_utils.six import iteritems, string_types, text_type -from ansible.plugins import lookup_loader, vars_loader +from ansible.plugins.loader import lookup_loader, vars_loader from ansible.plugins.cache import FactCache from ansible.template import Templar from ansible.utils.listify import listify_lookup_plugin_terms @@ -86,6 +86,21 @@ def strip_internal_keys(dirty): return clean +def remove_internal_keys(data): + ''' + More nuanced version of strip_internal_keys + ''' + for key in list(data.keys()): + if (key.startswith('_ansible_') and key != '_ansible_parsed') or key in C.INTERNAL_RESULT_KEYS: + display.warning("Removed unexpected internal key in module return: %s = %s" % (key, data[key])) + del data[key] + + # remove bad/empty internal keys + for key in ['warnings', 'deprecations']: + if key in data and not data[key]: + del data[key] + + class VariableManager: def __init__(self, loader=None, inventory=None): diff --git a/setup.py b/setup.py index 0262b7d3ef..4345e4206e 100644 --- a/setup.py +++ b/setup.py @@ -190,8 +190,7 @@ setup( 'galaxy/data/*/*/.*', 'galaxy/data/*/*/*.*', 'galaxy/data/*/tests/inventory', - 'config/data/*.yaml', - 'config/data/*.yml', + 'config/base.yml', ], }, classifiers=[ diff --git a/test/runner/lib/ansible_util.py b/test/runner/lib/ansible_util.py index 2cd7013fd8..c90772c8b3 100644 --- a/test/runner/lib/ansible_util.py +++ b/test/runner/lib/ansible_util.py @@ -21,20 +21,19 @@ def ansible_environment(args, color=True): if not path.startswith(ansible_path + os.pathsep): path = ansible_path + os.pathsep + path - ansible_config = '/dev/null' - if os.path.isfile('test/integration/%s.cfg' % args.command): - ansible_config = os.path.abspath('test/integration/%s.cfg' % args.command) - ansible = dict( ANSIBLE_FORCE_COLOR='%s' % 'true' if args.color and color else 'false', ANSIBLE_DEPRECATION_WARNINGS='false', - ANSIBLE_CONFIG=ansible_config, ANSIBLE_HOST_KEY_CHECKING='false', PYTHONPATH=os.path.abspath('lib'), PAGER='/bin/cat', PATH=path, ) + if os.path.isfile('test/integration/%s.cfg' % args.command): + ansible_config = os.path.abspath('test/integration/%s.cfg' % args.command) + ansible['ANSIBLE_CONFIG'] = ansible_config + env.update(ansible) if args.debug: diff --git a/test/sanity/pep8/legacy-files.txt b/test/sanity/pep8/legacy-files.txt index 6352547ef1..b6db79130e 100644 --- a/test/sanity/pep8/legacy-files.txt +++ b/test/sanity/pep8/legacy-files.txt @@ -1,4 +1,5 @@ lib/ansible/cli/config.py +lib/ansible/constants.py lib/ansible/config/data.py lib/ansible/config/manager.py lib/ansible/modules/cloud/amazon/_ec2_ami_search.py diff --git a/test/units/executor/test_task_executor.py b/test/units/executor/test_task_executor.py index 169e647db8..3b75636b82 100644 --- a/test/units/executor/test_task_executor.py +++ b/test/units/executor/test_task_executor.py @@ -24,7 +24,7 @@ from ansible.compat.tests.mock import patch, MagicMock from ansible.errors import AnsibleError, AnsibleParserError from ansible.executor.task_executor import TaskExecutor from ansible.playbook.play_context import PlayContext -from ansible.plugins import action_loader, lookup_loader +from ansible.plugins.loader import action_loader, lookup_loader from ansible.parsing.yaml.objects import AnsibleUnicode from units.mock.loader import DictDataLoader diff --git a/test/units/modules/cloud/amazon/test_data_pipeline.py b/test/units/modules/cloud/amazon/test_data_pipeline.py index 2ebc49ea4c..bf9ac230fa 100644 --- a/test/units/modules/cloud/amazon/test_data_pipeline.py +++ b/test/units/modules/cloud/amazon/test_data_pipeline.py @@ -16,14 +16,20 @@ # along with Ansible. If not, see . import pytest -import boto3 import os import json import collections from . placebo_fixtures import placeboify, maybe_sleep +from nose.plugins.skip import SkipTest + from ansible.modules.cloud.amazon import data_pipeline from ansible.module_utils._text import to_text +try: + import boto3 +except ImportError: + raise SkipTest("test_api_gateway.py requires the `boto3` and `botocore` modules") + @pytest.fixture(scope='module') def dp_setup(): diff --git a/test/units/modules/packaging/os/test_rhn_register.py b/test/units/modules/packaging/os/test_rhn_register.py index 3ee82d46f3..d3f92f0fe4 100644 --- a/test/units/modules/packaging/os/test_rhn_register.py +++ b/test/units/modules/packaging/os/test_rhn_register.py @@ -3,6 +3,7 @@ import json from ansible.compat.tests import unittest from ansible.compat.tests.mock import PropertyMock, patch, mock_open from ansible.module_utils import basic +from ansible.module_utils._text import to_native from ansible.module_utils.six.moves import xmlrpc_client from ansible.modules.packaging.os import rhn_register @@ -96,7 +97,7 @@ class TestRhnRegister(unittest.TestCase): orig_import = __import__ with patch('__builtin__.__import__', side_effect=mock_import): rhn = self.module.Rhn() - self.assertEqual('123456789', rhn.systemid) + self.assertEqual('123456789', to_native(rhn.systemid)) def test_without_required_parameters(self): """Failure must occurs when all parameters are missing""" diff --git a/test/units/plugins/filter/test_ipaddr.py b/test/units/plugins/filter/test_ipaddr.py index 5ef36db80f..9643a0d8bf 100644 --- a/test/units/plugins/filter/test_ipaddr.py +++ b/test/units/plugins/filter/test_ipaddr.py @@ -20,6 +20,11 @@ __metaclass__ = type from ansible.compat.tests import unittest from ansible.plugins.filter.ipaddr import (ipaddr, _netmask_query, nthhost, next_nth_usable, previous_nth_usable, network_in_usable, network_in_network) +try: + import netaddr +except ImportError: + from nose.plugins.skip import SkipTest + raise SkipTest("This test requires the `netaddr` python library") class TestIpFilter(unittest.TestCase): diff --git a/test/units/plugins/lookup/test_password.py b/test/units/plugins/lookup/test_password.py index 30f27a85fb..b19e3460fa 100644 --- a/test/units/plugins/lookup/test_password.py +++ b/test/units/plugins/lookup/test_password.py @@ -29,7 +29,7 @@ from ansible.compat.tests.mock import mock_open, patch from ansible.errors import AnsibleError from ansible.module_utils.six import text_type from ansible.module_utils.six.moves import builtins -from ansible.plugins import PluginLoader +from ansible.plugins.loader import PluginLoader from ansible.plugins.lookup import password from ansible.utils import encrypt diff --git a/test/units/plugins/test_plugins.py b/test/units/plugins/test_plugins.py index 64c12246c5..def2544ea8 100644 --- a/test/units/plugins/test_plugins.py +++ b/test/units/plugins/test_plugins.py @@ -24,7 +24,7 @@ import os from ansible.compat.tests import BUILTINS, unittest from ansible.compat.tests.mock import mock_open, patch, MagicMock -from ansible.plugins import MODULE_CACHE, PATH_CACHE, PLUGIN_PATH_CACHE, PluginLoader +from ansible.plugins.loader import MODULE_CACHE, PATH_CACHE, PLUGIN_PATH_CACHE, PluginLoader class TestErrors(unittest.TestCase):