Merge upstream changes
This commit is contained in:
parent
403f4881ee
commit
4e7542af37
1 changed files with 10 additions and 10 deletions
|
@ -523,13 +523,11 @@ class SSLValidationHandler(urllib2.BaseHandler):
|
||||||
# Rewrite of fetch_url to not require the module environment
|
# Rewrite of fetch_url to not require the module environment
|
||||||
def open_url(url, data=None, headers=None, method=None, use_proxy=True,
|
def open_url(url, data=None, headers=None, method=None, use_proxy=True,
|
||||||
force=False, last_mod_time=None, timeout=10, validate_certs=True,
|
force=False, last_mod_time=None, timeout=10, validate_certs=True,
|
||||||
url_username=None, url_password=None, http_agent=None):
|
url_username=None, url_password=None, http_agent=None, force_basic_auth=False):
|
||||||
force_basic_auth = dict(required=False, type='bool')
|
|
||||||
'''
|
'''
|
||||||
Fetches a file from an HTTP/FTP server using urllib2
|
Fetches a file from an HTTP/FTP server using urllib2
|
||||||
'''
|
'''
|
||||||
handlers = []
|
handlers = []
|
||||||
|
|
||||||
# FIXME: change the following to use the generic_urlparse function
|
# FIXME: change the following to use the generic_urlparse function
|
||||||
# to remove the indexed references for 'parsed'
|
# to remove the indexed references for 'parsed'
|
||||||
parsed = urlparse.urlparse(url)
|
parsed = urlparse.urlparse(url)
|
||||||
|
@ -556,7 +554,6 @@ def open_url(url, data=None, headers=None, method=None, use_proxy=True,
|
||||||
|
|
||||||
if parsed[0] != 'ftp':
|
if parsed[0] != 'ftp':
|
||||||
username = url_username
|
username = url_username
|
||||||
force_basic_auth = module.params.get('force_basic_auth', False)
|
|
||||||
|
|
||||||
if username:
|
if username:
|
||||||
password = url_password
|
password = url_password
|
||||||
|
@ -659,6 +656,8 @@ def url_argument_spec():
|
||||||
validate_certs = dict(default='yes', type='bool'),
|
validate_certs = dict(default='yes', type='bool'),
|
||||||
url_username = dict(required=False),
|
url_username = dict(required=False),
|
||||||
url_password = dict(required=False),
|
url_password = dict(required=False),
|
||||||
|
force_basic_auth = dict(required=False, type='bool', default='no'),
|
||||||
|
|
||||||
)
|
)
|
||||||
|
|
||||||
def fetch_url(module, url, data=None, headers=None, method=None,
|
def fetch_url(module, url, data=None, headers=None, method=None,
|
||||||
|
@ -678,6 +677,7 @@ def fetch_url(module, url, data=None, headers=None, method=None,
|
||||||
username = module.params.get('url_username', '')
|
username = module.params.get('url_username', '')
|
||||||
password = module.params.get('url_password', '')
|
password = module.params.get('url_password', '')
|
||||||
http_agent = module.params.get('http_agent', None)
|
http_agent = module.params.get('http_agent', None)
|
||||||
|
force_basic_auth = module.params.get('force_basic_auth', '')
|
||||||
|
|
||||||
r = None
|
r = None
|
||||||
info = dict(url=url)
|
info = dict(url=url)
|
||||||
|
@ -685,7 +685,7 @@ def fetch_url(module, url, data=None, headers=None, method=None,
|
||||||
r = open_url(url, data=data, headers=headers, method=method,
|
r = open_url(url, data=data, headers=headers, method=method,
|
||||||
use_proxy=use_proxy, force=force, last_mod_time=last_mod_time, timeout=timeout,
|
use_proxy=use_proxy, force=force, last_mod_time=last_mod_time, timeout=timeout,
|
||||||
validate_certs=validate_certs, url_username=username,
|
validate_certs=validate_certs, url_username=username,
|
||||||
url_password=password, http_agent=http_agent)
|
url_password=password, http_agent=http_agent, force_basic_auth=force_basic_auth)
|
||||||
info.update(r.info())
|
info.update(r.info())
|
||||||
info['url'] = r.geturl() # The URL goes in too, because of redirects.
|
info['url'] = r.geturl() # The URL goes in too, because of redirects.
|
||||||
info.update(dict(msg="OK (%s bytes)" % r.headers.get('Content-Length', 'unknown'), status=200))
|
info.update(dict(msg="OK (%s bytes)" % r.headers.get('Content-Length', 'unknown'), status=200))
|
||||||
|
|
Loading…
Reference in a new issue