1
0
Fork 0
mirror of https://github.com/ansible-collections/community.general.git synced 2026-03-22 13:19:13 +00:00
community.general/plugins/modules/maven_artifact.py
patchback[bot] baddfa5a80
[PR #11501/ed7ccbe3 backport][stable-12] maven_artifact: resolve SNAPSHOT to latest using snapshot metadata block (#11508)
maven_artifact: resolve SNAPSHOT to latest using snapshot metadata block (#11501)

* fix(maven_artifact): resolve SNAPSHOT to latest using snapshot metadata block

Prefer the <snapshot> block (timestamp + buildNumber) from maven-metadata.xml
which always points to the latest build, instead of scanning <snapshotVersions>
and returning on the first match. Repositories like GitHub Packages keep all
historical entries in <snapshotVersions> (oldest first), causing the module to
resolve to the oldest snapshot instead of the latest.

Fixes #5117
Fixes #11489

* fix(maven_artifact): address review feedback

- Check both timestamp and buildNumber before using snapshot block,
  preventing IndexError when buildNumber is missing
- Remove unreliable snapshotVersions scanning fallback; use literal
  -SNAPSHOT version for non-unique snapshot repos instead
- Add tests for incomplete snapshot block and non-SNAPSHOT versions

* fix(maven_artifact): restore snapshotVersions scanning with last-match

Restore <snapshotVersions> scanning as primary resolution (needed for
per-extension accuracy per MNG-5459), but collect the last match instead
of returning on the first. Fall back to <snapshot> block when no
<snapshotVersions> match is found, then to literal -SNAPSHOT version.

* docs: update changelog fragment to match final implementation

* fix(maven_artifact): use updated timestamp for snapshot resolution

Use the <updated> attribute to select the newest snapshotVersion entry
instead of relying on list order. This works independently of how the
repository manager sorts entries in maven-metadata.xml.

Also fix test docstring and update changelog fragment per reviewer
feedback.

* test(maven_artifact): shuffle entries to verify updated timestamp sorting

Reorder snapshotVersion entries so the newest JAR is in the middle,
not at the end. This ensures the test actually validates that resolution
uses the <updated> timestamp rather than relying on list position.

(cherry picked from commit ed7ccbe3d4)

Co-authored-by: Adam R. <ariwk@protonmail.com>
2026-02-14 21:14:36 +01:00

791 lines
30 KiB
Python

#!/usr/bin/python
# Copyright (c) 2014, Chris Schmidt <chris.schmidt () contrastsecurity.com>
#
# Built using https://github.com/hamnis/useful-scripts/blob/master/python/download-maven-artifact
# as a reference and starting point.
#
# GNU General Public License v3.0+ (see LICENSES/GPL-3.0-or-later.txt or https://www.gnu.org/licenses/gpl-3.0.txt)
# SPDX-License-Identifier: GPL-3.0-or-later
from __future__ import annotations
DOCUMENTATION = r"""
module: maven_artifact
short_description: Downloads an Artifact from a Maven Repository
description:
- Downloads an artifact from a maven repository given the maven coordinates provided to the module.
- Can retrieve snapshots or release versions of the artifact and resolve the latest available version if one is not available.
author: "Chris Schmidt (@chrisisbeef)"
requirements:
- lxml
- boto if using a S3 repository (V(s3://...))
attributes:
check_mode:
support: none
diff_mode:
support: none
options:
group_id:
type: str
description:
- The Maven groupId coordinate.
required: true
artifact_id:
type: str
description:
- The maven artifactId coordinate.
required: true
version:
type: str
description:
- The maven version coordinate.
- Mutually exclusive with O(version_by_spec).
version_by_spec:
type: str
description:
- The maven dependency version ranges.
- See supported version ranges on U(https://cwiki.apache.org/confluence/display/MAVENOLD/Dependency+Mediation+and+Conflict+Resolution).
- The range type V((,1.0],[1.2,\)) and V((,1.1\),(1.1,\)) is not supported.
- Mutually exclusive with O(version).
version_added: '0.2.0'
classifier:
type: str
description:
- The maven classifier coordinate.
default: ''
extension:
type: str
description:
- The maven type/extension coordinate.
default: jar
repository_url:
type: str
description:
- The URL of the Maven Repository to download from.
- Use V(s3://...) if the repository is hosted on Amazon S3.
- Use V(file://...) if the repository is local.
default: https://repo1.maven.org/maven2
username:
type: str
description:
- The username to authenticate as to the Maven Repository. Use AWS secret key of the repository is hosted on S3.
aliases: ["aws_secret_key"]
password:
type: str
description:
- The password to authenticate with to the Maven Repository. Use AWS secret access key of the repository is hosted on
S3.
aliases: ["aws_secret_access_key"]
headers:
description:
- Add custom HTTP headers to a request in hash/dict format.
type: dict
force_basic_auth:
description:
- C(httplib2), the library used by the URI module only sends authentication information when a webservice responds to
an initial request with a 401 status. Since some basic auth services do not properly send a 401, logins fail. This
option forces the sending of the Basic authentication header upon initial request.
default: false
type: bool
version_added: '0.2.0'
dest:
type: path
description:
- The path where the artifact should be written to.
- If file mode or ownerships are specified and destination path already exists, they affect the downloaded file.
required: true
state:
type: str
description:
- The desired state of the artifact.
default: present
choices: [present, absent]
timeout:
type: int
description:
- Specifies a timeout in seconds for the connection attempt.
default: 10
validate_certs:
description:
- If V(false), SSL certificates are not validated. This should only be set to V(false) when no other option exists.
type: bool
default: true
client_cert:
description:
- PEM formatted certificate chain file to be used for SSL client authentication.
- This file can also include the key as well, and if the key is included, O(client_key) is not required.
type: path
version_added: '1.3.0'
client_key:
description:
- PEM formatted file that contains your private key to be used for SSL client authentication.
- If O(client_cert) contains both the certificate and key, this option is not required.
type: path
version_added: '1.3.0'
keep_name:
description:
- If V(true), the downloaded artifact's name is preserved, in other words the version number remains part of it.
- This option only has effect when O(dest) is a directory and O(version) is set to V(latest) or O(version_by_spec) is
defined.
type: bool
default: false
verify_checksum:
type: str
description:
- If V(never), the MD5/SHA1 checksum is never downloaded and verified.
- If V(download), the MD5/SHA1 checksum is downloaded and verified only after artifact download. This is the default.
- If V(change), the MD5/SHA1 checksum is downloaded and verified if the destination already exist, to verify if they
are identical. This was the behaviour before 2.6. Since it downloads the checksum before (maybe) downloading the artifact,
and since some repository software, when acting as a proxy/cache, return a 404 error if the artifact has not been
cached yet, it may fail unexpectedly. If you still need it, you should consider using V(always) instead - if you deal
with a checksum, it is better to use it to verify integrity after download.
- V(always) combines V(download) and V(change).
default: 'download'
choices: ['never', 'download', 'change', 'always']
checksum_alg:
type: str
description:
- If V(md5), checksums use the MD5 algorithm. This is the default.
- If V(sha1), checksums use the SHA1 algorithm. This can be used on systems configured to use FIPS-compliant algorithms,
since MD5 is blocked on such systems.
default: 'md5'
choices: ['md5', 'sha1']
version_added: 3.2.0
unredirected_headers:
type: list
elements: str
version_added: 5.2.0
description:
- A list of headers that should not be included in the redirection. This headers are sent to the C(fetch_url) function.
- On ansible-core version 2.12 or later, the default of this option is V([Authorization, Cookie]).
- Useful if the redirection URL does not need to have sensitive headers in the request.
- Requires ansible-core version 2.12 or later.
directory_mode:
type: str
description:
- Filesystem permission mode applied recursively to O(dest) when it is a directory.
extends_documentation_fragment:
- ansible.builtin.files
- community.general.attributes
"""
EXAMPLES = r"""
- name: Download the latest version of the JUnit framework artifact from Maven Central
community.general.maven_artifact:
group_id: junit
artifact_id: junit
dest: /tmp/junit-latest.jar
- name: Download JUnit 4.11 from Maven Central
community.general.maven_artifact:
group_id: junit
artifact_id: junit
version: 4.11
dest: /tmp/junit-4.11.jar
- name: Download an artifact from a private repository requiring authentication
community.general.maven_artifact:
group_id: com.company
artifact_id: library-name
repository_url: 'https://repo.company.com/maven'
username: user
password: pass
dest: /tmp/library-name-latest.jar
- name: Download an artifact from a private repository requiring certificate authentication
community.general.maven_artifact:
group_id: com.company
artifact_id: library-name
repository_url: 'https://repo.company.com/maven'
client_cert: /path/to/cert.pem
client_key: /path/to/key.pem
dest: /tmp/library-name-latest.jar
- name: Download a WAR File to the Tomcat webapps directory to be deployed
community.general.maven_artifact:
group_id: com.company
artifact_id: web-app
extension: war
repository_url: 'https://repo.company.com/maven'
dest: /var/lib/tomcat7/webapps/web-app.war
- name: Keep a downloaded artifact's name, i.e. retain the version
community.general.maven_artifact:
version: latest
artifact_id: spring-core
group_id: org.springframework
dest: /tmp/
keep_name: true
- name: Download the latest version of the JUnit framework artifact from Maven local
community.general.maven_artifact:
group_id: junit
artifact_id: junit
dest: /tmp/junit-latest.jar
repository_url: "file://{{ lookup('env','HOME') }}/.m2/repository"
- name: Download the latest version between 3.8 and 4.0 (exclusive) of the JUnit framework artifact from Maven Central
community.general.maven_artifact:
group_id: junit
artifact_id: junit
version_by_spec: "[3.8,4.0)"
dest: /tmp/
"""
import hashlib
import os
import posixpath
import re
import shutil
import tempfile
import traceback
from re import match
from urllib.parse import urlparse
from ansible.module_utils.ansible_release import __version__ as ansible_version
LXML_ETREE_IMP_ERR = None
try:
from lxml import etree
HAS_LXML_ETREE = True
except ImportError:
LXML_ETREE_IMP_ERR = traceback.format_exc()
HAS_LXML_ETREE = False
BOTO_IMP_ERR = None
try:
import boto3
HAS_BOTO = True
except ImportError:
BOTO_IMP_ERR = traceback.format_exc()
HAS_BOTO = False
SEMANTIC_VERSION_IMP_ERR = None
try:
from semantic_version import Spec, Version
HAS_SEMANTIC_VERSION = True
except ImportError:
SEMANTIC_VERSION_IMP_ERR = traceback.format_exc()
HAS_SEMANTIC_VERSION = False
from ansible.module_utils.basic import AnsibleModule, missing_required_lib
from ansible.module_utils.common.text.converters import to_bytes, to_text
from ansible.module_utils.urls import fetch_url
def split_pre_existing_dir(dirname):
"""
Return the first pre-existing directory and a list of the new directories that will be created.
"""
head, tail = os.path.split(dirname)
b_head = to_bytes(head, errors="surrogate_or_strict")
if not os.path.exists(b_head):
if head == dirname:
return None, [head]
else:
(pre_existing_dir, new_directory_list) = split_pre_existing_dir(head)
else:
return head, [tail]
new_directory_list.append(tail)
return pre_existing_dir, new_directory_list
def adjust_recursive_directory_permissions(pre_existing_dir, new_directory_list, module, directory_args, changed):
"""
Walk the new directories list and make sure that permissions are as we would expect
"""
if new_directory_list:
first_sub_dir = new_directory_list.pop(0)
if not pre_existing_dir:
working_dir = first_sub_dir
else:
working_dir = os.path.join(pre_existing_dir, first_sub_dir)
directory_args["path"] = working_dir
changed = module.set_fs_attributes_if_different(directory_args, changed)
changed = adjust_recursive_directory_permissions(
working_dir, new_directory_list, module, directory_args, changed
)
return changed
class Artifact:
def __init__(self, group_id, artifact_id, version, version_by_spec, classifier="", extension="jar"):
if not group_id:
raise ValueError("group_id must be set")
if not artifact_id:
raise ValueError("artifact_id must be set")
self.group_id = group_id
self.artifact_id = artifact_id
self.version = version
self.version_by_spec = version_by_spec
self.classifier = classifier
if not extension:
self.extension = "jar"
else:
self.extension = extension
def is_snapshot(self):
return self.version and self.version.endswith("SNAPSHOT")
def path(self, with_version=True):
base = posixpath.join(self.group_id.replace(".", "/"), self.artifact_id)
if with_version and self.version:
timestamp_version_match = re.match("^(.*-)?([0-9]{8}\\.[0-9]{6}-[0-9]+)$", self.version)
if timestamp_version_match:
base = posixpath.join(base, f"{timestamp_version_match.group(1)}SNAPSHOT")
else:
base = posixpath.join(base, self.version)
return base
def _generate_filename(self):
filename = f"{self.artifact_id}-{self.classifier}.{self.extension}"
if not self.classifier:
filename = f"{self.artifact_id}.{self.extension}"
return filename
def get_filename(self, filename=None):
if not filename:
filename = self._generate_filename()
elif os.path.isdir(filename):
filename = os.path.join(filename, self._generate_filename())
return filename
def __str__(self):
result = f"{self.group_id}:{self.artifact_id}:{self.version}"
if self.classifier:
result = f"{self.group_id}:{self.artifact_id}:{self.extension}:{self.classifier}:{self.version}"
elif self.extension != "jar":
result = f"{self.group_id}:{self.artifact_id}:{self.extension}:{self.version}"
return result
@staticmethod
def parse(input):
parts = input.split(":")
if len(parts) >= 3:
g = parts[0]
a = parts[1]
v = parts[-1]
t = None
c = None
if len(parts) == 4:
t = parts[2]
if len(parts) == 5:
t = parts[2]
c = parts[3]
return Artifact(g, a, v, c, t)
else:
return None
class MavenDownloader:
def __init__(self, module, base, local=False, headers=None):
self.module = module
if base.endswith("/"):
base = base.rstrip("/")
self.base = base
self.local = local
self.headers = headers
self.user_agent = f"Ansible {ansible_version} maven_artifact"
self.latest_version_found = None
self.metadata_file_name = "maven-metadata-local.xml" if local else "maven-metadata.xml"
def find_version_by_spec(self, artifact):
path = f"/{artifact.path(False)}/{self.metadata_file_name}"
content = self._getContent(self.base + path, f"Failed to retrieve the maven metadata file: {path}")
xml = etree.fromstring(content)
original_versions = xml.xpath("/metadata/versioning/versions/version/text()")
versions = []
for version in original_versions:
try:
versions.append(Version.coerce(version))
except ValueError:
# This means that version string is not a valid semantic versioning
pass
parse_versions_syntax = {
# example -> (,1.0]
r"^\(,(?P<upper_bound>[0-9.]*)]$": "<={upper_bound}",
# example -> 1.0
r"^(?P<version>[0-9.]*)$": "~={version}",
# example -> [1.0]
r"^\[(?P<version>[0-9.]*)\]$": "=={version}",
# example -> [1.2, 1.3]
r"^\[(?P<lower_bound>[0-9.]*),\s*(?P<upper_bound>[0-9.]*)\]$": ">={lower_bound},<={upper_bound}",
# example -> [1.2, 1.3)
r"^\[(?P<lower_bound>[0-9.]*),\s*(?P<upper_bound>[0-9.]+)\)$": ">={lower_bound},<{upper_bound}",
# example -> [1.5,)
r"^\[(?P<lower_bound>[0-9.]*),\)$": ">={lower_bound}",
}
for regex, spec_format in parse_versions_syntax.items():
regex_result = match(regex, artifact.version_by_spec)
if regex_result:
spec = Spec(spec_format.format(**regex_result.groupdict()))
selected_version = spec.select(versions)
if not selected_version:
raise ValueError(f"No version found with this spec version: {artifact.version_by_spec}")
# To deal when repos on maven don't have patch number on first build (e.g. 3.8 instead of 3.8.0)
if str(selected_version) not in original_versions:
selected_version.patch = None
return str(selected_version)
raise ValueError(f"The spec version {artifact.version_by_spec} is not supported! ")
def find_latest_version_available(self, artifact):
if self.latest_version_found:
return self.latest_version_found
path = f"/{artifact.path(False)}/{self.metadata_file_name}"
content = self._getContent(self.base + path, f"Failed to retrieve the maven metadata file: {path}")
xml = etree.fromstring(content)
v = xml.xpath("/metadata/versioning/versions/version[last()]/text()")
if v:
self.latest_version_found = v[0]
return v[0]
def find_uri_for_artifact(self, artifact):
if artifact.version_by_spec:
artifact.version = self.find_version_by_spec(artifact)
if artifact.version == "latest":
artifact.version = self.find_latest_version_available(artifact)
if artifact.is_snapshot():
if self.local:
return self._uri_for_artifact(artifact, artifact.version)
path = f"/{artifact.path()}/{self.metadata_file_name}"
content = self._getContent(self.base + path, f"Failed to retrieve the maven metadata file: {path}")
xml = etree.fromstring(content)
candidates = []
for snapshot_artifact in xml.xpath("/metadata/versioning/snapshotVersions/snapshotVersion"):
classifier = snapshot_artifact.xpath("classifier/text()")
extension = snapshot_artifact.xpath("extension/text()")
if (classifier[0] if classifier else "") == artifact.classifier and (
extension[0] if extension else ""
) == artifact.extension:
value = snapshot_artifact.xpath("value/text()")
updated = snapshot_artifact.xpath("updated/text()")
if value:
candidates.append((updated[0] if updated else "", value[0]))
if candidates:
# updated is yyyymmddHHMMSS, so lexical max == newest
return self._uri_for_artifact(artifact, max(candidates, key=lambda item: item[0])[1])
timestamp_xmlpath = xml.xpath("/metadata/versioning/snapshot/timestamp/text()")
build_number_xmlpath = xml.xpath("/metadata/versioning/snapshot/buildNumber/text()")
if timestamp_xmlpath and build_number_xmlpath:
timestamp = timestamp_xmlpath[0]
build_number = build_number_xmlpath[0]
return self._uri_for_artifact(
artifact, artifact.version.replace("SNAPSHOT", f"{timestamp}-{build_number}")
)
return self._uri_for_artifact(artifact, artifact.version)
def _uri_for_artifact(self, artifact, version=None):
if artifact.is_snapshot() and not version:
raise ValueError(f"Expected uniqueversion for snapshot artifact {artifact}")
elif not artifact.is_snapshot():
version = artifact.version
if artifact.classifier:
return posixpath.join(
self.base,
artifact.path(),
f"{artifact.artifact_id}-{version}-{artifact.classifier}.{artifact.extension}",
)
return posixpath.join(self.base, artifact.path(), f"{artifact.artifact_id}-{version}.{artifact.extension}")
# for small files, directly get the full content
def _getContent(self, url, failmsg, force=True):
if self.local:
parsed_url = urlparse(url)
if os.path.isfile(parsed_url.path):
with open(parsed_url.path, "rb") as f:
return f.read()
if force:
raise ValueError(f"{failmsg} because can not find file: {url}")
return None
response = self._request(url, failmsg, force)
if response:
return response.read()
return None
# only for HTTP request
def _request(self, url, failmsg, force=True):
url_to_use = url
parsed_url = urlparse(url)
if parsed_url.scheme == "s3":
parsed_url = urlparse(url)
bucket_name = parsed_url.netloc
key_name = parsed_url.path[1:]
client = boto3.client(
"s3",
aws_access_key_id=self.module.params.get("username", ""),
aws_secret_access_key=self.module.params.get("password", ""),
)
url_to_use = client.generate_presigned_url(
"get_object", Params={"Bucket": bucket_name, "Key": key_name}, ExpiresIn=10
)
req_timeout = self.module.params.get("timeout")
# Hack to add parameters in the way that fetch_url expects
self.module.params["url_username"] = self.module.params.get("username", "")
self.module.params["url_password"] = self.module.params.get("password", "")
self.module.params["http_agent"] = self.user_agent
kwargs = {}
if self.module.params["unredirected_headers"]:
kwargs["unredirected_headers"] = self.module.params["unredirected_headers"]
response, info = fetch_url(self.module, url_to_use, timeout=req_timeout, headers=self.headers, **kwargs)
if info["status"] == 200:
return response
if force:
raise ValueError(f"{failmsg} because of {info['msg']}for URL {url_to_use}")
return None
def download(self, tmpdir, artifact, verify_download, filename=None, checksum_alg="md5"):
if (not artifact.version and not artifact.version_by_spec) or artifact.version == "latest":
artifact = Artifact(
artifact.group_id,
artifact.artifact_id,
self.find_latest_version_available(artifact),
None,
artifact.classifier,
artifact.extension,
)
url = self.find_uri_for_artifact(artifact)
tempfd, tempname = tempfile.mkstemp(dir=tmpdir)
try:
# copy to temp file
if self.local:
parsed_url = urlparse(url)
if os.path.isfile(parsed_url.path):
shutil.copy2(parsed_url.path, tempname)
else:
return f"Can not find local file: {parsed_url.path}"
else:
response = self._request(url, f"Failed to download artifact {artifact}")
with os.fdopen(tempfd, "wb") as f:
shutil.copyfileobj(response, f)
if verify_download:
invalid_checksum = self.is_invalid_checksum(tempname, url, checksum_alg)
if invalid_checksum:
# if verify_change was set, the previous file would be deleted
os.remove(tempname)
return invalid_checksum
except Exception as e:
os.remove(tempname)
raise e
# all good, now copy temp file to target
shutil.move(tempname, artifact.get_filename(filename))
return None
def is_invalid_checksum(self, file, remote_url, checksum_alg="md5"):
if os.path.exists(file):
local_checksum = self._local_checksum(checksum_alg, file)
if self.local:
parsed_url = urlparse(remote_url)
remote_checksum = self._local_checksum(checksum_alg, parsed_url.path)
else:
try:
remote_checksum = to_text(
self._getContent(f"{remote_url}.{checksum_alg}", "Failed to retrieve checksum", False),
errors="strict",
)
except UnicodeError as e:
return f"Cannot retrieve a valid {checksum_alg} checksum from {remote_url}: {e}"
if not remote_checksum:
return f"Cannot find {checksum_alg} checksum from {remote_url}"
try:
# Check if remote checksum only contains md5/sha1 or md5/sha1 + filename
_remote_checksum = remote_checksum.split(None, 1)[0]
remote_checksum = _remote_checksum
# remote_checksum is empty so we continue and keep original checksum string
# This should not happen since we check for remote_checksum before
except IndexError:
pass
if local_checksum.lower() == remote_checksum.lower():
return None
else:
return (
f"Checksum does not match: we computed {local_checksum} but the repository states {remote_checksum}"
)
return f"Path does not exist: {file}"
def _local_checksum(self, checksum_alg, file):
if checksum_alg.lower() == "md5":
hash = hashlib.md5()
elif checksum_alg.lower() == "sha1":
hash = hashlib.sha1()
else:
raise ValueError(f"Unknown checksum_alg {checksum_alg}")
with open(file, "rb") as f:
for chunk in iter(lambda: f.read(8192), b""):
hash.update(chunk)
return hash.hexdigest()
def main():
module = AnsibleModule(
argument_spec=dict(
group_id=dict(required=True),
artifact_id=dict(required=True),
version=dict(),
version_by_spec=dict(),
classifier=dict(default=""),
extension=dict(default="jar"),
repository_url=dict(default="https://repo1.maven.org/maven2"),
username=dict(aliases=["aws_secret_key"]),
password=dict(no_log=True, aliases=["aws_secret_access_key"]),
headers=dict(type="dict"),
force_basic_auth=dict(default=False, type="bool"),
state=dict(default="present", choices=["present", "absent"]),
timeout=dict(default=10, type="int"),
dest=dict(type="path", required=True),
validate_certs=dict(default=True, type="bool"),
client_cert=dict(type="path"),
client_key=dict(type="path"),
keep_name=dict(default=False, type="bool"),
verify_checksum=dict(default="download", choices=["never", "download", "change", "always"]),
checksum_alg=dict(default="md5", choices=["md5", "sha1"]),
unredirected_headers=dict(type="list", elements="str"),
directory_mode=dict(type="str"),
),
add_file_common_args=True,
mutually_exclusive=([("version", "version_by_spec")]),
)
if module.params["unredirected_headers"] is None:
# if the user did not supply unredirected params, we use the default
module.params["unredirected_headers"] = ["Authorization", "Cookie"]
if not HAS_LXML_ETREE:
module.fail_json(msg=missing_required_lib("lxml"), exception=LXML_ETREE_IMP_ERR)
if module.params["version_by_spec"] and not HAS_SEMANTIC_VERSION:
module.fail_json(msg=missing_required_lib("semantic_version"), exception=SEMANTIC_VERSION_IMP_ERR)
repository_url = module.params["repository_url"]
if not repository_url:
repository_url = "https://repo1.maven.org/maven2"
try:
parsed_url = urlparse(repository_url)
except AttributeError as e:
module.fail_json(msg=f"url parsing went wrong {e}")
local = parsed_url.scheme == "file"
if parsed_url.scheme == "s3" and not HAS_BOTO:
module.fail_json(
msg=missing_required_lib("boto3", reason="when using s3:// repository URLs"), exception=BOTO_IMP_ERR
)
group_id = module.params["group_id"]
artifact_id = module.params["artifact_id"]
version = module.params["version"]
version_by_spec = module.params["version_by_spec"]
classifier = module.params["classifier"]
extension = module.params["extension"]
headers = module.params["headers"]
state = module.params["state"]
dest = module.params["dest"]
b_dest = to_bytes(dest, errors="surrogate_or_strict")
keep_name = module.params["keep_name"]
verify_checksum = module.params["verify_checksum"]
verify_download = verify_checksum in ["download", "always"]
verify_change = verify_checksum in ["change", "always"]
checksum_alg = module.params["checksum_alg"]
downloader = MavenDownloader(module, repository_url, local, headers)
if not version_by_spec and not version:
version = "latest"
try:
artifact = Artifact(group_id, artifact_id, version, version_by_spec, classifier, extension)
except ValueError as e:
module.fail_json(msg=e.args[0])
changed = False
prev_state = "absent"
if dest.endswith(os.sep):
b_dest = to_bytes(dest, errors="surrogate_or_strict")
if not os.path.exists(b_dest):
(pre_existing_dir, new_directory_list) = split_pre_existing_dir(dest)
os.makedirs(b_dest)
directory_args = module.load_file_common_arguments(module.params)
directory_mode = module.params["directory_mode"]
if directory_mode is not None:
directory_args["mode"] = directory_mode
else:
directory_args["mode"] = None
changed = adjust_recursive_directory_permissions(
pre_existing_dir, new_directory_list, module, directory_args, changed
)
if os.path.isdir(b_dest):
version_part = version
if version == "latest":
version_part = downloader.find_latest_version_available(artifact)
elif version_by_spec:
version_part = downloader.find_version_by_spec(artifact)
filename = f"{artifact_id}{(f'-{version_part}' if keep_name else '')}{(f'-{classifier}' if classifier else '')}.{extension}"
dest = posixpath.join(dest, filename)
b_dest = to_bytes(dest, errors="surrogate_or_strict")
if os.path.lexists(b_dest) and (
(not verify_change)
or not downloader.is_invalid_checksum(dest, downloader.find_uri_for_artifact(artifact), checksum_alg)
):
prev_state = "present"
if prev_state == "absent":
try:
download_error = downloader.download(module.tmpdir, artifact, verify_download, b_dest, checksum_alg)
if download_error is None:
changed = True
else:
module.fail_json(msg=f"Cannot retrieve the artifact to destination: {download_error}")
except ValueError as e:
module.fail_json(msg=e.args[0])
file_args = module.load_file_common_arguments(module.params, path=dest)
changed = module.set_fs_attributes_if_different(file_args, changed)
if changed:
module.exit_json(
state=state,
dest=dest,
group_id=group_id,
artifact_id=artifact_id,
version=version,
classifier=classifier,
extension=extension,
repository_url=repository_url,
changed=changed,
)
else:
module.exit_json(state=state, dest=dest, changed=changed)
if __name__ == "__main__":
main()