aboutsummaryrefslogtreecommitdiffstats
path: root/resources/tools/presentation/input_data_files.py
diff options
context:
space:
mode:
authorTibor Frank <tifrank@cisco.com>2021-04-26 15:57:02 +0200
committerTibor Frank <tifrank@cisco.com>2021-05-03 08:20:06 +0000
commit37ea2ceb606bdfc338cc76330cb9289c12f63852 (patch)
treed67c7fafaaf165fc5fad3cb694d195a72c399fba /resources/tools/presentation/input_data_files.py
parentc290ed5fb0f768b76213e6eb34d56f16abfeb54e (diff)
PAL: Refactor the processing of spec and download
Change-Id: I204fdc3593a3702366727238f2b469d3d4320c2b Signed-off-by: Tibor Frank <tifrank@cisco.com>
Diffstat (limited to 'resources/tools/presentation/input_data_files.py')
-rw-r--r--resources/tools/presentation/input_data_files.py237
1 files changed, 124 insertions, 113 deletions
diff --git a/resources/tools/presentation/input_data_files.py b/resources/tools/presentation/input_data_files.py
index 63f91972e5..fc629bc218 100644
--- a/resources/tools/presentation/input_data_files.py
+++ b/resources/tools/presentation/input_data_files.py
@@ -32,8 +32,6 @@ from requests import codes
from urllib3.exceptions import HTTPError
-from pal_errors import PresentationError
-
# Chunk size used for file download
CHUNK_SIZE = 512
@@ -96,27 +94,19 @@ def _download_file(url, file_name, arch=False, verify=True, repeat=1):
repeat -= 1
session = None
try:
- logging.info(f" Connecting to {url} ...")
+ logging.info(f" Connecting to {url} ...")
session = requests_retry_session()
response = session.get(url, stream=True, verify=verify)
code = response.status_code
- logging.info(f" {code}: {responses[code]}")
+ logging.info(f" {code}: {responses[code]}")
if code != codes[u"OK"]:
if session:
session.close()
- url = url.replace(u"_info", u"")
- logging.info(f" Connecting to {url} ...")
- session = requests_retry_session()
- response = session.get(url, stream=True, verify=verify)
- code = response.status_code
- logging.info(f" {code}: {responses[code]}")
- if code != codes[u"OK"]:
- return False, file_name
- file_name = file_name.replace(u"_info", u"")
+ return False, file_name
dst_file_name = file_name.replace(u".gz", u"")
- logging.info(f" Downloading the file {url} to {dst_file_name}")
+ logging.info(f" Downloading the file {url} to {dst_file_name}")
with open(dst_file_name, u"wb") as file_handle:
for chunk in response.iter_content(chunk_size=CHUNK_SIZE):
if chunk:
@@ -125,7 +115,7 @@ def _download_file(url, file_name, arch=False, verify=True, repeat=1):
if arch and u".gz" in file_name:
if session:
session.close()
- logging.info(f" Downloading the file {url} to {file_name}")
+ logging.info(f" Downloading the file {url} to {file_name}")
session = requests_retry_session()
response = session.get(url, stream=True, verify=verify)
if response.status_code == codes[u"OK"]:
@@ -148,8 +138,6 @@ def _download_file(url, file_name, arch=False, verify=True, repeat=1):
finally:
if session:
session.close()
-
- logging.info(u" Download finished.")
return success, file_name
@@ -165,11 +153,7 @@ def _unzip_file(spec, build, pid):
"""
file_name = build[u"file-name"]
- if u".zip" in file_name:
- data_file = spec.input[u"zip-extract"]
- else:
- data_file = spec.input[u"extract"]
-
+ data_file = "robot-plugin/output.xml"
directory = spec.environment[u"paths"][u"DIR[WORKING,DATA]"]
tmp_dir = join(directory, str(pid))
try:
@@ -197,6 +181,95 @@ def _unzip_file(spec, build, pid):
return False
+def _download_json(source, job, build, w_dir, arch):
+ """
+
+ :param source:
+ :param job:
+ :param build:
+ :param w_dir: Path to working directory
+ :param arch:
+ :return:
+ """
+ success = False
+ downloaded_name = u""
+
+ return success, downloaded_name
+
+
+def _download_xml(source, job, build, w_dir, arch):
+ """
+
+ :param source:
+ :param job:
+ :param build:
+ :param w_dir: Path to working directory
+ :param arch:
+ :return:
+ """
+
+ file_name = source.get(u"file-name", u"")
+ new_name = join(
+ w_dir,
+ f"{job}{SEPARATOR}{build[u'build']}{SEPARATOR}{file_name}"
+ )
+ url = u"{0}/{1}".format(
+ source.get(u"url", u""),
+ source.get(u"path", u"").format(
+ job=job, build=build[u'build'], filename=file_name
+ )
+ )
+ verify = False if u"nginx" in url else True
+ logging.info(f" Trying to download {url}")
+ success, downloaded_name = _download_file(
+ url, new_name, arch=arch, verify=verify, repeat=3
+ )
+ return success, downloaded_name
+
+
+def _download_xml_docs(source, job, build, w_dir, arch):
+ """
+
+ :param source:
+ :param job:
+ :param build:
+ :param w_dir: Path to working directory
+ :param arch:
+ :return:
+ """
+
+ file_name = source.get(u"file-name", u"")
+ release = re.search(REGEX_RELEASE, job).group(2)
+ for rls in (release, u"master"):
+ try:
+ rls = f"rls{int(rls)}"
+ except ValueError:
+ pass # It is master
+ url = (
+ f"{source.get(u'url', u'')}/"
+ f"{rls}/"
+ f"{source.get(u'path', u'')}/"
+ f"{job}{SEPARATOR}{build[u'build']}{SEPARATOR}{file_name}"
+ )
+ new_name = join(
+ w_dir,
+ f"{job}{SEPARATOR}{build[u'build']}{SEPARATOR}{file_name}"
+ )
+
+ logging.info(f" Trying to download {url}")
+
+ success, downloaded_name = _download_file(url, new_name, arch=arch)
+ if success:
+ if file_name.endswith(u".gz"):
+ with gzip.open(downloaded_name[:-3], u"rb") as gzip_file:
+ file_content = gzip_file.read()
+ with open(downloaded_name[:-3], u"wb") as xml_file:
+ xml_file.write(file_content)
+ break
+
+ return success, downloaded_name
+
+
def download_and_unzip_data_file(spec, job, build, pid):
"""Download and unzip a source file.
@@ -212,108 +285,46 @@ def download_and_unzip_data_file(spec, job, build, pid):
:rtype: bool
"""
- success = False
+ download = {
+ "json": _download_json,
+ "xml": _download_xml,
+ "xml-docs": _download_xml_docs
+ }
- file_name = spec.input[u"file-name"]
- new_name = join(
- spec.environment[u"paths"][u"DIR[WORKING,DATA]"],
- f"{job}{SEPARATOR}{build[u'build']}{SEPARATOR}{file_name}"
- )
- arch = bool(spec.configuration.get(u"archive-inputs", True))
+ success = False
downloaded_name = u""
-
- # Try to download .gz from s3_storage
- for path in spec.input[u'download-path']:
- url = u"{0}/{1}".format(
- spec.environment[u'urls'][u'URL[S3_STORAGE,LOG]'],
- path.format(job=job, build=build[u'build'], filename=file_name)
- )
- logging.info(f"Trying to download {url}")
- success, downloaded_name = _download_file(
- url, new_name, arch=arch, verify=False, repeat=3
- )
- if success:
- break
-
- if not success:
- # Try to download .gz from logs.fd.io
- for path in spec.input[u'download-path']:
- url = u"{0}/{1}".format(
- spec.environment[u'urls'][u'URL[NEXUS,LOG]'],
- path.format(job=job, build=build[u'build'], filename=file_name)
- )
- logging.info(f"Trying to download {url}")
- success, downloaded_name = _download_file(
- url, new_name, arch=arch, verify=True, repeat=3
- )
- if success:
- break
-
- if not success:
- # Try to download .gz or .zip from docs.fd.io
- file_name = (spec.input[u"file-name"], spec.input[u"zip-file-name"])
- release = re.search(REGEX_RELEASE, job).group(2)
- for idx, rls in enumerate((release, u"master", )):
- try:
- rls = f"rls{int(rls)}"
- except ValueError:
- # It is master
- pass
- url = (
- f"{spec.environment[u'urls'][u'URL[NEXUS,DOC]']}/"
- f"{rls}/"
- f"{spec.environment[u'urls'][u'DIR[NEXUS,DOC]']}/"
- f"{job}{SEPARATOR}{build[u'build']}{SEPARATOR}{file_name[idx]}"
- )
-
- logging.info(f"Downloading {url}")
-
- new_name = join(
+ arch = bool(spec.environment.get(u"archive-inputs", True))
+
+ for source in spec.environment.get(u"data-sources", tuple()):
+ if not source.get(u"enabled", False):
+ continue
+ download_type = source.get(u"type", None)
+ if not download_type:
+ continue
+ success, downloaded_name = download[download_type](
+ source,
+ job,
+ build,
spec.environment[u"paths"][u"DIR[WORKING,DATA]"],
- f"{job}{SEPARATOR}{build[u'build']}{SEPARATOR}{file_name[idx]}"
+ arch
)
- success, downloaded_name = _download_file(url, new_name, arch=arch)
- if success:
- file_name = file_name[idx]
- if file_name.endswith(u".gz"):
- with gzip.open(downloaded_name[:-3], u"rb") as gzip_file:
- file_content = gzip_file.read()
- with open(downloaded_name[:-3], u"wb") as xml_file:
- xml_file.write(file_content)
- break
-
- # if not success:
- # # Try to download .zip from jenkins.fd.io
- # file_name = spec.input[u"zip-file-name"]
- # download_path = spec.input[u"zip-download-path"]
- # if job.startswith(u"csit-"):
- # url = spec.environment[u"urls"][u"URL[JENKINS,CSIT]"]
- # else:
- # raise PresentationError(f"No url defined for the job {job}.")
- #
- # full_name = download_path.format(
- # job=job, build=build[u"build"], filename=file_name
- # )
- # url = u"{0}/{1}".format(url, full_name)
- # new_name = join(
- # spec.environment[u"paths"][u"DIR[WORKING,DATA]"],
- # f"{job}{SEPARATOR}{build[u'build']}{SEPARATOR}{file_name}"
- # )
- # logging.info(f"Downloading {url}")
- # success, downloaded_name = _download_file(url, new_name)
+ if success:
+ source[u"successful-downloads"] += 1
+ build[u"source"] = source[u"type"]
+ break
+ # TODO: Remove when only .gz is used.
if success and downloaded_name.endswith(u".zip"):
if not is_zipfile(downloaded_name):
- logging.error(f"Zip file {new_name} is corrupted.")
+ logging.error(f"Zip file {downloaded_name} is corrupted.")
success = False
if success:
- build[u"file-name"] = downloaded_name
-
- if file_name.endswith(u".gz"):
+ if downloaded_name.endswith(u".gz"):
build[u"file-name"] = downloaded_name[:-3]
-
- if downloaded_name.endswith(u".zip"):
+ # TODO: Remove when only .gz is used.
+ elif downloaded_name.endswith(u".zip"):
+ build[u"file-name"] = downloaded_name
success = _unzip_file(spec, build, pid)
return success