diff options
Diffstat (limited to 'csit.infra.etl')
-rw-r--r-- | csit.infra.etl/coverage_device_rls2410.py (renamed from csit.infra.etl/coverage_device_rls2406.py) | 4 | ||||
-rw-r--r-- | csit.infra.etl/coverage_hoststack_rls2410.py (renamed from csit.infra.etl/coverage_hoststack_rls2406.py) | 4 | ||||
-rw-r--r-- | csit.infra.etl/coverage_mrr_rls2410.py (renamed from csit.infra.etl/coverage_mrr_rls2406.py) | 4 | ||||
-rw-r--r-- | csit.infra.etl/coverage_ndrpdr_rls2410.py (renamed from csit.infra.etl/coverage_ndrpdr_rls2406.py) | 4 | ||||
-rw-r--r-- | csit.infra.etl/coverage_reconf_rls2410.py (renamed from csit.infra.etl/coverage_reconf_rls2406.py) | 6 | ||||
-rw-r--r-- | csit.infra.etl/coverage_soak_rls2410.py (renamed from csit.infra.etl/coverage_soak_rls2406.py) | 4 | ||||
-rw-r--r-- | csit.infra.etl/iterative_hoststack_rls2410.py (renamed from csit.infra.etl/iterative_hoststack_rls2406.py) | 4 | ||||
-rw-r--r-- | csit.infra.etl/iterative_mrr_rls2410.py (renamed from csit.infra.etl/iterative_mrr_rls2406.py) | 4 | ||||
-rw-r--r-- | csit.infra.etl/iterative_ndrpdr_rls2410.py (renamed from csit.infra.etl/iterative_ndrpdr_rls2406.py) | 4 | ||||
-rw-r--r-- | csit.infra.etl/iterative_reconf_rls2410.py (renamed from csit.infra.etl/iterative_reconf_rls2406.py) | 4 | ||||
-rw-r--r-- | csit.infra.etl/iterative_soak_rls2410.py (renamed from csit.infra.etl/iterative_soak_rls2406.py) | 4 |
11 files changed, 23 insertions, 23 deletions
diff --git a/csit.infra.etl/coverage_device_rls2406.py b/csit.infra.etl/coverage_device_rls2410.py index 30a48247c0..48746daabb 100644 --- a/csit.infra.etl/coverage_device_rls2406.py +++ b/csit.infra.etl/coverage_device_rls2410.py @@ -141,7 +141,7 @@ paths = wr.s3.list_objects( ignore_empty=True ) -filtered_paths = [path for path in paths if "report-coverage-2406" in path] +filtered_paths = [path for path in paths if "report-coverage-2410" in path] out_sdf = process_json_to_dataframe("device", filtered_paths) out_sdf.printSchema() @@ -163,7 +163,7 @@ except KeyError: try: wr.s3.to_parquet( df=out_sdf.toPandas(), - path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/coverage_rls2406", + path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/coverage_rls2410", dataset=True, partition_cols=["test_type", "year", "month", "day"], compression="snappy", diff --git a/csit.infra.etl/coverage_hoststack_rls2406.py b/csit.infra.etl/coverage_hoststack_rls2410.py index 8a0a79b980..080e998b90 100644 --- a/csit.infra.etl/coverage_hoststack_rls2406.py +++ b/csit.infra.etl/coverage_hoststack_rls2410.py @@ -141,7 +141,7 @@ paths = wr.s3.list_objects( ignore_empty=True ) -filtered_paths = [path for path in paths if "report-coverage-2406" in path] +filtered_paths = [path for path in paths if "report-coverage-2410" in path] out_sdf = process_json_to_dataframe("hoststack", filtered_paths) out_sdf.printSchema() @@ -163,7 +163,7 @@ except KeyError: try: wr.s3.to_parquet( df=out_sdf.toPandas(), - path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/coverage_rls2406", + path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/coverage_rls2410", dataset=True, partition_cols=["test_type", "year", "month", "day"], compression="snappy", diff --git a/csit.infra.etl/coverage_mrr_rls2406.py b/csit.infra.etl/coverage_mrr_rls2410.py index 365159d171..9400d8d7fb 100644 --- a/csit.infra.etl/coverage_mrr_rls2406.py +++ b/csit.infra.etl/coverage_mrr_rls2410.py @@ -141,7 +141,7 @@ paths = wr.s3.list_objects( ignore_empty=True ) -filtered_paths = [path for path in paths if "report-coverage-2406" in path] +filtered_paths = [path for path in paths if "report-coverage-2410" in path] out_sdf = process_json_to_dataframe("mrr", filtered_paths) out_sdf.printSchema() @@ -163,7 +163,7 @@ except KeyError: try: wr.s3.to_parquet( df=out_sdf.toPandas(), - path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/coverage_rls2406", + path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/coverage_rls2410", dataset=True, partition_cols=["test_type", "year", "month", "day"], compression="snappy", diff --git a/csit.infra.etl/coverage_ndrpdr_rls2406.py b/csit.infra.etl/coverage_ndrpdr_rls2410.py index 8ad453e795..18b7627aa9 100644 --- a/csit.infra.etl/coverage_ndrpdr_rls2406.py +++ b/csit.infra.etl/coverage_ndrpdr_rls2410.py @@ -141,7 +141,7 @@ paths = wr.s3.list_objects( ignore_empty=True ) -filtered_paths = [path for path in paths if "report-coverage-2406" in path] +filtered_paths = [path for path in paths if "report-coverage-2410" in path] out_sdf = process_json_to_dataframe("ndrpdr", filtered_paths) out_sdf.printSchema() @@ -163,7 +163,7 @@ except KeyError: try: wr.s3.to_parquet( df=out_sdf.toPandas(), - path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/coverage_rls2406", + path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/coverage_rls2410", dataset=True, partition_cols=["test_type", "year", "month", "day"], compression="snappy", diff --git a/csit.infra.etl/coverage_reconf_rls2406.py b/csit.infra.etl/coverage_reconf_rls2410.py index 8685809c01..e60a3d66ad 100644 --- a/csit.infra.etl/coverage_reconf_rls2406.py +++ b/csit.infra.etl/coverage_reconf_rls2410.py @@ -139,8 +139,9 @@ paths = wr.s3.list_objects( last_modified_end=LAST_MODIFIED_END, ignore_suffix=IGNORE_SUFFIX, ignore_empty=True +) -filtered_paths = [path for path in paths if "report-coverage-2406" in path] +filtered_paths = [path for path in paths if "report-coverage-2410" in path] out_sdf = process_json_to_dataframe("reconf", filtered_paths) out_sdf.printSchema() @@ -158,12 +159,11 @@ try: ) except KeyError: boto3_session = session.Session() -) try: wr.s3.to_parquet( df=out_sdf.toPandas(), - path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/coverage_rls2406", + path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/coverage_rls2410", dataset=True, partition_cols=["test_type", "year", "month", "day"], compression="snappy", diff --git a/csit.infra.etl/coverage_soak_rls2406.py b/csit.infra.etl/coverage_soak_rls2410.py index 6f06f2c075..5dc9fb7ed0 100644 --- a/csit.infra.etl/coverage_soak_rls2406.py +++ b/csit.infra.etl/coverage_soak_rls2410.py @@ -141,7 +141,7 @@ paths = wr.s3.list_objects( ignore_empty=True ) -filtered_paths = [path for path in paths if "report-coverage-2406" in path] +filtered_paths = [path for path in paths if "report-coverage-2410" in path] out_sdf = process_json_to_dataframe("soak", filtered_paths) out_sdf.printSchema() @@ -163,7 +163,7 @@ except KeyError: try: wr.s3.to_parquet( df=out_sdf.toPandas(), - path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/coverage_rls2406", + path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/coverage_rls2410", dataset=True, partition_cols=["test_type", "year", "month", "day"], compression="snappy", diff --git a/csit.infra.etl/iterative_hoststack_rls2406.py b/csit.infra.etl/iterative_hoststack_rls2410.py index 7767cefa78..6d9e3d633f 100644 --- a/csit.infra.etl/iterative_hoststack_rls2406.py +++ b/csit.infra.etl/iterative_hoststack_rls2410.py @@ -141,7 +141,7 @@ paths = wr.s3.list_objects( ignore_empty=True ) -filtered_paths = [path for path in paths if "report-iterative-2406" in path] +filtered_paths = [path for path in paths if "report-iterative-2410" in path] out_sdf = process_json_to_dataframe("hoststack", filtered_paths) out_sdf.printSchema() @@ -163,7 +163,7 @@ except KeyError: try: wr.s3.to_parquet( df=out_sdf.toPandas(), - path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/iterative_rls2406", + path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/iterative_rls2410", dataset=True, partition_cols=["test_type", "year", "month", "day"], compression="snappy", diff --git a/csit.infra.etl/iterative_mrr_rls2406.py b/csit.infra.etl/iterative_mrr_rls2410.py index a1a8d96e26..d74e6a90f3 100644 --- a/csit.infra.etl/iterative_mrr_rls2406.py +++ b/csit.infra.etl/iterative_mrr_rls2410.py @@ -141,7 +141,7 @@ paths = wr.s3.list_objects( ignore_empty=True ) -filtered_paths = [path for path in paths if "report-iterative-2406" in path] +filtered_paths = [path for path in paths if "report-iterative-2410" in path] out_sdf = process_json_to_dataframe("mrr", filtered_paths) out_sdf.printSchema() @@ -163,7 +163,7 @@ except KeyError: try: wr.s3.to_parquet( df=out_sdf.toPandas(), - path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/iterative_rls2406", + path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/iterative_rls2410", dataset=True, partition_cols=["test_type", "year", "month", "day"], compression="snappy", diff --git a/csit.infra.etl/iterative_ndrpdr_rls2406.py b/csit.infra.etl/iterative_ndrpdr_rls2410.py index bb474f1d7f..0530dc8ef5 100644 --- a/csit.infra.etl/iterative_ndrpdr_rls2406.py +++ b/csit.infra.etl/iterative_ndrpdr_rls2410.py @@ -141,7 +141,7 @@ paths = wr.s3.list_objects( ignore_empty=True ) -filtered_paths = [path for path in paths if "report-iterative-2406" in path] +filtered_paths = [path for path in paths if "report-iterative-2410" in path] out_sdf = process_json_to_dataframe("ndrpdr", filtered_paths) out_sdf.printSchema() @@ -163,7 +163,7 @@ except KeyError: try: wr.s3.to_parquet( df=out_sdf.toPandas(), - path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/iterative_rls2406", + path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/iterative_rls2410", dataset=True, partition_cols=["test_type", "year", "month", "day"], compression="snappy", diff --git a/csit.infra.etl/iterative_reconf_rls2406.py b/csit.infra.etl/iterative_reconf_rls2410.py index 71e92b49f8..41d2aedcc1 100644 --- a/csit.infra.etl/iterative_reconf_rls2406.py +++ b/csit.infra.etl/iterative_reconf_rls2410.py @@ -141,7 +141,7 @@ paths = wr.s3.list_objects( ignore_empty=True ) -filtered_paths = [path for path in paths if "report-iterative-2406" in path] +filtered_paths = [path for path in paths if "report-iterative-2410" in path] out_sdf = process_json_to_dataframe("reconf", filtered_paths) out_sdf.show(truncate=False) @@ -164,7 +164,7 @@ except KeyError: try: wr.s3.to_parquet( df=out_sdf.toPandas(), - path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/iterative_rls2406", + path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/iterative_rls2410", dataset=True, partition_cols=["test_type", "year", "month", "day"], compression="snappy", diff --git a/csit.infra.etl/iterative_soak_rls2406.py b/csit.infra.etl/iterative_soak_rls2410.py index 966451539e..f581c7c1e9 100644 --- a/csit.infra.etl/iterative_soak_rls2406.py +++ b/csit.infra.etl/iterative_soak_rls2410.py @@ -141,7 +141,7 @@ paths = wr.s3.list_objects( ignore_empty=True ) -filtered_paths = [path for path in paths if "report-iterative-2406" in path] +filtered_paths = [path for path in paths if "report-iterative-2410" in path] out_sdf = process_json_to_dataframe("soak", filtered_paths) out_sdf.printSchema() @@ -163,7 +163,7 @@ except KeyError: try: wr.s3.to_parquet( df=out_sdf.toPandas(), - path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/iterative_rls2406", + path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/iterative_rls2410", dataset=True, partition_cols=["test_type", "year", "month", "day"], compression="snappy", |