diff options
Diffstat (limited to 'csit.infra.etl')
-rw-r--r-- | csit.infra.etl/coverage_device_rls2410.py (renamed from csit.infra.etl/coverage_device_rls2406.py) | 4 | ||||
-rw-r--r-- | csit.infra.etl/coverage_hoststack_rls2410.py (renamed from csit.infra.etl/coverage_hoststack_rls2406.py) | 4 | ||||
-rw-r--r-- | csit.infra.etl/coverage_mrr_rls2410.py (renamed from csit.infra.etl/coverage_mrr_rls2406.py) | 4 | ||||
-rw-r--r-- | csit.infra.etl/coverage_ndrpdr_rls2410.py (renamed from csit.infra.etl/coverage_ndrpdr_rls2406.py) | 4 | ||||
-rw-r--r-- | csit.infra.etl/coverage_reconf_rls2410.py (renamed from csit.infra.etl/coverage_reconf_rls2406.py) | 4 | ||||
-rw-r--r-- | csit.infra.etl/coverage_soak_rls2410.py (renamed from csit.infra.etl/coverage_soak_rls2406.py) | 4 | ||||
-rw-r--r-- | csit.infra.etl/iterative_hoststack_rls2410.py (renamed from csit.infra.etl/iterative_hoststack_rls2406.py) | 4 | ||||
-rw-r--r-- | csit.infra.etl/iterative_mrr_rls2410.py (renamed from csit.infra.etl/iterative_mrr_rls2406.py) | 4 | ||||
-rw-r--r-- | csit.infra.etl/iterative_ndrpdr_rls2410.py (renamed from csit.infra.etl/iterative_ndrpdr_rls2406.py) | 4 | ||||
-rw-r--r-- | csit.infra.etl/iterative_reconf_rls2410.py (renamed from csit.infra.etl/iterative_reconf_rls2406.py) | 4 | ||||
-rw-r--r-- | csit.infra.etl/iterative_soak_rls2410.py (renamed from csit.infra.etl/iterative_soak_rls2406.py) | 4 |
11 files changed, 22 insertions, 22 deletions
diff --git a/csit.infra.etl/coverage_device_rls2406.py b/csit.infra.etl/coverage_device_rls2410.py index 30a48247c0..48746daabb 100644 --- a/csit.infra.etl/coverage_device_rls2406.py +++ b/csit.infra.etl/coverage_device_rls2410.py @@ -141,7 +141,7 @@ paths = wr.s3.list_objects( ignore_empty=True ) -filtered_paths = [path for path in paths if "report-coverage-2406" in path] +filtered_paths = [path for path in paths if "report-coverage-2410" in path] out_sdf = process_json_to_dataframe("device", filtered_paths) out_sdf.printSchema() @@ -163,7 +163,7 @@ except KeyError: try: wr.s3.to_parquet( df=out_sdf.toPandas(), - path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/coverage_rls2406", + path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/coverage_rls2410", dataset=True, partition_cols=["test_type", "year", "month", "day"], compression="snappy", diff --git a/csit.infra.etl/coverage_hoststack_rls2406.py b/csit.infra.etl/coverage_hoststack_rls2410.py index 8a0a79b980..080e998b90 100644 --- a/csit.infra.etl/coverage_hoststack_rls2406.py +++ b/csit.infra.etl/coverage_hoststack_rls2410.py @@ -141,7 +141,7 @@ paths = wr.s3.list_objects( ignore_empty=True ) -filtered_paths = [path for path in paths if "report-coverage-2406" in path] +filtered_paths = [path for path in paths if "report-coverage-2410" in path] out_sdf = process_json_to_dataframe("hoststack", filtered_paths) out_sdf.printSchema() @@ -163,7 +163,7 @@ except KeyError: try: wr.s3.to_parquet( df=out_sdf.toPandas(), - path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/coverage_rls2406", + path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/coverage_rls2410", dataset=True, partition_cols=["test_type", "year", "month", "day"], compression="snappy", diff --git a/csit.infra.etl/coverage_mrr_rls2406.py b/csit.infra.etl/coverage_mrr_rls2410.py index 365159d171..9400d8d7fb 100644 --- a/csit.infra.etl/coverage_mrr_rls2406.py +++ b/csit.infra.etl/coverage_mrr_rls2410.py @@ -141,7 +141,7 @@ paths = wr.s3.list_objects( ignore_empty=True ) -filtered_paths = [path for path in paths if "report-coverage-2406" in path] +filtered_paths = [path for path in paths if "report-coverage-2410" in path] out_sdf = process_json_to_dataframe("mrr", filtered_paths) out_sdf.printSchema() @@ -163,7 +163,7 @@ except KeyError: try: wr.s3.to_parquet( df=out_sdf.toPandas(), - path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/coverage_rls2406", + path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/coverage_rls2410", dataset=True, partition_cols=["test_type", "year", "month", "day"], compression="snappy", diff --git a/csit.infra.etl/coverage_ndrpdr_rls2406.py b/csit.infra.etl/coverage_ndrpdr_rls2410.py index 8ad453e795..18b7627aa9 100644 --- a/csit.infra.etl/coverage_ndrpdr_rls2406.py +++ b/csit.infra.etl/coverage_ndrpdr_rls2410.py @@ -141,7 +141,7 @@ paths = wr.s3.list_objects( ignore_empty=True ) -filtered_paths = [path for path in paths if "report-coverage-2406" in path] +filtered_paths = [path for path in paths if "report-coverage-2410" in path] out_sdf = process_json_to_dataframe("ndrpdr", filtered_paths) out_sdf.printSchema() @@ -163,7 +163,7 @@ except KeyError: try: wr.s3.to_parquet( df=out_sdf.toPandas(), - path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/coverage_rls2406", + path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/coverage_rls2410", dataset=True, partition_cols=["test_type", "year", "month", "day"], compression="snappy", diff --git a/csit.infra.etl/coverage_reconf_rls2406.py b/csit.infra.etl/coverage_reconf_rls2410.py index 8685809c01..4e596c24c6 100644 --- a/csit.infra.etl/coverage_reconf_rls2406.py +++ b/csit.infra.etl/coverage_reconf_rls2410.py @@ -140,7 +140,7 @@ paths = wr.s3.list_objects( ignore_suffix=IGNORE_SUFFIX, ignore_empty=True -filtered_paths = [path for path in paths if "report-coverage-2406" in path] +filtered_paths = [path for path in paths if "report-coverage-2410" in path] out_sdf = process_json_to_dataframe("reconf", filtered_paths) out_sdf.printSchema() @@ -163,7 +163,7 @@ except KeyError: try: wr.s3.to_parquet( df=out_sdf.toPandas(), - path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/coverage_rls2406", + path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/coverage_rls2410", dataset=True, partition_cols=["test_type", "year", "month", "day"], compression="snappy", diff --git a/csit.infra.etl/coverage_soak_rls2406.py b/csit.infra.etl/coverage_soak_rls2410.py index 6f06f2c075..5dc9fb7ed0 100644 --- a/csit.infra.etl/coverage_soak_rls2406.py +++ b/csit.infra.etl/coverage_soak_rls2410.py @@ -141,7 +141,7 @@ paths = wr.s3.list_objects( ignore_empty=True ) -filtered_paths = [path for path in paths if "report-coverage-2406" in path] +filtered_paths = [path for path in paths if "report-coverage-2410" in path] out_sdf = process_json_to_dataframe("soak", filtered_paths) out_sdf.printSchema() @@ -163,7 +163,7 @@ except KeyError: try: wr.s3.to_parquet( df=out_sdf.toPandas(), - path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/coverage_rls2406", + path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/coverage_rls2410", dataset=True, partition_cols=["test_type", "year", "month", "day"], compression="snappy", diff --git a/csit.infra.etl/iterative_hoststack_rls2406.py b/csit.infra.etl/iterative_hoststack_rls2410.py index 7767cefa78..6d9e3d633f 100644 --- a/csit.infra.etl/iterative_hoststack_rls2406.py +++ b/csit.infra.etl/iterative_hoststack_rls2410.py @@ -141,7 +141,7 @@ paths = wr.s3.list_objects( ignore_empty=True ) -filtered_paths = [path for path in paths if "report-iterative-2406" in path] +filtered_paths = [path for path in paths if "report-iterative-2410" in path] out_sdf = process_json_to_dataframe("hoststack", filtered_paths) out_sdf.printSchema() @@ -163,7 +163,7 @@ except KeyError: try: wr.s3.to_parquet( df=out_sdf.toPandas(), - path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/iterative_rls2406", + path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/iterative_rls2410", dataset=True, partition_cols=["test_type", "year", "month", "day"], compression="snappy", diff --git a/csit.infra.etl/iterative_mrr_rls2406.py b/csit.infra.etl/iterative_mrr_rls2410.py index a1a8d96e26..d74e6a90f3 100644 --- a/csit.infra.etl/iterative_mrr_rls2406.py +++ b/csit.infra.etl/iterative_mrr_rls2410.py @@ -141,7 +141,7 @@ paths = wr.s3.list_objects( ignore_empty=True ) -filtered_paths = [path for path in paths if "report-iterative-2406" in path] +filtered_paths = [path for path in paths if "report-iterative-2410" in path] out_sdf = process_json_to_dataframe("mrr", filtered_paths) out_sdf.printSchema() @@ -163,7 +163,7 @@ except KeyError: try: wr.s3.to_parquet( df=out_sdf.toPandas(), - path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/iterative_rls2406", + path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/iterative_rls2410", dataset=True, partition_cols=["test_type", "year", "month", "day"], compression="snappy", diff --git a/csit.infra.etl/iterative_ndrpdr_rls2406.py b/csit.infra.etl/iterative_ndrpdr_rls2410.py index bb474f1d7f..0530dc8ef5 100644 --- a/csit.infra.etl/iterative_ndrpdr_rls2406.py +++ b/csit.infra.etl/iterative_ndrpdr_rls2410.py @@ -141,7 +141,7 @@ paths = wr.s3.list_objects( ignore_empty=True ) -filtered_paths = [path for path in paths if "report-iterative-2406" in path] +filtered_paths = [path for path in paths if "report-iterative-2410" in path] out_sdf = process_json_to_dataframe("ndrpdr", filtered_paths) out_sdf.printSchema() @@ -163,7 +163,7 @@ except KeyError: try: wr.s3.to_parquet( df=out_sdf.toPandas(), - path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/iterative_rls2406", + path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/iterative_rls2410", dataset=True, partition_cols=["test_type", "year", "month", "day"], compression="snappy", diff --git a/csit.infra.etl/iterative_reconf_rls2406.py b/csit.infra.etl/iterative_reconf_rls2410.py index 71e92b49f8..41d2aedcc1 100644 --- a/csit.infra.etl/iterative_reconf_rls2406.py +++ b/csit.infra.etl/iterative_reconf_rls2410.py @@ -141,7 +141,7 @@ paths = wr.s3.list_objects( ignore_empty=True ) -filtered_paths = [path for path in paths if "report-iterative-2406" in path] +filtered_paths = [path for path in paths if "report-iterative-2410" in path] out_sdf = process_json_to_dataframe("reconf", filtered_paths) out_sdf.show(truncate=False) @@ -164,7 +164,7 @@ except KeyError: try: wr.s3.to_parquet( df=out_sdf.toPandas(), - path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/iterative_rls2406", + path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/iterative_rls2410", dataset=True, partition_cols=["test_type", "year", "month", "day"], compression="snappy", diff --git a/csit.infra.etl/iterative_soak_rls2406.py b/csit.infra.etl/iterative_soak_rls2410.py index 966451539e..f581c7c1e9 100644 --- a/csit.infra.etl/iterative_soak_rls2406.py +++ b/csit.infra.etl/iterative_soak_rls2410.py @@ -141,7 +141,7 @@ paths = wr.s3.list_objects( ignore_empty=True ) -filtered_paths = [path for path in paths if "report-iterative-2406" in path] +filtered_paths = [path for path in paths if "report-iterative-2410" in path] out_sdf = process_json_to_dataframe("soak", filtered_paths) out_sdf.printSchema() @@ -163,7 +163,7 @@ except KeyError: try: wr.s3.to_parquet( df=out_sdf.toPandas(), - path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/iterative_rls2406", + path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/iterative_rls2410", dataset=True, partition_cols=["test_type", "year", "month", "day"], compression="snappy", |