aboutsummaryrefslogtreecommitdiffstats
path: root/csit.infra.etl
diff options
context:
space:
mode:
Diffstat (limited to 'csit.infra.etl')
-rw-r--r--csit.infra.etl/coverage_device_rls2410.py (renamed from csit.infra.etl/coverage_device_rls2406.py)4
-rw-r--r--csit.infra.etl/coverage_hoststack_rls2410.py (renamed from csit.infra.etl/coverage_hoststack_rls2406.py)4
-rw-r--r--csit.infra.etl/coverage_mrr_rls2410.py (renamed from csit.infra.etl/coverage_mrr_rls2406.py)4
-rw-r--r--csit.infra.etl/coverage_ndrpdr_rls2410.py (renamed from csit.infra.etl/coverage_ndrpdr_rls2406.py)4
-rw-r--r--csit.infra.etl/coverage_reconf_rls2410.py (renamed from csit.infra.etl/coverage_reconf_rls2406.py)4
-rw-r--r--csit.infra.etl/coverage_soak_rls2410.py (renamed from csit.infra.etl/coverage_soak_rls2406.py)4
-rw-r--r--csit.infra.etl/iterative_hoststack_rls2410.py (renamed from csit.infra.etl/iterative_hoststack_rls2406.py)4
-rw-r--r--csit.infra.etl/iterative_mrr_rls2410.py (renamed from csit.infra.etl/iterative_mrr_rls2406.py)4
-rw-r--r--csit.infra.etl/iterative_ndrpdr_rls2410.py (renamed from csit.infra.etl/iterative_ndrpdr_rls2406.py)4
-rw-r--r--csit.infra.etl/iterative_reconf_rls2410.py (renamed from csit.infra.etl/iterative_reconf_rls2406.py)4
-rw-r--r--csit.infra.etl/iterative_soak_rls2410.py (renamed from csit.infra.etl/iterative_soak_rls2406.py)4
11 files changed, 22 insertions, 22 deletions
diff --git a/csit.infra.etl/coverage_device_rls2406.py b/csit.infra.etl/coverage_device_rls2410.py
index 30a48247c0..48746daabb 100644
--- a/csit.infra.etl/coverage_device_rls2406.py
+++ b/csit.infra.etl/coverage_device_rls2410.py
@@ -141,7 +141,7 @@ paths = wr.s3.list_objects(
ignore_empty=True
)
-filtered_paths = [path for path in paths if "report-coverage-2406" in path]
+filtered_paths = [path for path in paths if "report-coverage-2410" in path]
out_sdf = process_json_to_dataframe("device", filtered_paths)
out_sdf.printSchema()
@@ -163,7 +163,7 @@ except KeyError:
try:
wr.s3.to_parquet(
df=out_sdf.toPandas(),
- path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/coverage_rls2406",
+ path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/coverage_rls2410",
dataset=True,
partition_cols=["test_type", "year", "month", "day"],
compression="snappy",
diff --git a/csit.infra.etl/coverage_hoststack_rls2406.py b/csit.infra.etl/coverage_hoststack_rls2410.py
index 8a0a79b980..080e998b90 100644
--- a/csit.infra.etl/coverage_hoststack_rls2406.py
+++ b/csit.infra.etl/coverage_hoststack_rls2410.py
@@ -141,7 +141,7 @@ paths = wr.s3.list_objects(
ignore_empty=True
)
-filtered_paths = [path for path in paths if "report-coverage-2406" in path]
+filtered_paths = [path for path in paths if "report-coverage-2410" in path]
out_sdf = process_json_to_dataframe("hoststack", filtered_paths)
out_sdf.printSchema()
@@ -163,7 +163,7 @@ except KeyError:
try:
wr.s3.to_parquet(
df=out_sdf.toPandas(),
- path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/coverage_rls2406",
+ path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/coverage_rls2410",
dataset=True,
partition_cols=["test_type", "year", "month", "day"],
compression="snappy",
diff --git a/csit.infra.etl/coverage_mrr_rls2406.py b/csit.infra.etl/coverage_mrr_rls2410.py
index 365159d171..9400d8d7fb 100644
--- a/csit.infra.etl/coverage_mrr_rls2406.py
+++ b/csit.infra.etl/coverage_mrr_rls2410.py
@@ -141,7 +141,7 @@ paths = wr.s3.list_objects(
ignore_empty=True
)
-filtered_paths = [path for path in paths if "report-coverage-2406" in path]
+filtered_paths = [path for path in paths if "report-coverage-2410" in path]
out_sdf = process_json_to_dataframe("mrr", filtered_paths)
out_sdf.printSchema()
@@ -163,7 +163,7 @@ except KeyError:
try:
wr.s3.to_parquet(
df=out_sdf.toPandas(),
- path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/coverage_rls2406",
+ path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/coverage_rls2410",
dataset=True,
partition_cols=["test_type", "year", "month", "day"],
compression="snappy",
diff --git a/csit.infra.etl/coverage_ndrpdr_rls2406.py b/csit.infra.etl/coverage_ndrpdr_rls2410.py
index 8ad453e795..18b7627aa9 100644
--- a/csit.infra.etl/coverage_ndrpdr_rls2406.py
+++ b/csit.infra.etl/coverage_ndrpdr_rls2410.py
@@ -141,7 +141,7 @@ paths = wr.s3.list_objects(
ignore_empty=True
)
-filtered_paths = [path for path in paths if "report-coverage-2406" in path]
+filtered_paths = [path for path in paths if "report-coverage-2410" in path]
out_sdf = process_json_to_dataframe("ndrpdr", filtered_paths)
out_sdf.printSchema()
@@ -163,7 +163,7 @@ except KeyError:
try:
wr.s3.to_parquet(
df=out_sdf.toPandas(),
- path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/coverage_rls2406",
+ path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/coverage_rls2410",
dataset=True,
partition_cols=["test_type", "year", "month", "day"],
compression="snappy",
diff --git a/csit.infra.etl/coverage_reconf_rls2406.py b/csit.infra.etl/coverage_reconf_rls2410.py
index 8685809c01..4e596c24c6 100644
--- a/csit.infra.etl/coverage_reconf_rls2406.py
+++ b/csit.infra.etl/coverage_reconf_rls2410.py
@@ -140,7 +140,7 @@ paths = wr.s3.list_objects(
ignore_suffix=IGNORE_SUFFIX,
ignore_empty=True
-filtered_paths = [path for path in paths if "report-coverage-2406" in path]
+filtered_paths = [path for path in paths if "report-coverage-2410" in path]
out_sdf = process_json_to_dataframe("reconf", filtered_paths)
out_sdf.printSchema()
@@ -163,7 +163,7 @@ except KeyError:
try:
wr.s3.to_parquet(
df=out_sdf.toPandas(),
- path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/coverage_rls2406",
+ path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/coverage_rls2410",
dataset=True,
partition_cols=["test_type", "year", "month", "day"],
compression="snappy",
diff --git a/csit.infra.etl/coverage_soak_rls2406.py b/csit.infra.etl/coverage_soak_rls2410.py
index 6f06f2c075..5dc9fb7ed0 100644
--- a/csit.infra.etl/coverage_soak_rls2406.py
+++ b/csit.infra.etl/coverage_soak_rls2410.py
@@ -141,7 +141,7 @@ paths = wr.s3.list_objects(
ignore_empty=True
)
-filtered_paths = [path for path in paths if "report-coverage-2406" in path]
+filtered_paths = [path for path in paths if "report-coverage-2410" in path]
out_sdf = process_json_to_dataframe("soak", filtered_paths)
out_sdf.printSchema()
@@ -163,7 +163,7 @@ except KeyError:
try:
wr.s3.to_parquet(
df=out_sdf.toPandas(),
- path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/coverage_rls2406",
+ path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/coverage_rls2410",
dataset=True,
partition_cols=["test_type", "year", "month", "day"],
compression="snappy",
diff --git a/csit.infra.etl/iterative_hoststack_rls2406.py b/csit.infra.etl/iterative_hoststack_rls2410.py
index 7767cefa78..6d9e3d633f 100644
--- a/csit.infra.etl/iterative_hoststack_rls2406.py
+++ b/csit.infra.etl/iterative_hoststack_rls2410.py
@@ -141,7 +141,7 @@ paths = wr.s3.list_objects(
ignore_empty=True
)
-filtered_paths = [path for path in paths if "report-iterative-2406" in path]
+filtered_paths = [path for path in paths if "report-iterative-2410" in path]
out_sdf = process_json_to_dataframe("hoststack", filtered_paths)
out_sdf.printSchema()
@@ -163,7 +163,7 @@ except KeyError:
try:
wr.s3.to_parquet(
df=out_sdf.toPandas(),
- path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/iterative_rls2406",
+ path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/iterative_rls2410",
dataset=True,
partition_cols=["test_type", "year", "month", "day"],
compression="snappy",
diff --git a/csit.infra.etl/iterative_mrr_rls2406.py b/csit.infra.etl/iterative_mrr_rls2410.py
index a1a8d96e26..d74e6a90f3 100644
--- a/csit.infra.etl/iterative_mrr_rls2406.py
+++ b/csit.infra.etl/iterative_mrr_rls2410.py
@@ -141,7 +141,7 @@ paths = wr.s3.list_objects(
ignore_empty=True
)
-filtered_paths = [path for path in paths if "report-iterative-2406" in path]
+filtered_paths = [path for path in paths if "report-iterative-2410" in path]
out_sdf = process_json_to_dataframe("mrr", filtered_paths)
out_sdf.printSchema()
@@ -163,7 +163,7 @@ except KeyError:
try:
wr.s3.to_parquet(
df=out_sdf.toPandas(),
- path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/iterative_rls2406",
+ path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/iterative_rls2410",
dataset=True,
partition_cols=["test_type", "year", "month", "day"],
compression="snappy",
diff --git a/csit.infra.etl/iterative_ndrpdr_rls2406.py b/csit.infra.etl/iterative_ndrpdr_rls2410.py
index bb474f1d7f..0530dc8ef5 100644
--- a/csit.infra.etl/iterative_ndrpdr_rls2406.py
+++ b/csit.infra.etl/iterative_ndrpdr_rls2410.py
@@ -141,7 +141,7 @@ paths = wr.s3.list_objects(
ignore_empty=True
)
-filtered_paths = [path for path in paths if "report-iterative-2406" in path]
+filtered_paths = [path for path in paths if "report-iterative-2410" in path]
out_sdf = process_json_to_dataframe("ndrpdr", filtered_paths)
out_sdf.printSchema()
@@ -163,7 +163,7 @@ except KeyError:
try:
wr.s3.to_parquet(
df=out_sdf.toPandas(),
- path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/iterative_rls2406",
+ path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/iterative_rls2410",
dataset=True,
partition_cols=["test_type", "year", "month", "day"],
compression="snappy",
diff --git a/csit.infra.etl/iterative_reconf_rls2406.py b/csit.infra.etl/iterative_reconf_rls2410.py
index 71e92b49f8..41d2aedcc1 100644
--- a/csit.infra.etl/iterative_reconf_rls2406.py
+++ b/csit.infra.etl/iterative_reconf_rls2410.py
@@ -141,7 +141,7 @@ paths = wr.s3.list_objects(
ignore_empty=True
)
-filtered_paths = [path for path in paths if "report-iterative-2406" in path]
+filtered_paths = [path for path in paths if "report-iterative-2410" in path]
out_sdf = process_json_to_dataframe("reconf", filtered_paths)
out_sdf.show(truncate=False)
@@ -164,7 +164,7 @@ except KeyError:
try:
wr.s3.to_parquet(
df=out_sdf.toPandas(),
- path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/iterative_rls2406",
+ path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/iterative_rls2410",
dataset=True,
partition_cols=["test_type", "year", "month", "day"],
compression="snappy",
diff --git a/csit.infra.etl/iterative_soak_rls2406.py b/csit.infra.etl/iterative_soak_rls2410.py
index 966451539e..f581c7c1e9 100644
--- a/csit.infra.etl/iterative_soak_rls2406.py
+++ b/csit.infra.etl/iterative_soak_rls2410.py
@@ -141,7 +141,7 @@ paths = wr.s3.list_objects(
ignore_empty=True
)
-filtered_paths = [path for path in paths if "report-iterative-2406" in path]
+filtered_paths = [path for path in paths if "report-iterative-2410" in path]
out_sdf = process_json_to_dataframe("soak", filtered_paths)
out_sdf.printSchema()
@@ -163,7 +163,7 @@ except KeyError:
try:
wr.s3.to_parquet(
df=out_sdf.toPandas(),
- path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/iterative_rls2406",
+ path=f"s3://{S3_DOCS_BUCKET}/csit/parquet/iterative_rls2410",
dataset=True,
partition_cols=["test_type", "year", "month", "day"],
compression="snappy",