aboutsummaryrefslogtreecommitdiffstats
path: root/resources/tools/presentation
diff options
context:
space:
mode:
authorTibor Frank <tifrank@cisco.com>2022-03-04 08:23:21 +0100
committerTibor Frank <tifrank@cisco.com>2022-03-04 08:23:21 +0100
commit528be8671f305c99c713a945f61d85e7ca21a077 (patch)
tree2a4be388caccf33855e9616be7f766dc10ece132 /resources/tools/presentation
parent819ba9a35bccb3b55b369aaf01805e427faeaf0c (diff)
Report: Add rls data
Change-Id: I7ba3d508b0c3b5718fc1b886d966c9bcac18023e Signed-off-by: Tibor Frank <tifrank@cisco.com>
Diffstat (limited to 'resources/tools/presentation')
-rw-r--r--resources/tools/presentation/specifications/report/data_sets.yaml56
-rw-r--r--resources/tools/presentation/specifications/report/elements.yaml488
-rw-r--r--resources/tools/presentation/sphinx_conf/report/conf.py2
3 files changed, 313 insertions, 233 deletions
diff --git a/resources/tools/presentation/specifications/report/data_sets.yaml b/resources/tools/presentation/specifications/report/data_sets.yaml
index 298fcc8ffd..702ee4a383 100644
--- a/resources/tools/presentation/specifications/report/data_sets.yaml
+++ b/resources/tools/presentation/specifications/report/data_sets.yaml
@@ -123,7 +123,7 @@
vpp-3n-skx-curr-cov:
csit-vpp-perf-report-coverage-2202-3n-skx:
- 35 # rls2202.rel NDRPDR cov env 9
- # - 36 # rls2202.rel NDRPDR cov env 9
+ - 36 # rls2202.rel NDRPDR cov env 9
- 37 # rls2202.rel NDRPDR cov env 9
- 38 # rls2202.rel NDRPDR cov env 9
- 39 # rls2202.rel NDRPDR cov env 9
@@ -137,19 +137,24 @@
- 47 # rls2202.rel NDRPDR cov env 9
- 48 # rls2202.rel NDRPDR cov env 9
- 49 # rls2202.rel NDRPDR cov env 9
- # - 50 # rls2202.rel NDRPDR cov env 9
- # - 51 # rls2202.rel NDRPDR cov env 9
+ - 50 # rls2202.rel NDRPDR cov env 9
+ - 51 # rls2202.rel NDRPDR cov env 9
- 52 # rls2202.rel NDRPDR cov env 9
- 53 # rls2202.rel NDRPDR cov env 9
- 54 # rls2202.rel NDRPDR cov env 9
- 55 # rls2202.rel NDRPDR cov env 9
# - 56 # rls2202.rel NDRPDR cov env 9
- # - 57 # rls2202.rel NDRPDR cov env 9
- # - 58 # rls2202.rel NDRPDR cov env 9
- # - 59 # rls2202.rel NDRPDR cov env 9
- # - 60 # rls2202.rel NDRPDR cov env 9
- # - 61 # rls2202.rel NDRPDR cov env 9
- # - 62 # rls2202.rel NDRPDR cov env 9
+ - 57 # rls2202.rel NDRPDR cov env 9
+ - 58 # rls2202.rel NDRPDR cov env 9
+ - 59 # rls2202.rel NDRPDR cov env 9
+ - 60 # rls2202.rel NDRPDR cov env 9
+ - 61 # rls2202.rel NDRPDR cov env 9
+ - 62 # rls2202.rel NDRPDR cov env 9
+ - 63 # rls2202.rel NDRPDR cov env 9
+ # - 64 # rls2202.rel NDRPDR cov env 9
+ # - 65 # rls2202.rel NDRPDR cov env 9
+ # - 66 # rls2202.rel NDRPDR cov env 9
+ # - 67 # rls2202.rel NDRPDR cov env 9
dpdk-3n-skx-hist-iter:
csit-dpdk-perf-report-iterative-2106-3n-skx:
@@ -475,7 +480,7 @@
csit-vpp-perf-report-iterative-2202-2n-zn2:
- 24 # rls2202.rel NDRPDR iter env 9
- 26 # rls2202.rel NDRPDR iter env 9
- # - 27 # rls2202.rel NDRPDR iter env 9
+ - 27 # rls2202.rel NDRPDR iter env 9
- 28 # rls2202.rel NDRPDR iter env 9
- 29 # rls2202.rel NDRPDR iter env 9
- 30 # rls2202.rel NDRPDR iter env 9
@@ -737,21 +742,21 @@
vpp-2n-clx-curr-cov:
csit-vpp-perf-report-coverage-2202-2n-clx:
- 18 # rls2202.rel NDRPDR cov env 9
- # - 19 # rls2202.rel NDRPDR cov env 9
+ - 19 # rls2202.rel NDRPDR cov env 9
- 20 # rls2202.rel NDRPDR cov env 9
- 21 # rls2202.rel NDRPDR cov env 9
- 22 # rls2202.rel NDRPDR cov env 9
- 23 # rls2202.rel NDRPDR cov env 9
- 24 # rls2202.rel NDRPDR cov env 9
- # - 25 # rls2202.rel NDRPDR cov env 9
+ - 25 # rls2202.rel NDRPDR cov env 9
- 26 # rls2202.rel NDRPDR cov env 9
- 27 # rls2202.rel NDRPDR cov env 9
- 28 # rls2202.rel NDRPDR cov env 9
# - 29 # rls2202.rel NDRPDR cov env 9
- # - 30 # rls2202.rel NDRPDR cov env 9
- # - 31 # rls2202.rel NDRPDR cov env 9
- # - 32 # rls2202.rel NDRPDR cov env 9
- # - 33 # rls2202.rel NDRPDR cov env 9
+ - 30 # rls2202.rel NDRPDR cov env 9
+ - 31 # rls2202.rel NDRPDR cov env 9
+ - 32 # rls2202.rel NDRPDR cov env 9
+ - 33 # rls2202.rel NDRPDR cov env 9
# - 34 # rls2202.rel NDRPDR cov env 9
dpdk-2n-clx-hist-iter:
@@ -845,7 +850,7 @@
# - 26 # rls2202.rel NDRPDR iter env 9
- 27 # rls2202.rel NDRPDR iter env 9
- 28 # rls2202.rel NDRPDR iter env 9
- # - 29 # rls2202.rel NDRPDR iter env 9
+ - 29 # rls2202.rel NDRPDR iter env 9
- 30 # rls2202.rel NDRPDR iter env 9
- 31 # rls2202.rel NDRPDR iter env 9
- 32 # rls2202.rel NDRPDR iter env 9
@@ -979,10 +984,13 @@
csit-vpp-perf-report-iterative-2202-2n-tx2:
- 26 # rls2202.rel NDRPDR iter env 9
- 27 # rls2202.rel NDRPDR iter env 9
+ - 28 # rls2202.rel NDRPDR iter env 9
- 29 # rls2202.rel NDRPDR iter env 9
- 30 # rls2202.rel NDRPDR iter env 9
- 31 # rls2202.rel NDRPDR iter env 9
- 32 # rls2202.rel NDRPDR iter env 9
+ # - 33 # rls2202.rel NDRPDR iter env 9
+ # - 34 # rls2202.rel NDRPDR iter env 9
- 35 # rls2202.rel NDRPDR iter env 9
vpp-2n-tx2-curr-iter-best:
@@ -1307,6 +1315,7 @@
vpp-2n-icx-curr-iter:
csit-vpp-perf-report-iterative-2202-2n-icx:
- 23 # rls2202.rel NDRPDR iter env 9
+ - 30 # rls2202.rel NDRPDR iter env 9
- 31 # rls2202.rel NDRPDR iter env 9
- 32 # rls2202.rel NDRPDR iter env 9
- 33 # rls2202.rel NDRPDR iter env 9
@@ -1314,7 +1323,7 @@
vpp-2n-icx-curr-iter-best:
csit-vpp-perf-report-iterative-2202-2n-icx:
- - 23 # rls2202.rel NDRPDR iter env 9
+ - 32 # rls2202.rel NDRPDR iter env 9
vpp-2n-icx-curr-iter-mrr:
csit-vpp-perf-report-iterative-2202-2n-icx:
@@ -1332,6 +1341,7 @@
csit-vpp-perf-report-iterative-2202-2n-icx:
- 26 # rls2202.rel NDRPDR reconf iter env 9
- 35 # rls2202.rel NDRPDR reconf iter env 9
+ - 38 # rls2202.rel NDRPDR reconf iter env 9
vpp-2n-icx-curr-soak:
csit-vpp-perf-report-iterative-2202-2n-icx:
@@ -1340,6 +1350,7 @@
vpp-soak-vs-ndr-2n-icx-ref:
csit-vpp-perf-report-iterative-2202-2n-icx:
- 23 # rls2202.rel NDRPDR iter env 9
+ - 30 # rls2202.rel NDRPDR iter env 9
- 31 # rls2202.rel NDRPDR iter env 9
- 32 # rls2202.rel NDRPDR iter env 9
- 33 # rls2202.rel NDRPDR iter env 9
@@ -1347,6 +1358,7 @@
vpp-soak-vs-ndr-2n-icx:
csit-vpp-perf-report-iterative-2202-2n-icx:
- 23 # rls2202.rel NDRPDR iter env 9
+ - 30 # rls2202.rel NDRPDR iter env 9
- 31 # rls2202.rel NDRPDR iter env 9
- 32 # rls2202.rel NDRPDR iter env 9
- 33 # rls2202.rel NDRPDR iter env 9
@@ -1420,14 +1432,14 @@
vpp-3n-icx-curr-cov:
csit-vpp-perf-report-coverage-2202-3n-icx:
- # - 35 # rls2202.rel NDRPDR cov env 9
- # - 36 # rls2202.rel NDRPDR cov env 9
- # - 37 # rls2202.rel NDRPDR cov env 9
+ - 35 # rls2202.rel NDRPDR cov env 9
+ - 36 # rls2202.rel NDRPDR cov env 9
+ - 37 # rls2202.rel NDRPDR cov env 9
# - 38 # rls2202.rel NDRPDR cov env 9
- 39 # rls2202.rel NDRPDR cov env 9
# - 40 # rls2202.rel NDRPDR cov env 9
- 41 # rls2202.rel NDRPDR cov env 9
- # - 42 # rls2202.rel NDRPDR cov env 9
+ - 42 # rls2202.rel NDRPDR cov env 9
# - 43 # rls2202.rel NDRPDR cov env 9
- 44 # rls2202.rel NDRPDR cov env 9
- 45 # rls2202.rel NDRPDR cov env 9
diff --git a/resources/tools/presentation/specifications/report/elements.yaml b/resources/tools/presentation/specifications/report/elements.yaml
index 364776c111..007f58d81b 100644
--- a/resources/tools/presentation/specifications/report/elements.yaml
+++ b/resources/tools/presentation/specifications/report/elements.yaml
@@ -48,13 +48,81 @@
- job-spec: "dpdk-mlr-00"
data-set: "dpdk-3n-icx-curr-iter"
-# - type: "table"
-# title: "Job Specification Duration: 3n-icx-coverage"
-# algorithm: "table_job_spec_duration"
-# output-file: "{DIR[STATIC,VPP]}/job-spec-duration-3n-icx-cov"
-# jb-type: "coverage"
-# data: "vpp-3n-icx-curr-cov"
-# lines:
+- type: "table"
+ title: "Job Specification Duration: 3n-icx-coverage"
+ algorithm: "table_job_spec_duration"
+ output-file: "{DIR[STATIC,VPP]}/job-spec-duration-3n-icx-cov"
+ jb-type: "coverage"
+ data: "vpp-3n-icx-curr-cov"
+ lines:
+ - job-spec: "crypto-00"
+ build: "35"
+ - job-spec: "crypto-01"
+ build: "36"
+ - job-spec: "crypto-02"
+ build: "37"
+ # - job-spec: "crypto-03" # Completly failed
+ # build: "38"
+ # - job-spec: "crypto-04"
+ # build: "39"
+ - job-spec: "ip4-00"
+ build: "39"
+ # - job-spec: "ip4-01"
+ # build: "41"
+ - job-spec: "ip4-02"
+ build: "41"
+ - job-spec: "ip4-03"
+ build: "42"
+ # - job-spec: "ip4-04"
+ # build: "43"
+ - job-spec: "ip4-05"
+ build: "44"
+ - job-spec: "ip4-06"
+ build: "45"
+ # - job-spec: "ip4-07"
+ # build: "46"
+ # - job-spec: "ip4tun-00"
+ # build: "47"
+ # - job-spec: "ip4tun-02"
+ # build: "48"
+ # - job-spec: "ip6-00"
+ # build: "49"
+ # - job-spec: "ip6-01"
+ # build: "50"
+ # - job-spec: "ip6tun-00"
+ # build: "51"
+ # - job-spec: "l2-00"
+ # build: "52"
+ # - job-spec: "l2-01"
+ # build: "53"
+ # - job-spec: "l2-02"
+ # build: "54"
+ # - job-spec: "l2-03"
+ # build: "55"
+ # - job-spec: "l2-04"
+ # build: "56"
+ # - job-spec: "l2-05"
+ # build: "57"
+ # - job-spec: "l2-06"
+ # build: "58"
+ # - job-spec: "l2-07"
+ # build: "59"
+ # - job-spec: "l2-08"
+ # build: "60"
+ # - job-spec: "l2-09"
+ # build: "61"
+ # - job-spec: "l2-10"
+ # build: "62"
+ # - job-spec: "memif-00"
+ # build: "63"
+ # - job-spec: "srv6-00"
+ # build: "64"
+ # - job-spec: "srv6-01"
+ # build: "65"
+ # - job-spec: "vhost-00"
+ # build: "66"
+ # - job-spec: "vhost-01"
+ # build: "67"
- type: "table"
title: "Job Specification Duration: 2n-clx-iterative"
@@ -81,47 +149,47 @@
- job-spec: "dpdk-mlr-00"
data-set: "dpdk-2n-clx-curr-iter"
-# - type: "table"
-# title: "Job Specification Duration: 2n-clx-coverage"
-# algorithm: "table_job_spec_duration"
-# output-file: "{DIR[STATIC,VPP]}/job-spec-duration-2n-clx-cov"
-# jb-type: "coverage"
-# data: "vpp-2n-clx-curr-cov"
-# lines:
-# - job-spec: "gso-00"
-# build: "18"
-# - job-spec: "ip4-00"
-# build: "19"
-# - job-spec: "ip4-01"
-# build: "20"
-# - job-spec: "ip4-02"
-# build: "21"
-# - job-spec: "ip4-03"
-# build: "22"
-# - job-spec: "ip4-04"
-# build: "23"
-# - job-spec: "ip4-05"
-# build: "24"
-# - job-spec: "ip4-06"
-# build: "25"
-# - job-spec: "ip4-07"
-# build: "26"
-# - job-spec: "ip4tun-00"
-# build: "27"
-# - job-spec: "ip6-00"
-# build: "28"
-# - job-spec: "ip6-01"
-# build: "29"
-# - job-spec: "l2-00"
-# build: "30"
-# - job-spec: "l2-01"
-# build: "31"
-# - job-spec: "lb-00"
-# build: "32"
-# - job-spec: "memif-00"
-# build: "33"
-# - job-spec: "vhost-00"
-# build: "34"
+- type: "table"
+ title: "Job Specification Duration: 2n-clx-coverage"
+ algorithm: "table_job_spec_duration"
+ output-file: "{DIR[STATIC,VPP]}/job-spec-duration-2n-clx-cov"
+ jb-type: "coverage"
+ data: "vpp-2n-clx-curr-cov"
+ lines:
+ - job-spec: "gso-00"
+ build: "18"
+ - job-spec: "ip4-00"
+ build: "19"
+ - job-spec: "ip4-01"
+ build: "20"
+ - job-spec: "ip4-02"
+ build: "21"
+ - job-spec: "ip4-03"
+ build: "22"
+ - job-spec: "ip4-04"
+ build: "23"
+ - job-spec: "ip4-05"
+ build: "24"
+ - job-spec: "ip4-06"
+ build: "25"
+ - job-spec: "ip4-07"
+ build: "26"
+ - job-spec: "ip4tun-00"
+ build: "27"
+ - job-spec: "ip6-00"
+ build: "28"
+ # - job-spec: "ip6-01"
+ # build: "29"
+ - job-spec: "l2-00"
+ build: "30"
+ - job-spec: "l2-01"
+ build: "31"
+ - job-spec: "lb-00"
+ build: "32"
+ - job-spec: "memif-00"
+ build: "33"
+ # - job-spec: "vhost-00"
+ # build: "34"
- type: "table"
title: "Job Specification Duration: 2n-dnv-iterative"
@@ -134,25 +202,25 @@
- job-spec: "vpp-mlr-00"
data-set: "vpp-2n-dnv-curr-iter"
-# - type: "table"
-# title: "Job Specification Duration: 2n-dnv-coverage"
-# algorithm: "table_job_spec_duration"
-# output-file: "{DIR[STATIC,VPP]}/job-spec-duration-2n-dnv-cov"
-# jb-type: "coverage"
-# data: "vpp-2n-dnv-curr-cov"
-# lines:
-# - job-spec: "ip4-00"
-# build: "7"
-# - job-spec: "ip4-01"
-# build: "8"
-# - job-spec: "ip6-00"
-# build: "9"
-# - job-spec: "ip6-01"
-# build: "10"
-# - job-spec: "l2-00"
-# build: "11"
-# - job-spec: "l2-01"
-# build: "12"
+- type: "table"
+ title: "Job Specification Duration: 2n-dnv-coverage"
+ algorithm: "table_job_spec_duration"
+ output-file: "{DIR[STATIC,VPP]}/job-spec-duration-2n-dnv-cov"
+ jb-type: "coverage"
+ data: "vpp-2n-dnv-curr-cov"
+ lines:
+ - job-spec: "ip4-00"
+ build: "7"
+ - job-spec: "ip4-01"
+ build: "8"
+ - job-spec: "ip6-00"
+ build: "14"
+ - job-spec: "ip6-01"
+ build: "10"
+ - job-spec: "l2-00"
+ build: "11"
+ - job-spec: "l2-01"
+ build: "12"
- type: "table"
title: "Job Specification Duration: 2n-aws-iterative"
@@ -221,47 +289,47 @@
- job-spec: "trex-mlr-00"
data-set: "trex-2n-skx-curr-iter"
-# - type: "table"
-# title: "Job Specification Duration: 2n-skx-coverage"
-# algorithm: "table_job_spec_duration"
-# output-file: "{DIR[STATIC,VPP]}/job-spec-duration-2n-skx-cov"
-# jb-type: "coverage"
-# data: "vpp-2n-skx-curr-cov"
-# lines:
-# - job-spec: "gso-00"
-# build: "18"
-# - job-spec: "ip4-00"
-# build: "19"
-# - job-spec: "ip4-01"
-# build: "20"
-# - job-spec: "ip4-02"
-# build: "21"
-# - job-spec: "ip4-03"
-# build: "22"
-# - job-spec: "ip4-04"
-# build: "23"
-# - job-spec: "ip4-05"
-# build: "24"
-# - job-spec: "ip4-06"
-# build: "25"
-# - job-spec: "ip4-07"
-# build: "26"
-# - job-spec: "ip4tun-00"
-# build: "27"
-# - job-spec: "ip6-00"
-# build: "28"
-# - job-spec: "ip6-01"
-# build: "29"
-# - job-spec: "l2-00"
-# build: "30"
-# - job-spec: "l2-01"
-# build: "31"
-# - job-spec: "lb-00"
-# build: "32"
-# - job-spec: "memif-00"
-# build: "33"
-# - job-spec: "vhost-00"
-# build: "34"
+- type: "table"
+ title: "Job Specification Duration: 2n-skx-coverage"
+ algorithm: "table_job_spec_duration"
+ output-file: "{DIR[STATIC,VPP]}/job-spec-duration-2n-skx-cov"
+ jb-type: "coverage"
+ data: "vpp-2n-skx-curr-cov"
+ lines:
+ - job-spec: "gso-00"
+ build: "18"
+ - job-spec: "ip4-00"
+ build: "19"
+ - job-spec: "ip4-01"
+ build: "20"
+ - job-spec: "ip4-02"
+ build: "21"
+ - job-spec: "ip4-03"
+ build: "22"
+ - job-spec: "ip4-04"
+ build: "23"
+ - job-spec: "ip4-05"
+ build: "24"
+ - job-spec: "ip4-06"
+ build: "25"
+ - job-spec: "ip4-07"
+ build: "26"
+ - job-spec: "ip4tun-00"
+ build: "27"
+ - job-spec: "ip6-00"
+ build: "28"
+ - job-spec: "ip6-01"
+ build: "29"
+ - job-spec: "l2-00"
+ build: "30"
+ - job-spec: "l2-01"
+ build: "31"
+ - job-spec: "lb-00"
+ build: "32"
+ - job-spec: "memif-00"
+ build: "33"
+ - job-spec: "vhost-00"
+ build: "34"
- type: "table"
title: "Job Specification Duration: 2n-tx2-iterative"
@@ -374,33 +442,33 @@
- job-spec: "vpp-mlr-00"
data-set: "vpp-3n-dnv-curr-iter"
-# - type: "table"
-# title: "Job Specification Duration: 3n-dnv-coverage"
-# algorithm: "table_job_spec_duration"
-# output-file: "{DIR[STATIC,VPP]}/job-spec-duration-3n-dnv-cov"
-# jb-type: "coverage"
-# data: "vpp-3n-dnv-curr-cov"
-# lines:
-# - job-spec: "crypto-00"
-# build: "12"
-# - job-spec: "crypto-01"
-# build: "13"
-# - job-spec: "crypto-02"
-# build: "14"
-# - job-spec: "ip4-00"
-# build: "15"
-# - job-spec: "ip4-01"
-# build: "16"
-# - job-spec: "ip4tun-00"
-# build: "17"
-# - job-spec: "ip6-00"
-# build: "18"
-# - job-spec: "ip6-01"
-# build: "19"
-# - job-spec: "l2-00"
-# build: "20"
-# - job-spec: "l2-01"
-# build: "21"
+- type: "table"
+ title: "Job Specification Duration: 3n-dnv-coverage"
+ algorithm: "table_job_spec_duration"
+ output-file: "{DIR[STATIC,VPP]}/job-spec-duration-3n-dnv-cov"
+ jb-type: "coverage"
+ data: "vpp-3n-dnv-curr-cov"
+ lines:
+ - job-spec: "crypto-00"
+ build: "11"
+ - job-spec: "crypto-01"
+ build: "12"
+ - job-spec: "crypto-02"
+ build: "13"
+ - job-spec: "ip4-00"
+ build: "14"
+ - job-spec: "ip4-01"
+ build: "15"
+ - job-spec: "ip4tun-00"
+ build: "16"
+ - job-spec: "ip6-00"
+ build: "17"
+ - job-spec: "ip6-01"
+ build: "18"
+ - job-spec: "l2-00"
+ build: "19"
+ - job-spec: "l2-01"
+ build: "20"
- type: "table"
title: "Job Specification Duration: 3n-skx-iterative"
@@ -417,81 +485,81 @@
- job-spec: "dpdk-mlr-00"
data-set: "dpdk-3n-skx-curr-iter"
-# - type: "table"
-# title: "Job Specification Duration: 3n-skx-coverage"
-# algorithm: "table_job_spec_duration"
-# output-file: "{DIR[STATIC,VPP]}/job-spec-duration-3n-skx-cov"
-# jb-type: "coverage"
-# data: "vpp-3n-skx-curr-cov"
-# lines:
-# - job-spec: "crypto-00"
-# build: "35"
-# - job-spec: "crypto-01"
-# build: "36"
-# - job-spec: "crypto-02"
-# build: "37"
-# - job-spec: "crypto-03" # Completly failed
-# build: "38"
-# - job-spec: "crypto-04"
-# build: "39"
-# - job-spec: "ip4-00"
-# build: "40"
-# - job-spec: "ip4-01"
-# build: "41"
-# - job-spec: "ip4-02"
-# build: "42"
-# - job-spec: "ip4-03"
-# build: "43"
-# - job-spec: "ip4-04"
-# build: "44"
-# - job-spec: "ip4-05"
-# build: "45"
-# - job-spec: "ip4-06"
-# build: "46"
-# - job-spec: "ip4-07"
-# build: "47"
-# - job-spec: "ip4tun-00"
-# build: "48"
-# - job-spec: "ip4tun-02"
-# build: "49"
-# - job-spec: "ip6-00"
-# build: "50"
-# - job-spec: "ip6-01"
-# build: "51"
-# - job-spec: "ip6tun-00"
-# build: "52"
-# - job-spec: "l2-00"
-# build: "53"
-# - job-spec: "l2-01"
-# build: "54"
-# - job-spec: "l2-02"
-# build: "55"
-# - job-spec: "l2-03"
-# build: "56"
-# - job-spec: "l2-04"
-# build: "57"
-# - job-spec: "l2-05"
-# build: "58"
-# - job-spec: "l2-06"
-# build: "59"
-# - job-spec: "l2-07"
-# build: "60"
-# - job-spec: "l2-08"
-# build: "61"
-# - job-spec: "l2-09"
-# build: "62"
-# - job-spec: "l2-10"
-# build: "63"
-# - job-spec: "memif-00"
-# build: "64"
-# - job-spec: "srv6-00"
-# build: "65"
-# - job-spec: "srv6-01"
-# build: "66"
-# - job-spec: "vhost-00"
-# build: "67"
-# - job-spec: "vhost-01"
-# build: "68"
+- type: "table"
+ title: "Job Specification Duration: 3n-skx-coverage"
+ algorithm: "table_job_spec_duration"
+ output-file: "{DIR[STATIC,VPP]}/job-spec-duration-3n-skx-cov"
+ jb-type: "coverage"
+ data: "vpp-3n-skx-curr-cov"
+ lines:
+ - job-spec: "crypto-00"
+ build: "35"
+ - job-spec: "crypto-01"
+ build: "36"
+ - job-spec: "crypto-02"
+ build: "37"
+ # - job-spec: "crypto-03" # Completly failed
+ # build: "38"
+ - job-spec: "crypto-04"
+ build: "38"
+ - job-spec: "ip4-00"
+ build: "39"
+ - job-spec: "ip4-01"
+ build: "40"
+ - job-spec: "ip4-02"
+ build: "41"
+ - job-spec: "ip4-03"
+ build: "42"
+ - job-spec: "ip4-04"
+ build: "43"
+ - job-spec: "ip4-05"
+ build: "44"
+ - job-spec: "ip4-06"
+ build: "45"
+ - job-spec: "ip4-07"
+ build: "46"
+ - job-spec: "ip4tun-00"
+ build: "47"
+ - job-spec: "ip4tun-02"
+ build: "48"
+ - job-spec: "ip6-00"
+ build: "49"
+ - job-spec: "ip6-01"
+ build: "50"
+ - job-spec: "ip6tun-00"
+ build: "51"
+ - job-spec: "l2-00"
+ build: "52"
+ - job-spec: "l2-01"
+ build: "53"
+ - job-spec: "l2-02"
+ build: "54"
+ - job-spec: "l2-03"
+ build: "55"
+ - job-spec: "l2-04"
+ build: "56"
+ - job-spec: "l2-05"
+ build: "57"
+ - job-spec: "l2-06"
+ build: "58"
+ - job-spec: "l2-07"
+ build: "59"
+ - job-spec: "l2-08"
+ build: "60"
+ - job-spec: "l2-09"
+ build: "61"
+ - job-spec: "l2-10"
+ build: "62"
+ - job-spec: "memif-00"
+ build: "63"
+ - job-spec: "srv6-00"
+ build: "64"
+ - job-spec: "srv6-01"
+ build: "65"
+ - job-spec: "vhost-00"
+ build: "66"
+ - job-spec: "vhost-01"
+ build: "67"
- type: "table"
title: "Job Specification Duration: 3n-tsh-iterative"
diff --git a/resources/tools/presentation/sphinx_conf/report/conf.py b/resources/tools/presentation/sphinx_conf/report/conf.py
index 3e3b779167..69ffbc240b 100644
--- a/resources/tools/presentation/sphinx_conf/report/conf.py
+++ b/resources/tools/presentation/sphinx_conf/report/conf.py
@@ -46,7 +46,7 @@ source_suffix = [u'.rst', u'.md']
master_doc = u'index'
# General information about the project.
-report_week = u'09'
+report_week = u'10'
project = u'FD.io CSIT-2202.{week}'.format(week=report_week)
copyright = u'2022, FD.io'
author = u'FD.io CSIT'
up_dport, u8 *lookup_protocol) { icmp46_header_t *icmp0; nat_icmp_echo_header_t *echo0, *inner_echo0 = 0; ip4_header_t *inner_ip0 = 0; void *l4_header = 0; icmp46_header_t *inner_icmp0; icmp0 = (icmp46_header_t *) ip4_next_header (ip0); echo0 = (nat_icmp_echo_header_t *) (icmp0 + 1); // avoid warning about unused variables in caller by setting to bogus values *lookup_sport = 0; *lookup_dport = 0; if (!icmp_type_is_error_message ( vnet_buffer (b)->ip.reass.icmp_type_or_tcp_flags)) { *lookup_protocol = IP_PROTOCOL_ICMP; lookup_saddr->as_u32 = ip0->src_address.as_u32; *lookup_sport = vnet_buffer (b)->ip.reass.l4_src_port; lookup_daddr->as_u32 = ip0->dst_address.as_u32; *lookup_dport = vnet_buffer (b)->ip.reass.l4_dst_port; } else { inner_ip0 = (ip4_header_t *) (echo0 + 1); l4_header = ip4_next_header (inner_ip0); *lookup_protocol = inner_ip0->protocol; lookup_saddr->as_u32 = inner_ip0->dst_address.as_u32; lookup_daddr->as_u32 = inner_ip0->src_address.as_u32; switch (inner_ip0->protocol) { case IP_PROTOCOL_ICMP: inner_icmp0 = (icmp46_header_t *) l4_header; inner_echo0 = (nat_icmp_echo_header_t *) (inner_icmp0 + 1); *lookup_sport = inner_echo0->identifier; *lookup_dport = inner_echo0->identifier; break; case IP_PROTOCOL_UDP: case IP_PROTOCOL_TCP: *lookup_sport = ((nat_tcp_udp_header_t *) l4_header)->dst_port; *lookup_dport = ((nat_tcp_udp_header_t *) l4_header)->src_port; break; default: return NAT_IN2OUT_ED_ERROR_UNSUPPORTED_PROTOCOL; } } return 0; } always_inline int nat44_ed_tcp_is_established (nat44_ed_tcp_state_e state) { return state == NAT44_ED_TCP_STATE_ESTABLISHED ? 1 : 0; } always_inline u32 nat44_session_get_timeout (snat_main_t *sm, snat_session_t *s) { switch (s->proto) { case IP_PROTOCOL_ICMP: /* fallthrough */ case IP_PROTOCOL_ICMP6: return sm->timeouts.icmp; case IP_PROTOCOL_UDP: return sm->timeouts.udp; case IP_PROTOCOL_TCP: { if (nat44_ed_tcp_is_established (s->tcp_state)) return sm->timeouts.tcp.established; else return sm->timeouts.tcp.transitory; } default: return sm->timeouts.udp; } return 0; } static_always_inline u8 nat44_ed_maximum_sessions_exceeded (snat_main_t *sm, u32 fib_index, u32 thread_index) { u32 translations; translations = pool_elts (sm->per_thread_data[thread_index].sessions); if (vec_len (sm->max_translations_per_fib) <= fib_index) fib_index = 0; return translations >= sm->max_translations_per_fib[fib_index]; } static_always_inline int nat_ed_lru_insert (snat_main_per_thread_data_t *tsm, snat_session_t *s, f64 now, u8 proto) { dlist_elt_t *lru_list_elt; pool_get (tsm->lru_pool, lru_list_elt); s->lru_index = lru_list_elt - tsm->lru_pool; switch (proto) { case IP_PROTOCOL_UDP: s->lru_head_index = tsm->udp_lru_head_index; break; case IP_PROTOCOL_TCP: s->lru_head_index = tsm->tcp_trans_lru_head_index; break; case IP_PROTOCOL_ICMP: s->lru_head_index = tsm->icmp_lru_head_index; break; default: s->lru_head_index = tsm->unk_proto_lru_head_index; break; } clib_dlist_addtail (tsm->lru_pool, s->lru_head_index, s->lru_index); lru_list_elt->value = s - tsm->sessions; s->last_lru_update = now; return 1; } static_always_inline void nat_6t_flow_to_ed_k (clib_bihash_kv_16_8_t *kv, nat_6t_flow_t *f) { init_ed_k (kv, f->match.saddr.as_u32, f->match.sport, f->match.daddr.as_u32, f->match.dport, f->match.fib_index, f->match.proto); } static_always_inline void nat_6t_flow_to_ed_kv (clib_bihash_kv_16_8_t *kv, nat_6t_flow_t *f, u32 thread_idx, u32 session_idx) { init_ed_kv (kv, f->match.saddr.as_u32, f->match.sport, f->match.daddr.as_u32, f->match.dport, f->match.fib_index, f->match.proto, thread_idx, session_idx); } static_always_inline int nat_ed_ses_i2o_flow_hash_add_del (snat_main_t *sm, u32 thread_idx, snat_session_t *s, int is_add) { snat_main_per_thread_data_t *tsm = vec_elt_at_index (sm->per_thread_data, thread_idx); clib_bihash_kv_16_8_t kv; if (0 == is_add) { nat_6t_flow_to_ed_k (&kv, &s->i2o); } else { nat_6t_flow_to_ed_kv (&kv, &s->i2o, thread_idx, s - tsm->sessions); nat_6t_l3_l4_csum_calc (&s->i2o); } ASSERT (thread_idx == s->thread_index); return clib_bihash_add_del_16_8 (&sm->flow_hash, &kv, is_add); } static_always_inline int nat_ed_ses_o2i_flow_hash_add_del (snat_main_t *sm, u32 thread_idx, snat_session_t *s, int is_add) { snat_main_per_thread_data_t *tsm = vec_elt_at_index (sm->per_thread_data, thread_idx); clib_bihash_kv_16_8_t kv; if (0 == is_add) { nat_6t_flow_to_ed_k (&kv, &s->o2i); } else { nat_6t_flow_to_ed_kv (&kv, &s->o2i, thread_idx, s - tsm->sessions); if (!(s->flags & SNAT_SESSION_FLAG_STATIC_MAPPING)) { if (nat44_ed_sm_o2i_lookup (sm, s->o2i.match.daddr, s->o2i.match.dport, 0, s->o2i.match.proto)) { return -1; } } nat_6t_l3_l4_csum_calc (&s->o2i); } ASSERT (thread_idx == s->thread_index); return clib_bihash_add_del_16_8 (&sm->flow_hash, &kv, is_add); } always_inline void nat_ed_session_delete (snat_main_t *sm, snat_session_t *ses, u32 thread_index, int lru_delete /* delete from global LRU list */) { snat_main_per_thread_data_t *tsm = vec_elt_at_index (sm->per_thread_data, thread_index); if (lru_delete) { clib_dlist_remove (tsm->lru_pool, ses->lru_index); } pool_put_index (tsm->lru_pool, ses->lru_index); if (nat_ed_ses_i2o_flow_hash_add_del (sm, thread_index, ses, 0)) nat_elog_warn (sm, "flow hash del failed"); if (nat_ed_ses_o2i_flow_hash_add_del (sm, thread_index, ses, 0)) nat_elog_warn (sm, "flow hash del failed"); pool_put (tsm->sessions, ses); vlib_set_simple_counter (&sm->total_sessions, thread_index, 0, pool_elts (tsm->sessions)); } static_always_inline int nat_lru_free_one_with_head (snat_main_t *sm, int thread_index, f64 now, u32 head_index) { snat_session_t *s = NULL; dlist_elt_t *oldest_elt; f64 sess_timeout_time; u32 oldest_index; snat_main_per_thread_data_t *tsm = &sm->per_thread_data[thread_index]; oldest_index = clib_dlist_remove_head (tsm->lru_pool, head_index); if (~0 != oldest_index) { oldest_elt = pool_elt_at_index (tsm->lru_pool, oldest_index); s = pool_elt_at_index (tsm->sessions, oldest_elt->value); sess_timeout_time = s->last_heard + (f64) nat44_session_get_timeout (sm, s); if (now >= sess_timeout_time) { nat44_ed_free_session_data (sm, s, thread_index, 0); nat_ed_session_delete (sm, s, thread_index, 0); return 1; } else { clib_dlist_addhead (tsm->lru_pool, head_index, oldest_index); } } return 0; } static_always_inline int nat_lru_free_one (snat_main_t *sm, int thread_index, f64 now) { snat_main_per_thread_data_t *tsm = &sm->per_thread_data[thread_index]; int rc = 0; #define _(p) \ if ((rc = nat_lru_free_one_with_head (sm, thread_index, now, \ tsm->p##_lru_head_index))) \ { \ return rc; \ } _ (tcp_trans); _ (udp); _ (unk_proto); _ (icmp); _ (tcp_estab); #undef _ return 0; } static_always_inline snat_session_t * nat_ed_session_alloc (snat_main_t *sm, u32 thread_index, f64 now, u8 proto) { snat_session_t *s; snat_main_per_thread_data_t *tsm = &sm->per_thread_data[thread_index]; nat_lru_free_one (sm, thread_index, now); pool_get (tsm->sessions, s); clib_memset (s, 0, sizeof (*s)); nat_ed_lru_insert (tsm, s, now, proto); s->ha_last_refreshed = now; vlib_set_simple_counter (&sm->total_sessions, thread_index, 0, pool_elts (tsm->sessions)); #if CLIB_ASSERT_ENABLE s->thread_index = thread_index; #endif return s; } // slow path static_always_inline void per_vrf_sessions_cleanup (u32 thread_index) { snat_main_t *sm = &snat_main; snat_main_per_thread_data_t *tsm = vec_elt_at_index (sm->per_thread_data, thread_index); per_vrf_sessions_t *per_vrf_sessions; u32 *to_free = 0, *i; pool_foreach (per_vrf_sessions, tsm->per_vrf_sessions_pool) { if (per_vrf_sessions->expired && per_vrf_sessions->ses_count == 0) vec_add1 (to_free, per_vrf_sessions - tsm->per_vrf_sessions_pool); } vec_foreach (i, to_free) { per_vrf_sessions = pool_elt_at_index (tsm->per_vrf_sessions_pool, *i); pool_put (tsm->per_vrf_sessions_pool, per_vrf_sessions); } vec_free (to_free); } // slow path static_always_inline void per_vrf_sessions_register_session (snat_session_t *s, u32 thread_index) { snat_main_t *sm = &snat_main; snat_main_per_thread_data_t *tsm = vec_elt_at_index (sm->per_thread_data, thread_index); per_vrf_sessions_t *per_vrf_sessions; per_vrf_sessions_cleanup (thread_index); // s->per_vrf_sessions_index == ~0 ... reuse of old session pool_foreach (per_vrf_sessions, tsm->per_vrf_sessions_pool) { // ignore already expired registrations if (per_vrf_sessions->expired) continue; if ((s->in2out.fib_index == per_vrf_sessions->rx_fib_index) && (s->out2in.fib_index == per_vrf_sessions->tx_fib_index)) { goto done; } if ((s->in2out.fib_index == per_vrf_sessions->tx_fib_index) && (s->out2in.fib_index == per_vrf_sessions->rx_fib_index)) { goto done; } } // create a new registration pool_get (tsm->per_vrf_sessions_pool, per_vrf_sessions); clib_memset (per_vrf_sessions, 0, sizeof (*per_vrf_sessions)); per_vrf_sessions->rx_fib_index = s->in2out.fib_index; per_vrf_sessions->tx_fib_index = s->out2in.fib_index; done: s->per_vrf_sessions_index = per_vrf_sessions - tsm->per_vrf_sessions_pool; per_vrf_sessions->ses_count++; } // fast path static_always_inline void per_vrf_sessions_unregister_session (snat_session_t *s, u32 thread_index) { snat_main_t *sm = &snat_main; snat_main_per_thread_data_t *tsm; per_vrf_sessions_t *per_vrf_sessions; ASSERT (s->per_vrf_sessions_index != ~0); tsm = vec_elt_at_index (sm->per_thread_data, thread_index); per_vrf_sessions = pool_elt_at_index (tsm->per_vrf_sessions_pool, s->per_vrf_sessions_index); ASSERT (per_vrf_sessions->ses_count != 0); per_vrf_sessions->ses_count--; s->per_vrf_sessions_index = ~0; } // fast path static_always_inline u8 per_vrf_sessions_is_expired (snat_session_t *s, u32 thread_index) { snat_main_t *sm = &snat_main; snat_main_per_thread_data_t *tsm; per_vrf_sessions_t *per_vrf_sessions; ASSERT (s->per_vrf_sessions_index != ~0); tsm = vec_elt_at_index (sm->per_thread_data, thread_index); per_vrf_sessions = pool_elt_at_index (tsm->per_vrf_sessions_pool, s->per_vrf_sessions_index); return per_vrf_sessions->expired; } static_always_inline void nat_6t_flow_init (nat_6t_flow_t *f, u32 thread_idx, ip4_address_t saddr, u16 sport, ip4_address_t daddr, u16 dport, u32 fib_index, u8 proto, u32 session_idx) { clib_memset (f, 0, sizeof (*f)); f->match.saddr = saddr; f->match.sport = sport; f->match.daddr = daddr; f->match.dport = dport; f->match.proto = proto; f->match.fib_index = fib_index; } static_always_inline void nat_6t_i2o_flow_init (snat_main_t *sm, u32 thread_idx, snat_session_t *s, ip4_address_t saddr, u16 sport, ip4_address_t daddr, u16 dport, u32 fib_index, u8 proto) { snat_main_per_thread_data_t *tsm = vec_elt_at_index (sm->per_thread_data, thread_idx); nat_6t_flow_init (&s->i2o, thread_idx, saddr, sport, daddr, dport, fib_index, proto, s - tsm->sessions); } static_always_inline void nat_6t_o2i_flow_init (snat_main_t *sm, u32 thread_idx, snat_session_t *s, ip4_address_t saddr, u16 sport, ip4_address_t daddr, u16 dport, u32 fib_index, u8 proto) { snat_main_per_thread_data_t *tsm = vec_elt_at_index (sm->per_thread_data, thread_idx); nat_6t_flow_init (&s->o2i, thread_idx, saddr, sport, daddr, dport, fib_index, proto, s - tsm->sessions); } static_always_inline int nat_6t_t_eq (nat_6t_t *t1, nat_6t_t *t2) { return t1->as_u64[0] == t2->as_u64[0] && t1->as_u64[1] == t2->as_u64[1]; } static inline uword nat_pre_node_fn_inline (vlib_main_t *vm, vlib_node_runtime_t *node, vlib_frame_t *frame, u32 def_next) { u32 n_left_from, *from; from = vlib_frame_vector_args (frame); n_left_from = frame->n_vectors; vlib_buffer_t *bufs[VLIB_FRAME_SIZE], **b = bufs; u16 nexts[VLIB_FRAME_SIZE], *next = nexts; vlib_get_buffers (vm, from, b, n_left_from); while (n_left_from >= 2) { u32 next0, next1; u32 arc_next0, arc_next1; vlib_buffer_t *b0, *b1; b0 = *b; b++; b1 = *b; b++; /* Prefetch next iteration. */ if (PREDICT_TRUE (n_left_from >= 4)) { vlib_buffer_t *p2, *p3; p2 = *b; p3 = *(b + 1); vlib_prefetch_buffer_header (p2, LOAD); vlib_prefetch_buffer_header (p3, LOAD); clib_prefetch_load (p2->data); clib_prefetch_load (p3->data); } next0 = def_next; next1 = def_next; vnet_feature_next (&arc_next0, b0); vnet_feature_next (&arc_next1, b1); vnet_buffer2 (b0)->nat.arc_next = arc_next0; vnet_buffer2 (b1)->nat.arc_next = arc_next1; if (PREDICT_FALSE ((node->flags & VLIB_NODE_FLAG_TRACE))) { if (b0->flags & VLIB_BUFFER_IS_TRACED) { nat_pre_trace_t *t = vlib_add_trace (vm, node, b0, sizeof (*t)); t->next_index = next0; t->arc_next_index = arc_next0; } if (b1->flags & VLIB_BUFFER_IS_TRACED) { nat_pre_trace_t *t = vlib_add_trace (vm, node, b0, sizeof (*t)); t->next_index = next1; t->arc_next_index = arc_next1; } } n_left_from -= 2; next[0] = next0; next[1] = next1; next += 2; } while (n_left_from > 0) { u32 next0; u32 arc_next0; vlib_buffer_t *b0; b0 = *b; b++; next0 = def_next; vnet_feature_next (&arc_next0, b0); vnet_buffer2 (b0)->nat.arc_next = arc_next0; if (PREDICT_FALSE ((node->flags & VLIB_NODE_FLAG_TRACE) && (b0->flags & VLIB_BUFFER_IS_TRACED))) { nat_pre_trace_t *t = vlib_add_trace (vm, node, b0, sizeof (*t)); t->next_index = next0; t->arc_next_index = arc_next0; } n_left_from--; next[0] = next0; next++; } vlib_buffer_enqueue_to_next (vm, node, from, (u16 *) nexts, frame->n_vectors); return frame->n_vectors; } static_always_inline u16 snat_random_port (u16 min, u16 max) { snat_main_t *sm = &snat_main; u32 rwide; u16 r; rwide = random_u32 (&sm->random_seed); r = rwide & 0xFFFF; if (r >= min && r <= max) return r; return min + (rwide % (max - min + 1)); } always_inline u8 is_interface_addr (snat_main_t *sm, vlib_node_runtime_t *node, u32 sw_if_index0, u32 ip4_addr) { snat_runtime_t *rt = (snat_runtime_t *) node->runtime_data; u8 ip4_addr_exists; if (PREDICT_FALSE (rt->cached_sw_if_index != sw_if_index0)) { ip_lookup_main_t *lm = &sm->ip4_main->lookup_main; ip_interface_address_t *ia; ip4_address_t *a; rt->cached_sw_if_index = ~0; hash_free (rt->cached_presence_by_ip4_address); foreach_ip_interface_address ( lm, ia, sw_if_index0, 1 /* honor unnumbered */, ({ a = ip_interface_address_get_address (lm, ia); hash_set (rt->cached_presence_by_ip4_address, a->as_u32, 1); rt->cached_sw_if_index = sw_if_index0; })); if (rt->cached_sw_if_index == ~0) return 0; } ip4_addr_exists = !!hash_get (rt->cached_presence_by_ip4_address, ip4_addr); if (PREDICT_FALSE (ip4_addr_exists)) return 1; else return 0; } always_inline void nat44_ed_session_reopen (u32 thread_index, snat_session_t *s) { nat_syslog_nat44_sdel (0, s->in2out.fib_index, &s->in2out.addr, s->in2out.port, &s->ext_host_nat_addr, s->ext_host_nat_port, &s->out2in.addr, s->out2in.port, &s->ext_host_addr, s->ext_host_port, s->proto, nat44_ed_is_twice_nat_session (s)); nat_ipfix_logging_nat44_ses_delete ( thread_index, s->in2out.addr.as_u32, s->out2in.addr.as_u32, s->proto, s->in2out.port, s->out2in.port, s->in2out.fib_index); nat_ipfix_logging_nat44_ses_create ( thread_index, s->in2out.addr.as_u32, s->out2in.addr.as_u32, s->proto, s->in2out.port, s->out2in.port, s->in2out.fib_index); nat_syslog_nat44_sadd (0, s->in2out.fib_index, &s->in2out.addr, s->in2out.port, &s->ext_host_nat_addr, s->ext_host_nat_port, &s->out2in.addr, s->out2in.port, &s->ext_host_addr, s->ext_host_port, s->proto, 0); s->total_pkts = 0; s->total_bytes = 0; } /* * "Some rise by SYN, and some by virtue FIN" - William Shakespeare * TCP state tracking patterned after RFC 7857 (and RFC 6146, which is * referenced by RFC 7857). In contrast to the state machine in RFC7857 we only * transition to ESTABLISHED state after seeing a full 3-way handshake (SYNs * and ACKs in both directions). RFC7857 as a means of protecting against * spurious RSTs closing a session, goes back to ESTABLISHED if a data packet * is received after the RST. This state machine will leave the state in * transitory if RST is seen. Our implementation also goes beyond by supporting * creation of a new session while old session is in transitory timeout after * seeing FIN packets from both sides. */ always_inline void nat44_set_tcp_session_state (snat_main_t *sm, f64 now, snat_session_t *ses, u8 tcp_flags, u32 thread_index, nat44_ed_dir_e dir) { snat_main_per_thread_data_t *tsm = &sm->per_thread_data[thread_index]; u8 old_flags = ses->tcp_flags[dir]; ses->tcp_flags[dir] |= tcp_flags & (TCP_FLAG_FIN | TCP_FLAG_SYN | TCP_FLAG_RST | TCP_FLAG_ACK); if (old_flags == ses->tcp_flags[dir]) return; u8 old_state = ses->tcp_state; switch (old_state) { case NAT44_ED_TCP_STATE_CLOSED: // ESTABLISHED when a SYN and ACK is seen from both sides if ((ses->tcp_flags[NAT44_ED_DIR_I2O] & ses->tcp_flags[NAT44_ED_DIR_O2I]) == (TCP_FLAG_SYN | TCP_FLAG_ACK)) { ses->tcp_state = NAT44_ED_TCP_STATE_ESTABLISHED; ses->lru_head_index = tsm->tcp_estab_lru_head_index; } break; case NAT44_ED_TCP_STATE_ESTABLISHED: // CLOSING when a FIN is seen from either side or session has been RST if ((ses->tcp_flags[dir] & TCP_FLAG_FIN) || (ses->tcp_flags[dir] & TCP_FLAG_RST)) { ses->tcp_state = NAT44_ED_TCP_STATE_CLOSING; ses->tcp_flags[NAT44_ED_DIR_I2O] = 0; ses->tcp_flags[NAT44_ED_DIR_O2I] = 0; // need to update last heard otherwise session might get // immediately timed out if it has been idle longer than // transitory timeout ses->last_heard = now; ses->lru_head_index = tsm->tcp_trans_lru_head_index; } break; case NAT44_ED_TCP_STATE_CLOSING: // Allow a transitory session to reopen if ((ses->tcp_flags[NAT44_ED_DIR_I2O] & ses->tcp_flags[NAT44_ED_DIR_O2I]) == (TCP_FLAG_SYN | TCP_FLAG_ACK)) { nat44_ed_session_reopen (thread_index, ses); ses->tcp_state = NAT44_ED_TCP_STATE_ESTABLISHED; ses->lru_head_index = tsm->tcp_estab_lru_head_index; } break; } if (old_state == ses->tcp_state) return; ses->last_lru_update = now; clib_dlist_remove (tsm->lru_pool, ses->lru_index); clib_dlist_addtail (tsm->lru_pool, ses->lru_head_index, ses->lru_index); } always_inline void nat44_set_tcp_session_state_i2o (snat_main_t *sm, f64 now, snat_session_t *ses, u8 tcp_flags, u32 thread_index) { return nat44_set_tcp_session_state (sm, now, ses, tcp_flags, thread_index, NAT44_ED_DIR_I2O); } always_inline void nat44_set_tcp_session_state_o2i (snat_main_t *sm, f64 now, snat_session_t *ses, u8 tcp_flags, u32 thread_index) { return nat44_set_tcp_session_state (sm, now, ses, tcp_flags, thread_index, NAT44_ED_DIR_O2I); } always_inline void nat44_session_update_counters (snat_session_t *s, f64 now, uword bytes, u32 thread_index) { // regardless of TCP state, reset the timer if data packet is seen. s->last_heard = now; s->total_pkts++; s->total_bytes += bytes; } /** \brief Per-user LRU list maintenance */ always_inline void nat44_session_update_lru (snat_main_t *sm, snat_session_t *s, u32 thread_index) { /* don't update too often - timeout is in magnitude of seconds anyway */ if (s->last_heard > s->last_lru_update + 1) { clib_dlist_remove (sm->per_thread_data[thread_index].lru_pool, s->lru_index); clib_dlist_addtail (sm->per_thread_data[thread_index].lru_pool, s->lru_head_index, s->lru_index); s->last_lru_update = s->last_heard; } } static_always_inline int nat44_ed_is_unk_proto (u8 proto) { static const int lookup_table[256] = { [IP_PROTOCOL_TCP] = 1, [IP_PROTOCOL_UDP] = 1, [IP_PROTOCOL_ICMP] = 1, [IP_PROTOCOL_ICMP6] = 1, }; return 1 - lookup_table[proto]; } #endif /* __included_nat44_ed_inlines_h__ */ /* * fd.io coding-style-patch-verification: ON * * Local Variables: * eval: (c-set-style "gnu") * End: */