diff options
Diffstat (limited to 'docs/content/infrastructure')
12 files changed, 339 insertions, 5 deletions
diff --git a/docs/content/infrastructure/fdio_csit_logical_topologies.md b/docs/content/infrastructure/fdio_csit_logical_topologies.md new file mode 100644 index 0000000000..5dd323d30c --- /dev/null +++ b/docs/content/infrastructure/fdio_csit_logical_topologies.md @@ -0,0 +1,138 @@ +--- +title: "FD.io CSIT Logical Topologies" +weight: 4 +--- + +# FD.io CSIT Logical Topologies + +CSIT VPP performance tests are executed on physical testbeds. Based on the +packet path thru server SUTs, three distinct logical topology types are used +for VPP DUT data plane testing: + +1. NIC-to-NIC switching topologies. +2. VM service switching topologies. +3. Container service switching topologies. + +## NIC-to-NIC Switching + +The simplest logical topology for software data plane application like +VPP is NIC-to-NIC switching. Tested topologies for 2-Node and 3-Node +testbeds are shown in figures below. + +{{< figure src="/cdocs/logical-2n-nic2nic.svg" >}} + +{{< figure src="/cdocs/logical-3n-nic2nic.svg" >}} + +Server Systems Under Test (SUT) run VPP application in Linux user-mode +as a Device Under Test (DUT). Server Traffic Generator (TG) runs T-Rex +application. Physical connectivity between SUTs and TG is provided using +different drivers and NIC models that need to be tested for performance +(packet/bandwidth throughput and latency). + +From SUT and DUT perspectives, all performance tests involve forwarding +packets between two (or more) physical Ethernet ports (10GE, 25GE, 40GE, +100GE). In most cases both physical ports on SUT are located on the same +NIC. The only exceptions are link bonding and 100GE tests. In the latter +case only one port per NIC can be driven at linerate due to PCIe Gen3 +x16 slot bandwidth limiations. 100GE NICs are not supported in PCIe Gen3 +x8 slots. + +Note that reported VPP DUT performance results are specific to the SUTs +tested. SUTs with other processors than the ones used in FD.io lab are +likely to yield different results. A good rule of thumb, that can be +applied to estimate VPP packet thoughput for NIC-to-NIC switching +topology, is to expect the forwarding performance to be proportional to +processor core frequency for the same processor architecture, assuming +processor is the only limiting factor and all other SUT parameters are +equivalent to FD.io CSIT environment. + +## VM Service Switching + +VM service switching topology test cases require VPP DUT to communicate +with Virtual Machines (VMs) over vhost-user virtual interfaces. + +Two types of VM service topologies are tested: + +1. "Parallel" topology with packets flowing within SUT from NIC(s) via + VPP DUT to VM, back to VPP DUT, then out thru NIC(s). +2. "Chained" topology (a.k.a. "Snake") with packets flowing within SUT + from NIC(s) via VPP DUT to VM, back to VPP DUT, then to the next VM, + back to VPP DUT and so on and so forth until the last VM in a chain, + then back to VPP DUT and out thru NIC(s). + +For each of the above topologies, VPP DUT is tested in a range of L2 +or IPv4/IPv6 configurations depending on the test suite. Sample VPP DUT +"Chained" VM service topologies for 2-Node and 3-Node testbeds with each +SUT running N of VM instances is shown in the figures below. + +{{< figure src="/cdocs/logical-2n-vm-vhost.svg" >}} + +{{< figure src="/cdocs/logical-3n-vm-vhost.svg" >}} + +In "Chained" VM topologies, packets are switched by VPP DUT multiple +times: twice for a single VM, three times for two VMs, N+1 times for N +VMs. Hence the external throughput rates measured by TG and listed in +this report must be multiplied by N+1 to represent the actual VPP DUT +aggregate packet forwarding rate. + +For "Parallel" service topology packets are always switched twice by VPP +DUT per service chain. + +Note that reported VPP DUT performance results are specific to the SUTs +tested. SUTs with other processor than the ones used in FD.io lab are +likely to yield different results. Similarly to NIC-to-NIC switching +topology, here one can also expect the forwarding performance to be +proportional to processor core frequency for the same processor +architecture, assuming processor is the only limiting factor. However +due to much higher dependency on intensive memory operations in VM +service chained topologies and sensitivity to Linux scheduler settings +and behaviour, this estimation may not always yield good enough +accuracy. + +## Container Service Switching + +Container service switching topology test cases require VPP DUT to +communicate with Containers (Ctrs) over memif virtual interfaces. + +Three types of VM service topologies are tested in |csit-release|: + +1. "Parallel" topology with packets flowing within SUT from NIC(s) via + VPP DUT to Container, back to VPP DUT, then out thru NIC(s). +2. "Chained" topology (a.k.a. "Snake") with packets flowing within SUT + from NIC(s) via VPP DUT to Container, back to VPP DUT, then to the + next Container, back to VPP DUT and so on and so forth until the + last Container in a chain, then back to VPP DUT and out thru NIC(s). +3. "Horizontal" topology with packets flowing within SUT from NIC(s) via + VPP DUT to Container, then via "horizontal" memif to the next + Container, and so on and so forth until the last Container, then + back to VPP DUT and out thru NIC(s). + +For each of the above topologies, VPP DUT is tested in a range of L2 +or IPv4/IPv6 configurations depending on the test suite. Sample VPP DUT +"Chained" Container service topologies for 2-Node and 3-Node testbeds +with each SUT running N of Container instances is shown in the figures +below. + +{{< figure src="/cdocs/logical-2n-container-memif.svg" >}} + +{{< figure src="/cdocs/logical-3n-container-memif.svg" >}} + +In "Chained" Container topologies, packets are switched by VPP DUT +multiple times: twice for a single Container, three times for two +Containers, N+1 times for N Containers. Hence the external throughput +rates measured by TG and listed in this report must be multiplied by N+1 +to represent the actual VPP DUT aggregate packet forwarding rate. + +For a "Parallel" and "Horizontal" service topologies packets are always +switched by VPP DUT twice per service chain. + +Note that reported VPP DUT performance results are specific to the SUTs +tested. SUTs with other processor than the ones used in FD.io lab are +likely to yield different results. Similarly to NIC-to-NIC switching +topology, here one can also expect the forwarding performance to be +proportional to processor core frequency for the same processor +architecture, assuming processor is the only limiting factor. However +due to much higher dependency on intensive memory operations in +Container service chained topologies and sensitivity to Linux scheduler +settings and behaviour, this estimation may not always yield good enough +accuracy. diff --git a/docs/content/infrastructure/fdio_csit_testbed_specifications.md b/docs/content/infrastructure/fdio_csit_testbed_specifications.md index 20bec11f9f..49b6d72fa9 100644 --- a/docs/content/infrastructure/fdio_csit_testbed_specifications.md +++ b/docs/content/infrastructure/fdio_csit_testbed_specifications.md @@ -801,6 +801,8 @@ connectivity and wiring across defined CSIT testbeds: #### 2-Node-Cascadelake Servers (2n-clx) PROD +{{< figure src="/cdocs/testbed-2n-clx.svg" >}} + ``` - SUT [Server-Type-C2]: - testbedname: testbed27. @@ -906,6 +908,8 @@ connectivity and wiring across defined CSIT testbeds: #### 2-Node-Zen2 Servers (2n-zn2) PROD +{{< figure src="/cdocs/testbed-2n-zn2.svg" >}} + ``` - SUT [Server-Type-D1]: - testbedname: testbed210. @@ -939,7 +943,7 @@ connectivity and wiring across defined CSIT testbeds: #### 2-Node-ThunderX2 Servers (2x-tx2) PROD -Note: Server19 (TG) is shared between testbed33 & testbed211 +{{< figure src="/cdocs/testbed-2n-tx2.svg" >}} ``` - SUT [Server-Type-E22]: @@ -972,6 +976,8 @@ Note: Server19 (TG) is shared between testbed33 & testbed211 #### 2-Node-Icelake Servers (2n-icx) PROD +{{< figure src="/cdocs/testbed-2n-icx.svg" >}} + ``` - SUT [Server-Type-F1]: - testbedname: testbed212. @@ -1131,7 +1137,7 @@ Note: There is no IPMI. Serial console is accessible via VIRL2 and VIRL3 USB. #### 3-Node-Taishan Servers (3n-tsh) PROD -Note: Server19 (TG) is shared between testbed33 & testbed211 +{{< figure src="/cdocs/testbed-3n-tsh.svg" >}} ``` - SUT [Server-Type-E21]: @@ -1176,6 +1182,8 @@ Note: Server19 (TG) is shared between testbed33 & testbed211 #### 3-Node-Altra Servers (3n-alt) PROD +{{< figure src="/cdocs/testbed-3n-alt.svg" >}} + ``` - SUT [Server-Type-E23]: - testbedname: testbed34. @@ -1213,6 +1221,8 @@ Note: Server19 (TG) is shared between testbed33 & testbed211 #### 3-Node-Icelake Servers (3n-icx) PROD +{{< figure src="/cdocs/testbed-3n-icx.svg" >}} + ``` - ServerF1 [Server-Type-F1]: - testbedname: testbed37. @@ -1302,6 +1312,8 @@ Note: Server19 (TG) is shared between testbed33 & testbed211 #### 3-Node-SnowRidge Servers (3n-snr) PROD +{{< figure src="/cdocs/testbed-3n-snr.svg" >}} + ``` - ServerG1 [Server-Type-G1]: - testbedname: testbed39. @@ -1337,6 +1349,8 @@ Note: Server19 (TG) is shared between testbed33 & testbed211 #### 2-Node-SapphireRapids Servers (2n-spr) PROD +{{< figure src="/cdocs/testbed-2n-spr.svg" >}} + ``` - SUT [Server-Type-H1]: - testbedname: testbed21. @@ -1827,4 +1841,4 @@ To be completed. - s59-t24-tg1-c7/p1 to s59-t24-tg1-c7/p2. - ring5 100GE-ports e810-2CQDA2-2p100GE: - s59-t24-tg1-c9/p1 to s59-t24-tg1-c9/p2. -```
\ No newline at end of file +``` diff --git a/docs/content/infrastructure/fdio_csit_testbed_versioning.md b/docs/content/infrastructure/fdio_csit_testbed_versioning.md new file mode 100644 index 0000000000..5185c787f7 --- /dev/null +++ b/docs/content/infrastructure/fdio_csit_testbed_versioning.md @@ -0,0 +1,121 @@ +--- +bookToc: true +title: "FD.io CSIT Testbed Versioning" +weight: 3 +--- + +# FD.io CSIT Testbed Versioning + +CSIT test environment versioning has been introduced to track modifications of +the test environment. + +Any benchmark anomalies (progressions, regressions) between releases of a DUT +application (e.g. VPP, DPDK), are determined by testing it in the same test +environment, to avoid test environment changes clouding the picture. +To beter distinguish impact of test environment changes, we also execute tests +without any SUT (just with TRex TG sending packets over a link looping back to +TG). + +A mirror approach is introduced to determine benchmarking anomalies due to the +test environment change. This is achieved by testing the same DUT application +version between releases of CSIT test system. This works under the assumption +that the behaviour of the DUT is deterministic under the test conditions. + +CSIT test environment versioning scheme ensures integrity of all the test system +components, including their HW revisions, compiled SW code versions and SW +source code, within a specific CSIT version. Components included in the CSIT +environment versioning include: + +- **HW** Server hardware firmware and BIOS (motherboard, processsor, + NIC(s), accelerator card(s)), tracked in CSIT branch. +- **Linux** Server Linux OS version and configuration, tracked in CSIT + Reports. +- **TRex** TRex Traffic Generator version, drivers and configuration + tracked in TG Settings. +- **CSIT** CSIT framework code tracked in CSIT release branches. + +Following is the list of CSIT versions to date: + +- Ver. 1 associated with CSIT rls1908 branch ( + [HW](https://git.fd.io/csit/tree/docs/lab?h=rls1908), + [Linux](https://docs.fd.io/csit/rls1908/report/vpp_performance_tests/test_environment.html#sut-settings-linux), + [TRex](https://docs.fd.io/csit/rls1908/report/vpp_performance_tests/test_environment.html#tg-settings-trex), + [CSIT](https://git.fd.io/csit/tree/?h=rls1908) + ). +- Ver. 2 associated with CSIT rls2001 branch ( + [HW](https://git.fd.io/csit/tree/docs/lab?h=rls2001), + [Linux](https://docs.fd.io/csit/rls2001/report/vpp_performance_tests/test_environment.html#sut-settings-linux), + [TRex](https://docs.fd.io/csit/rls2001/report/vpp_performance_tests/test_environment.html#tg-settings-trex), + [CSIT](https://git.fd.io/csit/tree/?h=rls2001) + ). +- Ver. 4 associated with CSIT rls2005 branch ( + [HW](https://git.fd.io/csit/tree/docs/lab?h=rls2005), + [Linux](https://docs.fd.io/csit/rls2005/report/vpp_performance_tests/test_environment.html#sut-settings-linux), + [TRex](https://docs.fd.io/csit/rls2005/report/vpp_performance_tests/test_environment.html#tg-settings-trex), + [CSIT](https://git.fd.io/csit/tree/?h=rls2005) + ). +- Ver. 5 associated with CSIT rls2009 branch ( + [HW](https://git.fd.io/csit/tree/docs/lab?h=rls2009), + [Linux](https://docs.fd.io/csit/rls2009/report/vpp_performance_tests/test_environment.html#sut-settings-linux), + [TRex](https://docs.fd.io/csit/rls2009/report/vpp_performance_tests/test_environment.html#tg-settings-trex), + [CSIT](https://git.fd.io/csit/tree/?h=rls2009) + ). + - The main change is TRex data-plane core resource adjustments: + [increase from 7 to 8 cores and pinning cores to interfaces](https://gerrit.fd.io/r/c/csit/+/28184) + for better TRex performance with symmetric traffic profiles. +- Ver. 6 associated with CSIT rls2101 branch ( + [HW](https://git.fd.io/csit/tree/docs/lab?h=rls2101), + [Linux](https://docs.fd.io/csit/rls2101/report/vpp_performance_tests/test_environment.html#sut-settings-linux), + [TRex](https://docs.fd.io/csit/rls2101/report/vpp_performance_tests/test_environment.html#tg-settings-trex), + [CSIT](https://git.fd.io/csit/tree/?h=rls2101) + ). + - The main change is TRex version upgrade: increase from 2.82 to 2.86. +- Ver. 7 associated with CSIT rls2106 branch ( + [HW](https://git.fd.io/csit/tree/docs/lab?h=rls2106), + [Linux](https://s3-docs.fd.io/csit/rls2106/report/vpp_performance_tests/test_environment.html#sut-settings-linux), + [TRex](https://s3-docs.fd.io/csit/rls2106/report/vpp_performance_tests/test_environment.html#tg-settings-trex), + [CSIT](https://git.fd.io/csit/tree/?h=rls2106) + ). + - TRex version upgrade: increase from 2.86 to 2.88. + - Ubuntu upgrade from 18.04 LTS to 20.04.2 LTS. +- Ver. 8 associated with CSIT rls2110 branch ( + [HW](https://git.fd.io/csit/tree/docs/lab?h=rls2110), + [Linux](https://s3-docs.fd.io/csit/rls2110/report/vpp_performance_tests/test_environment.html#sut-settings-linux), + [TRex](https://s3-docs.fd.io/csit/rls2110/report/vpp_performance_tests/test_environment.html#tg-settings-trex), + [CSIT](https://git.fd.io/csit/tree/?h=rls2110) + ). + - Intel NIC 700/800 series firmware upgrade based on DPDK compatibility + matrix. +- Ver. 9 associated with CSIT rls2202 branch ( + [HW](https://git.fd.io/csit/tree/docs/lab?h=rls2202), + [Linux](https://s3-docs.fd.io/csit/rls2202/report/vpp_performance_tests/test_environment.html#sut-settings-linux), + [TRex](https://s3-docs.fd.io/csit/rls2202/report/vpp_performance_tests/test_environment.html#tg-settings-trex), + [CSIT](https://git.fd.io/csit/tree/?h=rls2202) + ). + - Intel NIC 700/800 series firmware upgrade based on DPDK compatibility + matrix. +- Ver. 10 associated with CSIT rls2206 branch ( + [HW](https://git.fd.io/csit/tree/docs/lab?h=rls2206), + [Linux](https://s3-docs.fd.io/csit/rls2206/report/vpp_performance_tests/test_environment.html#sut-settings-linux), + [TRex](https://s3-docs.fd.io/csit/rls2206/report/vpp_performance_tests/test_environment.html#tg-settings-trex), + [CSIT](https://git.fd.io/csit/tree/?h=rls2206) + ). + - Intel NIC 700/800 series firmware upgrade based on DPDK compatibility + matrix. + - Mellanox 556A series firmware upgrade based on DPDK compatibility + matrix. + - Intel IceLake all core turbo frequency turned off. Current base frequency + is 2.6GHz. + - TRex version upgrade: increase from 2.88 to 2.97. +- Ver. 11 associated with CSIT rls2210 branch ( + [HW](https://git.fd.io/csit/tree/docs/lab?h=rls2210), + [Linux](https://s3-docs.fd.io/csit/rls2210/report/vpp_performance_tests/test_environment.html#sut-settings-linux), + [TRex](https://s3-docs.fd.io/csit/rls2210/report/vpp_performance_tests/test_environment.html#tg-settings-trex), + [CSIT](https://git.fd.io/csit/tree/?h=rls2210) + ). + - Intel NIC 700/800 series firmware upgrade based on DPDK compatibility + matrix. + - Mellanox 556A series firmware upgrade based on DPDK compatibility + matrix. + - Ubuntu upgrade from 20.04.2 LTS to 22.04.1 LTS. + - TRex version upgrade: increase from 2.97 to 3.00.
\ No newline at end of file diff --git a/docs/content/infrastructure/fdio_dc_vexxhost_inventory.md b/docs/content/infrastructure/fdio_dc_vexxhost_inventory.md index b1494f6f72..3bdca72dae 100644 --- a/docs/content/infrastructure/fdio_dc_vexxhost_inventory.md +++ b/docs/content/infrastructure/fdio_dc_vexxhost_inventory.md @@ -5,10 +5,9 @@ weight: 1 # FD.io DC Vexxhost Inventory -- for each DC location, per rack .csv table with server inventory - captured inventory data: name,oper-status,testbed-id,role,model,s/n,rackid,rackunit,mgmt-ip4,ipmi-ip4,new-rackid,new-rackunit,new-mgmt-ip4,new-ipmi-ip4 - name: CSIT functional server name as tracked in [CSIT testbed specification](https://git.fd.io/csit/tree/docs/lab/testbed_specifications.md), followed by "/" and the actual configured hostname, unless it is the same as CSIT name. - - oper-status: operational status (up|down|ipmi). + - oper-status: operational status (up|down). - testbed-id: CSIT testbed identifier. - role: 2n/3n-xxx performance testbed, nomad-client, nomad-server. - role exceptions: decommission, repurpose, spare. diff --git a/docs/content/infrastructure/testbed_configuration/_index.md b/docs/content/infrastructure/testbed_configuration/_index.md index ce023237c7..d0716003c5 100644 --- a/docs/content/infrastructure/testbed_configuration/_index.md +++ b/docs/content/infrastructure/testbed_configuration/_index.md @@ -1,4 +1,5 @@ --- +bookCollapseSection: true bookFlatSection: false title: "FD.io CSIT Testbed Configuration" weight: 3 diff --git a/docs/content/infrastructure/testbed_configuration/ami_alt_hw_bios_cfg.md b/docs/content/infrastructure/testbed_configuration/ami_alt_hw_bios_cfg.md new file mode 100644 index 0000000000..e02c642e6e --- /dev/null +++ b/docs/content/infrastructure/testbed_configuration/ami_alt_hw_bios_cfg.md @@ -0,0 +1,12 @@ +--- +bookToc: true +title: "MegaRac Altra" +--- + +# MegaRac Altra + +## Linux cmdline + +``` +BOOT_IMAGE=/boot/vmlinuz-5.15.0-46-generic root=UUID=7d1d0e77-4df0-43df-9619-a99db29ffb83 ro audit=0 default_hugepagesz=2M hugepagesz=1G hugepages=32 hugepagesz=2M hugepages=32768 iommu.passthrough=1 isolcpus=1-10,29-38 nmi_watchdog=0 nohz_full=1-10,29-38 nosoftlockup processor.max_cstate=1 rcu_nocbs=1-10,29-38 console=ttyAMA0,115200n8 quiet +``` diff --git a/docs/content/infrastructure/testbed_configuration/gigabyte_tx2_hw_bios_cfg.md b/docs/content/infrastructure/testbed_configuration/gigabyte_tx2_hw_bios_cfg.md new file mode 100644 index 0000000000..eb188d3bf9 --- /dev/null +++ b/docs/content/infrastructure/testbed_configuration/gigabyte_tx2_hw_bios_cfg.md @@ -0,0 +1,12 @@ +--- +bookToc: true +title: "GigaByte ThunderX2" +--- + +# GigaByte ThunderX2 + +## Linux cmdline + +``` +BOOT_IMAGE=/boot/vmlinuz-5.4.0-65-generic root=UUID=7d1d0e77-4df0-43df-9619-a99db29ffb83 ro audit=0 intel_iommu=on isolcpus=1-27,29-55 nmi_watchdog=0 nohz_full=1-27,29-55 nosoftlockup processor.max_cstate=1 rcu_nocbs=1-27,29-55 console=ttyAMA0,115200n8 quiet +``` diff --git a/docs/content/infrastructure/testbed_configuration/huawei_tsh_hw_bios_cfg.md b/docs/content/infrastructure/testbed_configuration/huawei_tsh_hw_bios_cfg.md new file mode 100644 index 0000000000..d9fd71b080 --- /dev/null +++ b/docs/content/infrastructure/testbed_configuration/huawei_tsh_hw_bios_cfg.md @@ -0,0 +1,12 @@ +--- +bookToc: true +title: "Huawei Taishan" +--- + +# Huawei Taishan + +## Linux cmdline + +``` +BOOT_IMAGE=/boot/vmlinuz-5.4.0-65-generic root=UUID=7d1d0e77-4df0-43df-9619-a99db29ffb83 ro audit=0 intel_iommu=on isolcpus=1-27,29-55 nmi_watchdog=0 nohz_full=1-27,29-55 nosoftlockup processor.max_cstate=1 rcu_nocbs=1-27,29-55 console=ttyAMA0,115200n8 quiet +``` diff --git a/docs/content/infrastructure/testbed_configuration/sm_clx_hw_bios_cfg.md b/docs/content/infrastructure/testbed_configuration/sm_clx_hw_bios_cfg.md index b2c859b11f..f4d9fbe475 100644 --- a/docs/content/infrastructure/testbed_configuration/sm_clx_hw_bios_cfg.md +++ b/docs/content/infrastructure/testbed_configuration/sm_clx_hw_bios_cfg.md @@ -1405,6 +1405,12 @@ pku ospke avx512_vnni md_clear flush_l1d arch_capabilities | High Precision Event Timer [Enabled] | | ``` +## Linux cmdline + +``` +$ cat /proc/cmdline +BOOT_IMAGE=/boot/vmlinuz-5.15.0-46-generic root=UUID=2d6f4d44-76b1-4343-bc73-c066a3e95b32 ro audit=0 default_hugepagesz=2M hugepagesz=1G hugepages=32 hugepagesz=2M hugepages=32768 hpet=disable intel_idle.max_cstate=1 intel_iommu=on intel_pstate=disable iommu=pt isolcpus=1-23,25-47,49-71,73-95 mce=off nmi_watchdog=0 nohz_full=1-23,25-47,49-71,73-95 nosoftlockup numa_balancing=disable processor.max_cstate=1 rcu_nocbs=1-23,25-47,49-71,73-95 tsc=reliable console=ttyS0,115200n8 quiet +``` ## Xeon Clx Server Firmware Inventory diff --git a/docs/content/infrastructure/testbed_configuration/sm_icx_hw_bios_cfg.md b/docs/content/infrastructure/testbed_configuration/sm_icx_hw_bios_cfg.md index 854d3d0418..97c7874d85 100644 --- a/docs/content/infrastructure/testbed_configuration/sm_icx_hw_bios_cfg.md +++ b/docs/content/infrastructure/testbed_configuration/sm_icx_hw_bios_cfg.md @@ -1092,6 +1092,13 @@ Memory Device |> Network Stack Configuration | | ``` +## Linux cmdline + +``` +$ cat /proc/cmdline +BOOT_IMAGE=/boot/vmlinuz-5.15.0-46-generic root=UUID=6ff26c8a-8c65-4025-a6e7-d97dee6025d0 ro audit=0 default_hugepagesz=2M hugepagesz=1G hugepages=32 hugepagesz=2M hugepages=32768 hpet=disable intel_idle.max_cstate=1 intel_iommu=on intel_pstate=disable iommu=pt isolcpus=1-31,33-63,65-95,97-127 mce=off nmi_watchdog=0 nohz_full=1-31,33-63,65-95,97-127 nosoftlockup numa_balancing=disable processor.max_cstate=1 rcu_nocbs=1-31,33-63,65-95,97-127 tsc=reliable console=ttyS0,115200n8 quiet +``` + ## Xeon ICX Server Firmware Inventory ``` diff --git a/docs/content/infrastructure/testbed_configuration/sm_spr_hw_bios_cfg.md b/docs/content/infrastructure/testbed_configuration/sm_spr_hw_bios_cfg.md index c2bf8fb795..a91fcfffb1 100644 --- a/docs/content/infrastructure/testbed_configuration/sm_spr_hw_bios_cfg.md +++ b/docs/content/infrastructure/testbed_configuration/sm_spr_hw_bios_cfg.md @@ -830,6 +830,12 @@ Memory Device Logical Size: None ``` +## Linux cmdline + +``` +BOOT_IMAGE=/boot/vmlinuz-5.15.0-46-generic root=UUID=b99a7749-d0ee-4afe-88a0-0be6c5873645 ro audit=0 default_hugepagesz=2M hugepagesz=1G hugepages=32 hugepagesz=2M hugepages=32768 hpet=disable intel_idle.max_cstate=1 intel_iommu=on intel_pstate=disable iommu=pt isolcpus=1-31,33-63,65-95,97-127 mce=off nmi_watchdog=0 nohz_full=1-31,33-63,65-95,97-127 nosoftlockup numa_balancing=disable processor.max_cstate=1 rcu_nocbs=1-31,33-63,65-95,97-127 tsc=reliable +``` + ## Xeon ICX Server Firmware Inventory ``` diff --git a/docs/content/infrastructure/testbed_configuration/sm_zn2_hw_bios_cfg.md b/docs/content/infrastructure/testbed_configuration/sm_zn2_hw_bios_cfg.md index 31335d5cc7..537fc9f42a 100644 --- a/docs/content/infrastructure/testbed_configuration/sm_zn2_hw_bios_cfg.md +++ b/docs/content/infrastructure/testbed_configuration/sm_zn2_hw_bios_cfg.md @@ -604,6 +604,12 @@ Flags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cm | | | ``` +## Linux cmdline + +``` +$ cat /proc/cmdline +BOOT_IMAGE=/boot/vmlinuz-5.15.0-46-generic root=UUID=cac1254f-9426-4ea6-a8db-2554f075db99 ro amd_iommu=on audit=0 default_hugepagesz=2M hugepagesz=1G hugepages=32 hugepagesz=2M hugepages=32768 hpet=disable iommu=pt isolcpus=1-15,17-31,33-47,49-63 nmi_watchdog=0 nohz_full=off nosoftlockup numa_balancing=disable processor.max_cstate=0 rcu_nocbs=1-15,17-31,33-47,49-63 tsc=reliable console=ttyS0,115200n8 quiet +``` ## EPYC zn2 Server Firmware Inventory |