Data Center Bridging Consortium 802.1Qaz Enhanced Transmission Selection Test Suite Version 1.2 Technical Document Last Updated: April 10th, 2012 Data Center Bridging Consortium HTTP://WWW.IOL.UNH.EDU/CONSORTIUMS/DCB InterOperability Laboratory 121 Technology Drive, Suite 2 University of New Hampshire Durham, NH 03824 Phone: +1-603-862-0701
Table of Contents Table of Contents...2 Modification Record...3 Acknowledgments...4 Introduction...5 References...7 Test Setup...8 Group 1: Traffic Class Mapping...9 TEST #37.1.1: BANDWIDTH UTILIZATION OF TWO TCS...10 TEST #37.1.2: BANDWIDTH UTILIZATION OF THREE TCS...11 TEST #37.1.3: BANDWIDTH UTILIZATION OF FOUR TCS...12 TEST #37.1.4: BANDWIDTH UTILIZATION OF EIGHT TCS...13 TEST #37.1.5: BANDWIDTH SHARING...14 TEST #37.1.6: LOSSY VS. LOSSLESS TRAFFIC PRIORITY...15 TEST #37.1.7: SERVICE STRICT PRIORITY FIRST...16
Modification Record 1. April 6 th, 2010 Initial Version (rzarick) 2. March 11 th, 2011 Updated to draft 2.4 of the standard (mhagen) 3. April 10 th, 2012 Updated to the final version of the standard (mhagen)
Acknowledgments The University of New Hampshire would like to acknowledge the efforts of the following individuals in the development of this test suite. Ryan Zarick Mike Hagen University of New Hampshire University of New Hampshire
Introduction Overview The University of New Hampshire s InterOperability Laboratory (IOL) is an institution designed to improve the interoperability of standards based products by providing an environment where a product can be tested against other implementations of a standard. These tests are designed to determine if a DCB product conforms to specifications defined in IEEE Std P802.1Qaz-2011 Media Access Control (MAC) Bridges and Virtual Bridged Local Area Networks Amendment 18: Enhanced Transmission Selection for Bandwidth Sharing Between Traffic Classes (hereafter referred to as ETS ). This suite of tests has been developed to help implementers evaluate the functioning of their Data Center Bridging based products. The tests do not determine if a product conforms to the IEEE standard, nor are they purely interoperability tests. Rather, they provide one method to isolate problems within a Data Center Bridging device. Successful completion of all tests contained in this suite does not guarantee that the tested device will operate with other Data Center Bridging devices. However, combined with satisfactory operation in the IOL s semi-production environment, these tests provide a reasonable level of confidence that the Device Under Test (DUT) will function well in most multi-vendor Data Center Bridging environments. Organization of Tests The tests contained in this document are organized to simplify the identification of information related to a test and to facilitate in the actual testing process. Each test contains an identification section that describes the test and provides cross-reference information. The detailed section discusses the background information and specifies how the test is to be performed. Tests are grouped in order to reduce setup time in the lab environment. Each test contains the following information: Test Number The Test Number associated with each test follows a simple grouping structure. Listed first is the Clause followed by the Test Group Number followed by the test's number within the group. This allows for the addition of future tests to the appropriate groups of the test suite without requiring the renumbering of the subsequent tests. Purpose The purpose is a short statement describing what the test attempts to achieve. The test is written at the functional level. References The references section lists cross-references to the IEEE standards and other documentation that might be helpful in understanding and evaluating the test and results. Resource Requirements The requirements section specifies the software, hardware, and test equipment that will be needed to perform the test. The items contained in this section are special test devices, software that must reside on the DUT, or other facilities, which may not be available on all devices. Last Modification This specifies the date of the last modification to this test. Discussion The discussion covers the assumptions made in the design or implementation of the test as well as known limitations. Other items specific to the test are covered here.
Test Setup The setup section describes in detail the configuration of the test environment and includes a block diagram for clarification as well as information such as the interconnection of devices, what monitoring equipment should capture, what the generation equipment should send, and any other configuration information vital to carrying out the test. Small changes in the configuration should be included in the test procedure. Procedure The procedure section of the test description contains the step-by-step instructions for carrying out the test. It provides a cookbook approach to testing, and will often be interspersed with observable results. Observable Results The observable results section lists observables that can be examined by the tester to verify that the DUT is operating properly. When multiple values are possible for an observable, this section provides a short discussion on how to interpret them. Note that complete delineation between the observables in the Procedure and Observable Results is virtually impossible. As such a careful note should be made of the requirements in both sections. In certain cases, it may be necessary to modify certain steps in the Procedure section while doing the actual tests so as to be able to perform the tests. In such cases, the modifications will be noted in the summary report. Possible Problems This section provides some clues to look for if the test does not yield the expected results. Legend For reasons of brevity, the following abbreviation has been used in the Test Suite: DCBX Data Center Bridging Capability Exchange DUT Device Under Test ETS Enhanced Transmission Selection PFC Priority-based Flow Control TC Traffic Class TS Testing Station
References The following documents are referenced in this text: IEEE Std P802.1Qaz-2011 Media Access Control (MAC) Bridges and Virtual Bridged Local Area Networks Amendment 18: Enhanced Transmission Selection for Bandwidth Sharing Between Traffic Classes
Test Setup The following test setups are used in this test suite: Test Setup 1: The DUT is a Source or a Sink Testing Station Monitor DUT Test Setup 2: The DUT is a Switch Testing Station Sink DUT Switch Testing Station Source Initiator1........... Testing Station Source N
Group 1: Traffic Class Mapping Overview: These tests observe the behavior of the DUT during transmission and reception of frames using ETS. They ensure that the DUT properly transmits frames according to the ETS configurations set before transmission.
Test #37.1.1: Bandwidth Utilization of Two TCs Purpose: To verify that the DUT properly disperses traffic to within 10% of it's ETS configured settings. References: [1] ETS Sub-clause 37.3 Resource Requirements: A resource for traffic generation. Monitor capable of capturing Ethernet traffic. Last Modification: March 8, 2011. Discussion: The DUT should have the ability to set the bandwidth allocation for at least three TCs. When all configured TCs consume their share of the available bandwidth, each TC should be within +/-10% of the allocated bandwidth over a period of time as measured by 10,000,000 bit times during transmission of max sized frames without any PFC frames received. Test Setup: Case 1: Test Setup 1. Connect the DUT to a TS. Connect a monitor between the DUT and the TS. Case 2: Test Setup 2. Connect two TS Sources and one TS Sink to the DUT. Connect a monitor between the DUT and the TS Sink. Procedure: Test Run TC 1 TC 2 1 50.0% 50.0% 2 15.0% 85.0% 3 85.0% 15.0% Case 1: The DUT is an End Device 1. Bring the links between the TS and DUT up. 2. Configure the DUT and the TS to have the ETS settings for Test Run 1. 3. Transmit at least 500 max sized frames at 100% utilization from the TS. 4. Capture the results. 5. Repeat for Test Run 2 and 3. Case 2: The DUT is a Switch 1. Bring the links between the TSs and DUT up. 2. Configure the DUT to have the ETS settings for Test Run 1. 3. Configure each Source to transmit a unique TC from the table above. 4. Transmit at least 500 max sized frames at 100% utilization from Source 1 and 2. 5. Capture the results. 6. Repeat for Test Run 2 and 3. Observable Results: On the Monitor, verify that the DUT properly disperses the frames to within 10% of each of the ETS settings. Possible Problems: none.
Test #37.1.2: Bandwidth Utilization of Three TCs Purpose: To verify that the DUT properly disperses traffic to within 10% of it's ETS configured settings. References: [1] ETS Sub-clause 37.3 Resource Requirements: A resource for traffic generation. Monitor capable of capturing Ethernet traffic. Last Modification: March 8, 2011. Discussion: The DUT should have the ability to set the bandwidth allocation for at least three TCs. When all configured TCs consume their share of the available bandwidth, each TC should be within +/-10% of the allocated bandwidth over a period of time as measured by 10,000,000 bit times during transmission of max sized frames without any PFC frames received. Test Setup: Case 1: Test Setup 1. Connect the DUT to a TS. Connect a monitor between the DUT and the TS. Case 2: Test Setup 2. Connect three TS Sources and one TS Sink to the DUT. Connect a monitor between the DUT and the TS Sink. Procedure: Test Run TC 1 TC 2 TC 3 1 33.3% 33.3% 33.4% 2 70.0% 15.0% 15.0% 3 15.0% 70.0% 15.0% 4 15.0% 15.0% 70.0% Case 1: The DUT is an End Device 1. Bring the links between the TS and DUT up. 2. Configure the DUT and the TS to have the ETS settings for Test Run 1. 3. Transmit at least 500 max sized frames at 100% utilization from the TS. 4. Capture the results. 5. Repeat for Test Run 2, 3, and 4. Case 2: The DUT is a Switch 1. Bring the links between the TSs and DUT up. 2. Configure the DUT to have the ETS settings for Test Run 1. 3. Configure each Source to transmit a unique TC from the table above. 4. Transmit at least 500 max sized frames at 100% utilization from Source 1, 2, and 3. 5. Capture the results. 6. Repeat for Test Run 2, 3, and 4. Observable Results: On the Monitor, verify that the DUT properly disperses the frames to within 10% of each of the TC settings. Possible Problems: none.
Test #37.1.3: Bandwidth Utilization of Four TCs Purpose: To verify that the DUT properly disperses traffic to within 10% of it's ETS configured settings. References: [1] ETS Sub-clause 37.3 Resource Requirements: A resource for traffic generation. Monitor capable of capturing Ethernet traffic. Last Modification: March 8, 2011. Discussion: The DUT should have the ability to set the bandwidth allocation for at least three TCs. When all configured TCs consume their share of the available bandwidth, each TC should be within +/-10% of the allocated bandwidth over a period of time as measured by 10,000,000 bit times during transmission of max sized frames without any PFC frames received. Test Setup: Case 1: Test Setup 1. Connect the DUT to a TS. Connect a monitor between the DUT and the TS. Case 2: Test Setup 2. Connect four TS Sources and one TS Sink to the DUT. Connect a monitor between the DUT and the TS Sink. Procedure: Case 1: The DUT is an End Device 1. Bring the links between the TS and DUT up. 2. Configure the DUT and the TS to have the ETS settings for Test Run 1. 3. Transmit at least 500 max sized frames at 100% utilization from the TS. 4. Capture the results. 5. Repeat for Test Run 2-5 Case 2: The DUT is a Switch 1. Bring the links between the TSs and DUT up. 2. Configure the DUT to have the ETS settings for Test Run 1. 3. Configure each Source to transmit a unique TC from the table above. 4. Transmit at least 500 max sized frames at 100% utilization from Source 1-4. 5. Capture the results. 6. Repeat for Test Run 2-5. Observable Results: Test Run TC 1 TC 2 TC 3 TC 4 1 25.0% 25.0% 25.0% 25.0% 2 55.0% 15.0% 15.0% 15.0% 3 15.0% 55.0% 15.0% 15.0% 4 15.0% 15.0% 55.0% 15.0% 5 15.0% 15.0% 15.0% 55.0% On the Monitor, verify that the DUT properly disperses the frames to within 10% of each of the TC settings. Possible Problems: If the DUT does not support more than 3 TCs, this test cannot be tested.
Test #37.1.4: Bandwidth Utilization of Eight TCs Purpose: To verify that the DUT properly disperses traffic to within 10% of it's ETS configured settings. References: [1] ETS Sub-clause 37.3 Resource Requirements: A resource for traffic generation. Monitor capable of capturing Ethernet traffic. Last Modification: March 8, 2011. Discussion: The DUT should have the ability to set the bandwidth allocation for at least three TCs. When all configured TCs consume their share of the available bandwidth, each TC should be within +/-10% of the allocated bandwidth over a period of time as measured by 10,000,000 bit times during transmission of max sized frames without any PFC frames received. Test Setup: Case 1: Test Setup 1. Connect the DUT to a TS. Connect a monitor between the DUT and the TS. Case 2: Test Setup 2. Connect four TS Sources and one TS Sink to the DUT. Connect a monitor between the DUT and the TS Sink. Procedure: Test Run TC 1 TC 2 TC 3 TC 4 TC 5 TC 6 TC 7 TC8 1 12.5% 12.5% 12.5% 12.5% 12.5% 12.5% 12.5% 12.5% Case 1: The DUT is an End Device 1. Bring the link between the TS and DUT up. 2. Configure the DUT and the TS to have the ETS settings for Test Run 1. 3. Transmit at least 500 max sized frames at 100% utilization from the TS. 4. Capture the results. Case 2: The DUT is a Switch 1. Bring the links between the TSs and DUT up. 2. Configure the DUT to have the ETS settings for Test Run 1. 3. Configure each Source to transmit a unique TC from the table above. 4. Transmit frames at 100% utilization from Source 1-8. 5. Capture the results. Observable Results: On the Monitor, verify that the DUT properly disperses the frames to within 10% of each of the TC settings. Possible Problems: If the DUT does not support more than 3 TCs, this test cannot be tested.
Test #37.1.5: Bandwidth Sharing Purpose: To verify that the DUT properly shares bandwidth with other TCs when one TC 's bandwidth is not fully allocated. References: [1] ETS Sub-clause 37.3 Resource Requirements: A resource for traffic generation. Monitor capable of capturing Ethernet traffic. Last Modification: March 8, 2011. Discussion: When a TC does not consume all of it's allocated bandwidth the DUT is to disperse the remaining bandwidth amongst the other TCs. Test Setup: Case 1: Test Setup 1. Connect the DUT to a TS. Connect a monitor between the DUT and the TS. Case 2: Test Setup 2. Connect two TS Sources and one TS Sink to the DUT. Connect a monitor between the DUT and the TS Sink. Procedure: Test Run TC 1 TC 2 1 50.0% 50.0% 2 15.0% 85.0% Case 1: The DUT is an End Device 1. Bring the links between the TS and DUT up. 2. Configure the DUT to have the ETS settings for Test Run 1. 3. Configure the TS to have the ETS settings for Test Run 2. 4. Transmit at least 500 max sized frames at 100% utilization from the TS. 5. Capture the results. Case 2: The DUT is a Switch 1. Bring the links between the TSs and DUT up. 2. Configure the DUT to have the ETS settings for Test Run 1. 3. Configure each Source to transmit a unique TC from the table above. 4. Transmit at least 500 max sized frames at the utilization described in Test Run 2 from Source 1 and 2. 5. Capture the results. Observable Results: On the Monitor, verify that the DUT properly utilizes the excess bandwidth. Possible Problems: none.
Test #37.1.6: Lossy vs. Lossless Traffic Priority Purpose: To verify that when the DUT is overloaded with lossy and lossless traffic, it drops the lossy traffic and properly uses PFC to regulate the rates of incoming lossless traffic. References: [1] ETS Sub-clause 37.3 Resource Requirements: A resource for traffic generation. Monitor capable of capturing Ethernet traffic. Last Modification: March 8, 2011. Discussion: When lossy and lossless traffic is transmitted to the DUT at faster rates than the DUT can handle, the DUT is required to drop the lossy traffic and use PFC to regulate the rates of the incoming lossless traffic. Regardless of the state of lossy or lossless, the bandwidth allocations should be maintained. Test Setup: Case 1: Test Setup 1. Connect the DUT to a TS. Connect a monitor between the DUT and the TS. Case 2: Test Setup 2. Connect two TS Sources and one TS Sink to the DUT. Connect a monitor between the DUT and the TS Sink. Procedure: Test Run TC 1 TC 2 1 50.0% 50.0% 2 15.0% 85.0% 3 85.0% 15.0% Case 1: The DUT is an End Device 1. Bring the link between the TS and DUT up. 2. Configure the DUT and the TS to have the ETS settings for Test Run 1. 3. Configure TC 1 to be a lossless class and TC 2 to be a lossy class. 4. Transmit at least 500 maxed sized frames at 100% utilization from the TS. 5. Capture the results. 6. Repeat for Test Run 2 and 3. Case 2: The DUT is a Switch 1. Bring the links between the TSs and DUT up. 2. Configure the DUT to have the ETS settings for Test Run 1. 3. Configure each Source to transmit a unique TC from the table above. 4. Configure TC 1 to be a lossless class and TC 2 to be a lossy class. 5. Transmit at least 500 maxed sized frames at 100% utilization from Source 1 and 2. 6. Capture the results. 7. Repeat for Test Run 2 and 3. Observable Results: On the Monitor, verify that the DUT properly disperses the frames to within 10% of each of the ETS settings. Possible Problems: According to the standard, the observable result can only be measured while PFC is not seen, so a DUT cannot fail this test it is only a recommended behavior.
Test #37.1.7: Service Strict Priority First Purpose: To verify that the DUT properly services traffic classes within the strict priority first, then allocates the remaining bandwidth according to the ETS values assigned. References: [1] ETS Sub-clause 37.3 Resource Requirements: A resource for traffic generation. Monitor capable of capturing Ethernet traffic. Last Modification: March 11, 2011. Discussion: A DUT must service all traffic classes not assigned to the ETS class and Vendor Specific class first. Once they have serviced the other traffic classes, the remaining bandwidth, referred to as available bandwidth, is allocated according to the values assigned. Test Setup: Case 1: Test Setup 1. Connect the DUT to a TS. Connect a monitor between the DUT and the TS. Case 2: Test Setup 2. Connect two TS Sources and one TS Sink to the DUT. Connect a monitor between the DUT and the TS Sink. Procedure: Test Run TC 1 TC 2 TC 3 1 50.0% 50.0% s trict 2 15.0% 85.0% s trict 3 85.0% 15.0% s trict Case 1: The DUT is an End Device 1. Bring the link between the TS and DUT up. 2. Configure the DUT and the TS to have the ETS settings for Test Run 1. 3. Configure TC 3 to utilize 50% of the available bandwidth on the network. 4. Transmit at least 500 maxed sized frames at 100% utilization from the TS for TC 1 and TC 2. 5. Capture the results. 6. Repeat for Test Run 2 and 3. Case 2: The DUT is a Switch 1. Bring the links between the TSs and DUT up. 2. Configure the DUT to have the ETS settings for Test Run 1. 3. Configure each Source to transmit a unique TC from the table above. 4. Configure TC 3 to utilize 50% of the available bandwidth on the network. 5. Transmit at least 500 maxed sized frames at 100% utilization from Source 1 and 2. 6. Capture the results. 7. Repeat for Test Run 2 and 3. Observable Results: On the Monitor, verify that the DUT properly disperses the frames of TC 1 and TC 2 to within 10% of each of the ETS settings after servicing the bandwidth requirements of TC 3. Possible Problems: None.