ThinkSystem Mellanox ConnectX-6 HDR100/100GbE VPI AdaptersProduct Guide

Author
Updated
8 Jun 2021
Form Number
LP1170
PDF size
15 pages, 242 KB

Abstract

The ThinkSystem Mellanox ConnectX-6 HDR100/100GbE VPI Adapters offer 100 Gb/s Ethernet and InfiniBand connectivity for high-performance connectivity when running HPC, cloud, storage and machine learning applications.

This product guide provides essential presales information to understand the adapter and its key features, specifications, and compatibility. This guide is intended for technical specialists, sales specialists, sales engineers, IT architects, and other IT professionals who want to learn more about the ConnectX-6 HDR100 adapters and consider their use in IT solutions.

Change History

Changes in the June 8, 2021 update:

  • New cables supported - Supported transceivers and cables section:
    • 3m Mellanox HDR IB Active Copper QSFP56 Cable, 4X97A12610
    • 4m Mellanox HDR IB Active Copper QSFP56 Cable, 4X97A12611

Introduction

The ThinkSystem Mellanox ConnectX-6 HDR100 Adapters offer 100 Gb/s Ethernet and InfiniBand connectivity for high-performance connectivity when running HPC, cloud, storage and machine learning applications.

The following figure shows the 2-port ConnectX-6 HDR100/100GbE VPI Adapter (the standard heat sink has been removed in this photo).

ThinkSystem Mellanox ConnectX-6 HDR100 QSFP56 2-port PCIe InfiniBand Adapter
Figure 1. ThinkSystem Mellanox ConnectX-6 HDR100/100GbE QSFP56 1-port PCIe VPI Adapter

Did you know?

Mellanox ConnectX-6 brings new acceleration engines for maximizing High Performance, Machine Learning, Storage, Web 2.0, Cloud, Data Analytics and Telecommunications platforms. ConnectX-6 HDR100 adapters support up to 100G total bandwidth at sub-600 nanosecond latency, and NVMe over Fabric offloads, providing the highest performance and most flexible solution for the most demanding applications and markets. ThinkSystem servers with Mellanox adapters and switches deliver the most intelligent fabrics for High Performance Computing clusters.

Part number information

The following table shows the part numbers for the adapters.

Table 1. Ordering information
Part
number
Feature
code
Mellanox
equivalent
Description
4C57A14177 B4R9 MCX653105A-ECAT ThinkSystem Mellanox ConnectX-6 HDR100/100GbE QSFP56 1-port PCIe VPI Adapter
4C57A14178 B4RA MCX653106A-ECAT ThinkSystem Mellanox ConnectX-6 HDR100/100GbE QSFP56 2-port PCIe VPI Adapter

The part numbers include the following:

  • One Mellanox adapter
  • Low-profile (2U) and full-height (3U) adapter brackets
  • Documentation

Note: These adapters were previously named as follows:

  • ThinkSystem Mellanox ConnectX-6 HDR100 QSFP56 1-port PCIe InfiniBand Adapter
  • ThinkSystem Mellanox ConnectX-6 HDR100 QSFP56 2-port PCIe InfiniBand Adapter

Supported transceivers and cables

The adapter has an empty QSFP56 cage for connectivity. The following table lists the supported transceivers.

Table 2. Transceivers
Part number Feature code Description
100Gb transceivers
7G17A03539 AV1D Lenovo 100GBase-SR4 QSFP28 Transceiver
4M27A67042 BFH1 Lenovo 100Gb SR4 QSFP28 Ethernet Transceiver
Converters
4G17A10853 B306 Mellanox QSA 100G to 25G Cable Adapter

The following table lists the supported fiber optic cables and Active Optical Cables.

Table 3. Optical cables
Part number Feature code Description
QSFP28 EDR InfiniBand Optical Cables
00MP563 ASRN 3m Mellanox EDR IB Optical QSFP28 Cable
00MP540 ASQZ 5m Mellanox EDR IB Optical QSFP28 Cable
00MP544 ASR0 10m Mellanox EDR IB Optical QSFP28 Cable
00MP548 ASR1 15m Mellanox EDR IB Optical QSFP28 Cable
00MP552 ASR2 20m Mellanox EDR IB Optical QSFP28 Cable
00MP556 ASR3 30m Mellanox EDR IB Optical QSFP28 Cable
00MP566 ASRP 50m Mellanox EDR IB Optical QSFP28 Cable
QSFP 40Gb Active Optical Cables
7Z57A04256 AX42 Lenovo 1m 40G QSFP+ Active Optical Cable
00YL652 ATZ3 Lenovo 3m 40G QSFP+ to QSFP+ Active Optical Cable
00YL655 ATZ4 Lenovo 5m 40G QSFP+ to QSFP+ Active Optical Cable
00YL658 ATZ5 Lenovo 7m 40G QSFP+ to QSFP+ Active Optical Cable
00YL661 ATZ6 Lenovo 15m 40G QSFP+ to QSFP+ Active Optical Cable
00YL664 ATZ7 Lenovo 20m 40G QSFP+ to QSFP+ Active Optical Cable
QSFP 40Gb Optical Cables (these cables require a transceiver)
00VX003 AT2U Lenovo 10m QSFP+ MPO-MPO OM3 MMF Cable
00VX005 AT2V Lenovo 30m QSFP+ MPO-MPO OM3 MMF Cable
QSFP28 100Gb Ethernet Active Optical Cables
4Z57A10844 B2UZ Lenovo 1m 100G QSFP28 Active Optical Cable
7Z57A03546 AV1L Lenovo 3m 100G QSFP28 Active Optical Cable
7Z57A03547 AV1M Lenovo 5m 100G QSFP28 Active Optical Cable
7Z57A03548 AV1N Lenovo 10m 100G QSFP28 Active Optical Cable
7Z57A03549 AV1P Lenovo 15m 100G QSFP28 Active Optical Cable
7Z57A03550 AV1Q Lenovo 20m 100G QSFP28 Active Optical Cable
100G MPO OM4 MMF Cables (these cables require a transceiver)
7Z57A03567 AV25 Lenovo 5m MPO-MPO OM4 MMF Cable
7Z57A03568 AV26 Lenovo 7m MPO-MPO OM4 MMF Cable
7Z57A03569 AV27 Lenovo 10m MPO-MPO OM4 MMF Cable
7Z57A03570 AV28 Lenovo 15m MPO-MPO OM4 MMF Cable
7Z57A03571 AV29 Lenovo 20m MPO-MPO OM4 MMF Cable
7Z57A03572 AV2A Lenovo 30m MPO-MPO OM4 MMF Cable
QSFP56 HDR IB Optical Cables
4Z57A14188 B4QW 3m Mellanox HDR IB Optical QSFP56 Cable
4Z57A14189 B4QX 5m Mellanox HDR IB Optical QSFP56 Cable
4Z57A14190 B4QY 10m Mellanox HDR IB Optical QSFP56 Cable
4Z57A14191 B4QZ 15m Mellanox HDR IB Optical QSFP56 Cable
4Z57A14192 B4R0 20m Mellanox HDR IB Optical QSFP56 Cable
4Z57A16016 B68P 30m Mellanox HDR IB Active Optical QSFP56 Cable
4Z57A16017 B68N 50m Mellanox HDR IB Active Optical QSFP56 Cable
4Z57A16018 B68M 100m Mellanox HDR IB Active Optical QSFP56 Cable
QSFP56 HDR IB to 2x HDR100 Optical Splitter Cables
4Z57A14196 B4R4 3m Mellanox HDR IB to 2x HDR100 Splitter Optical QSFP56 Cable
4Z57A14197 B4R5 5m Mellanox HDR IB to 2x HDR100 Splitter Optical QSFP56 Cable
4Z57A14198 B4R6 10m Mellanox HDR IB to 2x HDR100 Splitter Optical QSFP56 Cable
4Z57A14199 B4R7 15m Mellanox HDR IB to 2x HDR100 Splitter Optical QSFP56 Cable
4Z57A14214 B4R8 20m Mellanox HDR IB to 2x HDR100 Splitter Optical QSFP56 Cable
4Z57A11490 B68K 30m Mellanox HDR to 2x HDR100 IB Active Optical QSFP56 Splitter Cable

The following table lists the supported direct-attach copper (DAC) cables.

Table 4. Copper cables
Part number Feature code Description
QSFP28 EDR InfiniBand Passive Copper Cables
00MP516 ASQT 0.5m Mellanox EDR IB Passive Copper QSFP28 Cable
00MP520 ASQU 0.75m Mellanox EDR IB Passive Copper QSFP28 Cable
00MP524 ASQV 1m Mellanox EDR IB Passive Copper QSFP28 Cable
00MP528 ASQW 1.25m Mellanox EDR IB Passive Copper QSFP28 Cable
00MP532 ASQX 1.5m Mellanox EDR IB Passive Copper QSFP28 Cable
00MP536 ASQY 2m Mellanox EDR IB Passive Copper QSFP28 Cable
00MP560 ASRM 3m Mellanox EDR IB Passive Copper QSFP28 Cable
QSFP28 100Gb Ethernet Passive DAC Cables
7Z57A03561 AV1Z Lenovo 1m Passive 100G QSFP28 DAC Cable
7Z57A03562 AV20 Lenovo 3m Passive 100G QSFP28 DAC Cable
7Z57A03563 AV21 Lenovo 5m Passive 100G QSFP28 DAC Cable
QSFP56 HDR InfiniBand Passive DAC Cables
4Z57A14182 B4QQ 0.5m Mellanox HDR IB Passive Copper QSFP56 Cable
4Z57A14183 B4QR 1m Mellanox HDR IB Passive Copper QSFP56 Cable
4Z57A14184 B4QS 1.5m Mellanox HDR IB Passive Copper QSFP56 Cable
4Z57A14185 B4QT 2m Mellanox HDR IB Passive Copper QSFP56 Cable
QSFP56 HDR InfiniBand to 2x HDR100 Passive DAC Splitter Cables
4Z57A14193 B4R1 1m Mellanox HDR IB to 2x HDR100 Splitter Passive Copper QSFP56 Cable
4Z57A14194 B4R2 1.5m Mellanox HDR IB to 2x HDR100 Splitter Passive Copper QSFP56 Cable
4Z57A11477 B68L 2m Mellanox HDR to 2x HDR100 IB Passive DAC QSFP56 Splitter Cable
QSFP56 200Gb Ethernet Passive DAC Cables
4X97A11113 BF6W Lenovo 1m 200G QSFP56 DAC Cable
4X97A12613 BF92 Lenovo 3m 200G QSFP56 DAC Cable
QSFP56 HDR InfiniBand Active DAC Cables
4X97A12610 BCQW 3m Mellanox HDR IB Active Copper QSFP56 Cable
4X97A12611 BCQX 4m Mellanox HDR IB Active Copper QSFP56 Cable

Features

Machine learning and big data environments

Data analytics has become an essential function within many enterprise data centers, clouds and hyperscale platforms. Machine learning relies on especially high throughput and low latency to train deep neural networks and to improve recognition and classification accuracy. ConnectX-6 offers an excellent solution to provide machine learning applications with the levels of performance and scalability that they require.

ConnectX-6 utilizes the RDMA technology to deliver low-latency and high performance. ConnectX-6 enhances RDMA network capabilities even further by delivering end-to-end packet level flow control.

Security

ConnectX-6 block-level encryption offers a critical innovation to network security. As data in transit is stored or retrieved, it undergoes encryption and decryption. The ConnectX-6 hardware offloads the IEEE AES-XTS encryption/decryption from the CPU, saving latency and CPU utilization. It also guarantees protection for users sharing the same resources through the use of dedicated encryption keys.

By performing block-storage encryption in the adapter, ConnectX-6 excludes the need for self-encrypted disks. This allows customers the freedom to choose their preferred storage device, including those that traditionally do not provide encryption. ConnectX-6 can support Federal Information Processing Standards (FIPS) compliance.

ConnectX-6 also includes a hardware Root-of-Trust (RoT), which uses HMAC relying on a device-unique key. This provides both a secure boot as well as cloning-protection. Delivering best-in-class device and firmware protection, ConnectX-6 also provides secured debugging capabilities, without the need for physical access.

Storage environments

NVMe storage devices offer very fast access to storage media. The evolving NVMe over Fabric (NVMe-oF) protocol leverages RDMA connectivity to remotely access NVMe storage devices efficiently, while keeping the end-to-end NVMe model at lowest latency. With its NVMe-oF target and initiator offloads, ConnectX-6 brings further optimization to NVMe-oF, enhancing CPU utilization and scalability.

Cloud and Web 2.0 environments

Telco, Cloud and Web 2.0 customers developing their platforms on software-defined network (SDN) environments are leveraging the Virtual Switching capabilities of server operating systems to enable maximum flexibility in the management and routing protocols of their networks.

Open V-Switch (OVS) is an example of a virtual switch that allows virtual machines to communicate among themselves and with the outside world. Software-based virtual switches, traditionally residing in the hypervisor, are CPU intensive, affecting system performance and preventing full utilization of available CPU for compute functions.

To address such performance issues, ConnectX-6 offers Mellanox Accelerated Switching and Packet Processing (ASAP2) Direct technology. ASAP2 offloads the vSwitch/vRouter by handling the data plane in the NIC hardware while maintaining the control plane unmodified. As a result, significantly higher vSwitch/vRouter performance is achieved minus the associated CPU load.

The vSwitch/vRouter offload functions supported by ConnectX-5 and ConnectX-6 include encapsulation and de-capsulation of overlay network headers, as well as stateless offloads of inner packets, packet headers re-write (enabling NAT functionality), hairpin, and more.

In addition, ConnectX-6 offers intelligent flexible pipeline capabilities, including programmable flexible parser and flexible match-action tables, which enable hardware offloads for future protocols.

Technical specifications

The adapters have the following technical specifications.

PCI Express Interface

  • PCIe 4.0 x16 host interface (also supports a PCIe 3.0 host interface)
  • Support for PCIe x1, x2, x4, x8, and x16 configurations
  • PCIe Atomic
  • TLP (Transaction Layer Packet) Processing Hints (TPH)
  • PCIe switch Downstream Port Containment (DPC) enablement for PCIe hot-plug
  • Advanced Error Reporting (AER)
  • Access Control Service (ACS) for peer-to-peer secure communication
  • Process Address Space ID (PASID) Address Translation Services (ATS)
  • IBM CAPIv2 (Coherent Accelerator Processor Interface)
  • Support for MSI/MSI-X mechanisms

Connectivity

  • One or two QSFP56 ports
  • Supports passive copper cables with ESD protection
  • Powered connectors for optical and active cable support

InfiniBand

  • Supports interoperability with InfiniBand switches (up to HDR100)
  • When used in a PCIe 3.0 slot, total connectivity is up to 100 Gb/s:
    • One port adapter supports a single 100 Gb/s link
    • Two-port adapter supports two connections of 50 Gb/s each or one 100 Gb/s active link and the other a standby link
  • When used in a PCIe 4.0 slot, total connectivity is up to 200 Gb/s:
    • One port adapter supports a single 100 Gb/s link
    • Two-port adapter supports two connections of 100 Gb/s each
  • HDR100 / EDR / FDR / QDR / DDR / SDR
  • IBTA Specification 1.3 compliant
  • RDMA, Send/Receive semantics
  • Hardware-based congestion control
  • Atomic operations
  • 16 million I/O channels
  • 256 to 4Kbyte MTU, 2Gbyte messages
  • 8 virtual lanes + VL15

Ethernet (requires firmware 20.28.1002 or later)

  • Support interoperability with Ethernet switches (up to 100GbE, as 2 lanes of 50Gb/s data rate)
  • When used in the PCIe 3.0 slot, total connectivity is 100 Gb/s:
    • One port adapter supports a single 100 Gb/s link
    • Two-port adapter supports two connections of 50 Gb/s each or one 100 Gb/s active link and the other a standby link
  • When used in the PCIe 4.0 slot, total connectivity is 200 Gb/s:
    • One port adapter supports a single 100 Gb/s link
    • Two-port adapter supports two connections of 100 Gb/s each
  • Supports 100GbE / 50GbE / 40GbE / 25GbE / 10GbE / 1GbE
  • Ethernet speed must be set; auto-negotiation is currently not supported (planned for a later firmware update)
  • IEEE 802.3bj, 802.3bm 100 Gigabit Ethernet
  • IEEE 802.3by, Ethernet Consortium 25, 50 Gigabit Ethernet, supporting all FEC modes
  • IEEE 802.3ba 40 Gigabit Ethernet
  • IEEE 802.3ae 10 Gigabit Ethernet
  • IEEE 802.3az Energy Efficient Ethernet
  • IEEE 802.3ap based auto-negotiation and KR startup (planned for a later firmware update)
  • IEEE 802.3ad, 802.1AX Link Aggregation
  • IEEE 802.1Q, 802.1P VLAN tags and priority
  • IEEE 802.1Qau (QCN) – Congestion Notification
  • IEEE 802.1Qaz (ETS)
  • IEEE 802.1Qbb (PFC)
  • IEEE 802.1Qbg
  • IEEE 1588v2
  • Jumbo frame support (9.6KB)

Enhanced Features

  • Hardware-based reliable transport
  • Collective operations offloads
  • Vector collective operations offloads
  • PeerDirect RDMA (GPUDirect) communication acceleration
  • 64/66 encoding
  • Enhanced Atomic operations
  • Advanced memory mapping support, allowing user mode registration and remapping of memory (UMR)
  • Extended Reliable Connected transport (XRC)
  • Dynamically Connected transport (DCT)
  • On demand paging (ODP)
  • MPI Tag Matching
  • Rendezvous protocol offload
  • Out-of-order RDMA supporting Adaptive Routing
  • Burst buffer offload
  • In-Network Memory registration-free RDMA memory access

CPU Offloads

  • RDMA over Converged Ethernet (RoCE)
  • TCP/UDP/IP stateless offload
  • LSO, LRO, checksum offload
  • RSS (also on encapsulated packet), TSS, HDS, VLAN and MPLS tag insertion/stripping, Receive flow steering
  • Data Plane Development Kit (DPDK) for kernel bypass applications
  • Open VSwitch (OVS) offload using ASAP2
    • Flexible match-action flow tables
    • Tunneling encapsulation / de-capsulation
  • Intelligent interrupt coalescence
  • Header rewrite supporting hardware offload of NAT router

Storage Offloads

  • Block-level encryption: XTS-AES 256/512 bit key
  • NVMe over Fabric offloads for target machine
  • Erasure Coding offload - offloading Reed-Solomon calculations
  • T10 DIF - signature handover operation at wire speed, for ingress and egress traffic
  • Storage Protocols: SRP, iSER, NFS RDMA, SMB Direct, NVMe-oF

Overlay Networks

  • RoCE over overlay networks
  • Stateless offloads for overlay network tunneling protocols
  • Hardware offload of encapsulation and decapsulation of VXLAN, NVGRE, and GENEVE overlay networks

Hardware-Based I/O Virtualization

  • Single Root IOV
  • Address translation and protection
  • VMware NetQueue support
  • SR-IOV: Up to 512 Virtual Functions
  • SR-IOV: Up to 16 Physical Functions per host
  • Virtualization hierarchies (network partitioning, NPAR)
    • Virtualizing Physical Functions on a physical port
    • SR-IOV on every Physical Function
  • Configurable and user-programmable QoS
  • Guaranteed QoS for VMs

HPC Software Libraries

  • HPC-X, OpenMPI, MVAPICH, MPICH, OpenSHMEM, PGAS and varied commercial packages

Management and Control

  • NC-SI, MCTP over SMBus and MCTP over PCIe - BMC interface
  • PLDM for Monitor and Control DSP0248
  • PLDM for Firmware Update DSP0267
  • SDN management interface for managing the eSwitch
  • I2C interface for device control and configuration
  • General Purpose I/O pins
  • SPI interface to Flash
  • JTAG IEEE 1149.1 and IEEE 1149.6

Remote Boot

  • Remote boot over InfiniBand
  • Remote boot over Ethernet
  • Remote boot over iSCSI
  • Unified Extensible Firmware Interface (UEFI)
  • Pre-execution Environment (PXE)

Server support

The following servers offer a PCIe 4.0 host interface. All other supported servers have a PCIe 3.0 host interface.

  • ThinkSystem SR635
  • ThinkSystem SR655
  • ThinkSystem SR645
  • ThinkSystem SR665

The following tables list the ThinkSystem servers that are compatible.

Table 5. Server support (Part 1 of 2)
Part Number Description E 2S Intel V2 AMD Dense V2 4S V2 8S
SE350 (7Z46 / 7D1X)
ST650 V2 (7Z75 / 7Z74)
SR630 V2 (7Z70 / 7Z71)
SR650 V2 (7Z72 / 7Z73)
SR670 V2 (7Z22 / 7Z23)
SR635 (7Y98 / 7Y99)
SR655 (7Y00 / 7Z01)
SR645 (7D2Y / 7D2X)
SR665 (7D2W / 7D2V)
SD630 V2 (7D1K)
SD650 V2 (7D1M)
SD650-N V2 (7D1N)
SN550 V2 (7Z69)
SR850 V2 (7D31 / 7D32)
SR860 V2 (7Z59 / 7Z60)
SR950 (7X11 / 7X12)
4C57A14177 ThinkSystem Mellanox ConnectX-6 HDR100/100GbE QSFP56 1-port PCIe VPI Adapter N N Y Y Y Y Y Y Y Y N N N Y Y Y
4C57A14178 ThinkSystem Mellanox ConnectX-6 HDR100/100GbE QSFP56 2-port PCIe VPI Adapter N N Y Y Y Y Y Y Y N N N N Y Y Y
Table 6. Server support (Part 2 of 2)
Part Number Description 1S Intel 2S Intel V1 Dense V1 4S V1
ST50 (7Y48 / 7Y50)
ST250 (7Y45 / 7Y46)
SR150 (7Y54)
SR250 (7Y52 / 7Y51)
ST550 (7X09 / 7X10)
SR530 (7X07 / 7X08)
SR550 (7X03 / 7X04)
SR570 (7Y02 / 7Y03)
SR590 (7X98 / 7X99)
SR630 (7X01 / 7X02)
SR650 (7X05 / 7X06)
SR670 (7Y36 / 7Y37)
SD530 (7X21)
SD650 (7X58)
SN550 (7X16)
SN850 (7X15)
SR850 (7X18 / 7X19)
SR850P (7D2F / 2D2G)
SR860 (7X69 / 7X70)
4C57A14177 ThinkSystem Mellanox ConnectX-6 HDR100/100GbE QSFP56 1-port PCIe VPI Adapter N N N N N N N N N Y Y Y Y N N N Y Y Y
4C57A14178 ThinkSystem Mellanox ConnectX-6 HDR100/100GbE QSFP56 2-port PCIe VPI Adapter N N N N N N N N N Y Y Y Y N N N Y Y Y

Operating system support

The adapters support the operating systems listed in the following tables:

Tip: These tables are automatically generated based on data from Lenovo ServerProven.

Table 7. Operating system support for ThinkSystem Mellanox ConnectX-6 HDR100/100GbE QSFP56 1-port PCIe VPI Adapter, 4C57A14177
Operating systems
SR635
SR645
SR655
SR665
SD530 (Gen 2)
SR630 (Gen 2)
SR650 (Gen 2)
SR670 (Gen 2)
SR850 (Gen 2)
SR850P
SR860 (Gen 2)
SR950 (Gen 2)
SR670 V2
SR850 V2
SD530 (Gen 1)
SR630 (Gen 1)
SR650 (Gen 1)
SR850 (Gen 1)
SR860 (Gen 1)
SR950 (Gen 1)
Microsoft Windows Server 2016 Y Y Y Y Y Y Y N Y Y Y Y Y Y Y Y Y Y Y Y
Microsoft Windows Server 2019 Y Y Y Y Y Y Y N Y Y Y Y Y Y Y Y Y Y Y Y
Red Hat Enterprise Linux 7.5 N N N N N N N Y N N N N N N Y Y Y Y Y Y
Red Hat Enterprise Linux 7.6 Y Y Y Y Y Y Y Y Y Y Y Y N N Y Y Y Y Y Y
Red Hat Enterprise Linux 7.7 Y Y Y Y N N N N N Y N N N N N N N N N N
Red Hat Enterprise Linux 7.8 Y Y Y Y Y Y Y Y Y Y Y Y N N Y Y Y Y Y Y
Red Hat Enterprise Linux 7.9 Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y
Red Hat Enterprise Linux 8.1 Y N Y N N N N N N Y N N N N N N N N N N
Red Hat Enterprise Linux 8.2 Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y
Red Hat Enterprise Linux 8.3 Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y
Red Hat Enterprise Linux 8.4 Y Y Y Y Y Y Y Y Y Y Y Y N Y Y Y Y Y Y Y
SUSE Linux Enterprise Server 12 SP3 N N N N N N N N N Y N N N N Y Y Y Y Y Y
SUSE Linux Enterprise Server 12 SP4 Y N Y N Y Y Y N Y Y Y Y N N Y Y Y Y Y Y
SUSE Linux Enterprise Server 12 SP5 Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y
SUSE Linux Enterprise Server 15 N N N N Y Y Y N Y Y Y Y N N Y Y Y Y Y Y
SUSE Linux Enterprise Server 15 SP1 Y Y Y Y Y Y Y Y Y Y Y Y N N Y Y Y Y Y Y
SUSE Linux Enterprise Server 15 SP2 Y 1 Y Y 1 Y Y 1 Y 1 Y 1 Y 1 Y 1 Y 1 Y 1 Y 1 Y Y Y 1 Y 1 Y 1 Y 1 Y 1 Y 1
Ubuntu 18.04 LTS N N N N N N N N N N N N Y N N N N N N N
VMware vSphere Hypervisor (ESXi) 6.7 N N N N N N N N N N N N N N Y Y Y Y Y Y
VMware vSphere Hypervisor (ESXi) 6.7 U1 N N N N Y Y Y N Y Y Y Y N N Y Y Y Y Y Y
VMware vSphere Hypervisor (ESXi) 6.7 U2 N N N N Y Y Y N Y Y Y Y N N Y Y Y Y Y Y
VMware vSphere Hypervisor (ESXi) 6.7 U3 Y Y Y Y Y Y Y Y Y Y Y Y Y N Y Y Y Y Y Y
VMware vSphere Hypervisor (ESXi) 7.0 Y Y Y Y Y Y Y Y Y Y Y Y N N Y Y Y Y Y Y
VMware vSphere Hypervisor (ESXi) 7.0 U1 Y Y Y Y Y Y Y Y Y Y Y Y N Y Y Y Y Y Y Y
VMware vSphere Hypervisor (ESXi) 7.0 U2 Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y

1 Need out of box driver to support infiniband feature

Table 8. Operating system support for ThinkSystem Mellanox ConnectX-6 HDR100/100GbE QSFP56 2-port PCIe VPI Adapter, 4C57A14178
Operating systems
SR635
SR645
SR655
SR665
SD530 (Gen 2)
SR630 (Gen 2)
SR650 (Gen 2)
SR670 (Gen 2)
SR850 (Gen 2)
SR850P
SR860 (Gen 2)
SR950 (Gen 2)
SR670 V2
SR850 V2
SD530 (Gen 1)
SR630 (Gen 1)
SR650 (Gen 1)
SR850 (Gen 1)
SR860 (Gen 1)
SR950 (Gen 1)
Microsoft Windows Server 2016 Y Y Y Y Y Y Y N Y Y Y Y Y Y Y Y Y Y Y Y
Microsoft Windows Server 2019 Y Y Y Y Y Y Y N Y Y Y Y Y Y Y Y Y Y Y Y
Red Hat Enterprise Linux 7.5 N N N N N N N Y N N N N N N Y Y Y Y Y Y
Red Hat Enterprise Linux 7.6 Y Y Y Y Y Y Y Y Y Y Y Y N N Y Y Y Y Y Y
Red Hat Enterprise Linux 7.7 Y Y Y Y N N N N N Y N N N N N N N N N N
Red Hat Enterprise Linux 7.8 Y Y Y Y Y Y Y Y Y Y Y Y N N Y Y Y Y Y Y
Red Hat Enterprise Linux 7.9 Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y
Red Hat Enterprise Linux 8.1 Y N Y N N N N N N Y N N N N N N N N N N
Red Hat Enterprise Linux 8.2 Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y
Red Hat Enterprise Linux 8.3 Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y
Red Hat Enterprise Linux 8.4 Y Y Y Y Y Y Y Y Y Y Y Y N Y Y Y Y Y Y Y
SUSE Linux Enterprise Server 12 SP3 N N N N N N N N N Y N N N N Y Y Y Y Y Y
SUSE Linux Enterprise Server 12 SP4 Y N Y N Y Y Y N Y Y Y Y N N Y Y Y Y Y Y
SUSE Linux Enterprise Server 12 SP5 Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y
SUSE Linux Enterprise Server 15 N N N N Y Y Y N Y Y Y Y N N Y Y Y Y Y Y
SUSE Linux Enterprise Server 15 SP1 Y Y Y Y Y Y Y Y Y Y Y Y N N Y Y Y Y Y Y
SUSE Linux Enterprise Server 15 SP2 Y 1 Y Y 1 Y Y 1 Y 1 Y 1 Y 1 Y 1 Y 1 Y 1 Y 1 Y Y Y 1 Y 1 Y 1 Y 1 Y 1 Y 1
Ubuntu 18.04 LTS N N N N N N N N N N N N Y N N N N N N N
VMware vSphere Hypervisor (ESXi) 6.7 N N N N N N N N N N N N N N Y Y Y Y Y Y
VMware vSphere Hypervisor (ESXi) 6.7 U1 N N N N Y Y Y N Y Y Y Y N N Y Y Y Y Y Y
VMware vSphere Hypervisor (ESXi) 6.7 U2 N N N N Y Y Y N Y Y Y Y N N Y Y Y Y Y Y
VMware vSphere Hypervisor (ESXi) 6.7 U3 Y Y Y Y Y Y Y Y Y Y Y Y Y N Y Y Y Y Y Y
VMware vSphere Hypervisor (ESXi) 7.0 Y Y Y Y Y Y Y Y Y Y Y Y N N Y Y Y Y Y Y
VMware vSphere Hypervisor (ESXi) 7.0 U1 Y Y Y Y Y Y Y Y Y Y Y Y N Y Y Y Y Y Y Y
VMware vSphere Hypervisor (ESXi) 7.0 U2 Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y

1 Need out of box driver to support infiniband feature

Regulatory approvals

The adapters have the following regulatory approvals:

  • Safety: CB / cTUVus / CE
  • EMC: CE / FCC / VCCI / ICES / RCM / KC
  • RoHS: RoHS Compliant

Operating environment

The adapters have the following operating characteristics:

  • Typical power consumption (passive cables): 15.6W
  • Maximum power available through QSFP56 port: 5W
  • Temperature
    • Operational: 0°C to 55°C
    • Non-operational: -40°C to 70°C
  • Humidity: 90% relative humidity

Warranty

One year limited warranty. When installed in a Lenovo server, the adapter assumes the server’s base warranty and any warranty upgrades.

Related product families

Product families related to this document are the following:

Trademarks

Lenovo and the Lenovo logo are trademarks or registered trademarks of Lenovo in the United States, other countries, or both. A current list of Lenovo trademarks is available on the Web at https://www.lenovo.com/us/en/legal/copytrade/.

The following terms are trademarks of Lenovo in the United States, other countries, or both:
Lenovo®
ServerProven®
ThinkSystem

The following terms are trademarks of other companies:

Intel® is a trademark of Intel Corporation or its subsidiaries.

Linux® is the trademark of Linus Torvalds in the U.S. and other countries.

Microsoft®, Windows Server®, and Windows® are trademarks of Microsoft Corporation in the United States, other countries, or both.

Other company, product, or service names may be trademarks or service marks of others.