Jump to: navigation, search

Difference between revisions of "Mellanox-Neutron-ML2-Kilo"

 
(15 intermediate revisions by 2 users not shown)
Line 1: Line 1:
  
<big>'''This page is still under construction''' </big>
+
Mellanox ML2 Mechanism Driver implements the ML2 Plugin Mechanism Driver API.
 +
 
 +
<big>'''This page is deprecated.'''</big> Please refer to [[Mellanox-Neutron-Kilo-InfiniBand]]
 +
 
 +
 
 +
 
 +
This driver supports Mellanox embedded switch functionality as part of the InfiniBand HCA. Mellanox ML2 Mechanism Driver provides functional parity with Mellanox Neutron plugin.
  
 +
Mellanox ML2 Mechanism Driver supports DIRECT (pci passthrough) vnic type. For vnic type configuration API details, please refer to configuration reference guide. Hardware vNICs mapped to the guest VMs allow higher performance and advanced features such as RDMA (remote direct memory access).
  
 +
The driver supports VLAN network type to facilitate virtual networks on InfiniBand fabrics.
 +
* Mellanox OpenStack Neutron Agent (L2 Agent) runs on each compute node.
 +
* Agent should apply VIF connectivity based on mapping between a VIF (VM vNIC) and Embedded Switch port.
  
  
Mellanox ML2 Mechanism Driver implements the ML2 Plugin Mechanism Driver API.  
+
'''Prerequisites'''
 +
* All nodes equipped with Mellanox ConnectX®-3/Mellanox ConnectX®-3 PRO [http://www.mellanox.com/page/products_dyn?product_family=119 Network Adapter]
 +
* Mellanox OFED 2.4 or greater installed on all nodes. Please refer to Mellanox website for the latest [http://www.mellanox.com/page/products_dyn?product_family=26&mtag=linux_sw_drivers OFED]
  
This driver supports Mellanox embedded switch functionality as part of the VPI (Ethernet/InfiniBand) HCA. Mellanox ML2 Mechanism Driver provides functional parity with Mellanox Neutron plugin.
 
  
Mellanox ML2 Mechanism Driver supports DIRECT (pci passthrough) vnic type. For vnic type configuration API details, please refer to configuration reference guide (click here). Hardware vNICs mapped to the guest VMs allow higher performance and advanced features such as RDMA (remote direct memory access).
+
For '''Kilo''' release Information refer to the relevant OS as follows:
  
The driver supports VLAN network type to facilitate virtual networks either on Ethernet or InfiniBand fabrics.
+
Infiniband
* Mellanox OpenStack Neutron Agent (L2 Agent) runs on each compute node.
+
* [[Mellanox-Neutron-Kilo-Redhat-InfiniBand | Redhat7]]
* Agent should apply VIF connectivity based on mapping between a VIF (VM vNIC) and Embedded Switch port.  
+
* [[Mellanox-Neutron-Kilo-Ubuntu-InfiniBand | Ubuntu14.04]]
  
* Redhat7 [[Mellanox-Neutron-Kilo-Redhat-Ethernet|Ethernet]] / [[Mellanox-Neutron-Kilo-Redhat-InfiniBand|InfiniBand]]
+
Ethernet(SR-IOV)
* ( Currently unsupported )  Ubuntu [[Mellanox-Neutron-Kilo-Ubuntu-Ethernet|Ethernet(SR-IOV)]] / [[Mellanox-Neutron-Kilo-Ubuntu-InfiniBand|InfiniBand]]
+
* [[Mellanox-Neutron-Kilo-Redhat-Ethernet |Redhat7/Ubuntu14.04]]

Latest revision as of 09:04, 6 October 2016

Mellanox ML2 Mechanism Driver implements the ML2 Plugin Mechanism Driver API.

This page is deprecated. Please refer to Mellanox-Neutron-Kilo-InfiniBand


This driver supports Mellanox embedded switch functionality as part of the InfiniBand HCA. Mellanox ML2 Mechanism Driver provides functional parity with Mellanox Neutron plugin.

Mellanox ML2 Mechanism Driver supports DIRECT (pci passthrough) vnic type. For vnic type configuration API details, please refer to configuration reference guide. Hardware vNICs mapped to the guest VMs allow higher performance and advanced features such as RDMA (remote direct memory access).

The driver supports VLAN network type to facilitate virtual networks on InfiniBand fabrics.

  • Mellanox OpenStack Neutron Agent (L2 Agent) runs on each compute node.
  • Agent should apply VIF connectivity based on mapping between a VIF (VM vNIC) and Embedded Switch port.


Prerequisites

  • All nodes equipped with Mellanox ConnectX®-3/Mellanox ConnectX®-3 PRO Network Adapter
  • Mellanox OFED 2.4 or greater installed on all nodes. Please refer to Mellanox website for the latest OFED


For Kilo release Information refer to the relevant OS as follows:

Infiniband

Ethernet(SR-IOV)