Jump to: navigation, search

Difference between revisions of "Mellanox-Neutron-ML2-Juno"

 
(14 intermediate revisions by the same user not shown)
Line 1: Line 1:
  
<big>'''This page is still under construction''' </big>
+
Mellanox ML2 Mechanism Driver implements the ML2 Plugin Mechanism Driver API.
  
 +
<big>'''This page is deprecated.'''</big> Please refer to [[Mellanox-Neutron-Juno-InfiniBand]]
  
 +
This driver supports Mellanox embedded switch functionality as part of the VPI (Ethernet/InfiniBand) HCA. Mellanox ML2 Mechanism Driver provides functional parity with Mellanox Neutron plugin.
  
 +
Mellanox ML2 Mechanism Driver supports DIRECT (pci passthrough) vnic type. For vnic type configuration API details, please refer to configuration reference guide (click here). Hardware vNICs mapped to the guest VMs allow higher performance and advanced features such as RDMA (remote direct memory access).
  
Mellanox ML2 Mechanism Driver implements the ML2 Plugin Mechanism Driver API.  
+
The driver supports VLAN network type to facilitate virtual networks either on Ethernet or InfiniBand fabrics.
 +
* Mellanox OpenStack Neutron Agent (L2 Agent) runs on each compute node.
 +
* Agent should apply VIF connectivity based on mapping between a VIF (VM vNIC) and Embedded Switch port.
  
This driver supports Mellanox embedded switch functionality as part of the VPI (Ethernet/InfiniBand) HCA. Mellanox ML2 Mechanism Driver provides functional parity with Mellanox Neutron plugin.
 
  
Mellanox ML2 Mechanism Driver supports DIRECT (pci passthrough) and MACVTAP (virtual interface with a tap-like software interface) vnic types. For vnic type configuration API details, please refer to configuration reference guide (click [http://docs.openstack.org/api/openstack-network/2.0/content/binding_ext_ports.html here]). Hardware vNICs mapped to the guest VMs allow higher performance and advanced features such as RDMA (remote direct memory access).
+
'''Prerequisites'''
 +
* All nodes equipped with Mellanox ConnectX®-3/Mellanox ConnectX®-3 PRO [http://www.mellanox.com/page/products_dyn?product_family=119 Network Adapter]
 +
* Mellanox OFED 2.4 or greater installed on all nodes. Please refer to Mellanox website for the latest [http://www.mellanox.com/page/products_dyn?product_family=26&mtag=linux_sw_drivers OFED]
  
The driver supports VLAN network type to facilitate virtual networks either on Ethernet or InfiniBand fabrics.
 
  
* Mellanox Openstack Neutron Agent (L2 Agent) runs on each compute node.
+
For '''Juno''' release Information refer to the relevant OS as follows:
* Agent should apply VIF connectivity based on mapping between a VIF (VM vNIC) and Embedded Switch port.
 
  
 +
Infiniband
 +
* [[Mellanox-Neutron-Juno-Redhat-InfiniBand|Redhat7]]
 +
* [[Mellanox-Neutron-Juno-Ubuntu-InfiniBand|Ubuntu14.04]] [ Under construction ]
  
For '''Juno''' release Information refer to the relevant OS as follows:
 
  
* Redhat [[Mellanox-Neutron-Juno-Redhat-Ethernet|Ethernet]] / [[Mellanox-Neutron-Juno-Redhat-InfiniBand|InfiniBand]]
+
Ethernet(SR-IOV)
* Ubuntu [[Mellanox-Neutron-Juno-Ubuntu-Ethernet|Ethernet]] / [[Mellanox-Neutron-Juno-Ubuntu-InfiniBand|InfiniBand]]
+
* [[Mellanox-Neutron-Juno-Redhat-Ethernet|Redhat7/Ubuntu14.04]]

Latest revision as of 09:06, 6 October 2016

Mellanox ML2 Mechanism Driver implements the ML2 Plugin Mechanism Driver API.

This page is deprecated. Please refer to Mellanox-Neutron-Juno-InfiniBand

This driver supports Mellanox embedded switch functionality as part of the VPI (Ethernet/InfiniBand) HCA. Mellanox ML2 Mechanism Driver provides functional parity with Mellanox Neutron plugin.

Mellanox ML2 Mechanism Driver supports DIRECT (pci passthrough) vnic type. For vnic type configuration API details, please refer to configuration reference guide (click here). Hardware vNICs mapped to the guest VMs allow higher performance and advanced features such as RDMA (remote direct memory access).

The driver supports VLAN network type to facilitate virtual networks either on Ethernet or InfiniBand fabrics.

  • Mellanox OpenStack Neutron Agent (L2 Agent) runs on each compute node.
  • Agent should apply VIF connectivity based on mapping between a VIF (VM vNIC) and Embedded Switch port.


Prerequisites

  • All nodes equipped with Mellanox ConnectX®-3/Mellanox ConnectX®-3 PRO Network Adapter
  • Mellanox OFED 2.4 or greater installed on all nodes. Please refer to Mellanox website for the latest OFED


For Juno release Information refer to the relevant OS as follows:

Infiniband


Ethernet(SR-IOV)