Jump to: navigation, search

Difference between revisions of "Mellanox-Neutron-ML2-Kilo"

Line 22: Line 22:
  
 
* Redhat7 [[Mellanox-Neutron-Kilo-Redhat-Ethernet|Ethernet]] / [[Mellanox-Neutron-Kilo-Redhat-InfiniBand|InfiniBand]]
 
* Redhat7 [[Mellanox-Neutron-Kilo-Redhat-Ethernet|Ethernet]] / [[Mellanox-Neutron-Kilo-Redhat-InfiniBand|InfiniBand]]
 +
* ( Currently unsupported )  Ubuntu [[Mellanox-Neutron-Kilo-Ubuntu-Ethernet|Ethernet(SR-IOV)]] / [[Mellanox-Neutron-Kilo-Ubuntu-InfiniBand|InfiniBand]]

Revision as of 14:35, 14 July 2015

This page is still under construction



Mellanox ML2 Mechanism Driver implements the ML2 Plugin Mechanism Driver API.

This driver supports Mellanox embedded switch functionality as part of the VPI (Ethernet/InfiniBand) HCA. Mellanox ML2 Mechanism Driver provides functional parity with Mellanox Neutron plugin.

Mellanox ML2 Mechanism Driver supports DIRECT (pci passthrough) vnic type. For vnic type configuration API details, please refer to configuration reference guide (click here). Hardware vNICs mapped to the guest VMs allow higher performance and advanced features such as RDMA (remote direct memory access).

The driver supports VLAN network type to facilitate virtual networks either on Ethernet or InfiniBand fabrics.

  • Mellanox OpenStack Neutron Agent (L2 Agent) runs on each compute node.
  • Agent should apply VIF connectivity based on mapping between a VIF (VM vNIC) and Embedded Switch port.