Jump to: navigation, search

Difference between revisions of "Tricircle"

(Overview)
(Video Resources)
 
(85 intermediate revisions by 4 users not shown)
Line 1: Line 1:
Tricircle provides an OpenStack API gateway and networking automation to allow multiple OpenStack instances, spanning in one site or multiple sites or in hybrid cloud, to be managed as a single OpenStack cloud
+
The Tricircle is to provide networking automation across Neutron  in multi-region OpenStack deployments.
  
 
==Use Cases==
 
==Use Cases==
==== Massive distributed edge cloud ====
+
To understand the motivation of the Tricircle project, we should understand the use cases with end user, and it would help you understand this project, please refer to presentation material: https://docs.google.com/presentation/d/1Zkoi4vMOGN713Vv_YO0GP6YLyjLpQ7fRbHlirpq6ZK4/
  
Now building massive distributed edge clouds in edge data centers with computing and storage close to end users is emerging for enterprise application, NFV service and personal service.
+
==== 0. Telecom application cloud level redundancy in OPNFV Beijing Summit ====
 +
Shared Networks to Support VNF High Availability Across OpenStack Multi Region Deployment
 +
This slides was presented in OPNFV Beijing Summit, 2017.
 +
slides: https://docs.google.com/presentation/d/1WBdra-ZaiB-K8_m3Pv76o_jhylEqJXTTxzEZ-cu8u2A/
 +
Video: https://www.youtube.com/watch?v=tbcc7-eZnkY
 +
OPNFV Summit vIMS Multisite Demo(Youtube): https://www.youtube.com/watch?v=zS0wwPHmDWs
 +
Video Conference High Availability across multiple OpenStack clouds(Youtube): https://www.youtube.com/watch?v=nK1nWnH45gI
  
===== Enterprise Application=====
+
==== 1. Application high availability running over cloud ====
Some enterprises also found issues for applications running in remote  centralized cloud, for example for video editing, 3D modeling application and IoT service etc which bandwidth and latency are sensitive.  
+
In telecom and other industries(for example, MySQL Galera cluster), normally applications tend to be designed as Active-Active, Active-Passive or N-to-1 node configurations. Because those applications need to be available as much as possible. Once those applications are planned to be migrated to cloud infrastructure, active instance(s) need(s) to be deployed in one OpenStack instance first. After that, passive or active instances(s) will be deployed in another (other) OpenStack instance(s) in order for achieving 99.999% availability, if it’s required.
  
The high bandwidth and low latency provided by the edge cloud are critical for enterprise level applications like video editing, 3D modeling, IoT service
+
The reason why this deployment style is required is that in general cloud system only achieve 99.99% availability. [1] [2]
  
For Enterprise, most of the employee will work in different branches, and access to the nearby edge cloud, and collaboration among employee from different branch leads to the requirement on cross edge cloud functionalities, like tenant level networking, data distribution and migration.
+
To achieve required high availability, design of network architecture (especially Layer 2 and Layer 3) needs across Neutron to be considered for application state replication or heartbeat.
 +
<br />
 +
[[File:Tricircle_usecase2.png|frameless|center|600px|Tricircle usecase2]]
 +
<br />
 +
The above picture shows an application with Galera DB cluster as backend which are geographically distributed into multiple OpenStack instances.
  
===== NFV and Edge Cloud Service=====
+
[1] https://aws.amazon.com/cn/ec2/sla/<br />
NFV(network function virtualization) will provide more flexible and better customized networking capability, for example, dynamic customized network bandwidth management,  and also help to move the computing and storage close to end users. With shortest path from the end users to the storage and computing, the uplink speed could be larger and terminate the bandwidth consumption as quick as possible,  will definitely bring better user experience, and change the way of content generation and store: real time, all data in cloud.  
+
[2] https://news.ycombinator.com/item?id=2470298<br />
  
For example, an user/enterprise can dynamically ask for high bandwidth/storage requirement for streaming the HD video/AR/VR data into the cloud temporarily, after finishing streaming, ask for more computing resources to do the post processing, and re-distribute the video to other sites. And when a user want to move/re-distribute the application and data from one edge cloud to another one, should be able to dynamically ask for cross edge cloud bandwidth managed by NFV.  
+
====  2. Dual ISPs Load Balancing for internet link ====
 +
Deploy application in separate OpenStack instance with dual ISPs for internet link redundancy, load balancing, east-west traffic isolation for data/state replication is needed.  
 +
<br />
 +
[[File:Tricircle_usecase5.png|frameless|center|600px|Tricircle usecase5]]
 +
<br />
  
For VNF(telecom virtualized application), distributed designed VNF will be placed to multiple edge data centers for higher reliability/availability. To provide this support, typically requires state replication between application instances (directly or via replicated database services, or via private designed message format), tenant level isolated networking plane across data centers is needed for application state replication.
+
==== 3. Isolation of East-West traffic ====
 +
In financial industry, more than one OpenStack instances will be deployed, some OpenStack instance will be put in DMZ zone, and the other ones in trust zone, so that application or part of the application could be put in different level security zone. Although the tenant’s resources are provisioned in different OpenStack instances, east-west traffic among the tenant’s resources should be isolated.
 +
<br />
 +
[[File:Tricircle_usecase3.png|frameless|center|600px|Tricircle usecase3]]
 +
<br />
 +
Bandwidth sensitive, heavy load App like CAD modeling asked for the cloud close to the end user in distributed edge site for better user experience, multiple OpenStack instances will be deployed into edge sites, but the east-west communication with isolation between the tenant’s resources are also required.
 +
<br />
 +
[[File:Tricircle_usecase4.png|frameless|center|600px|Tricircle usecase4]]
 +
<br />
  
===== Personal service =====
+
Similar requirements could also be found in the ops session in OpenStack Barcelona summit "Control Plane Design (multi-region)" Line 25~26, 47~50: https://etherpad.openstack.org/p/BCN-ops-control-plane-design
Current Internet is good at processing down-link service. All contents are stored in remote centralized data centers and to some extent the access is accelerated with CDN.<br />
 
  
As more and more users generate content uploaded/streamed to the cloud and web site, these contents and data still have to be uploaded/streamed to some centralized data centers, the path is long and the bandwidth is limited and slow. For example, it’s very slow to uploading/streaming HD/2k/4k video for every user concurrently. For pictures or videos, they have to be uploaded with quality loss, and slow, using cloud as the first storage for users data has not the choice yet, currently it’s mainly for backup, and for none time/latency sensitive data. Some video captured and stored with quality loss even lead to the difficulty to provide the crime evidence or other purpose. The last mile of network access (fix or mobile) is wide enough, the main hindrance is that bandwidth in MAN(Metropolitan Area Network) and Backbone and WAN is limited and expensive. Real time video/data uploading/streaming from end user/terminal to the local edge cloud is quite attractive cloud service.
+
==== 4. Cross Neutron L2 network for NFV area ====
 
+
In NFV(Network Function Virtualization) area, network functions like router, NAT, LB etc will be virtualized, The cross Neutron  L2 networking capability like IP address space management, IP allocation and L2 network segment global management provided by the Tricircle can help VNFs(virtualized network function) across site inter-connection. For example, vRouter1 in site1, but vRouter2 in site2, these two VNFs could be in one L2 network across site.
From family or personal point of view, the movement or distribution of App/Storage from one edge data center to another one is also needed. For example, all video will be stored and processed in Hawaii locally when I am taking video in travelling, but I want the video after processing to be moved to Shenzhen China when I come back to China. But in Shenzhen, I want to share the video with streaming service not only in Shenzhen but to friends in Shanghai Beijing, so the data and the streaming service can be built in Shenzhen/Shanghai/Beijing too. The dynamically bandwidth incremental and app/data movement/replication can be helped through NFV edge cloud.
 
 
 
===== Requirements =====
 
The emerging massive distributed edge clouds will be not only some independent clouds, but also some new requirements are needed:
 
* Tenant level L2/L3 networking across data centers for isolation to tenant's E-W traffic
 
* Tenant level Volume/VM/object storage backup/migration/distribution
 
* Distributed image management
 
* Distributed quota management
 
* ...
 
 
 
==== Large scale cloud ====
 
 
 
Compared with Amazon, the scalability of OpenStack is still not good enough. One Amazon AZ can supports >50000 servers(http://www.slideshare.net/AmazonWebServices/spot301-aws-innovation-at-scale-aws-reinvent-2014).
 
 
 
Cells is a good enhancement, but the shortage of Cells are: 1) only nova supports cells. 2) using RPC for inter-datacenter communication will bring the difficulty in inter-dc troubleshooting and maintenance, no CLI or other tools to manage a child cell directly, if the link between the API cell and child cells is broken, then the child cell is unmanageable. 3) upgrade has to deal with DB and RPC change. 4)difficult for multi-vendor integration for different cells.
 
 
 
From the experience of production large scale public cloud point of view, the large scale cloud can only be built by capacity expansion step by step (intra-AZ and inter-AZ). And the challenge in capacity expansion is how to do the sizing:
 
* Number of Nova-API Server...
 
* Number of Cinder-API Server..
 
* Number of Neutron-API Server…
 
* Number of Scheduler..
 
* Number of Conductor…
 
* Specification of physical switch…
 
* Size of storage for Image..
 
* Size of management plane bandwidth…
 
* Size of data plane bandwidth…
 
* Reservation of rack space …
 
* Reservation of networking slots…
 
* ….<br />
 
 
 
You have to estimate, calculate, monitor, simulate, test, online grey expansion for controller nodes and network nodes…whenever you add new machines to the cloud. The difficulty is that you can’t test and verify in all size.
 
 
 
The feasible way to expand one large scale cloud is to add some already tested building block. That means we would prefer to build large scale public cloud by adding tested OpenStack instance (including controller and compute nodes) one by one, but would not prefer to unconditionally enlarge the capacity of one OpenStack instance. This way put the cloud construction under control.
 
 
 
Building large scale cloud by adding tested OpenStack instance one by one, but tenant's VMs may need to to be added to same network even if you add a new OpenStack building, or networks will be added into same router even if these networks of the tenant located in different OpenStack building blocks. But from the end user and PaaS point of view, they still want to use OpenStack API for already developed CLI, SDK, Portal, PaaS, Heat, Maganum, Murano etc. This way of building large scale public cloud also brings some new requirement to OpenStack based cloud which is quite similar like that in massive distributed edge clouds:
 
 
 
* Tenant level L2/L3 networking across OpenStack instances for isolation to tenant's E-W traffic
 
* Distributed quota management
 
* Global resource view of the tenant
 
* Tenant level Volume/VM migration/backup
 
* Multi-DC image import/clone/export
 
* ...
 
 
 
====  OpenStack API enabled hybrid cloud ====
 
Refer to https://wiki.openstack.org/wiki/Jacket
 
  
 +
==== 5. Cloud Capacity Expansion ====
 +
When an OpenStack cloud is created and many resources has been provisioned on it, the capacity of one of the OpenStack deployments may not be enough. A new OpenStack deployment needs to be added to this cloud. But tenants still want to add VMs to existing network. And they don’t want to add a new network, router or whatever required resources.
 +
<br />
 +
[[File:Tricircle_usecase1.png|frameless|center|600px|Tricircle usecase1]]
 
<br />
 
<br />
The detailed use cases could be found in this presentation: https://docs.google.com/presentation/d/1UQWeAMIJgJsWw-cyz9R7NvcAuSWUnKvaZFXLfRAQ6fI/edit?usp=sharing
 
 
And also can meet the demand for several working group: Telco WG [https://wiki.openstack.org/wiki/TelcoWorkingGroup/UseCases documents], [https://etherpad.openstack.org/p/Network_Segmentation_Usecases Large Deployment Team Use Cases], and [https://wiki.opnfv.org/multisite/use_cases OPNFV Multisite Use Cases]
 
 
==Overview==
 
 
The Tricircle provides an OpenStack API gateway and networking automation to allow multiple OpenStack instances, spanning in one site or multiple sites or in hybrid cloud, to be managed as a single OpenStack cloud.
 
 
The Tricircle and these managed OpenStack instances will use shared KeyStone (with centralized or distributed deployment) or federated KeyStones for identity management. The Tricircle presents one big region to the end user in KeyStone. And each OpenStack instance called a pod is a sub-region of the Tricircle in KeyStone, and usually not visible to end user directly.
 
 
The Tricircle acts as OpenStack API gateway, can handle OpenStack API calls, schedule one proper OpenStack instance if needed during the API calls handling, forward the API calls to the appropriate OpenStack instance, and deal with tenant level L2/L3 networking across OpenStack instances automatically, so that the VMs of the tenant, no matter located in which bottom OpenStack instance, can communicate with each other via L2 or L3.
 
 
The end user can see availability zone(AZ) and use AZ to provision VM, Volume, even Network through the Tricircle. One AZ can include many OpenStack instances, the Tricircle can schedule and bind OpenStack instance for the tenant inside one AZ. A tenant's resources could be bound to multiple specific bottom OpenStack instances in one or multiple AZs automatically.
 
 
The Tricircle is the formal open source project for OpenStack cascading solution ( https://wiki.openstack.org/wiki/OpenStack_cascading_solution ) but with enhanced and decoupled design.
 
 
The Tricircle could be extended to support more powerful capabilities such as support the  Tricircle instance being virtually splitted into multiple micro instances which could enable user to have a more fine granularity on the tenancy and service. And the Tricircle also enables OpenStack based hybrid cloud.
 
  
 
==Architecture==
 
==Architecture==
The cascading solution based on PoC design with enhancement is running in several production clouds like Huawei Enterprise Cloud in China, which brings the confidence of the value of cascading model, here the focus is on how to design and develop a perfect solution in open source.<br />
 
  
The initial architectural in the PoC is stateful, which could be found in https://wiki.openstack.org/wiki/OpenStack_cascading_solution, and the major headache part identified in the PoC are status synchronization for VM,Volume, etc, UUID mapping and coupling with OpenSatck existing services like Nova, Cinder.
+
Now the Tricircle is dedicated for networking automation across Neutron in multi-region OpenStack deployments. The design blueprint has been developed with ongoing improvement in https://docs.google.com/document/d/1zcxwl8xMEpxVCqLTce2-dUOtB-ObmzJTbV1uSQ6qTsY/,  
  
Now the Tricircle is being developed with stateless design to remove the challenges, and fully decoupled from OpenStack services. An  improved design blueprint is being developed in https://docs.google.com/document/d/18kZZ1snMOCD9IQvUKI5NVDzSASpw-QKj7l2zNqMEd3g/edit?usp=sharing,  
+
From the control plane view (cloud management view ), Tricircle is to make Neutron(s) in multi-region OpenStack clouds working as one cluster, and enable the creation of global  network/router etc abstract networking resources across multiple OpenStack clouds. From the data plane view (end user resources view), all VMs(also could be bare metal servers or containers) are provisioned in different cloud but can be inter-connected via the global abstract networking resources, of course, with tenant level isolation.
  
<big>Stateless Architecture</big>
 
 
<br />
 
<br />
  
[[File:Tricircle_Stateless_proposal.png|frameless|center|x400px|Tricircle improved architecture design - stateless]]
+
[[File:Tricircle_view.png|frameless|center|x400px|Tricircle_view]]
 
<br />
 
<br />
  
* Nova API-GW
+
The software design is as follows:
# A standalone web service to receive all nova api request, and routing the request to appropriate bottom OpenStack instance according to Availability Zone ( during creation ) or resource id ( during operation and query ). If more than one pod in one Availability Zone, schedule and forward the request to proper pod, and build the tenant ID and pod binding relationship.
 
# Nova APIGW is the functionality to trigger automatic networking when new VMs are being provisioned.
 
# work as stateless service, and could run with processes distributed in multi-hosts.
 
* Cinder API-GW
 
# a standalone web service to receive all cinder api request, and route the request to appropriate bottom OpenStack according to Availability Zone ( during creation ) or resource id ( during operation and query ).  If more than one pod in one Availability Zone, schedule and forward the request to proper pod, and build the tenant ID and pod binding relationship.
 
# Cinder APIGW and Nova APIGW will make sure the volumes for the same VM will co-locate in same OpenStack instance.
 
# work as stateless service, and could run with processes distributed in multi-hosts.
 
* Neutron API Server
 
# Neutron API Server is reused from Neutron to receive and handle Neutron API request.
 
# Neutron Tricircle Plugin. It runs under Neutron API server in the same process like OVN Neutron plugin. The Tricircle plugin serve for tenant level L2/L3 networking automation across multi-OpenStack instances. It will use driver interface to call L2GW api, especially for cross OpenStack mixed VLAN / VxLAN L2 networking.
 
* Admin API
 
# manage sites(bottom OpenStack instances) and availability zone mappings.
 
# Retrieve object uuid routing.
 
# Expose api for maintenance.
 
* XJob
 
# Receive and process cross OpenStack functionalities and other async jobs from Nova API-GW, or Cinder API-GW, Admin API or Neutron Tricircle Plugin. For example, when booting a VM for the first time for the project, router, security group rule, FIP and other resources may have not already been created in the bottom OpenStack instance. But it’s required. Not like network,security group, ssh keypair, other resources they must be created before a VM booting. These resources could be created in async way to accelerate response for the first VM booting request.
 
# Cross OpenStack networking also will be done in async jobs.
 
# Any of Admin API, Nova API-GW, Cinder API-GW, Neutron Tricircle plugin could send an async job to XJob through message bus with RPC API provided by XJob.
 
* Database
 
# The Tricircle has its own database to store pods, pod-bindings, jobs, resource routing tables.
 
# And Neutron Tricircle plugin reuse DB of Neutron, for one tenant’s network, router will be spread into multiple OpenStack instances, and managing tenant level IP/mac address to avoid conflict.
 
  
==FAQ==
+
<br />
  
'''Q: What is the different between Tricircle and OpenStack Cascading?''' <br />
+
[[File:Tricircle_architecture.png|frameless|center|x400px|Tricircle_architecture]]
 +
<br />
  
OpenStack Cascading was mainly an implementation method used in a PoC done in late 2014 and early 2015, which aims to test out the idea that multiple OpenStack instances '''COULD''' be deployed across multiple geo-diverse sites. After the PoC was carried out successfully, the team then planned to contribute the core idea to the community.
+
Shared KeyStone (centralized or distributed deployment) or federated KeyStones could be used for identity management for the Tricrcle and multiple OpenStack instances.
  
Tricircle Project was born out of that idea, however got a different shape and focus. Unlike what is usually part of in a PoC, which has plenty twists and plumbers of feature enhancements, Tricircle in its earliest stage tries to build a clean architecture that is extendable, pluggable and reusable in nature.
+
* Tricircle Local Neutron Plugin
 +
# It runs under Neutron server in the same process like OVN/Dragonflow Neutron plugin.
 +
# The Tricircle Local Neutron Plugin serve for cross Neutron networking automation triggering. It is a shim layer between real core plugin and Neutron API server
 +
# In the Tricircle context, The Tricircle Local Neutron Plugin could be called “local plugin” for short.
 +
# The Neutron server with the Tricircle Local Neutron Plugin installed is also called “local Neutron” or “local Neutron server”. Nova/Cinder which work together with local Neutron also could be called local Nova/local Cinder, compared to the central Neutron.  
  
In short, OpenStack Cascading is a specific deployment solution used for production purpose, while Tricircle represents an idea of one type of services, like Neutron or Murano, that in the future could be applied to OpenStack Ecosystem.
+
* Tricircle Central Neutron Plugin
 +
# It runs under Neutron server in the same process like OVN/Dragonflow Neutron plugin.
 +
# The Tricircle Central Neutron Plugin serve for tenant level L2/L3 networking automation across multi-OpenStack instances.
 +
# In the Tricircle context, The Tricircle Central Neutron Plugin could be called “central plugin” for short.
 +
# The Neutron server with the Tricircle Central Neutron Plugin installed is also called “central Neutron” or “central Neutron server”.
  
'''Q: What is the goal of Tricircle?'''<br />
+
* Admin API
 +
# Manage the mappings between OpenStack instances and availability zone.
 +
# Retrieve object uuid routing.
 +
# Expose API for maintenance.
  
In short term, Tricircle would focus on developing a robust architecture and related features, in a long run, we hope we could successfully establish a paradigm that could be applied to the whole OpenStack community
+
* XJob
 +
# Receive and process cross OpenStack functionalities and other async jobs from Admin API or Tricircle Central Neutron Plugin.
 +
# For example, when booting an instance for the first time for the project, router, security group rule, FIP and other resources may have not already been created in the bottom OpenStack instance, these resources could be created in async way to accelerate response for the first instance booting request. Not like network, subnet, security group resources they must be created before an instance booting.
 +
# Admin API, Tricircle Central Neutron plugin could send an async job to XJob through message bus with RPC API provided by XJob.
  
'''Q: How can I set up Tricircle hand by hand ?'''<br />
+
* Database
 +
# The Tricircle has its own database to store pods, jobs and resource routing tables for the Tricircle Central Neutron plugin, Admin API and XJob.
 +
# The Tricircle Central Neutron Plugin will also reuse database of central Neutron server to do the global management of tenant's resources like IP/MAC/network/router.
 +
# The Tricircle Local Neutron Plugin is a shim layer between real core plugin and Neutron API server, so Neutron DB will be still there for Neutron API server and real core plugin.
  
Yes, some volunteers sucessfully set up the Tricircle in 3 VMs with virtualbox in Ubuntu 14.04 LTS. The blog can be found in [http://shipengfei92.cn/play_tricircle_with_virtualbox this]
+
For Glance deployment, there are several choice:
 +
* Shared Glance, if all OpenStack instances are located inside a high bandwidth, low latency site.
 +
* Shared Glance with distributed back-end, if OpenStack instances are located in several sites.
 +
* Distributed Glance deployment, Glance service is deployed distributed in multiple site with distributed back-end
 +
* Separate Glance deployment, each site is installed with separate Glance instance and back-end, no cross site image sharing is needed.
  
Or refer to the README.md in https://github.com/openstack/tricircle for single node setup using devstack.
+
==Value==
 +
The motivation to develop the Tricircle open source project is to meet the demands which are required in the use cases mentioned above:
  
==How to read source code==
+
* Leverage Neutron API for cross Neutron networking automation, eco-system like CLI, SDK, Heat, Murano, Magum etc, all of these could be reserved seamlessly.
To read the source code, it's much easier if you follow this blueprint:
+
* support modularized capacity expansion in large scale cloud, just add more and more OpenStack instance, and these OpenStack instances are inter-connected in tenant level.
 +
* L2/L3 networking automation across OpenStack instances.
 +
* Tenant's VMs communicate with each other via L2 or L3 networking across OpenStack instances.
 +
* Security group applied across OpenStack instances.
 +
* Tenant level IP/mac addresses management to avoid conflict across OpenStack instances.
 +
* Tenant level quota control across OpenStack instances.
  
Implement Stateless Architecture: https://blueprints.launchpad.net/tricircle/+spec/implement-stateless
+
==Installation and Play==
 +
refer to the docuementation in https://docs.openstack.org/developer/tricircle/ for single node/multi-nodes setup and networking guide.
  
This blueprint is to build Tricircle from scratch
+
==How to read the source code==
 +
To read the source code, there is one guide: https://wiki.openstack.org/wiki/TricircleHowToReadCode
  
 
==Resources==
 
==Resources==
* Design documentation: [https://docs.google.com/document/d/18kZZ1snMOCD9IQvUKI5NVDzSASpw-QKj7l2zNqMEd3g/edit?usp=sharing Tricircle Design Blueprint]
+
* Design documentation: [https://docs.google.com/document/d/1zcxwl8xMEpxVCqLTce2-dUOtB-ObmzJTbV1uSQ6qTsY/ Tricircle Design Blueprint]
 
* Wiki: https://wiki.openstack.org/wiki/tricircle
 
* Wiki: https://wiki.openstack.org/wiki/tricircle
 +
* Documentation(Installation, Configuration, Networking guide): https://docs.openstack.org/tricircle/latest/
 
* Source: https://github.com/openstack/tricircle
 
* Source: https://github.com/openstack/tricircle
 
* Bugs: http://bugs.launchpad.net/tricircle
 
* Bugs: http://bugs.launchpad.net/tricircle
 
* Blueprints: https://launchpad.net/tricircle
 
* Blueprints: https://launchpad.net/tricircle
 
* Review Board: https://review.openstack.org/#/q/project:openstack/tricircle
 
* Review Board: https://review.openstack.org/#/q/project:openstack/tricircle
* Weekly meetings: https://wiki.openstack.org/wiki/Meetings/Tricircle at #openstack-meeting on every Wednesday starting from UTC 13:00
+
* Announcements: https://wiki.openstack.org/wiki/Meetings/Tricircle#Announcements
 +
* Weekly meeting IRC channel: #openstack-meeting, irc.freenode.net on every Wednesday starting from UTC 1:00 to UTC 2:00
 +
* Weekly meeting IRC log: https://wiki.openstack.org/wiki/Meetings/Tricircle#Meeting_minutes_and_logs
 +
* Tricircle project IRC channel: #openstack-tricircle on irc.freenode.net
 +
* Tricircle project IRC channel log: http://eavesdrop.openstack.org/irclogs/%23openstack-tricircle/
 
* Mail list:  openstack-dev@lists.openstack.org, with [openstack-dev][tricircle] in the mail subject
 
* Mail list:  openstack-dev@lists.openstack.org, with [openstack-dev][tricircle] in the mail subject
 
* New contributor's guide: http://docs.openstack.org/infra/manual/developers.html
 
* New contributor's guide: http://docs.openstack.org/infra/manual/developers.html
 
* Documentation: http://docs.openstack.org/developer/tricircle
 
* Documentation: http://docs.openstack.org/developer/tricircle
 +
* Tricircle discuss zone: https://wiki.openstack.org/wiki/TricircleDiscuzZone
 +
 +
* Tricircle big-tent application defense: https://review.openstack.org/#/c/338796 (A lots of comment and discussion to learn about Tricircle from many aspects)
  
 
Tricircle is designed to use the same tools for submission and review as other OpenStack projects.  As such we follow the [http://docs.openstack.org/infra/manual/developers.html#development-workflow OpenStack development workflow]. New contributors should follow the [http://docs.openstack.org/infra/manual/developers.html#getting-started getting started] steps before proceeding, as a Launchpad ID and signed contributor license are required to add new entries.
 
Tricircle is designed to use the same tools for submission and review as other OpenStack projects.  As such we follow the [http://docs.openstack.org/infra/manual/developers.html#development-workflow OpenStack development workflow]. New contributors should follow the [http://docs.openstack.org/infra/manual/developers.html#getting-started getting started] steps before proceeding, as a Launchpad ID and signed contributor license are required to add new entries.
 +
 +
==Video Resources==
 +
* Tricircle project update(Pike, OpenStack Sydney Summit), video: https://www.youtube.com/watch?v=baSu-eoUE1E
 +
* Tricircle project update(Pike, OpenStack Sydney Summit), slides: https://docs.google.com/presentation/d/1JlGaMPDvnv42QV5isUl7Y1JHuTjBmAqR5fGzIcizlvg
 +
* Move mission critical application to multi-site, what we learned: https://www.youtube.com/watch?v=l4Q2EoblDnY
 +
* Shared Networks to Support VNF High Availability Across OpenStack Multi Region Deployment  slides: https://docs.google.com/presentation/d/1WBdra-ZaiB-K8_m3Pv76o_jhylEqJXTTxzEZ-cu8u2A/ Video: https://www.youtube.com/watch?v=tbcc7-eZnkY
 +
* OPNFV Summit vIMS Multisite Demo(Youtube): https://www.youtube.com/watch?v=zS0wwPHmDWs
 +
* Video Conference High Availability across multiple OpenStack clouds(Youtube): https://www.youtube.com/watch?v=nK1nWnH45gI
 +
 +
==History==
 +
During the big-tent application of Tricircle: https://review.openstack.org/#/c/338796/, the proposal is to move API-gateway part away from Tricircle, and form two independent and decoupled projects:<br />
 +
 +
Tricircle: Dedicated for cross Neutron networking automation in multi-region OpenStack deployments, run without or with Trio2o. <br />
 +
 +
Trio2o: Dedicated to provide API gateway for those who need single Nova/Cinder API endpoint in multi-region OpenStack deployment, run without or with Tricircle.<br />
 +
 +
Splitting blueprint: https://blueprints.launchpad.net/tricircle/+spec/make-tricircle-dedicated-for-networking-automation-across-neutron
 +
 +
The wiki for Tricircle before splitting is linked here: https://wiki.openstack.org/wiki/tricircle_before_splitting
  
 
==Meeting minutes and logs==
 
==Meeting minutes and logs==
 
all meeting logs and minutes could be found in <br />
 
all meeting logs and minutes could be found in <br />
 +
2017: http://eavesdrop.openstack.org/meetings/tricircle/2017/
 +
<br />
 
2016: http://eavesdrop.openstack.org/meetings/tricircle/2016/
 
2016: http://eavesdrop.openstack.org/meetings/tricircle/2016/
 
<br />
 
<br />
Line 174: Line 173:
  
 
==To do list==
 
==To do list==
To do list is in the etherpad: https://etherpad.openstack.org/p/TricircleToDo
 
  
==Team Member==
+
Queens:<br />
Contact team members in IRC channel: #openstack-tricircle
+
# Queens Etherpad: https://etherpad.openstack.org/p/tricircle-queens-ptg
  
===Current active participants===
+
Pike:<br />
 +
# Pike Etherpad: https://etherpad.openstack.org/p/tricircle-pike-design-topics
 +
# Boston On-boarding Etherpad: https://etherpad.openstack.org/p/BOS-forum-tricircle-onboarding
  
Joe Huang, Huawei
+
Ocata:<br />
 +
# To do list in Ocata: https://etherpad.openstack.org/p/ocata-tricircle-work-session
  
Khayam Gondal, Dell
+
Newton:<br />
 +
# To do list is in the etherpad: https://etherpad.openstack.org/p/TricircleToDo<br />
 +
# Splitting Tricircle into two projects: https://etherpad.openstack.org/p/TricircleSplitting
  
Shinobu Kinjo, RedHat
+
==Team Member==
 
+
Contact team members in IRC channel: #openstack-tricircle
Ge Li, China UnionPay
 
 
 
Vega Cai, Huawei
 
 
 
Pengfei Shi, OMNI Lab
 
 
 
Bean Zhang, OMNI Lab
 
  
Yipei Niu, Huazhong University of Science and Technology
+
===Current active contributors===
  
Ronghui Chao, Hunan University
+
You can find active contributors from http://stackalytics.com:
  
Xiongqiu Long, Hunan University
+
Review: http://stackalytics.com/?release=all&module=tricircle-group&metric=marks<br />
  
Zhuo Tang, Hunan University
+
Code of lines: http://stackalytics.com/?release=all&module=tricircle-group&metric=loc<br />
  
Howard Huang, Huawei
+
Email: http://stackalytics.com/?release=all&module=tricircle-group&metric=emails<br />

Latest revision as of 06:18, 11 December 2017

The Tricircle is to provide networking automation across Neutron in multi-region OpenStack deployments.

Use Cases

To understand the motivation of the Tricircle project, we should understand the use cases with end user, and it would help you understand this project, please refer to presentation material: https://docs.google.com/presentation/d/1Zkoi4vMOGN713Vv_YO0GP6YLyjLpQ7fRbHlirpq6ZK4/

0. Telecom application cloud level redundancy in OPNFV Beijing Summit

Shared Networks to Support VNF High Availability Across OpenStack Multi Region Deployment This slides was presented in OPNFV Beijing Summit, 2017. slides: https://docs.google.com/presentation/d/1WBdra-ZaiB-K8_m3Pv76o_jhylEqJXTTxzEZ-cu8u2A/ Video: https://www.youtube.com/watch?v=tbcc7-eZnkY OPNFV Summit vIMS Multisite Demo(Youtube): https://www.youtube.com/watch?v=zS0wwPHmDWs Video Conference High Availability across multiple OpenStack clouds(Youtube): https://www.youtube.com/watch?v=nK1nWnH45gI

1. Application high availability running over cloud

In telecom and other industries(for example, MySQL Galera cluster), normally applications tend to be designed as Active-Active, Active-Passive or N-to-1 node configurations. Because those applications need to be available as much as possible. Once those applications are planned to be migrated to cloud infrastructure, active instance(s) need(s) to be deployed in one OpenStack instance first. After that, passive or active instances(s) will be deployed in another (other) OpenStack instance(s) in order for achieving 99.999% availability, if it’s required.

The reason why this deployment style is required is that in general cloud system only achieve 99.99% availability. [1] [2]

To achieve required high availability, design of network architecture (especially Layer 2 and Layer 3) needs across Neutron to be considered for application state replication or heartbeat.

Tricircle usecase2


The above picture shows an application with Galera DB cluster as backend which are geographically distributed into multiple OpenStack instances.

[1] https://aws.amazon.com/cn/ec2/sla/
[2] https://news.ycombinator.com/item?id=2470298

2. Dual ISPs Load Balancing for internet link

Deploy application in separate OpenStack instance with dual ISPs for internet link redundancy, load balancing, east-west traffic isolation for data/state replication is needed.

Tricircle usecase5


3. Isolation of East-West traffic

In financial industry, more than one OpenStack instances will be deployed, some OpenStack instance will be put in DMZ zone, and the other ones in trust zone, so that application or part of the application could be put in different level security zone. Although the tenant’s resources are provisioned in different OpenStack instances, east-west traffic among the tenant’s resources should be isolated.

Tricircle usecase3


Bandwidth sensitive, heavy load App like CAD modeling asked for the cloud close to the end user in distributed edge site for better user experience, multiple OpenStack instances will be deployed into edge sites, but the east-west communication with isolation between the tenant’s resources are also required.

Tricircle usecase4


Similar requirements could also be found in the ops session in OpenStack Barcelona summit "Control Plane Design (multi-region)" Line 25~26, 47~50: https://etherpad.openstack.org/p/BCN-ops-control-plane-design

4. Cross Neutron L2 network for NFV area

In NFV(Network Function Virtualization) area, network functions like router, NAT, LB etc will be virtualized, The cross Neutron L2 networking capability like IP address space management, IP allocation and L2 network segment global management provided by the Tricircle can help VNFs(virtualized network function) across site inter-connection. For example, vRouter1 in site1, but vRouter2 in site2, these two VNFs could be in one L2 network across site.

5. Cloud Capacity Expansion

When an OpenStack cloud is created and many resources has been provisioned on it, the capacity of one of the OpenStack deployments may not be enough. A new OpenStack deployment needs to be added to this cloud. But tenants still want to add VMs to existing network. And they don’t want to add a new network, router or whatever required resources.

Tricircle usecase1


Architecture

Now the Tricircle is dedicated for networking automation across Neutron in multi-region OpenStack deployments. The design blueprint has been developed with ongoing improvement in https://docs.google.com/document/d/1zcxwl8xMEpxVCqLTce2-dUOtB-ObmzJTbV1uSQ6qTsY/,

From the control plane view (cloud management view ), Tricircle is to make Neutron(s) in multi-region OpenStack clouds working as one cluster, and enable the creation of global network/router etc abstract networking resources across multiple OpenStack clouds. From the data plane view (end user resources view), all VMs(also could be bare metal servers or containers) are provisioned in different cloud but can be inter-connected via the global abstract networking resources, of course, with tenant level isolation.


Tricircle_view


The software design is as follows:


Tricircle_architecture


Shared KeyStone (centralized or distributed deployment) or federated KeyStones could be used for identity management for the Tricrcle and multiple OpenStack instances.

  • Tricircle Local Neutron Plugin
  1. It runs under Neutron server in the same process like OVN/Dragonflow Neutron plugin.
  2. The Tricircle Local Neutron Plugin serve for cross Neutron networking automation triggering. It is a shim layer between real core plugin and Neutron API server
  3. In the Tricircle context, The Tricircle Local Neutron Plugin could be called “local plugin” for short.
  4. The Neutron server with the Tricircle Local Neutron Plugin installed is also called “local Neutron” or “local Neutron server”. Nova/Cinder which work together with local Neutron also could be called local Nova/local Cinder, compared to the central Neutron.
  • Tricircle Central Neutron Plugin
  1. It runs under Neutron server in the same process like OVN/Dragonflow Neutron plugin.
  2. The Tricircle Central Neutron Plugin serve for tenant level L2/L3 networking automation across multi-OpenStack instances.
  3. In the Tricircle context, The Tricircle Central Neutron Plugin could be called “central plugin” for short.
  4. The Neutron server with the Tricircle Central Neutron Plugin installed is also called “central Neutron” or “central Neutron server”.
  • Admin API
  1. Manage the mappings between OpenStack instances and availability zone.
  2. Retrieve object uuid routing.
  3. Expose API for maintenance.
  • XJob
  1. Receive and process cross OpenStack functionalities and other async jobs from Admin API or Tricircle Central Neutron Plugin.
  2. For example, when booting an instance for the first time for the project, router, security group rule, FIP and other resources may have not already been created in the bottom OpenStack instance, these resources could be created in async way to accelerate response for the first instance booting request. Not like network, subnet, security group resources they must be created before an instance booting.
  3. Admin API, Tricircle Central Neutron plugin could send an async job to XJob through message bus with RPC API provided by XJob.
  • Database
  1. The Tricircle has its own database to store pods, jobs and resource routing tables for the Tricircle Central Neutron plugin, Admin API and XJob.
  2. The Tricircle Central Neutron Plugin will also reuse database of central Neutron server to do the global management of tenant's resources like IP/MAC/network/router.
  3. The Tricircle Local Neutron Plugin is a shim layer between real core plugin and Neutron API server, so Neutron DB will be still there for Neutron API server and real core plugin.

For Glance deployment, there are several choice:

  • Shared Glance, if all OpenStack instances are located inside a high bandwidth, low latency site.
  • Shared Glance with distributed back-end, if OpenStack instances are located in several sites.
  • Distributed Glance deployment, Glance service is deployed distributed in multiple site with distributed back-end
  • Separate Glance deployment, each site is installed with separate Glance instance and back-end, no cross site image sharing is needed.

Value

The motivation to develop the Tricircle open source project is to meet the demands which are required in the use cases mentioned above:

  • Leverage Neutron API for cross Neutron networking automation, eco-system like CLI, SDK, Heat, Murano, Magum etc, all of these could be reserved seamlessly.
  • support modularized capacity expansion in large scale cloud, just add more and more OpenStack instance, and these OpenStack instances are inter-connected in tenant level.
  • L2/L3 networking automation across OpenStack instances.
  • Tenant's VMs communicate with each other via L2 or L3 networking across OpenStack instances.
  • Security group applied across OpenStack instances.
  • Tenant level IP/mac addresses management to avoid conflict across OpenStack instances.
  • Tenant level quota control across OpenStack instances.

Installation and Play

refer to the docuementation in https://docs.openstack.org/developer/tricircle/ for single node/multi-nodes setup and networking guide.

How to read the source code

To read the source code, there is one guide: https://wiki.openstack.org/wiki/TricircleHowToReadCode

Resources

Tricircle is designed to use the same tools for submission and review as other OpenStack projects. As such we follow the OpenStack development workflow. New contributors should follow the getting started steps before proceeding, as a Launchpad ID and signed contributor license are required to add new entries.

Video Resources

History

During the big-tent application of Tricircle: https://review.openstack.org/#/c/338796/, the proposal is to move API-gateway part away from Tricircle, and form two independent and decoupled projects:

Tricircle: Dedicated for cross Neutron networking automation in multi-region OpenStack deployments, run without or with Trio2o.

Trio2o: Dedicated to provide API gateway for those who need single Nova/Cinder API endpoint in multi-region OpenStack deployment, run without or with Tricircle.

Splitting blueprint: https://blueprints.launchpad.net/tricircle/+spec/make-tricircle-dedicated-for-networking-automation-across-neutron

The wiki for Tricircle before splitting is linked here: https://wiki.openstack.org/wiki/tricircle_before_splitting

Meeting minutes and logs

all meeting logs and minutes could be found in
2017: http://eavesdrop.openstack.org/meetings/tricircle/2017/
2016: http://eavesdrop.openstack.org/meetings/tricircle/2016/
2015: http://eavesdrop.openstack.org/meetings/tricircle/2015/

To do list

Queens:

  1. Queens Etherpad: https://etherpad.openstack.org/p/tricircle-queens-ptg

Pike:

  1. Pike Etherpad: https://etherpad.openstack.org/p/tricircle-pike-design-topics
  2. Boston On-boarding Etherpad: https://etherpad.openstack.org/p/BOS-forum-tricircle-onboarding

Ocata:

  1. To do list in Ocata: https://etherpad.openstack.org/p/ocata-tricircle-work-session

Newton:

  1. To do list is in the etherpad: https://etherpad.openstack.org/p/TricircleToDo
  2. Splitting Tricircle into two projects: https://etherpad.openstack.org/p/TricircleSplitting

Team Member

Contact team members in IRC channel: #openstack-tricircle

Current active contributors

You can find active contributors from http://stackalytics.com:

Review: http://stackalytics.com/?release=all&module=tricircle-group&metric=marks

Code of lines: http://stackalytics.com/?release=all&module=tricircle-group&metric=loc

Email: http://stackalytics.com/?release=all&module=tricircle-group&metric=emails