[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

[Octavia]-Seeking performance numbers on Octavia

Thanks so much for the valuable insights Michael! Appreciate it and keep up the good work, as I ramp up with more dev know how hopefully I would start making contributions and can maybe convince my team to start as well.

Thanks & Regards

Prabhjit Singh 

-----Original Message-----
From: Michael Johnson <johnsomor at gmail.com> 
Sent: Monday, July 22, 2019 5:48 PM
To: Singh, Prabhjit <Prabhjit.Singh22 at T-Mobile.com>
Cc: openstack-discuss at lists.openstack.org
Subject: Re: [Octavia]-Seeking performance numbers on Octavia


Hi Prabhjit,

Comments in-line below.


On Sun, Jul 21, 2019 at 5:24 PM Singh, Prabhjit <Prabhjit.Singh22 at t-mobile.com> wrote:
> Hi Michael,
> Thanks for taking the time out to send me your inputs and valuable suggestions. I do remember meeting you at the Denver Summit and hearing to a couple of your sessions.
> If you wouldn't mind, I do have a few more questions and your answers would help me understand that should I continue to invest in having Octavia as one of our available LBs.
> 1. Based on your response and the amount of time you are investing in 
> supporting Octavia, what are some of the use cases, like for e.g. if load balancing web traffic how many transactions/connections minimum can be expected. I do understand you mentioned that it's hard to performance test Octavia but some real time situations from your testing and how customers have adopted Octavia would help me level set some expectations.

This is really cloud and application specific. I would recommend you fire up an Octavia install and use your preferred tool to measure it.
Some good tools are tsung, weighttp, and iperf3.

> 2. We are thinking of Octavia as one of the offerings, that offers a self-serve type model. Do you know of any customers who have been able to use Octavia as one of their primary load balancers and any encouraging feedback you have gotten on Octavia.

There are examples of organizations using Octavia available if you google Octavia.

> 3.  You suggested increasing the Ram size, I could go about making a whole new Flavor.

Yes, to increase the allocated RAM for a load balancer, you would create an additional nova flavor with the specifications you would like. You can then either set this as the default nova flavor for amphora (amp_flavor_id is the setting) or you can create an Octavia flavor that specifies the nova compute flavor to use (See
https://nam02.safelinks.protection.outlook.com/?url=https%3A%2F%2Fdocs.openstack.org%2Foctavia%2Flatest%2Fadmin%2Fflavors.html&amp;data=02%7C01%7CPrabhjit.Singh22%40t-mobile.com%7Cfb41388d6020453d92c908d70eee4a72%7Cbe0f980bdd994b19bd7bbc71a09b026c%7C0%7C0%7C636994288931593870&amp;sdata=FDlAK3%2FKh0DNo%2BMSJQ8kJ8lSnn01TJXASS6AHd1kRoA%3D&amp;reserved=0 for more information on Octavia flavors).

> 4. I also noticed on the haproxy.conf the maxconns is set to 2000, should I increase this, does this affect the connection per server, which you said 64000 conns per server, so if I have 10 servers can I expect somewhere close to 640000 sessions?

I think you are looking at the haproxy.conf file provided by your operating system package. Octavia does not use this file, it creates it's own HAProxy configuration files as needed under /var/lib/octavia inside the amphora. The default, if the user does not specify one at listener creation, is 1,000,000.

> 5. Based on some of the limitations and the dev work in progress, I think the most important feature that would make Octavia a real solid offering would be the Active-Active and Autoscaling feature. I brought this up with you in our brief conversation at the summit, and you did mention that its not a top priority at this time and you are looking for some help. I have noticed a lot of documentation has been updated on this feature, do you think with the available document and progress I could spin up a distributor and manage sessions between Amphora or it's not complete yet.

Active/Active is still on our roadmap, but unfortunately the people that were working on it had to stop for personal reasons. There may be some folks picking up this work again soon. At this point the Active/Active patches up for review are non-functional and still a work in progress.

> 6. We have a Triple O setup, do you think I can make the above tweaks with the Triple O setup.

I think you are able to make various adjustments to Octavia with Triple O, but I do not have specifics on that.

> Thanks & Regards
> Prabhjit Singh
> Systems Design and Strategy - Magentabox
> | O: (973) 397-4819 | M: (973) 563-4445
> -----Original Message-----
> From: Michael Johnson <johnsomor at gmail.com>
> Sent: Friday, July 19, 2019 6:00 PM
> To: Singh, Prabhjit <Prabhjit.Singh22 at T-Mobile.com>
> Cc: openstack-discuss at lists.openstack.org
> Subject: Re: [Octavia]-Seeking performance numbers on Octavia
> [External]
> Hi Prabhjit,
> As you have mentioned, it is very challenging to get accurate performance results in cloud environments. There are a large number(very large in fact) of factors that can impact the overall performance of OpenStack and Octavia.
> In our OpenDev testing environment, we only have software emulation virtual machines available (Qemu running with the TCG engine) which performs extremely poorly. This means that the testing environment does not reflect how the software is used in real world deployments.
> An example of this is simply booting a VM can take up to ten minutes on Qemu with TCG when it takes about twenty seconds on a real OpenStack deployment.
> With this resource limitation, we cannot effectively run performance benchmarking test jobs on the OpenDev environment.
> Because of this, we don't publish performance numbers as they will not reflect what you can achieve in your environment.
> Let me try to speak to your bullet points:
> 1. The Octavia team has never (to my knowledge) claimed the Amphora driver is "carrier grade". We do consider the Amphora driver to be "operator grade", which speaks to a cloud operator's perspective versus the previous offering that did not support high availability, have appropriate maintenance tooling, upgrade paths, performance, etc.
> To me, "carrier grade" has an additional level of requirements including performance, latency, scale, and availability SLAs. This is not what the Octavia Amphora driver is currently ready for. That said, third party provider drivers for Octavia may be able to provide a "carrier grade" level of load balancing for OpenStack.
> 2. As for performance tuning, much of this is either automatically handled by Octavia or are dependent on the application you are load balancing and your cloud deployment. For example we have many configuration settings to tune how many retries we attempt when interacting with other services. In performing and stable clouds, these can be tuned down, in others the defaults may be appropriate. If you would like faster failover, at the expense of slightly more network traffic, you can tune the health monitoring and keepalived_vrrp settings. We do not currently have a performance tuning guide for Octavia but would support someone authoring one.
> 3. We do not currently have a guide for this. I will say with the version of HAproxy currently being shipped with the distributions, going beyond the 1vCPU per amphora does not gain you much. With the release of HAProxy 2.0 this has changed and we expect to be adding support for vertically scaling the Amphora in future releases. Disk space is only necessary if you are storing the flow logs locally, which I would not recommend for a performance load balancer (See the notes in the log offloading guide:
> https://nam02.safelinks.protection.outlook.com/?url=https%3A%2F%2Fdocs.openstack.org%2Foctavia%2Flatest%2Fadmin%2Flog-offloading.html&amp;data=02%7C01%7CPrabhjit.Singh22%40t-mobile.com%7Cfb41388d6020453d92c908d70eee4a72%7Cbe0f980bdd994b19bd7bbc71a09b026c%7C0%7C0%7C636994288931593870&amp;sdata=qyX1BM6wR6v804WCYB2HY6IRmDfeQS1zi38FS34kB1U%3D&amp;reserved=0).
> Finally, the RAM usage is a factor of the number of concurrent connections and if you are enabling TLS on the load balancer. For typical load balancing loads, the default is typically fine. However, if you have high connection counts and/or TLS offloading, you may want to experiment with increasing the available RAM.
> 4. The source IP issue is a known issue 
> (https://nam02.safelinks.protection.outlook.com/?url=https%3A%2F%2Fstoryboard.openstack.org%2F%23!%2Fstory%2F1629066&amp;data=02%7C01%7CPrabhjit.Singh22%40t-mobile.com%7Cfb41388d6020453d92c908d70eee4a72%7Cbe0f980bdd994b19bd7bbc71a09b026c%7C0%7C0%7C636994288931593870&amp;sdata=GkTPXRmOfjpMYXDYZ9t5xH1aEq0E%2BWDZRhK8ux%2FnrUQ%3D&amp;reserved=0). We have not prioritized addressing this as we have not had anyone come forward that they needed this in their deployment. If this is an issue impacting your use case, please comment on the story to that effect and provide a use case. This will help the team prioritize this work.
> Also, patches are welcome! If you are interested in working on this issue, I can help you with information about how this could be added.
> It should also be noted that it is a limitation of 64,000 connections per-backend server, not per load balancer.
> 5. The team uses the #openstack-lbaas IRC channel on freenode and is happy to answer questions, etc.
> To date, we have had limited resources (people and equipment) available to do performance evaluation and tuning. There are definitely kernel and HAProxy tuning settings we have evaluated and added to the Amphora driver, but I know there is more work that can be done. If you are interested in help us with this work, please let us know.
> Michael
> P.S. Here are just a few considerations that can/will impact the performance of an Octavia Amphora load balancer:
> Hardware used for the compute nodes
> Network Interface Cards (NICs) used in the compute nodes Number of 
> network ports enabled on the compute hosts Network switch 
> configurations (Jumbo frames, and so on) Cloud network topology 
> (leafâ??spine, fatâ??tree, and so on) The OpenStack Neutron networking 
> configuration (ML2 and ML3 drivers) Tenant networking configuration 
> (VXLAN, VLANS, GRE, and so on) Colocation of applications and Octavia 
> amphorae Over subscription of the compute and networking resources 
> Protocols being load balanced Configuration settings used when 
> creating the load balancer (connection limits, and so on) Version of 
> OpenStack services (nova, neutron, and so on) Version of OpenStack 
> Octavia Flavor of the OpenStack Octavia load balancer OS and 
> hypervisor versions used Deployed security mitigations (Spectre, 
> Meltdown, and so on) Customer application performance Health of the 
> customer application
> On Fri, Jul 19, 2019 at 8:52 AM Singh, Prabhjit <Prabhjit.Singh22 at t-mobile.com> wrote:
> >
> > Hi
> >
> >
> >
> > I have been trying to test Octavia with some traffic generators and 
> > my tests are inconclusive. Appreciate your inputs on the following
> >
> >
> >
> > It would be really nice to have some performance numbers that you guys have been able to achieve for this to be termed as carrier grade.
> > Would also appreciate if you could share any inputs on performance 
> > tuning Octavia Any recommended flavor sizes for spinning up Amphorae, the default size of 1 core, 2 Gb disk and 1 Gig RAM does not seem enough.
> > Also I noticed when the Amphorae are spun up, at one time only one 
> > master is talking to the backend servers and has one IP that its 
> > using, it has to run out of ports after 64000 TCP concurrent 
> > sessions, id there a way to add more IPs or is this the limitation 
> > If I needed some help with Octavia and some guidance around 
> > performance tuning can someone from the community help
> >
> >
> >
> > Thanks & Regards
> >
> >
> >
> > Prabhjit Singh
> >
> >
> >
> >
> >
> >