[openstack-dev] [magnum] 2 million requests / sec, 100s of nodes
Ton Ngo
ton at us.ibm.com
Fri Aug 5 15:11:22 UTC 2016
Hi Ricardo,
For your question 1, you can modify the Heat template to not create
the Cinder volume and tweak the call to
configure-docker-storage.sh to use local storage. It should be fairly
straightforward. You just need to make
sure the local storage of the flavor is sufficient to host the containers
in the benchmark.
If you think this is a common scenario, we can open a blueprint for
this option.
Ton,
From: Ricardo Rocha <rocha.porto at gmail.com>
To: "OpenStack Development Mailing List (not for usage questions)"
<openstack-dev at lists.openstack.org>
Date: 08/05/2016 04:51 AM
Subject: Re: [openstack-dev] [magnum] 2 million requests / sec, 100s of
nodes
Hi.
Quick update is 1000 nodes and 7 million reqs/sec :) - and the number of
requests should be higher but we had some internal issues. We have a
submission for barcelona to provide a lot more details.
But a couple questions came during the exercise:
1. Do we really need a volume in the VMs? On large clusters this is a
burden, and local storage only should be enough?
2. We observe a significant delay (~10min, which is half the total time to
deploy the cluster) on heat when it seems to be crunching the kube_minions
nested stacks. Once it's done, it still adds new stacks gradually, so it
doesn't look like it precomputed all the info in advance
Anyone tried to scale Heat to stacks this size? We end up with a stack
with:
* 1000 nested stacks (depth 2)
* 22000 resources
* 47008 events
And already changed most of the timeout/retrial values for rpc to get this
working.
This delay is already visible in clusters of 512 nodes, but 40% of the time
in 1000 nodes seems like something we could improve. Any hints on Heat
configuration optimizations for large stacks very welcome.
Cheers,
Ricardo
On Sun, Jun 19, 2016 at 10:59 PM, Brad Topol <btopol at us.ibm.com> wrote:
Thanks Ricardo! This is very exciting progress!
--Brad
Brad Topol, Ph.D.
IBM Distinguished Engineer
OpenStack
(919) 543-0646
Internet: btopol at us.ibm.com
Assistant: Kendra Witherspoon (919) 254-0680
Inactive hide details for Ton Ngo---06/17/2016 12:10:33 PM---Thanks
Ricardo for sharing the data, this is really encouraging! TTon
Ngo---06/17/2016 12:10:33 PM---Thanks Ricardo for sharing the data, this
is really encouraging! Ton,
From: Ton Ngo/Watson/IBM at IBMUS
To: "OpenStack Development Mailing List \(not for usage questions\)" <
openstack-dev at lists.openstack.org>
Date: 06/17/2016 12:10 PM
Subject: Re: [openstack-dev] [magnum] 2 million requests / sec, 100s of
nodes
Thanks Ricardo for sharing the data, this is really encouraging!
Ton,
Inactive hide details for Ricardo Rocha ---06/17/2016 08:16:15 AM---Hi.
Just thought the Magnum team would be happy to hear :)Ricardo Rocha
---06/17/2016 08:16:15 AM---Hi. Just thought the Magnum team would be
happy to hear :)
From: Ricardo Rocha <rocha.porto at gmail.com>
To: "OpenStack Development Mailing List (not for usage questions)" <
openstack-dev at lists.openstack.org>
Date: 06/17/2016 08:16 AM
Subject: [openstack-dev] [magnum] 2 million requests / sec, 100s of nodes
Hi.
Just thought the Magnum team would be happy to hear :)
We had access to some hardware the last couple days, and tried some
tests with Magnum and Kubernetes - following an original blog post
from the kubernetes team.
Got a 200 node kubernetes bay (800 cores) reaching 2 million requests /
sec.
Check here for some details:
https://openstack-in-production.blogspot.ch/2016/06/scaling-magnum-and-kubernetes-2-million.html
We'll try bigger in a couple weeks, also using the Rally work from
Winnie, Ton and Spyros to see where it breaks. Already identified a
couple issues, will add bugs or push patches for those. If you have
ideas or suggestions for the next tests let us know.
Magnum is looking pretty good!
Cheers,
Ricardo
__________________________________________________________________________
OpenStack Development Mailing List (not for usage questions)
Unsubscribe:
OpenStack-dev-request at lists.openstack.org?subject:unsubscribe
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
__________________________________________________________________________
OpenStack Development Mailing List (not for usage questions)
Unsubscribe:
OpenStack-dev-request at lists.openstack.org?subject:unsubscribe
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
__________________________________________________________________________
OpenStack Development Mailing List (not for usage questions)
Unsubscribe:
OpenStack-dev-request at lists.openstack.org?subject:unsubscribe
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
__________________________________________________________________________
OpenStack Development Mailing List (not for usage questions)
Unsubscribe: OpenStack-dev-request at lists.openstack.org?subject:unsubscribe
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.openstack.org/pipermail/openstack-dev/attachments/20160805/cc42cd5d/attachment.html>
-------------- next part --------------
A non-text attachment was scrubbed...
Name: graycol.gif
Type: image/gif
Size: 105 bytes
Desc: not available
URL: <http://lists.openstack.org/pipermail/openstack-dev/attachments/20160805/cc42cd5d/attachment.gif>
More information about the OpenStack-dev
mailing list