[Openstack] [keystone] memcache token backend performance
Jay Pipes
jaypipes at gmail.com
Mon Jan 6 16:32:13 UTC 2014
On Mon, 2014-01-06 at 10:10 -0500, Adam Young wrote:
>
> On 01/03/2014 11:38 PM, Xu (Simon) Chen wrote:
>
> > Hi folks,
> >
> >
> > I am having trouble with using memcache as the keystone token
> > backend. I have three keystone nodes running active/active. Each is
> > running keystone on apache (for kerberos auth). I recently switched
> > from using sql backend to memcache, while have memcached running on
> > all three of the keystone nodes.
> >
>
> This triggers a memory of htere being something wonky with
> Greenthreads, the threading override in Eventlet, and Memcached. But
> you said Apache, so I think that you are not running with
> greenthreads?
>
> There are numerous things out there about apache and memcached
> performance. For example, one article talks about filling up
> partitions, etc.
>
>
> >
> >
> > This setup would run well for a while, but then apache would start
> > to hog CPUs, and memcached would increase to 30% or so. I tried to
> > increase memcached cluster from 3 to 6 nodes, but in general the
> > performance is much worse compared to sql backend.
> Probably due to the need for replication. In order to keep it
> anywhere close to in sync, it is going to require some non-trivial
> subset of fully connectedness.
Instead of doing replication of memcache, instead just tell your
loadbalancer to have sticky sessions, and have each keystone server have
its own dedicated memcache instance.
-jay
More information about the Openstack
mailing list