[openstack-dev] [QA][gate][all] dsvm gate stability and scenario tests
Jordan Pittier
jordan.pittier at scality.com
Fri Mar 17 15:23:11 UTC 2017
On Fri, Mar 17, 2017 at 3:11 PM, Sean Dague <sean at dague.net> wrote:
> On 03/17/2017 09:24 AM, Jordan Pittier wrote:
> >
> >
> > On Fri, Mar 17, 2017 at 1:58 PM, Sean Dague <sean at dague.net
> > <mailto:sean at dague.net>> wrote:
> >
> > On 03/17/2017 08:27 AM, Jordan Pittier wrote:
> > > The patch that reduced the number of Tempest Scenarios we run in
> every
> > > job and also reduce the test run concurrency [0] was merged 13
> days ago.
> > > Since, the situation (i.e the high number of false negative job
> results)
> > > has not improved significantly. We need to keep looking
> collectively at
> > > this.
> >
> > While the situation hasn't completely cleared out -
> > http://tinyurl.com/mdmdxlk - since we've merged this we've not seen
> that
> > job go over 25% failure rate in the gate, which it was regularly
> > crossing in the prior 2 week period. That does feel like progress. In
> > spot checking I we are also rarely failing in scenario tests now, but
> > the fails tend to end up inside heavy API tests running in parallel.
> >
> >
> > > There seems to be an agreement that we are hitting some memory
> limit.
> > > Several of our most frequent failures are memory related [1]. So we
> > > should either reduce our memory usage or ask for bigger VMs, with
> more
> > > than 8GB of RAM.
> > >
> > > There was/is several attempts to reduce our memory usage, by
> reducing
> > > the Mysql memory consumption ([2] but quickly reverted [3]),
> reducing
> > > the number of Apache workers ([4], [5]), more apache2 tuning [6].
> If you
> > > have any crazy idea to help in this regard, please help. This is
> high
> > > priority for the whole openstack project, because it's plaguing
> many
> > > projects.
> >
> > Interesting, I hadn't seen the revert. It is also curious that it was
> > largely limitted to the neutron-api test job. It's also notable that
> the
> > sort buffers seem to have been set to the minimum allowed limit of
> mysql
> > -
> > https://dev.mysql.com/doc/refman/5.6/en/innodb-parameters.
> html#sysvar_innodb_sort_buffer_size
> > <https://dev.mysql.com/doc/refman/5.6/en/innodb-parameters.
> html#sysvar_innodb_sort_buffer_size>
> > - and is over an order of magnitude decrease from the existing
> default.
> >
> > I wonder about redoing the change with everything except it and
> seeing
> > how that impacts the neutron-api job.
> >
> > Yes, that would be great because mysql is by far our biggest memory
> > consumer so we should target this first.
>
> While it is the single biggest process, weighing in at 500 MB, the
> python services are really our biggest memory consumers. They are
> collectively far outweighing either mysql or rabbit, and are the reason
> that even with 64MB guests we're running out of memory. So we want to
> keep that under perspective.
>
Absolutely. I have https://review.openstack.org/#/c/446986/ in that vain.
And if someone wants to start the work of not running the several Swift
*auditor*, *updater*, *reaper*, *replicator* services, in case the Swift
Replication factor is set to 1, that's also a good memory saving.
>
> -Sean
>
> --
> Sean Dague
> http://dague.net
>
> __________________________________________________________________________
> OpenStack Development Mailing List (not for usage questions)
> Unsubscribe: OpenStack-dev-request at lists.openstack.org?subject:unsubscribe
> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.openstack.org/pipermail/openstack-dev/attachments/20170317/abff9aec/attachment.html>
More information about the OpenStack-dev
mailing list