[openstack-dev] [Magnum] Consistent functional test failures
Steven Dake (stdake)
stdake at cisco.com
Fri Aug 21 01:14:53 UTC 2015
Kai,
This sounds like a good solution. The actual VM doesn’t need to be super large given our present tests.
Regards
-steve
From: Kai Qiang Wu <wkqwu at cn.ibm.com<mailto:wkqwu at cn.ibm.com>>
Reply-To: "OpenStack Development Mailing List (not for usage questions)" <openstack-dev at lists.openstack.org<mailto:openstack-dev at lists.openstack.org>>
Date: Friday, August 14, 2015 at 3:46 AM
To: "OpenStack Development Mailing List (not for usage questions)" <openstack-dev at lists.openstack.org<mailto:openstack-dev at lists.openstack.org>>
Subject: Re: [openstack-dev] [Magnum] Consistent functional test failures
I have checked with infra team members. For two instances, 10GB each should be OK.
So I add some steps to create magnum specific flavor(8 GB disk), instead of use the existed devstack flavors (m1.small needs 20GB, m1.tiny can not be used)
Magnum creates one for jenkins job and delete it when tests finished.
Thanks
Best Wishes,
--------------------------------------------------------------------------------
Kai Qiang Wu (吴开强 Kennan)
IBM China System and Technology Lab, Beijing
E-mail: wkqwu at cn.ibm.com<mailto:wkqwu at cn.ibm.com>
Tel: 86-10-82451647
Address: Building 28(Ring Building), ZhongGuanCun Software Park,
No.8 Dong Bei Wang West Road, Haidian District Beijing P.R.China 100193
--------------------------------------------------------------------------------
Follow your heart. You are miracle!
[Inactive hide details for Clark Boylan ---08/14/2015 08:05:15 AM---On Thu, Aug 13, 2015, at 03:13 AM, Tom Cammann wrote: > Hi T]Clark Boylan ---08/14/2015 08:05:15 AM---On Thu, Aug 13, 2015, at 03:13 AM, Tom Cammann wrote: > Hi Team,
From: Clark Boylan <cboylan at sapwetik.org<mailto:cboylan at sapwetik.org>>
To: openstack-dev at lists.openstack.org<mailto:openstack-dev at lists.openstack.org>
Date: 08/14/2015 08:05 AM
Subject: Re: [openstack-dev] [Magnum] Consistent functional test failures
________________________________
On Thu, Aug 13, 2015, at 03:13 AM, Tom Cammann wrote:
> Hi Team,
>
> Wanted to let you know why we are having consistent functional test
> failures in the gate.
>
> This is being caused by Nova returning "No valid host" to heat:
>
> 2015-08-13 08:26:16.303 31543 INFO heat.engine.resource [-] CREATE:
> Server "kube_minion" [12ab45ef-0177-4118-9ba0-3fffbc3c1d1a] Stack
> "testbay-y366b2atg6mm-kube_minions-cdlfyvhaximr-0-dufsjliqfoet"
> [b40f0c9f-cb54-4d75-86c3-8a9f347a27a6]
> 2015-08-13 08:26:16.303 31543 ERROR heat.engine.resource Traceback (most
> recent call last):
> 2015-08-13 08:26:16.303 31543 ERROR heat.engine.resource File
> "/opt/stack/new/heat/heat/engine/resource.py", line 625, in
> _action_recorder
> 2015-08-13 08:26:16.303 31543 ERROR heat.engine.resource yield
> 2015-08-13 08:26:16.303 31543 ERROR heat.engine.resource File
> "/opt/stack/new/heat/heat/engine/resource.py", line 696, in _do_action
> 2015-08-13 08:26:16.303 31543 ERROR heat.engine.resource yield
> self.action_handler_task(action, args=handler_args)
> 2015-08-13 08:26:16.303 31543 ERROR heat.engine.resource File
> "/opt/stack/new/heat/heat/engine/scheduler.py", line 320, in wrapper
> 2015-08-13 08:26:16.303 31543 ERROR heat.engine.resource step =
> next(subtask)
> 2015-08-13 08:26:16.303 31543 ERROR heat.engine.resource File
> "/opt/stack/new/heat/heat/engine/resource.py", line 670, in
> action_handler_task
> 2015-08-13 08:26:16.303 31543 ERROR heat.engine.resource while not
> check(handler_data):
> 2015-08-13 08:26:16.303 31543 ERROR heat.engine.resource File
> "/opt/stack/new/heat/heat/engine/resources/openstack/nova/server.py",
> line 759, in check_create_complete
> 2015-08-13 08:26:16.303 31543 ERROR heat.engine.resource return
> self.client_plugin()._check_active(server_id)
> 2015-08-13 08:26:16.303 31543 ERROR heat.engine.resource File
> "/opt/stack/new/heat/heat/engine/clients/os/nova.py", line 232, in
> _check_active
> 2015-08-13 08:26:16.303 31543 ERROR heat.engine.resource 'code':
> fault.get('code', _('Unknown'))
> 2015-08-13 08:26:16.303 31543 ERROR heat.engine.resource
> ResourceInError: Went to status ERROR due to "Message: No valid host was
> found. There are not enough hosts available., Code: 500"
>
> And this in turn is being caused by the compute instance running out of
> disk space:
>
> 2015-08-13 08:26:15.216 DEBUG nova.filters
> [req-e5bb52cb-387e-4638-911e-8c72aa1b6400 admin admin] Starting with 1
> host(s) get_filtered_objects /opt/stack/new/nova/nova/filters.py:70
> 2015-08-13 08:26:15.217 DEBUG nova.filters
> [req-e5bb52cb-387e-4638-911e-8c72aa1b6400 admin admin] Filter
> RetryFilter returned 1 host(s) get_filtered_objects
> /opt/stack/new/nova/nova/filters.py:84
> 2015-08-13 08:26:15.217 DEBUG nova.filters
> [req-e5bb52cb-387e-4638-911e-8c72aa1b6400 admin admin] Filter
> AvailabilityZoneFilter returned 1 host(s) get_filtered_objects
> /opt/stack/new/nova/nova/filters.py:84
> 2015-08-13 08:26:15.217 DEBUG nova.filters
> [req-e5bb52cb-387e-4638-911e-8c72aa1b6400 admin admin] Filter RamFilter
> returned 1 host(s) get_filtered_objects
> /opt/stack/new/nova/nova/filters.py:84
> 2015-08-13 08:26:15.218 DEBUG nova.scheduler.filters.disk_filter
> [req-e5bb52cb-387e-4638-911e-8c72aa1b6400 admin admin]
> (devstack-trusty-rax-dfw-4299602, devstack-trusty-rax-dfw-4299602)
> ram:5172 disk:17408 io_ops:0 instances:1 does not have 20480 MB usable
> disk, it only has 17408.0 MB usable disk. host_passes
> /opt/stack/new/nova/nova/scheduler/filters/disk_filter.py:60
> 2015-08-13 08:26:15.218 INFO nova.filters
> [req-e5bb52cb-387e-4638-911e-8c72aa1b6400 admin admin] Filter DiskFilter
> returned 0 hosts
>
> For now a recheck seems to work about 1 in 2, so we can still land
> patches.
>
> The fix for this could be to clean up our Magnum devstack install more
> aggressively, which might be as simple as cleaning up the images we use,
> or get infra to provide our tests with a larger disk size. I will
> probably test out a patch today which cleans up the images we use in
> devstack to see if that helps.
>
It is not trivial to provide your tests with more disk as we are using
the flavors appropriate for our RAM and CPU needs and are constrained by
quotas in the clouds we use. Do you really need 20GB nested test
instances? The VMs these jobs run on have ~13GB images which is almost
half the size of the instances you are trying to boot there. I would
definitely look into trimming the disk requirements for the nested VMs
before anything else.
As for working ~50% of the time hpcloud gives us more disk than
rackspace which is likely why you see about half fail and half pass. The
runs that pass probably run on hpcloud VMs.
Clark
__________________________________________________________________________
OpenStack Development Mailing List (not for usage questions)
Unsubscribe: OpenStack-dev-request at lists.openstack.org<mailto:OpenStack-dev-request at lists.openstack.org>?subject:unsubscribe
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.openstack.org/pipermail/openstack-dev/attachments/20150821/39142f7f/attachment.html>
-------------- next part --------------
A non-text attachment was scrubbed...
Name: graycol.gif
Type: image/gif
Size: 105 bytes
Desc: graycol.gif
URL: <http://lists.openstack.org/pipermail/openstack-dev/attachments/20150821/39142f7f/attachment.gif>
More information about the OpenStack-dev
mailing list