[openstack-dev] Reply: [Neutron][IPv6] tox run forever

Shixiong Shang sparkofwisdom.cloud at gmail.com
Fri Feb 28 14:20:04 UTC 2014


I started thinking whether passing the test of Mr. Jenkins is a mindless dreaming….:)  What does the “TOX” say? :)

Shixiong



Begin forwarded message:

> From: Shixiong Shang <sparkofwisdom.cloud at gmail.com>
> Subject: Re: [openstack-dev] [Neutron] tox run forever
> Date: February 27, 2014 at 9:41:34 PM EST
> To: "OpenStack Development Mailing List (not for usage questions)" <openstack-dev at lists.openstack.org>
> 
> Hi, Clark:
> 
> Thanks a lot for the prompt response! I added the OS_TEST_TIMEOUT value (300 sec) and was tailing the tmp file. It turned out that the TOX run stopped at the following point. My machine was tossed so badly that it became unresponsive and I had to hard reboot it….I am pulling my teeth off now…..Is it normal to see Traceback?
> 
> 2014-02-27 21:33:51,212     INFO [neutron.api.extensions] Extension 'agent' provides no backward compatibility map for extended attributes
> 2014-02-27 21:33:51,212     INFO [neutron.api.extensions] Extension 'Allowed Address Pairs' provides no backward compatibility map for extended attributes
> 2014-02-27 21:33:51,212     INFO [neutron.api.extensions] Extension 'Neutron Extra Route' provides no backward compatibility map for extended attributes
> 2014-02-27 21:33:51,522    ERROR [neutron.api.rpc.agentnotifiers.dhcp_rpc_agent_api] No DHCP agents are associated with network '397fab50-26aa-4cb7-8aa4-c4d43909a00b'. Unable to send notification for 'network_create_end' with payload: {'network': {'status': 'ACTIVE', 'subnets': [], 'name': 'net1', 'provider:physical_network': u'physnet1', 'admin_state_up': True, 'tenant_id': 'test-tenant', 'provider:network_type': 'vlan', 'shared': False, 'id': '397fab50-26aa-4cb7-8aa4-c4d43909a00b', 'provider:segmentation_id': 1000}}
> 2014-02-27 21:33:51,567    ERROR [neutron.api.v2.resource] create failed
> Traceback (most recent call last):
>   File "neutron/api/v2/resource.py", line 84, in resource
>     result = method(request=request, **args)
>   File "neutron/api/v2/base.py", line 347, in create
>     allow_bulk=self._allow_bulk)
>   File "neutron/api/v2/base.py", line 600, in prepare_request_body
>     raise webob.exc.HTTPBadRequest(msg)
> HTTPBadRequest: Invalid input for cidr. Reason: '10.0.2.0' isn't a recognized IP subnet cidr, '10.0.2.0/32' is recommended.
> 
> 
> Thanks again!
> 
> Shixiong
> 
> 
> 
> 
> 
> Shixiong Shang
> 
> !--- Stay Hungry, Stay Foolish ---!
> 
> On Feb 27, 2014, at 8:28 PM, Clark Boylan <clark.boylan at gmail.com> wrote:
> 
>> On Thu, Feb 27, 2014 at 4:43 PM, Shixiong Shang
>> <sparkofwisdom.cloud at gmail.com> wrote:
>>> Hi, guys:
>>> 
>>> I created a fresh local repository and pulled the most recent Neutron code. Before I put in my own code, I did TOX run. However, seems like it is stuck to the following condition for over a hour and didn't go any further. Yesterday, the TOX had been running with a fresh copy of Neutron, but didn't return SUCCESS after the entire night.
>>> 
>>> I assume the copy from MASTER BRANCH should already be sanitized.....However, what I saw in the past 48 hours told me different story. Did I do anything wrong?
>>> 
>>> 
>>> shshang at net-ubuntu2:~/github/neutron$ tox -e py27
>>> py27 create: /home/shshang/github/neutron/.tox/py27
>>> py27 installdeps: -r/home/shshang/github/neutron/requirements.txt, -r/home/shshang/github/neutron/test-requirements.txt, setuptools_git>=0.4
>>> py27 develop-inst: /home/shshang/github/neutron
>>> py27 runtests: commands[0] | python -m neutron.openstack.common.lockutils python setup.py testr --slowest --testr-args=
>>> [pbr] Excluding argparse: Python 2.6 only dependency
>>> running testr
>>> running=OS_STDOUT_CAPTURE=1 OS_STDERR_CAPTURE=1 OS_LOG_CAPTURE=1 ${PYTHON:-python} -m subunit.run discover -t ./ ${OS_TEST_PATH:-./neutron/tests/unit} --list
>>> running=OS_STDOUT_CAPTURE=1 OS_STDERR_CAPTURE=1 OS_LOG_CAPTURE=1 ${PYTHON:-python} -m subunit.run discover -t ./ ${OS_TEST_PATH:-./neutron/tests/unit}  --load-list /tmp/tmpbZwLwg
>>> running=OS_STDOUT_CAPTURE=1 OS_STDERR_CAPTURE=1 OS_LOG_CAPTURE=1 ${PYTHON:-python} -m subunit.run discover -t ./ ${OS_TEST_PATH:-./neutron/tests/unit}  --load-list /tmp/tmp39qJYM
>>> running=OS_STDOUT_CAPTURE=1 OS_STDERR_CAPTURE=1 OS_LOG_CAPTURE=1 ${PYTHON:-python} -m subunit.run discover -t ./ ${OS_TEST_PATH:-./neutron/tests/unit}  --load-list /tmp/tmpppXiTc
>>> running=OS_STDOUT_CAPTURE=1 OS_STDERR_CAPTURE=1 OS_LOG_CAPTURE=1 ${PYTHON:-python} -m subunit.run discover -t ./ ${OS_TEST_PATH:-./neutron/tests/unit}  --load-list /tmp/tmpPhJZDc
>>> 
>>> Thanks!
>>> 
>>> Shixiong
>>> _______________________________________________
>>> OpenStack-dev mailing list
>>> OpenStack-dev at lists.openstack.org
>>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
>> 
>> I think there are two potential problems here. Either a test is
>> deadlocking due to something it has done or
>> neutron.openstack.common.lockutils is deadlocking. In either case
>> OS_TEST_TIMEOUT is not set in .testr.conf so the test suite will not
>> timeout individual tests if necessary. I would start by setting that
>> in the .testr.conf next to OS_STDOUT_CAPTURE and you probably want a
>> value of like 300 (that is seconds).
>> 
>> The other thing you can do to debug this is grab the subunit log file
>> out of .testrepository. While tests are running it will have a tmp
>> random generated name. After tests have run it will be moved to a file
>> named after the most recent test run eg 1 for the first run. The
>> subunit log should offer clues to what was running at the time of
>> deadlocking.
>> 
>> Clark
>> 
>> _______________________________________________
>> OpenStack-dev mailing list
>> OpenStack-dev at lists.openstack.org
>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
> 

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.openstack.org/pipermail/openstack-dev/attachments/20140228/af0f8825/attachment.html>


More information about the OpenStack-dev mailing list