[openstack-dev] [Neutron] tox run forever

Shixiong Shang sparkofwisdom.cloud at gmail.com
Fri Feb 28 02:41:34 UTC 2014


Hi, Clark:

Thanks a lot for the prompt response! I added the OS_TEST_TIMEOUT value (300 sec) and was tailing the tmp file. It turned out that the TOX run stopped at the following point. My machine was tossed so badly that it became unresponsive and I had to hard reboot it….I am pulling my teeth off now…..Is it normal to see Traceback?

2014-02-27 21:33:51,212     INFO [neutron.api.extensions] Extension 'agent' provides no backward compatibility map for extended attributes
2014-02-27 21:33:51,212     INFO [neutron.api.extensions] Extension 'Allowed Address Pairs' provides no backward compatibility map for extended attributes
2014-02-27 21:33:51,212     INFO [neutron.api.extensions] Extension 'Neutron Extra Route' provides no backward compatibility map for extended attributes
2014-02-27 21:33:51,522    ERROR [neutron.api.rpc.agentnotifiers.dhcp_rpc_agent_api] No DHCP agents are associated with network '397fab50-26aa-4cb7-8aa4-c4d43909a00b'. Unable to send notification for 'network_create_end' with payload: {'network': {'status': 'ACTIVE', 'subnets': [], 'name': 'net1', 'provider:physical_network': u'physnet1', 'admin_state_up': True, 'tenant_id': 'test-tenant', 'provider:network_type': 'vlan', 'shared': False, 'id': '397fab50-26aa-4cb7-8aa4-c4d43909a00b', 'provider:segmentation_id': 1000}}
2014-02-27 21:33:51,567    ERROR [neutron.api.v2.resource] create failed
Traceback (most recent call last):
  File "neutron/api/v2/resource.py", line 84, in resource
    result = method(request=request, **args)
  File "neutron/api/v2/base.py", line 347, in create
    allow_bulk=self._allow_bulk)
  File "neutron/api/v2/base.py", line 600, in prepare_request_body
    raise webob.exc.HTTPBadRequest(msg)
HTTPBadRequest: Invalid input for cidr. Reason: '10.0.2.0' isn't a recognized IP subnet cidr, '10.0.2.0/32' is recommended.


Thanks again!

Shixiong





Shixiong Shang

!--- Stay Hungry, Stay Foolish ---!

On Feb 27, 2014, at 8:28 PM, Clark Boylan <clark.boylan at gmail.com> wrote:

> On Thu, Feb 27, 2014 at 4:43 PM, Shixiong Shang
> <sparkofwisdom.cloud at gmail.com> wrote:
>> Hi, guys:
>> 
>> I created a fresh local repository and pulled the most recent Neutron code. Before I put in my own code, I did TOX run. However, seems like it is stuck to the following condition for over a hour and didn't go any further. Yesterday, the TOX had been running with a fresh copy of Neutron, but didn't return SUCCESS after the entire night.
>> 
>> I assume the copy from MASTER BRANCH should already be sanitized.....However, what I saw in the past 48 hours told me different story. Did I do anything wrong?
>> 
>> 
>> shshang at net-ubuntu2:~/github/neutron$ tox -e py27
>> py27 create: /home/shshang/github/neutron/.tox/py27
>> py27 installdeps: -r/home/shshang/github/neutron/requirements.txt, -r/home/shshang/github/neutron/test-requirements.txt, setuptools_git>=0.4
>> py27 develop-inst: /home/shshang/github/neutron
>> py27 runtests: commands[0] | python -m neutron.openstack.common.lockutils python setup.py testr --slowest --testr-args=
>> [pbr] Excluding argparse: Python 2.6 only dependency
>> running testr
>> running=OS_STDOUT_CAPTURE=1 OS_STDERR_CAPTURE=1 OS_LOG_CAPTURE=1 ${PYTHON:-python} -m subunit.run discover -t ./ ${OS_TEST_PATH:-./neutron/tests/unit} --list
>> running=OS_STDOUT_CAPTURE=1 OS_STDERR_CAPTURE=1 OS_LOG_CAPTURE=1 ${PYTHON:-python} -m subunit.run discover -t ./ ${OS_TEST_PATH:-./neutron/tests/unit}  --load-list /tmp/tmpbZwLwg
>> running=OS_STDOUT_CAPTURE=1 OS_STDERR_CAPTURE=1 OS_LOG_CAPTURE=1 ${PYTHON:-python} -m subunit.run discover -t ./ ${OS_TEST_PATH:-./neutron/tests/unit}  --load-list /tmp/tmp39qJYM
>> running=OS_STDOUT_CAPTURE=1 OS_STDERR_CAPTURE=1 OS_LOG_CAPTURE=1 ${PYTHON:-python} -m subunit.run discover -t ./ ${OS_TEST_PATH:-./neutron/tests/unit}  --load-list /tmp/tmpppXiTc
>> running=OS_STDOUT_CAPTURE=1 OS_STDERR_CAPTURE=1 OS_LOG_CAPTURE=1 ${PYTHON:-python} -m subunit.run discover -t ./ ${OS_TEST_PATH:-./neutron/tests/unit}  --load-list /tmp/tmpPhJZDc
>> 
>> Thanks!
>> 
>> Shixiong
>> _______________________________________________
>> OpenStack-dev mailing list
>> OpenStack-dev at lists.openstack.org
>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
> 
> I think there are two potential problems here. Either a test is
> deadlocking due to something it has done or
> neutron.openstack.common.lockutils is deadlocking. In either case
> OS_TEST_TIMEOUT is not set in .testr.conf so the test suite will not
> timeout individual tests if necessary. I would start by setting that
> in the .testr.conf next to OS_STDOUT_CAPTURE and you probably want a
> value of like 300 (that is seconds).
> 
> The other thing you can do to debug this is grab the subunit log file
> out of .testrepository. While tests are running it will have a tmp
> random generated name. After tests have run it will be moved to a file
> named after the most recent test run eg 1 for the first run. The
> subunit log should offer clues to what was running at the time of
> deadlocking.
> 
> Clark
> 
> _______________________________________________
> OpenStack-dev mailing list
> OpenStack-dev at lists.openstack.org
> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.openstack.org/pipermail/openstack-dev/attachments/20140227/85024cf5/attachment.html>


More information about the OpenStack-dev mailing list