[openstack-dev] [Openstack] No one replying on tempest issue?Please share your experience
GHANSHYAM MANN
ghanshyammann at gmail.com
Tue Sep 30 00:29:59 UTC 2014
Its present as you mentioned. you can look screen-n-cpu.*.log. All running
services logs files will be @ /opt/stack/logs/screen/, which you can
analyze and find where issue is.
Such query can be asked on IRC ( https://wiki.openstack.org/wiki/IRC) for
quick reply instead of waiting on mail.
For further discussion on mail please change the subject now :).
On Mon, Sep 29, 2014 at 7:51 PM, Nikesh Kumar Mahalka <
nikeshmahalka at vedams.com> wrote:
> How to get nova-compute logs in juno devstack?
> Below are nova services:
> vedams at vedams-compute-fc:/opt/stack/tempest$ ps -aef | grep nova
> vedams 15065 14812 0 10:56 pts/10 00:00:52 /usr/bin/python
> /usr/local/bin/nova-compute --config-file /etc/nova/nova.conf
> vedams 15077 14811 0 10:56 pts/9 00:02:06 /usr/bin/python
> /usr/local/bin/nova-api
> vedams 15086 14818 0 10:56 pts/12 00:00:09 /usr/bin/python
> /usr/local/bin/nova-cert --config-file /etc/nova/nova.conf
> vedams 15095 14836 0 10:56 pts/17 00:00:09 /usr/bin/python
> /usr/local/bin/nova-consoleauth --config-file /etc/nova/nova.conf
> vedams 15096 14821 0 10:56 pts/13 00:00:09 /usr/bin/python
> /usr/local/bin/nova-network --config-file /etc/nova/nova.conf
> vedams 15100 14844 0 10:56 pts/18 00:00:00 /usr/bin/python
> /usr/local/bin/nova-objectstore --config-file /etc/nova/nova.conf
> vedams 15101 14826 0 10:56 pts/15 00:00:05 /usr/bin/python
> /usr/local/bin/nova-novncproxy --config-file /etc/nova/nova.conf --web
> /opt/stack/noVNC
> vedams 15103 14814 0 10:56 pts/11 00:02:02 /usr/bin/python
> /usr/local/bin/nova-conductor --config-file /etc/nova/nova.conf
> vedams 15104 14823 0 10:56 pts/14 00:00:11 /usr/bin/python
> /usr/local/bin/nova-scheduler --config-file /etc/nova/nova.conf
> vedams 15117 14831 0 10:56 pts/16 00:00:00 /usr/bin/python
> /usr/local/bin/nova-xvpvncproxy --config-file /etc/nova/nova.conf
> vedams 15195 15103 0 10:56 pts/11 00:00:24 /usr/bin/python
> /usr/local/bin/nova-conductor --config-file /etc/nova/nova.conf
> vedams 15196 15103 0 10:56 pts/11 00:00:25 /usr/bin/python
> /usr/local/bin/nova-conductor --config-file /etc/nova/nova.conf
> vedams 15197 15103 0 10:56 pts/11 00:00:24 /usr/bin/python
> /usr/local/bin/nova-conductor --config-file /etc/nova/nova.conf
> vedams 15198 15103 0 10:56 pts/11 00:00:24 /usr/bin/python
> /usr/local/bin/nova-conductor --config-file /etc/nova/nova.conf
> vedams 15208 15077 0 10:56 pts/9 00:00:00 /usr/bin/python
> /usr/local/bin/nova-api
> vedams 15209 15077 0 10:56 pts/9 00:00:00 /usr/bin/python
> /usr/local/bin/nova-api
> vedams 15238 15077 0 10:56 pts/9 00:00:03 /usr/bin/python
> /usr/local/bin/nova-api
> vedams 15239 15077 0 10:56 pts/9 00:00:01 /usr/bin/python
> /usr/local/bin/nova-api
> vedams 15240 15077 0 10:56 pts/9 00:00:03 /usr/bin/python
> /usr/local/bin/nova-api
> vedams 15241 15077 0 10:56 pts/9 00:00:03 /usr/bin/python
> /usr/local/bin/nova-api
> vedams 15248 15077 0 10:56 pts/9 00:00:00 /usr/bin/python
> /usr/local/bin/nova-api
> vedams 15249 15077 0 10:56 pts/9 00:00:00 /usr/bin/python
> /usr/local/bin/nova-api
> vedams 21850 14712 0 16:16 pts/0 00:00:00 grep --color=auto nova
>
>
> Below are nova logs files:
> vedams at vedams-compute-fc:/opt/stack/tempest$ ls
> /opt/stack/logs/screen/screen-n-
> screen-n-api.2014-09-28-101810.log
> screen-n-cond.log screen-n-net.2014-09-28-101810.log
> screen-n-obj.log
> screen-n-api.log
> screen-n-cpu.2014-09-28-101810.log screen-n-net.log
> screen-n-sch.2014-09-28-101810.log
> screen-n-cauth.2014-09-28-101810.log
> screen-n-cpu.log screen-n-novnc.2014-09-28-101810.log
> screen-n-sch.log
> screen-n-cauth.log
> screen-n-crt.2014-09-28-101810.log screen-n-novnc.log
> screen-n-xvnc.2014-09-28-101810.log
> screen-n-cond.2014-09-28-101810.log
> screen-n-crt.log screen-n-obj.2014-09-28-101810.log
> screen-n-xvnc.log
>
>
> Below are nova screen-seesions:
> 6-$(L) n-api 7$(L) n-cpu 8$(L) n-cond 9$(L) n-crt 10$(L) n-net 11$(L)
> n-sch 12$(L) n-novnc 13$(L) n-xvnc 14$(L) n-cauth 15$(L) n-obj
>
>
>
>
> Regards
> Nikesh
>
>
> On Tue, Sep 23, 2014 at 3:10 PM, Nikesh Kumar Mahalka <
> nikeshmahalka at vedams.com> wrote:
>
>> Hi,
>> I am able to do all volume operations through dashboard and cli commands.
>> But when i am running tempest tests,some tests are getting failed.
>> For contributing cinder volume driver for my client in cinder,do all
>> tempest tests should passed?
>>
>> Ex:
>> 1)
>> ./run_tempest.sh tempest.api.volume.test_volumes_snapshots : 1or 2 tests
>> are getting failed
>>
>> But when i am running individual tests in "test_volumes_snapshots",all
>> tests are getting passed.
>>
>> 2)
>> ./run_tempest.sh
>> tempest.api.volume.test_volumes_actions.VolumesV2ActionsTest.test_volume_upload:
>> This is also getting failed.
>>
>>
>>
>> Regards
>> Nikesh
>>
>> On Mon, Sep 22, 2014 at 4:12 PM, Ken'ichi Ohmichi <ken1ohmichi at gmail.com>
>> wrote:
>>
>>> Hi Nikesh,
>>>
>>> > -----Original Message-----
>>> > From: Nikesh Kumar Mahalka [mailto:nikeshmahalka at vedams.com]
>>> > Sent: Saturday, September 20, 2014 9:49 PM
>>> > To: openstack at lists.openstack.org; OpenStack Development Mailing List
>>> (not for usage questions)
>>> > Subject: Re: [Openstack] No one replying on tempest issue?Please share
>>> your experience
>>> >
>>> > Still i didnot get any reply.
>>>
>>> Jay has already replied to this mail, please check the nova-compute
>>> and cinder-volume log as he said[1].
>>>
>>> [1]:
>>> http://lists.openstack.org/pipermail/openstack-dev/2014-September/046147.html
>>>
>>> > Now i ran below command:
>>> > ./run_tempest.sh
>>> tempest.api.volume.test_volumes_snapshots.VolumesSnapshotTest.test_volume_from_snapshot
>>> >
>>> > and i am getting test failed.
>>> >
>>> >
>>> > Actually,after analyzing tempest.log,i found that:
>>> > during creation of a volume from snapshot,tearDownClass is called and
>>> it is deleting snapshot bfore creation of volume
>>> > and my test is getting failed.
>>>
>>> I guess the failure you mentioned at the above is:
>>>
>>> 2014-09-20 00:42:12.519 10684 INFO tempest.common.rest_client
>>> [req-d4dccdcd-bbfa-4ddf-acd8-5a7dcd5b15db None] Request
>>> (VolumesSnapshotTest:tearDownClass): 404 GET
>>>
>>> http://192.168.2.153:8776/v1/ff110b66c98d455092c6f2a2577b4c80/snapshots/71d3cad4-440d-4fbb-8758-76da17b6ace6
>>> 0.029s
>>>
>>> and
>>>
>>> 2014-09-20 00:42:22.511 10684 INFO tempest.common.rest_client
>>> [req-520a54ad-7e0a-44ba-95c0-17f4657bc3b0 None] Request
>>> (VolumesSnapshotTest:tearDownClass): 404 GET
>>>
>>> http://192.168.2.153:8776/v1/ff110b66c98d455092c6f2a2577b4c80/volumes/7469271a-d2a7-4ee6-b54a-cd0bf767be6b
>>> 0.034s
>>>
>>> right?
>>> If so, that is not a problem.
>>> VolumesSnapshotTest creates two volumes, and the tearDownClass checks
>>> these
>>> volumes deletions by getting volume status until 404(NotFound) [2].
>>>
>>> [2]:
>>> https://github.com/openstack/tempest/blob/master/tempest/api/volume/base.py#L128
>>>
>>> > I deployed a juno devstack setup for a cinder volume driver.
>>> > I changed cinder.conf file and tempest.conf file for single backend
>>> and restarted cinder services.
>>> > Now i ran tempest test as below:
>>> > /opt/stack/tempest/run_tempest.sh
>>> tempest.api.volume.test_volumes_snapshots
>>> >
>>> > I am getting below output:
>>> > Traceback (most recent call last):
>>> > File
>>> "/opt/stack/tempest/tempest/api/volume/test_volumes_snapshots.py", line
>>> 176, in test_volume_from_snapshot
>>> > snapshot = self.create_snapshot(self.volume_origin['id'])
>>> > File "/opt/stack/tempest/tempest/api/volume/base.py", line 112, in
>>> create_snapshot
>>> > 'available')
>>> > File
>>> "/opt/stack/tempest/tempest/services/volume/json/snapshots_client.py", line
>>> 126, in wait_for_snapshot_status
>>> > value = self._get_snapshot_status(snapshot_id)
>>> > File
>>> "/opt/stack/tempest/tempest/services/volume/json/snapshots_client.py", line
>>> 99, in _get_snapshot_status
>>> > snapshot_id=snapshot_id)
>>> > SnapshotBuildErrorException: Snapshot
>>> 6b1eb319-33ef-4357-987a-58eb15549520 failed to build and is in
>>> > ERROR status
>>>
>>> What happens if running the same operation as Tempest by hands on your
>>> environment like the following ?
>>>
>>> [1] $ cinder create 1
>>> [2] $ cinder snapshot-create <id of the created volume at [1]>
>>> [3] $ cinder create --snapshot-id <id of the created snapshot at [2]> 1
>>> [4] $ cinder show <id of the created volume at [3]>
>>>
>>> Please check whether the status of created volume at [3] is "available"
>>> or not.
>>>
>>> Thanks
>>> Ken'ichi Ohmichi
>>>
>>> _______________________________________________
>>> Mailing list:
>>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
>>> Post to : openstack at lists.openstack.org
>>> Unsubscribe :
>>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
>>>
>>
>>
>
> _______________________________________________
> OpenStack-dev mailing list
> OpenStack-dev at lists.openstack.org
> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
>
>
--
Thanks & Regards
Ghanshyam Mann
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.openstack.org/pipermail/openstack-dev/attachments/20140930/d1c168cd/attachment.html>
More information about the OpenStack-dev
mailing list