[Openstack] [CINDER]Cinder Volume Creation

John Griffith john.griffith at solidfire.com
Wed Dec 4 21:41:29 UTC 2013


On Wed, Dec 4, 2013 at 2:26 PM, Byron Briggs <byron.briggs at ait.com> wrote:

> root at Compute01-CodeNode:/etc/cinder# vgs
>
> File descriptor 3 (/usr/share/bash-completion/completions) leaked on vgs
> invocation. Parent PID 10778: -bash
>
>   VG             #PV #LV #SN Attr   VSize   VFree
>
>   cinder-volumes   1   0   0 wz--n- 100.00g 100.00g
>
>
>
>
>
> My control nodes don’t but that shouldn’t matter from my understanding.
>
> root at control01:/etc/cinder# vgs
>
>   VG           #PV #LV #SN Attr   VSize   VFree
>
>   control01-vg   1   2   0 wz--n- 931.27g 44.00m
>
>
>
> *From:* John Griffith [mailto:john.griffith at solidfire.com]
> *Sent:* Wednesday, December 04, 2013 3:20 PM
> *To:* Byron Briggs
> *Cc:* openstack at lists.openstack.org; SYSADMIN
>
> *Subject:* Re: [Openstack] [CINDER]Cinder Volume Creation
>
>
>
>
>
>
>
> On Wed, Dec 4, 2013 at 12:43 PM, Byron Briggs <byron.briggs at ait.com>
> wrote:
>
> When running
>
>
>
> *cinder create --display_name test 10*
>
>
>
> From the compute01NovaCompute.dmz-pod2 Node listed below. It is also
> running xenapi for a xenserver.
>
> ( Control01,2,3 are an HA cluster sitting behind haproxy/keepalivd running
> all the communication and schedulers  -all that is working well.)
>
>
>
>
>
> I get this error in my cinder-volume.log on the control nodes(since they
> run schedulers) There is nothing else to go off of other then “ERROR” on
> the volume status.
>
>
>
> 2013-12-04 12:55:32  WARNING [cinder.scheduler.host_manager] service is
> down or disabled.
>
> 2013-12-04 12:55:32    ERROR [cinder.scheduler.manager] Failed to
> schedule_create_volume: No valid host was found.
>
>
>
>
>
> In case you can’t see excel  below.
>
> 192.168.220.40 Is the proxy being distributed into the three control nodes.
>
>
>
> control01.dmz-pod2(192.168.220.41) ->cinder-api,cinder-scheduler
>
> control02.dmz-pod2(192.168.220.42) ->cinder-api,cinder-scheduler
>
> control03.dmz-pod2(192.168.220.43) ->cinder-api,cinder-scheduler
>
> compute01NovaCompute.dmz-pod2(192.168.220.101) ->cinder-volume
>
>
>
>
>
>
>
> Server Name
>
> cinder-api
>
> cinder-scheduler
>
> cinder-volume
>
> control01.dmz-pod2
>
> YES
>
> YES
>
> NO
>
> control02.dmz-pod2
>
> YES
>
> YES
>
> NO
>
> control03.dmz-pod2
>
> YES
>
> YES
>
> NO
>
> compute01NovaCompute.dmz-pod2
>
> NO
>
> NO
>
> NO
>
>
>
> All services are running with no log errors.
>
>
>
> Here is my configs
>
> *compute01NovaCompute.dmz-pod2*
>
> */etc/cinder/api-paste.ini*
>
> [filter:authtoken]
>
> paste.filter_factory = keystoneclient.middleware.auth_token:filter_factory
>
> service_protocol = http
>
> service_host = %SERVICE_TENANT_NAME%
>
> service_port = 5000
>
> auth_host = 127.0.0.1
>
> auth_port = 35357
>
> auth_protocol = http
>
> admin_tenant_name = services
>
> admin_user = %SERVICE_USER%
>
> admin_password = %SERVICE_PASSWORD%
>
> signing_dir = /var/lib/cinder
>
>
>
> */etc/cinder/cinder.conf*
>
> [DEFAULT]
>
> iscsi_ip_address=192.168.220.101
>
> rabbit_ha_queues=True
>
> rabbit_hosts=control01:5672,control02:5672,control03:5672
>
> rabbit_userid=openstack_rabbit_user
>
> rabbit_password=openstack_rabbit_password
>
> sql_connection = mysql://cinder:cinder_pass@192.168.220.40/cinder
>
> rootwrap_config = /etc/cinder/rootwrap.conf
>
> api_paste_confg = /etc/cinder/api-paste.ini
>
> iscsi_helper = tgtadm
>
> volume_name_template = volume-%s
>
> volume_group = cinder-volumes
>
> verbose = True
>
> auth_strategy = keystone
>
> state_path = /var/lib/cinder
>
> lock_path = /var/lock/cinder
>
> volumes_dir = /var/lib/cinder/volumes
>
>
>
> *pvscan*
>
> File descriptor 3 (/usr/share/bash-completion/completions) leaked on
> pvscan invocation. Parent PID 10778: -bash
>
>   PV /dev/xvdb   VG cinder-volumes   lvm2 [100.00 GiB / 100.00 GiB free]
>
>   Total: 1 [100.00 GiB] / in use: 1 [100.00 GiB] / in no VG: 0 [0   ]
>
>
>
>
>
> *control01,2,3*
>
> */etc/cinder/api-paste.ini*
>
> [filter:authtoken]
>
> paste.filter_factory = keystoneclient.middleware.auth_token:filter_factory
>
> service_protocol = http
>
> service_host = 192.168.220.40
>
> service_port = 5000
>
> auth_host = 192.168.220.40
>
> auth_port = 35357
>
> auth_protocol = http
>
> admin_tenant_name = services
>
> admin_user = cinder
>
> admin_password = keystone_admin
>
> signing_dir = /var/lib/cinder
>
>
>
> */etc/cinder/conder.conf*
>
> [DEFAULT]
>
> sql_idle_timeout=30
>
> rabbit_ha_queues=True
>
> rabbit_hosts=control01:5672,control02:5672,control03:5672
>
> rabbit_userid=openstack_rabbit_user
>
> rabbit_password=openstack_rabbit_password
>
> sql_connection = mysql://cinder:cinder_pass@192.168.220.40/cinder
>
> osapi_volume_listen = 192.168.220.41
>
> rootwrap_config = /etc/cinder/rootwrap.conf
>
> api_paste_confg = /etc/cinder/api-paste.ini
>
> iscsi_helper = tgtadm
>
> volume_name_template = volume-%s
>
> volume_group = nova-volumes
>
> verbose = True
>
> auth_strategy = keystone
>
> state_path = /var/lib/cinder
>
> lock_path = /var/lock/cinder
>
> volumes_dir = /var/lib/cinder/volumes
>
>
>
>
>
> Grizzly Release
>
>
>
>
>
> Any ideas on where to look more into the issue or something with my config?
>
>
>
>
> _______________________________________________
> Mailing list:
> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
> Post to     : openstack at lists.openstack.org
> Unsubscribe :
> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
>
>
>
> So this happens in a couple of situations, the most common is when there's
> not enough space/capacity being reported by the configured backend driver
> to allocate the amount of space you're requesting.  Try a "sudo vgs" and
> verify that you have enough capacity on your backing store (nova-volumes)
> to deploy a 10G volume.
>
>
>

Ooops, sorry I didn't catch the separate controller node and only saw the
volume_group setting there.  Any chance you could link a paste-bin to the
cinder-scheduler logs?
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.openstack.org/pipermail/openstack/attachments/20131204/8795de65/attachment.html>


More information about the Openstack mailing list