[Openstack] Multiples storages

John Griffith john.griffith at solidfire.com
Wed Nov 13 15:55:17 UTC 2013


On Wed, Nov 13, 2013 at 5:50 AM, Guilherme Russi
<luisguilherme.cr at gmail.com> wrote:
> Hello Razique, I'm here opening this thread again, I've done some cinder
> delete but when I try to create another storeges it returns there's no space
> to create a new volume.
>
> Here is part of my lvdisplay output:
>
> Alloc PE / Size       52224 / 204,00 GiB
> Free  PE / Size       19350 / 75,59 GiB
>
> And here is my lvdisplay:
>
>
>   --- Logical volume ---
>   LV Name
> /dev/cinder-volumes/volume-06ccd141-91c4-45e4-b21f-595f4a36779b
>   VG Name                cinder-volumes
>   LV UUID                wdqxVd-GgUQ-21O4-OWlR-sRT3-HvUA-Q8j9kL
>   LV Write Access        read/write
>   LV snapshot status     source of
>
> /dev/cinder-volumes/_snapshot-04e8414e-2c0e-4fc2-8bff-43dd80ecca09 [active]
>   LV Status              available
>   # open                 0
>   LV Size                10,00 GiB
>   Current LE             2560
>   Segments               1
>   Allocation             inherit
>   Read ahead sectors     auto
>   - currently set to     256
>   Block device           252:1
>
>   --- Logical volume ---
>   LV Name
> /dev/cinder-volumes/_snapshot-04e8414e-2c0e-4fc2-8bff-43dd80ecca09
>   VG Name                cinder-volumes
>   LV UUID                EZz1lC-a8H2-1PlN-pJTN-XAIm-wW0q-qtUQOc
>   LV Write Access        read/write
>   LV snapshot status     active destination for
> /dev/cinder-volumes/volume-06ccd141-91c4-45e4-b21f-595f4a36779b
>   LV Status              available
>   # open                 0
>   LV Size                10,00 GiB
>   Current LE             2560
>   COW-table size         10,00 GiB
>   COW-table LE           2560
>   Allocated to snapshot  0,00%
>   Snapshot chunk size    4,00 KiB
>   Segments               1
>   Allocation             inherit
>   Read ahead sectors     auto
>   - currently set to     256
>   Block device           252:3
>
>   --- Logical volume ---
>   LV Name
> /dev/cinder-volumes/volume-ca36920e-938e-4ad1-b9c4-74c1e28abd31
>   VG Name                cinder-volumes
>   LV UUID                b40kQV-P8N4-R6jt-k97Z-I2a1-9TXm-5GXqfz
>   LV Write Access        read/write
>   LV Status              available
>   # open                 1
>   LV Size                60,00 GiB
>   Current LE             15360
>   Segments               1
>   Allocation             inherit
>   Read ahead sectors     auto
>   - currently set to     256
>   Block device           252:4
>
>   --- Logical volume ---
>   LV Name
> /dev/cinder-volumes/volume-70be4f36-10bd-4877-b841-80333ccfe985
>   VG Name                cinder-volumes
>   LV UUID                2YDrMs-BrYo-aQcZ-8AlX-A4La-HET1-9UQ0gV
>   LV Write Access        read/write
>   LV Status              available
>   # open                 1
>   LV Size                1,00 GiB
>   Current LE             256
>   Segments               1
>   Allocation             inherit
>   Read ahead sectors     auto
>   - currently set to     256
>   Block device           252:5
>
>   --- Logical volume ---
>   LV Name
> /dev/cinder-volumes/volume-00c532bd-91fb-4a38-b340-4389fb7f0ed5
>   VG Name                cinder-volumes
>   LV UUID                MfVOuB-5x5A-jne3-H4Ul-4NP8-eI7b-UYSYE7
>   LV Write Access        read/write
>   LV Status              available
>   # open                 0
>   LV Size                1,00 GiB
>   Current LE             256
>   Segments               1
>   Allocation             inherit
>   Read ahead sectors     auto
>   - currently set to     256
>   Block device           252:6
>
>   --- Logical volume ---
>   LV Name
> /dev/cinder-volumes/volume-ae133dbc-6141-48cf-beeb-9d6576e57a45
>   VG Name                cinder-volumes
>   LV UUID                53w8j3-WT4V-8m52-r6LK-ZYd3-mMHA-FtuyXV
>   LV Write Access        read/write
>   LV Status              available
>   # open                 0
>   LV Size                1,00 GiB
>   Current LE             256
>   Segments               1
>   Allocation             inherit
>   Read ahead sectors     auto
>   - currently set to     256
>   Block device           252:7
>
>   --- Logical volume ---
>   LV Name
> /dev/cinder-volumes/volume-954d2f1b-837b-4ba5-abfd-b3610597be5e
>   VG Name                cinder-volumes
>   LV UUID                belquE-WxQ2-gt6Y-WlPE-Hmq3-B9Am-zcYD3P
>   LV Write Access        read/write
>   LV Status              available
>   # open                 0
>   LV Size                60,00 GiB
>   Current LE             15360
>   Segments               1
>   Allocation             inherit
>   Read ahead sectors     auto
>   - currently set to     256
>   Block device           252:8
>
>   --- Logical volume ---
>   LV Name
> /dev/cinder-volumes/volume-05d037d1-4e61-4419-929a-fe340e00e1af
>   VG Name                cinder-volumes
>   LV UUID                Pt61e7-l3Nu-1IdX-T2sb-0GQD-PhS6-XtIIUj
>   LV Write Access        read/write
>   LV Status              available
>   # open                 1
>   LV Size                1,00 GiB
>   Current LE             256
>   Segments               1
>   Allocation             inherit
>   Read ahead sectors     auto
>   - currently set to     256
>   Block device           252:9
>
>   --- Logical volume ---
>   LV Name
> /dev/cinder-volumes/volume-316f77c6-bf13-4ea4-9b98-028198f3922f
>   VG Name                cinder-volumes
>   LV UUID                e46mBx-CRps-HYKk-aJsc-XFRd-B1Rv-UVk8gT
>   LV Write Access        read/write
>   LV Status              available
>   # open                 1
>   LV Size                60,00 GiB
>   Current LE             15360
>   Segments               1
>   Allocation             inherit
>   Read ahead sectors     auto
>   - currently set to     256
>   Block device           252:10
>
>
> Do you know how I can remove all of them? When I try lvremove
> /dev/cinder-volumes/volume-316f77c6-bf13-4ea4-9b98-028198f3922f, for
> example, I get:
>
>  Can't remove open logical volume
> "volume-316f77c6-bf13-4ea4-9b98-028198f3922f"
>
> Thank you again.
>
>
>
> 2013/11/8 Guilherme Russi <luisguilherme.cr at gmail.com>
>>
>> Very thanks again.
>>
>> Best regards.
>>
>>
>> 2013/11/8 Razique Mahroua <razique.mahroua at gmail.com>
>>>
>>> Oh yah true!
>>> not sure “conductors” exist yet for Cinder, meaning meanwhile, every node
>>> needs a direct access to the database
>>> glad to hear it’s working :)
>>>
>>> On 08 Nov 2013, at 08:53, Guilherme Russi <luisguilherme.cr at gmail.com>
>>> wrote:
>>>
>>> Hello again Razique, I've found the problem, I need to add the grants on
>>> the mysql to my another IP. Now it's working really good :D
>>> I've found this link too if someone needs:
>>> http://docs.openstack.org/admin-guide-cloud/content//managing-volumes.html
>>>
>>> Thank you so much, and if you need me just let me know.
>>>
>>> Best regards.
>>>
>>> Guilherme.
>>>
>>>
>>>
>>> 2013/11/8 Guilherme Russi <luisguilherme.cr at gmail.com>
>>>>
>>>> Hello Razique, I got a couple of doubts, do you know if I need to do
>>>> something else that's is not on the link you sent me? I'm asking because I
>>>> followed the configuration but it's not working, here is what I get: I've
>>>> installed the cinder-volume at the second computer that have the HD, and
>>>> I've changed it's cinder.conf. I've changed too the master's cinder.conf
>>>> like is following:
>>>>
>>>>
>>>> [DEFAULT]
>>>> rootwrap_config = /etc/cinder/rootwrap.conf
>>>> sql_connection = mysql://cinder:password@localhost/cinder
>>>> api_paste_confg = /etc/cinder/api-paste.ini
>>>> #iscsi_helper=iscsiadm
>>>> #iscsi_helper = ietadm
>>>> iscsi_helper = tgtadm
>>>> volume_name_template = volume-%s
>>>> #volume_group = cinder-volumes
>>>> verbose = True
>>>> auth_strategy = keystone
>>>> iscsi_ip_address = 192.168.3.1
>>>> scheduler_driver=cinder.scheduler.filter_scheduler.FilterScheduler
>>>>
>>>> # Rabbit authorization
>>>> rabbit_host = localhost
>>>> rabbit_port = 5672
>>>> rabbit_hosts = $rabbit_host:$rabbit_port
>>>> rabbit_use_ssl = false
>>>> rabbit_userid = guest
>>>> rabbit_password = password
>>>> #rabbit_virtual_host = /nova
>>>>
>>>> state_path = /var/lib/cinder
>>>> lock_path = /var/lock/cinder
>>>> volumes_dir = /var/lib/cinder/volumes
>>>> #rpc_backend = cinder.rpc.impl_kombu
>>>>
>>>> enabled_backends=orion-1,orion-4
>>>> [orion-1]
>>>> volume_group=cinder-volumes
>>>> volume_driver=cinder.volume.drivers.lvm.LVMISCSIDriver
>>>> volume_backend_name=LVM_iSCSI
>>>> [orion-4]
>>>> volume_group=cinder-volumes-2
>>>> volume_driver=cinder.volume.drivers.lvm.LVMISCSIDriver
>>>> volume_backend_name=LVM_iSCSI
>>>>
>>>> The cinder.conf on the second computer is like this but the IPs are
>>>> changed with the controller IP (It has the cinder-api), and when I run
>>>> service cinder-volume restart at the second computer it's status is
>>>> stop/waiting.
>>>>
>>>> Any ideas?
>>>>
>>>> Thanks :)
>>>>
>>>>
>>>> 2013/11/8 Razique Mahroua <razique.mahroua at gmail.com>
>>>>>
>>>>> sure :)
>>>>>
>>>>> On 08 Nov 2013, at 05:39, Guilherme Russi <luisguilherme.cr at gmail.com>
>>>>> wrote:
>>>>>
>>>>> Oh great! I'll try here and send you the results.
>>>>>
>>>>> Very thanks :)
>>>>>
>>>>>
>>>>> 2013/11/8 Razique Mahroua <razique.mahroua at gmail.com>
>>>>>>
>>>>>> If I’m not mistaken, you only need to install the “cinder-volume’
>>>>>> service that will update its status to your main node
>>>>>> :)
>>>>>>
>>>>>> On 08 Nov 2013, at 05:34, Guilherme Russi <luisguilherme.cr at gmail.com>
>>>>>> wrote:
>>>>>>
>>>>>> Great! I was reading the link and I have one question, do I need to
>>>>>> install cinder at the other computer too?
>>>>>>
>>>>>> Thanks :)
>>>>>>
>>>>>>
>>>>>> 2013/11/8 Razique Mahroua <razique.mahroua at gmail.com>
>>>>>>>
>>>>>>> Ok in that case, with Grizzly you can use the “multi-backends”
>>>>>>> feature:
>>>>>>> https://wiki.openstack.org/wiki/Cinder-multi-backend
>>>>>>>
>>>>>>> and that should do it :)
>>>>>>>
>>>>>>> On 08 Nov 2013, at 05:29, Guilherme Russi
>>>>>>> <luisguilherme.cr at gmail.com> wrote:
>>>>>>>
>>>>>>> It is a hard disk, my scenario is one Controller (where I have my
>>>>>>> storage cinder and my network quantum) and four compute nodes.
>>>>>>>
>>>>>>>
>>>>>>> 2013/11/8 Razique Mahroua <razique.mahroua at gmail.com>
>>>>>>>>
>>>>>>>> ok !
>>>>>>>> what is your actual Cinder backend? Is it a hard disk, a SAN, a
>>>>>>>> network volume, etc…
>>>>>>>>
>>>>>>>> On 08 Nov 2013, at 05:20, Guilherme Russi
>>>>>>>> <luisguilherme.cr at gmail.com> wrote:
>>>>>>>>
>>>>>>>> Hi Razique, thank you for answering, I want to expand my cinder
>>>>>>>> storage, is it the block storage? I'll use the storage to allow VMs to have
>>>>>>>> more hard disk space.
>>>>>>>>
>>>>>>>> Regards.
>>>>>>>>
>>>>>>>> Guilherme.
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> 2013/11/8 Razique Mahroua <razique.mahroua at gmail.com>
>>>>>>>>>
>>>>>>>>> Hi Guilherme !
>>>>>>>>> Which storage do you precisely want to expand?
>>>>>>>>>
>>>>>>>>> Regards,
>>>>>>>>> Razique
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> On 08 Nov 2013, at 04:52, Guilherme Russi
>>>>>>>>> <luisguilherme.cr at gmail.com> wrote:
>>>>>>>>>
>>>>>>>>> > Hello guys, I have a Grizzly deployment running fine with 5
>>>>>>>>> > nodes, and I want to add more storage on it. My question is, can I install a
>>>>>>>>> > new HD on another computer thats not the controller and link this HD with my
>>>>>>>>> > cinder that it can be a storage too?
>>>>>>>>> > The computer I will install my new HD is at the same network as
>>>>>>>>> > my cloud is. I'm asking because I haven't seen a question like that here.
>>>>>>>>> > Does anybody knows how to do that? Have a clue? Any help is welcome.
>>>>>>>>> >
>>>>>>>>> > Thank you all.
>>>>>>>>> >
>>>>>>>>> > Best regards.
>>>>>>>>> > _______________________________________________
>>>>>>>>> > Mailing list:
>>>>>>>>> > http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
>>>>>>>>> > Post to     : openstack at lists.openstack.org
>>>>>>>>> > Unsubscribe :
>>>>>>>>> > http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
>>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>
>>>>>>
>>>>>
>>>>>
>>>>
>>>
>>>
>>
>
>
> _______________________________________________
> Mailing list: http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
> Post to     : openstack at lists.openstack.org
> Unsubscribe : http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
>

It appears that you have in fact exceeded your VG space.  As far as
the errors on lvremove;  Is your system in a state where you can use
Cinder (cinder delete xxxx) to remove the volume?

There are a number of ways to try and clean up the busy device on
lvremove.  You can try using the -f flag and retrying, sometimes udev
settle is helpful (but I don't think this applies in your case) maybe
"dmsetup remove" will work in your case.  I'd be curious if your
volumes are actually in use currently and the system is trying to
"help" you here :)




More information about the Openstack mailing list