[Openstack] Multiples storages

Guilherme Russi luisguilherme.cr at gmail.com
Mon Nov 18 13:11:44 UTC 2013


Hello again guys, I'm trying to overcome another issue with cinder, I'm
here trying to create e 500 GB volume, this is my disk:

pvscan
  PV /dev/sdb1   VG cinder-volumes-2   lvm2 [931,51 GiB / 531,51 GiB free]
  Total: 1 [931,51 GiB] / in use: 1 [931,51 GiB] / in no VG: 0 [0   ]

But when I try:
cinder create --volume_type lvm_one --display_name v2st3-500 500

I get:
ERROR: VolumeSizeExceedsAvailableQuota: Requested volume or snapshot
exceeds allowed Gigabytes quota

Does anybody know where I begin to fix it?

The outputs on cinder-api.log are:

ERROR [cinder.api.middleware.fault] Caught error: Requested volume or
snapshot exceeds allowed Gigabytes quota
Traceback (most recent call last):
  File "/usr/lib/python2.7/dist-packages/cinder/api/middleware/fault.py",
line 73, in __call__
    return req.get_response(self.application)
  File "/usr/lib/python2.7/dist-packages/webob/request.py", line 1296, in
send
    application, catch_exc_info=False)
  File "/usr/lib/python2.7/dist-packages/webob/request.py", line 1260, in
call_application
    app_iter = application(self.environ, start_response)
  File "/usr/lib/python2.7/dist-packages/webob/dec.py", line 144, in
__call__
    return resp(environ, start_response)
  File
"/usr/lib/python2.7/dist-packages/keystoneclient/middleware/auth_token.py",
line 450, in __call__
    return self.app(env, start_response)
  File "/usr/lib/python2.7/dist-packages/webob/dec.py", line 144, in
__call__
    return resp(environ, start_response)
  File "/usr/lib/python2.7/dist-packages/webob/dec.py", line 144, in
__call__
    return resp(environ, start_response)
  File "/usr/lib/python2.7/dist-packages/routes/middleware.py", line 131,
in __call__
    response = self.app(environ, start_response)
  File "/usr/lib/python2.7/dist-packages/webob/dec.py", line 144, in
__call__
    return resp(environ, start_response)
  File "/usr/lib/python2.7/dist-packages/webob/dec.py", line 130, in
__call__
    resp = self.call_func(req, *args, **self.kwargs)
  File "/usr/lib/python2.7/dist-packages/webob/dec.py", line 195, in
call_func
    return self.func(req, *args, **kwargs)
  File "/usr/lib/python2.7/dist-packages/cinder/api/openstack/wsgi.py",
line 803, in __call__
    content_type, body, accept)
  File "/usr/lib/python2.7/dist-packages/cinder/api/openstack/wsgi.py",
line 851, in _process_stack
    action_result = self.dispatch(meth, request, action_args)
  File "/usr/lib/python2.7/dist-packages/cinder/api/openstack/wsgi.py",
line 927, in dispatch
    return method(req=request, **action_args)
  File "/usr/lib/python2.7/dist-packages/cinder/api/v1/volumes.py", line
358, in create
    **kwargs)
  File "/usr/lib/python2.7/dist-packages/cinder/volume/api.py", line 165,
in create
    raise exception.VolumeSizeExceedsAvailableQuota()
VolumeSizeExceedsAvailableQuota: Requested volume or snapshot exceeds
allowed Gigabytes quota

Thank you all.



2013/11/15 Razique Mahroua <razique.mahroua at gmail.com>

> Awesome :)
>
> Razique
>
> On 14 Nov 2013, at 15:27, Guilherme Russi wrote:
>
>  That's right, I've stopped the open-iscsi and tgt process, and the
>> lvremove
>> worked. Thank you all.
>>
>> Regards.
>>
>>
>> 2013/11/13 Razique Mahroua <razique.mahroua at gmail.com>
>>
>>  Hey :)
>>> that means the volume is still in use. (lvopen : 1) make sure it's not by
>>> checking the process, qemu-nbd, etc...
>>>
>>> On 13 Nov 2013, at 4:50, Guilherme Russi wrote:
>>>
>>> Hello Razique, I'm here opening this thread again, I've done some cinder
>>> delete but when I try to create another storeges it returns there's no
>>> space to create a new volume.
>>>
>>> Here is part of my lvdisplay output:
>>>
>>> Alloc PE / Size 52224 / 204,00 GiB
>>> Free PE / Size 19350 / 75,59 GiB
>>>
>>> And here is my lvdisplay:
>>>
>>> --- Logical volume ---
>>> LV Name
>>> /dev/cinder-volumes/volume-06ccd141-91c4-45e4-b21f-595f4a36779b
>>> VG Name cinder-volumes
>>> LV UUID wdqxVd-GgUQ-21O4-OWlR-sRT3-HvUA-Q8j9kL
>>> LV Write Access read/write
>>> LV snapshot status source of
>>>
>>> /dev/cinder-volumes/_snapshot-04e8414e-2c0e-4fc2-8bff-43dd80ecca09
>>> [active]
>>> LV Status available
>>> open 0
>>>
>>> LV Size 10,00 GiB
>>> Current LE 2560
>>> Segments 1
>>> Allocation inherit
>>> Read ahead sectors auto
>>> - currently set to 256
>>> Block device 252:1
>>>
>>> --- Logical volume ---
>>> LV Name
>>> /dev/cinder-volumes/_snapshot-04e8414e-2c0e-4fc2-8bff-43dd80ecca09
>>> VG Name cinder-volumes
>>> LV UUID EZz1lC-a8H2-1PlN-pJTN-XAIm-wW0q-qtUQOc
>>> LV Write Access read/write
>>> LV snapshot status active destination for
>>> /dev/cinder-volumes/volume-06ccd141-91c4-45e4-b21f-595f4a36779b
>>> LV Status available
>>> open 0
>>>
>>> LV Size 10,00 GiB
>>> Current LE 2560
>>> COW-table size 10,00 GiB
>>> COW-table LE 2560
>>> Allocated to snapshot 0,00%
>>> Snapshot chunk size 4,00 KiB
>>> Segments 1
>>> Allocation inherit
>>> Read ahead sectors auto
>>> - currently set to 256
>>> Block device 252:3
>>>
>>> --- Logical volume ---
>>> LV Name
>>> /dev/cinder-volumes/volume-ca36920e-938e-4ad1-b9c4-74c1e28abd31
>>> VG Name cinder-volumes
>>> LV UUID b40kQV-P8N4-R6jt-k97Z-I2a1-9TXm-5GXqfz
>>> LV Write Access read/write
>>> LV Status available
>>> open 1
>>>
>>> LV Size 60,00 GiB
>>> Current LE 15360
>>> Segments 1
>>> Allocation inherit
>>> Read ahead sectors auto
>>> - currently set to 256
>>> Block device 252:4
>>>
>>> --- Logical volume ---
>>> LV Name
>>> /dev/cinder-volumes/volume-70be4f36-10bd-4877-b841-80333ccfe985
>>> VG Name cinder-volumes
>>> LV UUID 2YDrMs-BrYo-aQcZ-8AlX-A4La-HET1-9UQ0gV
>>> LV Write Access read/write
>>> LV Status available
>>> open 1
>>>
>>> LV Size 1,00 GiB
>>> Current LE 256
>>> Segments 1
>>> Allocation inherit
>>> Read ahead sectors auto
>>> - currently set to 256
>>> Block device 252:5
>>>
>>> --- Logical volume ---
>>> LV Name
>>> /dev/cinder-volumes/volume-00c532bd-91fb-4a38-b340-4389fb7f0ed5
>>> VG Name cinder-volumes
>>> LV UUID MfVOuB-5x5A-jne3-H4Ul-4NP8-eI7b-UYSYE7
>>> LV Write Access read/write
>>> LV Status available
>>> open 0
>>>
>>> LV Size 1,00 GiB
>>> Current LE 256
>>> Segments 1
>>> Allocation inherit
>>> Read ahead sectors auto
>>> - currently set to 256
>>> Block device 252:6
>>>
>>> --- Logical volume ---
>>> LV Name
>>> /dev/cinder-volumes/volume-ae133dbc-6141-48cf-beeb-9d6576e57a45
>>> VG Name cinder-volumes
>>> LV UUID 53w8j3-WT4V-8m52-r6LK-ZYd3-mMHA-FtuyXV
>>> LV Write Access read/write
>>> LV Status available
>>> open 0
>>>
>>> LV Size 1,00 GiB
>>> Current LE 256
>>> Segments 1
>>> Allocation inherit
>>> Read ahead sectors auto
>>> - currently set to 256
>>> Block device 252:7
>>>
>>> --- Logical volume ---
>>> LV Name
>>> /dev/cinder-volumes/volume-954d2f1b-837b-4ba5-abfd-b3610597be5e
>>> VG Name cinder-volumes
>>> LV UUID belquE-WxQ2-gt6Y-WlPE-Hmq3-B9Am-zcYD3P
>>> LV Write Access read/write
>>> LV Status available
>>> open 0
>>>
>>> LV Size 60,00 GiB
>>> Current LE 15360
>>> Segments 1
>>> Allocation inherit
>>> Read ahead sectors auto
>>> - currently set to 256
>>> Block device 252:8
>>>
>>> --- Logical volume ---
>>> LV Name
>>> /dev/cinder-volumes/volume-05d037d1-4e61-4419-929a-fe340e00e1af
>>> VG Name cinder-volumes
>>> LV UUID Pt61e7-l3Nu-1IdX-T2sb-0GQD-PhS6-XtIIUj
>>> LV Write Access read/write
>>> LV Status available
>>> open 1
>>>
>>> LV Size 1,00 GiB
>>> Current LE 256
>>> Segments 1
>>> Allocation inherit
>>> Read ahead sectors auto
>>> - currently set to 256
>>> Block device 252:9
>>>
>>> --- Logical volume ---
>>> LV Name
>>> /dev/cinder-volumes/volume-316f77c6-bf13-4ea4-9b98-028198f3922f
>>> VG Name cinder-volumes
>>> LV UUID e46mBx-CRps-HYKk-aJsc-XFRd-B1Rv-UVk8gT
>>> LV Write Access read/write
>>> LV Status available
>>> open 1
>>>
>>> LV Size 60,00 GiB
>>> Current LE 15360
>>> Segments 1
>>> Allocation inherit
>>> Read ahead sectors auto
>>> - currently set to 256
>>> Block device 252:10
>>>
>>> Do you know how I can remove all of them? When I try lvremove
>>> /dev/cinder-volumes/volume-316f77c6-bf13-4ea4-9b98-028198f3922f, for
>>> example, I get:
>>>
>>> Can't remove open logical volume
>>> "volume-316f77c6-bf13-4ea4-9b98-028198f3922f"
>>>
>>> Thank you again.
>>>
>>> 2013/11/8 Guilherme Russi luisguilherme.cr at gmail.com
>>>
>>> Very thanks again.
>>>
>>> Best regards.
>>>
>>> 2013/11/8 Razique Mahroua razique.mahroua at gmail.com
>>>
>>> Oh yah true!
>>> not sure “conductors” exist yet for Cinder, meaning meanwhile, every node
>>> needs a direct access to the database
>>> glad to hear it’s working :)
>>>
>>> On 08 Nov 2013, at 08:53, Guilherme Russi luisguilherme.cr at gmail.com
>>> wrote:
>>>
>>> Hello again Razique, I've found the problem, I need to add the grants on
>>> the mysql to my another IP. Now it's working really good :D
>>> I've found this link too if someone needs:
>>> http://docs.openstack.org/admin-guide-cloud/content//
>>> managing-volumes.html
>>>
>>> Thank you so much, and if you need me just let me know.
>>>
>>> Best regards.
>>>
>>> Guilherme.
>>>
>>> 2013/11/8 Guilherme Russi luisguilherme.cr at gmail.com
>>>
>>> Hello Razique, I got a couple of doubts, do you know if I need to do
>>> something else that's is not on the link you sent me? I'm asking because
>>> I
>>> followed the configuration but it's not working, here is what I get: I've
>>> installed the cinder-volume at the second computer that have the HD, and
>>> I've changed it's cinder.conf. I've changed too the master's cinder.conf
>>> like is following:
>>>
>>> [DEFAULT]
>>> rootwrap
>>> *config = /etc/cinder/rootwrap.conf sql*connection =
>>> mysql://cinder:password@localhost/cinder
>>> api*paste*confg = /etc/cinder/api-paste.ini
>>>
>>> iscsi_helper=iscsiadm iscsi_helper = ietadm
>>>
>>> iscsi
>>> *helper = tgtadm volume*name_template = volume-%s
>>>
>>> volume_group = cinder-volumes
>>>
>>> verbose = True
>>> auth
>>> *strategy = keystone iscsi*ip
>>> *address = 192.168.3.1 scheduler*
>>> driver=cinder.scheduler.filter_scheduler.FilterScheduler
>>> Rabbit authorization
>>>
>>> rabbit
>>> *host = localhost rabbit*port = 5672
>>> rabbit*hosts = $rabbit*host:$rabbit
>>> *port rabbit*use
>>> *ssl = false rabbit*userid = guest
>>> rabbit_password = password
>>> rabbit*virtual*host = /nova
>>>
>>> state
>>> *path = /var/lib/cinder lock*path = /var/lock/cinder
>>> volumes_dir = /var/lib/cinder/volumes
>>> rpc*backend = cinder.rpc.impl*kombu
>>>
>>> enabled
>>>
>>> *backends=orion-1,orion-4 [orion-1] volume*group=cinder-volumes
>>> volume
>>> *driver=cinder.volume.drivers.lvm.LVMISCSIDriver
>>> volume*backend*name=LVM*
>>> iSCSI
>>> [orion-4]
>>> volume
>>> *group=cinder-volumes-2 volume*
>>> driver=cinder.volume.drivers.lvm.LVMISCSIDriver
>>> volume*backend*name=LVM_iSCSI
>>>
>>>
>>> The cinder.conf on the second computer is like this but the IPs are
>>> changed with the controller IP (It has the cinder-api), and when I run
>>> service cinder-volume restart at the second computer it's status is
>>> stop/waiting.
>>>
>>> Any ideas?
>>>
>>> Thanks :)
>>>
>>> 2013/11/8 Razique Mahroua razique.mahroua at gmail.com
>>>
>>> sure :)
>>>
>>> On 08 Nov 2013, at 05:39, Guilherme Russi luisguilherme.cr at gmail.com
>>> wrote:
>>>
>>> Oh great! I'll try here and send you the results.
>>>
>>> Very thanks :)
>>>
>>> 2013/11/8 Razique Mahroua razique.mahroua at gmail.com
>>>
>>> If I’m not mistaken, you only need to install the “cinder-volume’
>>> service that will update its status to your main node
>>> :)
>>>
>>> On 08 Nov 2013, at 05:34, Guilherme Russi luisguilherme.cr at gmail.com
>>> wrote:
>>>
>>> Great! I was reading the link and I have one question, do I need to
>>> install cinder at the other computer too?
>>>
>>> Thanks :)
>>>
>>> 2013/11/8 Razique Mahroua razique.mahroua at gmail.com
>>>
>>> Ok in that case, with Grizzly you can use the “multi-backends”
>>> feature:
>>> https://wiki.openstack.org/wiki/Cinder-multi-backend
>>>
>>> and that should do it :)
>>>
>>> On 08 Nov 2013, at 05:29, Guilherme Russi luisguilherme.cr at gmail.com
>>> wrote:
>>>
>>> It is a hard disk, my scenario is one Controller (where I have my
>>> storage cinder and my network quantum) and four compute nodes.
>>>
>>> 2013/11/8 Razique Mahroua razique.mahroua at gmail.com
>>>
>>> ok !
>>> what is your actual Cinder backend? Is it a hard disk, a SAN, a
>>> network volume, etc…
>>>
>>> On 08 Nov 2013, at 05:20, Guilherme Russi <
>>> luisguilherme.cr at gmail.com> wrote:
>>>
>>> Hi Razique, thank you for answering, I want to expand my cinder
>>> storage, is it the block storage? I'll use the storage to allow VMs to
>>> have
>>> more hard disk space.
>>>
>>> Regards.
>>>
>>> Guilherme.
>>>
>>> 2013/11/8 Razique Mahroua razique.mahroua at gmail.com
>>>
>>> Hi Guilherme !
>>> Which storage do you precisely want to expand?
>>>
>>> Regards,
>>> Razique
>>>
>>> On 08 Nov 2013, at 04:52, Guilherme Russi <
>>> luisguilherme.cr at gmail.com> wrote:
>>>
>>> Hello guys, I have a Grizzly deployment running fine with 5
>>>
>>> nodes, and I want to add more storage on it. My question is, can I
>>> install
>>> a new HD on another computer thats not the controller and link this HD
>>> with
>>> my cinder that it can be a storage too?
>>>
>>> The computer I will install my new HD is at the same network as
>>>
>>> my cloud is. I'm asking because I haven't seen a question like that here.
>>> Does anybody knows how to do that? Have a clue? Any help is welcome.
>>>
>>> Thank you all.
>>>
>>> Best regards.
>>> ------------------------------
>>>
>>>
>>> Mailing list:
>>>
>>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
>>>
>>> Post to : openstack at lists.openstack.org
>>> Unsubscribe :
>>>
>>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
>>>
>>>
>>>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.openstack.org/pipermail/openstack/attachments/20131118/ae5a58b5/attachment.html>


More information about the Openstack mailing list