[Openstack-operators] Shared storage HA question
Stephane Boisvert
stephane.boisvert at gameloft.com
Wed Jul 24 15:30:07 UTC 2013
Thanks for the quick answer.. I already did it but it seems not to be
taken in account... I'll test it again and open a new thread if I fail.
Thanks Jacob
On 13-07-24 11:20 AM, Jacob Godin wrote:
> Hi Stephane,
>
> This is actually done in Nova with the config
> directive disk_cachemodes="file=writeback"
>
>
> On Wed, Jul 24, 2013 at 11:47 AM, Stephane Boisvert
> <stephane.boisvert at gameloft.com
> <mailto:stephane.boisvert at gameloft.com>> wrote:
>
> sorry to interfere in that thread but I did set cache=true in my
> ceph config... but where I can set cache=writeback ?
>
>
> thanks for your help
>
> On 13-07-24 10:40 AM, Jacob Godin wrote:
>> A few things I found were key for I/O performance:
>>
>> 1. Make sure your network can sustain the traffic. We are using
>> a 10G backbone with 2 bonded interfaces per node.
>> 2. Use high speed drives. SATA will not cut it.
>> 3. Look into tuning settings. Razique, thanks for sending these
>> along to me a little while back. A couple that I found were
>> useful:
>> * KVM cache=writeback (a little risky, but WAY faster)
>> * Gluster write-behind-window-size (set to 4MB in our setup)
>> * Gluster cache-size (ideal values in our setup were
>> 96MB-128MB)
>>
>> Hope that helps!
>>
>>
>>
>> On Wed, Jul 24, 2013 at 11:32 AM, Razique Mahroua
>> <razique.mahroua at gmail.com <mailto:razique.mahroua at gmail.com>> wrote:
>>
>> I had much performance issues myself with Windows instances,
>> and I/O demanding instances. Make sure it fits your env.
>> first before deploying it in production
>>
>> Regards,
>> Razique
>>
>> *Razique Mahroua** - **Nuage & Co*
>> razique.mahroua at gmail.com <mailto:razique.mahroua at gmail.com>
>> Tel : +33 9 72 37 94 15
>>
>>
>> Le 24 juil. 2013 à 16:25, Jacob Godin <jacobgodin at gmail.com
>> <mailto:jacobgodin at gmail.com>> a écrit :
>>
>>> Hi Denis,
>>>
>>> I would take a look into GlusterFS with a distributed,
>>> replicated volume. We have been using it for several months
>>> now, and it has been stable. Nova will need to have the
>>> volume mounted to its instances directory (default
>>> /var/lib/nova/instances), and Cinder has direct support for
>>> Gluster as of Grizzly I believe.
>>>
>>>
>>>
>>> On Wed, Jul 24, 2013 at 11:11 AM, Denis Loshakov
>>> <dloshakov at gmail.com <mailto:dloshakov at gmail.com>> wrote:
>>>
>>> Hi all,
>>>
>>> I have issue with creating shared storage for Openstack.
>>> Main idea is to create 100% redundant shared storage
>>> from two servers (kind of network RAID from two servers).
>>> I have two identical servers with many disks inside.
>>> What solution can any one provide for such schema? I
>>> need shared storage for running VMs (so live migration
>>> can work) and also for cinder-volumes.
>>>
>>> One solution is to install Linux on both servers and use
>>> DRBD + OCFS2, any comments on this?
>>> Also I heard about Quadstor software and it can create
>>> network RAID and present it via iSCSI.
>>>
>>> Thanks.
>>>
>>> P.S. Glance uses swift and is setuped on another servers
>>>
>>> _______________________________________________
>>> OpenStack-operators mailing list
>>> OpenStack-operators at lists.openstack.org
>>> <mailto:OpenStack-operators at lists.openstack.org>
>>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators
>>>
>>>
>>> _______________________________________________
>>> OpenStack-operators mailing list
>>> OpenStack-operators at lists.openstack.org
>>> <mailto:OpenStack-operators at lists.openstack.org>
>>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators
>>
>>
>>
>>
>> _______________________________________________
>> OpenStack-operators mailing list
>> OpenStack-operators at lists.openstack.org <mailto:OpenStack-operators at lists.openstack.org>
>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators
>
>
> --
> *Stéphane Boisvert*
> GNS-Shop Technical Coordinator
> 5800 St-Denis suite 1001
> Montreal (QC), H2S 3L5
> *MSN:*stephane.boisvert at gameloft.com
> <mailto:stephane.boisvert at gameloft.com>
> *E-mail:*stephane.boisvert at gameloft.com
> <mailto:stephane.boisvert at gameloft.com>
>
>
> _______________________________________________
> OpenStack-operators mailing list
> OpenStack-operators at lists.openstack.org
> <mailto:OpenStack-operators at lists.openstack.org>
> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators
>
>
--
*Stéphane Boisvert*
GNS-Shop Technical Coordinator
5800 St-Denis suite 1001
Montreal (QC), H2S 3L5
*MSN:*stephane.boisvert at gameloft.com
*E-mail:*stephane.boisvert at gameloft.com
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.openstack.org/pipermail/openstack-operators/attachments/20130724/b03d2008/attachment.html>
-------------- next part --------------
A non-text attachment was scrubbed...
Name: not available
Type: image/jpeg
Size: 10122 bytes
Desc: not available
URL: <http://lists.openstack.org/pipermail/openstack-operators/attachments/20130724/b03d2008/attachment.jpe>
-------------- next part --------------
A non-text attachment was scrubbed...
Name: not available
Type: image/jpeg
Size: 8437 bytes
Desc: not available
URL: <http://lists.openstack.org/pipermail/openstack-operators/attachments/20130724/b03d2008/attachment-0001.jpe>
-------------- next part --------------
A non-text attachment was scrubbed...
Name: Inbox.jpg
Type: image/jpeg
Size: 8437 bytes
Desc: not available
URL: <http://lists.openstack.org/pipermail/openstack-operators/attachments/20130724/b03d2008/attachment.jpg>
More information about the OpenStack-operators
mailing list