[Openstack-operators] Shared storage HA question
Sylvain Bauza
sylvain.bauza at bull.net
Thu Jul 25 07:22:51 UTC 2013
Hi Denis,
As per my short testings, I would assume anything but FUSE-mounted would
match your needs.
On the performance glance, here is what I could suggest :
- if using GlusterFS, wait for this BP [1] to be implemented. I do
agree with Razique on the issues you could face with GlusterFS, this is
mainly due to the Windows caching system mixed with QCOW2 copy-on-write
images relying on a FUSE mountpoint.
- if using Ceph, use RADOS to boot from Cinder volumes. Don't use FUSE
mountpoint, again.
-Sylvain
[1] https://blueprints.launchpad.net/nova/+spec/glusterfs-native-support
Le 25/07/2013 08:16, Denis Loshakov a écrit :
> So, first i'm going to try Ceph.
> Thanks for advices and lets RTFM begin :)
>
> On 24.07.2013 23:18, Razique Mahroua wrote:
>> +1 :)
>>
>>
>> Le 24 juil. 2013 à 21:08, Joe Topjian <joe.topjian at cybera.ca
>> <mailto:joe.topjian at cybera.ca>> a écrit :
>>
>>> Hi Jacob,
>>>
>>> Are you using SAS or SSD drives for Gluster? Also, do you have one
>>> large Gluster volume across your entire cloud or is it broke up into a
>>> few different ones? I've wondered if there's a benefit to doing the
>>> latter so distribution activity is isolated to only a few nodes. The
>>> downside to that, of course, is you're limited to what compute nodes
>>> instances can migrate to.
>>>
>>> I use Gluster for instance storage in all of my "controlled"
>>> environments like internal and sandbox clouds, but I'm hesitant to
>>> introduce it into production environments as I've seen the same issues
>>> that Razique describes -- especially with Windows instances. My guess
>>> is due to how NTFS writes to disk.
>>>
>>> I'm curious if you could report the results of the following test: in
>>> a Windows instance running on Gluster, copy a 3-4gb file to it from
>>> the local network so it comes in at a very high speed. When I do this,
>>> the first few gigs come in very fast, but then slows to a crawl and
>>> the Gluster processes on all nodes spike.
>>>
>>> Thanks,
>>> Joe
>>>
>>>
>>>
>>> On Wed, Jul 24, 2013 at 12:37 PM, Jacob Godin <jacobgodin at gmail.com
>>> <mailto:jacobgodin at gmail.com>> wrote:
>>>
>>> Oh really, you've done away with Gluster all together? The fast
>>> backbone is definitely needed, but I would think that was the case
>>> with any distributed filesystem.
>>>
>>> MooseFS looks promising, but apparently it has a few reliability
>>> problems.
>>>
>>>
>>> On Wed, Jul 24, 2013 at 3:31 PM, Razique Mahroua
>>> <razique.mahroua at gmail.com <mailto:razique.mahroua at gmail.com>>
>>> wrote:
>>>
>>> :-)
>>> Actually I had to remove all my instances running on it
>>> (especially the windows ones), yah unfortunately my network
>>> backbone wasn't fast enough to support the load induced by GFS
>>> - especially the numerous operations performed by the
>>> self-healing agents :(
>>>
>>> I'm currently considering MooseFS, it has the advantage to
>>> have a pretty long list of companies using it in production
>>>
>>> take care
>>>
>>>
>>> Le 24 juil. 2013 à 16:40, Jacob Godin <jacobgodin at gmail.com
>>> <mailto:jacobgodin at gmail.com>> a écrit :
>>>
>>>> A few things I found were key for I/O performance:
>>>>
>>>> 1. Make sure your network can sustain the traffic. We are
>>>> using a 10G backbone with 2 bonded interfaces per node.
>>>> 2. Use high speed drives. SATA will not cut it.
>>>> 3. Look into tuning settings. Razique, thanks for sending
>>>> these along to me a little while back. A couple that I
>>>> found were useful:
>>>> * KVM cache=writeback (a little risky, but WAY faster)
>>>> * Gluster write-behind-window-size (set to 4MB in our
>>>> setup)
>>>> * Gluster cache-size (ideal values in our setup were
>>>> 96MB-128MB)
>>>>
>>>> Hope that helps!
>>>>
>>>>
>>>>
>>>> On Wed, Jul 24, 2013 at 11:32 AM, Razique Mahroua
>>>> <razique.mahroua at gmail.com
>>>> <mailto:razique.mahroua at gmail.com>> wrote:
>>>>
>>>> I had much performance issues myself with Windows
>>>> instances, and I/O demanding instances. Make sure it fits
>>>> your env. first before deploying it in production
>>>>
>>>> Regards,
>>>> Razique
>>>>
>>>> *Razique Mahroua** - **Nuage & Co*
>>>> razique.mahroua at gmail.com
>>>> <mailto:razique.mahroua at gmail.com>
>>>> Tel : +33 9 72 37 94 15
>>>> <tel:%2B33%209%2072%2037%2094%2015>
>>>>
>>>> <NUAGECO-LOGO-Fblan_petit.jpg>
>>>>
>>>> Le 24 juil. 2013 à 16:25, Jacob Godin
>>>> <jacobgodin at gmail.com <mailto:jacobgodin at gmail.com>> a
>>>> écrit :
>>>>
>>>>> Hi Denis,
>>>>>
>>>>> I would take a look into GlusterFS with a distributed,
>>>>> replicated volume. We have been using it for several
>>>>> months now, and it has been stable. Nova will need to
>>>>> have the volume mounted to its instances directory
>>>>> (default /var/lib/nova/instances), and Cinder has direct
>>>>> support for Gluster as of Grizzly I believe.
>>>>>
>>>>>
>>>>>
>>>>> On Wed, Jul 24, 2013 at 11:11 AM, Denis Loshakov
>>>>> <dloshakov at gmail.com <mailto:dloshakov at gmail.com>> wrote:
>>>>>
>>>>> Hi all,
>>>>>
>>>>> I have issue with creating shared storage for
>>>>> Openstack. Main idea is to create 100% redundant
>>>>> shared storage from two servers (kind of network
>>>>> RAID from two servers).
>>>>> I have two identical servers with many disks inside.
>>>>> What solution can any one provide for such schema? I
>>>>> need shared storage for running VMs (so live
>>>>> migration can work) and also for cinder-volumes.
>>>>>
>>>>> One solution is to install Linux on both servers and
>>>>> use DRBD + OCFS2, any comments on this?
>>>>> Also I heard about Quadstor software and it can
>>>>> create network RAID and present it via iSCSI.
>>>>>
>>>>> Thanks.
>>>>>
>>>>> P.S. Glance uses swift and is setuped on another
>>>>> servers
>>>>>
>>>>> _________________________________________________
>>>>> OpenStack-operators mailing list
>>>>> OpenStack-operators at lists.__openstack.org
>>>>> <mailto:OpenStack-operators at lists.openstack.org>
>>>>> http://lists.openstack.org/__cgi-bin/mailman/listinfo/__openstack-operators
>>>>> <http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators>
>>>>>
>>>>>
>>>>> _______________________________________________
>>>>> OpenStack-operators mailing list
>>>>> OpenStack-operators at lists.openstack.org
>>>>> <mailto:OpenStack-operators at lists.openstack.org>
>>>>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators
>>>>>
>>>>
>>>>
>>>
>>>
>>>
>>> _______________________________________________
>>> OpenStack-operators mailing list
>>> OpenStack-operators at lists.openstack.org
>>> <mailto:OpenStack-operators at lists.openstack.org>
>>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators
>>>
>>>
>>>
>>>
>>> --
>>> Joe Topjian
>>> Systems Architect
>>> Cybera Inc.
>>>
>>> www.cybera.ca <http://www.cybera.ca/>
>>>
>>> Cybera is a not-for-profit organization that works to spur and support
>>> innovation, for the economic benefit of Alberta, through the use
>>> of cyberinfrastructure.
>>
>>
>>
>> _______________________________________________
>> OpenStack-operators mailing list
>> OpenStack-operators at lists.openstack.org
>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators
>>
>
> _______________________________________________
> OpenStack-operators mailing list
> OpenStack-operators at lists.openstack.org
> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.openstack.org/pipermail/openstack-operators/attachments/20130725/e40ad146/attachment.html>
More information about the OpenStack-operators
mailing list