[openstack-hpc] Infiniband + Openstack?

Narayan Desai narayan.desai at gmail.com
Thu Jan 24 17:25:51 UTC 2013


I think that this sounds sensible, so long as you get one of the
switches that can run in either mode, so that you don't need to change
hardware configurations later.
 -nld

On Thu, Jan 24, 2013 at 2:16 AM, zebra z zebra <zebra.x0r at gmail.com> wrote:
> Hi.
>
> So, we've had a further idea for our build out process.
>
> We'd like a "sanity check" from the list to see if it makes sense.
>
> Based upon what JP has said below, it gives us confidence that, in time,
> things will work "end to end" from an OFED + RDMA + SRV-IO perspective
> (and IPoIB will work too, one day). Currently though, we aren't there yet.
>
> So, to that end, we've had an idea.
>
> What if, for the time being, we run the ConnectX-3 IB cards in Ethernet
> personality, reap all the benefits of a 40Gbit/sec ethernet plumbing
> environment, and then, when the time comes, and OpenStack is at the point
> where it'll really hum nicely in terms of full IB integration, we can
> simply flash the personality of the IB card to IB mode and run with that.
>
> Seems like a sensible way to wade into Infiniband, but have a functional
> system from the get go, then have all the benefits of IB later.
>
> Comments/thoughts?
>
> Thank you, all.
>
> z
>
>
>
> On 24/01/13 7:02 AM, "John Paul Walters" <jwalters at isi.edu> wrote:
>
>>Z,
>>
>>We're working on integrating IB into OpenStack via SR-IOV.  It's not
>>quite there yet, but we expect to have it working within a month or so.
>>Right now, I'm unaware of anyone who has IB working inside of a VM,
>>that's what we're developing.  To do this, you need a few things: 1) a
>>host that supports SR-IOV, 2) a ConnectX-2 or ConnectX-3 IB card, and 3)
>>Alpha SR-IOV-enabled firmware for the cards as well as the corresponding
>>SR-IOV-enabled OFED.  Both of those (the firmware and OFED) need to come
>>from Mellanox at the moment - they're not yet released.  We've so far
>>been successful in getting RDMA working inside of the VM; however, oddly,
>>IPoIB doesn't work (the ports are recognized as being connected).  We
>>hope that this will get worked out in later firmware/OFED releases.
>>
>>Be aware that this isn't a Quantum integration, at least not yet.  We're
>>going to manage SR-IOV VIFs as resources and schedule accordingly.  I
>>know that there are folks using IB for image distribution and perhaps
>>also for block storage.  Our target is to get IB into the VM.
>>
>>JP
>>
>>
>>On Jan 23, 2013, at 3:36 PM, zebra z zebra <zebra.x0r at gmail.com> wrote:
>>
>>> Hi.
>>>
>>> I've been trying to get some simple/straight answers on IB for OpenStack
>>> for a few days now. It could well be that I've just been posting to the
>>> wrong list! ;)
>>>
>>> I'm considering a reasonably sized OpenStack deployment shortly (several
>>> thousand cores). One of the decisions I need to make centres around the
>>> type of interconnect I will use to connect the compute nodes, Swift
>>> objectstorage, shared storage, swift proxies et al.
>>>
>>> I have been heavily considering FDR infiniband for this task, but, from
>>> what I can see, IB has only recently become something one might consider
>>> stable for use in OpenStack. My peers are suggesting to me that I should
>>> just stick with 10GbE interconnects and be happy.
>>>
>>> In doing some background reading, it looks like somebody asked similar
>>> questions a year or two ago, and the general sentiment is that it'd work
>>> in IPoIB mode as just a very fast ethernet device, but it wouldn't
>>>really
>>> have any of the benefits of RDMA communication, or any such capability -
>>> and the issue at the time was that no python libraries had any
>>> understanding or concept of it.
>>>
>>> I'd like to hear the experiences of people who have or are using IB, or
>>> who have experiences with IB deployments.
>>>
>>> Thank you.
>>>
>>> --z
>>>
>>>
>>>
>>> _______________________________________________
>>> OpenStack-HPC mailing list
>>> OpenStack-HPC at lists.openstack.org
>>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-hpc
>>
>
>
>
> _______________________________________________
> OpenStack-HPC mailing list
> OpenStack-HPC at lists.openstack.org
> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-hpc



More information about the OpenStack-HPC mailing list