[openstack-dev] [nova] [neutron] PCI pass-through network support
john at johngarbutt.com
Thu Dec 19 11:34:34 UTC 2013
Apologies for being late onto this thread, and not making the meeting
the other day.
Also apologies this is almost totally a top post.
On 17 December 2013 15:09, Ian Wells <ijw.ubuntu at cack.org.uk> wrote:
> Firstly, I disagree that
> https://wiki.openstack.org/wiki/PCI_passthrough_SRIOV_support is an accurate
> reflection of the current state. It's a very unilateral view, largely
> because the rest of us had been focussing on the google document that we've
> been using for weeks.
I haven't seen the google doc. I got involved through the blueprint
review of this:
I assume its this one?
On a quick read, my main concern is separating out the "user" more:
* administration (defines pci-flavor, defines which hosts can provide
it, defines server flavor...)
* person who boots server (picks server flavor, defines neutron ports)
Note, I don't see the person who boots the server ever seeing the
pci-flavor, only understanding the server flavor.
We might also want a "nic-flavor" that tells neutron information it
requires, but lets get to that later...
> Secondly, I totally disagree with this approach. This assumes that
> description of the (cloud-internal, hardware) details of each compute node
> is best done with data stored centrally and driven by an API. I don't agree
> with either of these points.
Possibly, but I would like to first agree on the use cases and data
model we want.
Nova has generally gone for APIs over config in recent times.
Mostly so you can do run-time configuration of the system.
But lets just see what makes sense when we have the use cases agreed.
>> On 2013年12月16日 22:27, Robert Li (baoli) wrote:
>> I'd like to give you guy a summary of current state, let's discuss it
>> 1) fade out alias ( i think this ok for all)
>> 2) white list became pic-flavor ( i think this ok for all)
>> 3) address simply regular expression support: only * and a number range
>> is support [hex-hex]. ( i think this ok?)
>> 4) aggregate : now it's clear enough, and won't impact SRIOV. ( i think
>> this irrelevant to SRIOV now)
So... this means we have:
* i.e. standardGPU, standardGPUnew, fastGPU, hdFlash1TB etc
* decide which hosts you allow a particular flavor to be used
* note, the scheduler still needs to find out if any devices are "free"
flavor (of the server):
* usual RAM, CPU, Storage
* use extra specs to add PCI devices
** add one PCI device, choice of standardGPU or standardGPUnew
** also add: one hdFlash1TB
Now, the other bit is SRIOV... At a high level:
* user wants to connect to a particular neutron network
* user wants a super-fast SRIOV connection
* needs to map PCI device to what neutron network the connect to
The big question is:
* is this a specific SRIOV only (provider) network
* OR... are other non-SRIOV connections also made to that same network
I feel we have to go for that latter. Imagine a network on VLAN 42,
you might want some SRIOV into that network, and some OVS connecting
into the same network. The user might have VMs connected using both
methods, so wants the same IP address ranges and same network id
If we go for that latter new either need:
* some kind of nic-flavor
** boot ... -nic nic-id:"public-id:,nic-flavor:"10GBpassthrough"
** but neutron could store nic-flavor, and pass it through to VIF
driver, and user says port-id
* OR add NIC config into the server flavor
** extra spec to say, tell VIF driver it could use on of this list of
PCI devices: (list pci-flavors)
* OR do both
I vote for nic-flavor only, because it matches the volume-type we have
However, it does suggest that Nova should leave all the SRIOV work to
the VIF driver.
So the VIF driver, as activate by neutron, will understand which PCI
devices to passthrough.
Similar to the plan with brick, we could have an oslo lib that helps
you attach SRIOV devices that could be used by the neturon VIF drivers
and the nova PCI passthrough code.
More information about the OpenStack-dev