<div dir="ltr">On Thu, May 9, 2013 at 8:45 AM, Sean Dague <span dir="ltr"><<a href="mailto:sean@dague.net" target="_blank">sean@dague.net</a>></span> wrote:<br><div class="gmail_extra"><div class="gmail_quote"><blockquote class="gmail_quote" style="margin:0 0 0 .8ex;border-left:1px #ccc solid;padding-left:1ex">
<div class="im">On 05/09/2013 10:53 AM, Russell Bryant wrote:<br>
<blockquote class="gmail_quote" style="margin:0 0 0 .8ex;border-left:1px #ccc solid;padding-left:1ex">
Greetings,<br>
<br>
I've been growing concerned with the evolution of Nova's architecture in<br>
terms of the virt drivers and the impact they have on the rest of Nova.<br>
I've heard these concerns from others in private conversation. Another<br>
thread on the list today pushed me to where I think it's time we talk<br>
about it:<br>
<br>
<a href="http://lists.openstack.org/pipermail/openstack-dev/2013-May/008801.html" target="_blank">http://lists.openstack.org/<u></u>pipermail/openstack-dev/2013-<u></u>May/008801.html</a><br>
<br>
At our last design summit, there was a discussion of adding a new virt<br>
driver to support oVirt (RHEVM). That seems inappropriate for Nova to<br>
me. oVirt is a full virt management system and uses libvirt+KVM<br>
hypervisors. We use libvirt+KVM directly. Punting off to yet another<br>
management system that wants to manage all of the same things as<br>
OpenStack seems like a broken architecture. In fact, oVirt has done a<br>
lot of work to *consume* OpenStack resources (glance, quantum), which<br>
seems completely appropriate.<br>
</blockquote>
<br></div>
+1<div class="im"><br>
<br>
<blockquote class="gmail_quote" style="margin:0 0 0 .8ex;border-left:1px #ccc solid;padding-left:1ex">
Things get more complicated if we take that argument and apply it to<br>
other drivers that we already have in Nova. In particular, I think this<br>
applies to the VMware (vCenter mode, not ESX mode) and Hyper-V drivers.<br>
I'm not necessarily proposing that those drivers work significantly<br>
different. I don't think that's practical if we want to support these<br>
systems.<br>
<br>
We now have two different types of drivers: those that manage individual<br>
hypervisor nodes, and those that proxy to much more complex systems.<br>
<br>
We need to be very aware of what's going on in all virt drivers, even<br>
the ones we don't care about as much because we don't use them. We also<br>
need to continue to solidify the virt driver interface and be extremely<br>
cautious when these drivers require changes to other parts of Nova.<br>
Above all, let's make sure that evolution in this area is well thought<br>
out and done by conscious decision.<br>
<br>
Comments airing more specific concerns in this area would be appreciated.<br>
</blockquote>
<br></div>
I think we learned a really important lesson in baremetal: putting a different complex management system underneath the virt driver interface is a bad fit, requires nova to do unnatural things, and just doesn't make anyone happy at the end of the day. That's since resulted in baremetal spinning out to a new incubated project, Ironic, which I think is really the right long term approach.<br>
<br>
I think we need to take that lesson for what it was, and realize these virt cluster drivers are very much the same kind of problem. They are better served living in some new incubated effort instead of force fitting into the nova-compute virt layer and driving a lot more complexity into nova.</blockquote>
<div><br></div><div style>I don't feel like a new project is needed here -- the ongoing discussion about moving scheduling/orchestration logic out of nova-compute and into conductor-or-something-else seems to frame this discussion, too. </div>
<div style><br></div><div style>The biggest change to Nova that I recall around adding the Baremetal code was the addition of the "node" aka "hypervisor_hostname" concept -- that a single nova compute host might control more than one discrete thing, which thereby need to be identified as (host, node). That change opened the door for other cluster drivers to fit under the virt interface. IMBW, but I believe this is exactly what the vCenter and Hyper-V folks are looking at. It's also my current plan for Ironic. However, I also believe that this logic doesn't necessarily have to live under the virt API layer; I think it's a really good fit for the orchestration/conductor discussions....</div>
<div style><br></div><div style>We were talking about this a few days ago in -nova, particularly how moving some of the ComputeManager logic out to conductor might fit together with simplifying the (host, node) complexities, and help make nova-compute just a thin virt API layer. Here is a very poor summary of what I recall...</div>
<div style>* AMQP topic is based on "nodename", not "hostname"</div><div style>* for local hypervisors (KVM, etc), the topic identifies the local host, and the local nova-compute agent subscribes to it</div>
<div style>* for clustered hypervisors (ironic, vCenter, etc), the topic identifies the unique resource, and any nova-compute which can manage that resource subscribes to the topic.</div><div style><br></div><div style>This would also remove the SPoF that nova-compute currently has for any cluster-of-discrete-things it manages today (eg, baremetal).</div>
<div style><br></div><div style><br></div><div style>Thoughts?</div><div style><br></div><div style>-Devananda</div><div style><br></div></div></div></div>