[openstack-dev] [nova] Do we assume that the value of instances_path is the same on all compute nodes?

Kashyap Chamarthy kchamart at redhat.com
Wed May 15 13:10:47 UTC 2013


On 05/15/2013 05:38 AM, Michael Still wrote:
> On Tue, May 14, 2013 at 11:26 PM, Russell Bryant <rbryant at redhat.com> wrote:
>> On 05/14/2013 02:52 AM, Michael Still wrote:
>>> Hi.
>>>
>>> Russell -- I was reviewing https://review.openstack.org/#/c/28950/,
>>> and I think I found a new bug. This code assumes that both the source
>>> and destination for a migration have the same value set for
>>> CONF.instances_path.
>>>
>>> So... Do we think that's true? Do we assume that this value is set to
>>> the same value deployment wide? I can see cases where that might be
>>> surprising to deployers.
>>>
>>> I think we need a policy discussion here, so I'm hoping this thread
>>> will help us decide.
>>
>> Is there ever a case where *not* setting it to the same thing is useful?
> 
> I'm not sure. I can think of cases where admins might have different
> sized filesystems depending on how they setup machines, but surely if
> you're managing a cluster of many machines you'd standardise them all
> as much as possible.
> 
>> So, let's start with this:
>>
>>   We assume that the value of the instances_path configuration option
>>   is the same across an entire migration domain, meaning that we expect
>>   it to be the same for the source and destination of any migration.  In
>>   practice this may mean across a host aggregate, a cell, or an entire
>>   deployment, depending on the specifics of the deployment.
>>
>> Now, poke holes in it.  Is it bad?  Why?  What code makes this
>> assumption now that would have to change?
> 
> I think the biggest reason its bad is that its an opportunity for
> confusion -- migrations fail if this isn't true, but its not obvious
> to a deployer that they're failing for this reason. Perhaps we should
> test this assumption in the code as part of the shared storage
> detection?

Interesting I see this discussion here - just now, I changed the instances_path to an NFS
share, added the right SELinux context, made nova: owner of it, restarted all instances,
ensured qpid is up, etc.

Setup: All services on a single host, Grizzly, RDO.

Since then, it's been failing w/ errors like below:
======
[ ~(keystone_admin)]# tail  /var/log/nova/conductor.log
2012-04-18 07:31:56.893 11196 CRITICAL nova [-] need more than 0 values to unpack
2012-04-18 09:41:03.738 41546 CRITICAL nova [-] need more than 0 values to unpack
======

I'm trying to debug this issue w/ Flavio Percoco.

> 
>> In the mean time, it would be good to document this expectation in the
>> config.
> 
> For sure.
> 
> Michael
> 
> _______________________________________________
> OpenStack-dev mailing list
> OpenStack-dev at lists.openstack.org
> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
> 


-- 
/kashyap



More information about the OpenStack-dev mailing list