[openstack-dev] [TripleO] Introspection rules aka advances profiles replacement: next steps
John Trowbridge
trown at redhat.com
Wed Oct 14 17:18:06 UTC 2015
On 10/14/2015 10:57 AM, Ben Nemec wrote:
> On 10/14/2015 06:38 AM, Dmitry Tantsur wrote:
>> Hi OoO'ers :)
>>
>> It's going to be a long letter, fasten your seat-belts (and excuse my
>> bad, as usual, English)!
>>
>> In RDO Manager we used to have a feature called advanced profiles
>> matching. It's still there in the documentation at
>> http://docs.openstack.org/developer/tripleo-docs/advanced_deployment/profile_matching.html
>> but the related code needed reworking and didn't quite make it upstream
>> yet. This mail is an attempt to restart the discussion on this topic.
>>
>> Short explanation for those unaware of this feature: we used detailed
>> data from introspection (acquired using hardware-detect utility [1]) to
>> provide scheduling hints, which we called profiles. A profile is
>> essentially a flavor, but calculated using much more data. E.g. you
>> could sat that a profile "foo" will be assigned to nodes with 1024 <=
>> RAM <= 4096 and with GPU devices present (an artificial example).
>> Profile was put on an Ironic as a capability as a result of
>> introspection. Please read the documentation linked above for more details.
>>
>> This feature had a bunch of problems with it, to name a few:
>> 1. It didn't have an API
>> 2. It required a user to modify files by hand to use it
>> 3. It was tied to a pretty specific syntax of the hardware [1] library
>>
>> So we decided to split this thing into 2 parts, which are of value one
>> their own:
>>
>> 1. Pluggable introspection ramdisk - so that we don't force dependency
>> on hardware-detect on everyone.
>> 2. User-defined introspection rules - some DSL that will allow a user to
>> define something like a specs file (see link above) via an API. The
>> outcome would be something, probably capabilit(y|ies) set on a node.
>> 3. Scheduler helper - an utility that will take capabilities set by the
>> previous step, and turn them into exactly one profile to use.
>>
>> Long story short, we got 1 and 2 implemented in appropriate projects
>> (ironic-python-agent and ironic-inspector) during the Liberty time
>> frame. Now it's time to figure out what we do in TripleO about this, namely:
>>
>> 1. Do we need some standard way to define introspection rules for
>> TripleO? E.g. a JSON file like we have for ironic nodes?
>
> Yes, please.
>
>>
>> 2. Do we need a scheduler helper at all? We could use only capabilities
>> for scheduling, but then we can end up with the following situation:
>> node1 has capabilities C1 and C2, node2 has capability C1. First we
>> deploy a flavor with capability C1, it goes to node1. Then we deploy a
>> flavor with capability C2 and it fails, despite us having 2 correct
>> nodes initially. This is what state files were solving in [1] (again,
>> please refer to the documentation).
>
> It sounds like the answer is yes. If the existing scheduler can't
> handle a valid use case then we need some sort of solution.
>
>>
>> 3. If we need, where does it go? tripleo-common? Do we need an HTTP API
>> for it, or do we just do it in place where we need it? After all, it's a
>> pretty trivial manipulation with ironic nodes..
>
> I think that would depend on what the helper ends up being. I can't see
> it needing a REST API, but presumably it will have to plug into Nova
> somehow. If it's something that would be generally useful (which it
> sounds like it might be - Ironic capabilities aren't a TripleO-specific
> thing), then it belongs in Nova itself IMHO.
>
>>
>> 4. Finally, we need an option to tell introspection to use
>> python-hardware. I don't think it should be on by default, but it will
>> require rebuilding of IPA (due to a new dependency).
>
> Can we not just build it in always, but only use it when desired? Is
> the one extra dependency that much of a burden?
It pulls in python-numpy and python-pandas, which are pretty large.
>
>>
>> Looking forward to your opinions.
>> Dmitry.
>>
>> [1] https://github.com/redhat-cip/hardware
>>
>> __________________________________________________________________________
>> OpenStack Development Mailing List (not for usage questions)
>> Unsubscribe: OpenStack-dev-request at lists.openstack.org?subject:unsubscribe
>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
>>
>
>
> __________________________________________________________________________
> OpenStack Development Mailing List (not for usage questions)
> Unsubscribe: OpenStack-dev-request at lists.openstack.org?subject:unsubscribe
> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
>
More information about the OpenStack-dev
mailing list