[openstack-dev] [fuel][ConfigDB] Separating node and cluster serialized data
xarses at gmail.com
Thu Mar 31 20:19:38 UTC 2016
One of the problems we've faced with trying to plug-in ConfigDB is trying
to separate the cluster attributes from the node attributes in the
serialized output (ie astute.yaml)
I started talking with Alex S about how we could separate them after
astute.yaml is prepared trying to ensure which was which we came back
uncertain that the results would be accurate.
So I figured I'd go back to the source and see if there was a way to know
which keys belonged where. It turns out that we could solve the problem in
a simpler and more precise way than cutting them back apart later.
Looking over the deployment_serializers.py  the serialized data follows
a simple work flow
iterate over every node in cluster
if node is customized:
serialized_data = node.replaced_deployment_data
serialized_data = dict_merge(
Taking this into mind, we can simply construct an extension to expose these
as an APIs so that we can consume them as a task in the deployment graph.
We can simply expose
This would then be plumbed to the cluster level in ConfigDB
if a Node has customized data, then we can return that at the node level,
this continues to work at the same as native since it most likely has
Cluster merged into it.
otherwise we can return the serialized node with whichever of the first
'role' the node has
We would expose DeploymentMultinodeSerializer().serialize_node(node,
for our usage, we don't need to worry about the normal node role
combination as the data only influences 'role' and 'fail_if_error'
attributes, both are not consumed in the library.
Fuel Community Ambassador
-------------- next part --------------
An HTML attachment was scrubbed...
More information about the OpenStack-dev