[Openstack-operators] Small openstack (part 2), distributed glance
George Shuklin
george.shuklin at gmail.com
Thu Jan 15 23:16:45 UTC 2015
We do not using centralized storages (all instances running with local
drives). And I just can't express my happiness about this. Every time
monitoring send me '** PROBLEM ALERT bla-bla-bla', I know it not a big
deal. Just one server.
I do not want to turn gray prematurely. Just light glance on
https://www.google.com/search?q=ceph+crash+corruption give me strong
feeling I don't want to centralize points of failures.
Btw: If I sold nodes designated for Ceph as normal compute nodes, it
will be more effective than sell only space from them (and buy more
compute nodes for actual work).
On 01/16/2015 12:31 AM, Abel Lopez wrote:
> That specific bottleneck can be solved by running glance on ceph, and
> running ephemeral instances also on ceph. Snapshots are a quick
> backend operation then. But you've made your installation on a house
> of cards.
>
> On Thursday, January 15, 2015, George Shuklin
> <george.shuklin at gmail.com <mailto:george.shuklin at gmail.com>> wrote:
>
> Hello everyone.
>
> One more thing in the light of small openstack.
>
> I really dislike tripple network load caused by current glance
> snapshot operations. When compute do snapshot, it playing with
> files locally, than it sends them to glance-api, and (if glance
> API is linked to swift), glance sends them to swift. Basically,
> for each 100Gb disk there is 300Gb on network operations. It is
> specially painful for glance-api, which need to get more CPU and
> network bandwidth than we want to spend on it.
>
> So idea: put glance-api on each compute node without cache.
>
> To help compute to go to the proper glance, endpoint points to
> fqdn, and on each compute that fqdn is pointing to localhost
> (where glance-api is live). Plus normal glance-api on
> API/controller node to serve dashboard/api clients.
>
> I didn't test it yet.
>
> Any ideas on possible problems/bottlenecks? And how many
> glance-registry I need for this?
>
> _______________________________________________
> OpenStack-operators mailing list
> OpenStack-operators at lists.openstack.org
> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.openstack.org/pipermail/openstack-operators/attachments/20150116/f033f938/attachment.html>
More information about the OpenStack-operators
mailing list