[openstack-ansible][magnum]

Ignazio Cassano ignaziocassano at gmail.com
Tue Jan 29 15:25:04 UTC 2019


Hello,
in openstack there are a lot of newtroks .
Your 172.29 blabla bla network is probably the network where openstack
endpoint are exposed , right ?
If yes, that is not the network where virtual machine are attached.
In your openstack ym must have also networks for virtual machines.
When you create a magnum cluster, yum must specify an external netowrok
used by virtual machine for download packages from internet and to be
contacted .
Magnum create private netowrk (probablly your 10.1.8 network) which is
connected to the external network by a virtual router created by magnum
heat template.
Try to see your network topology in openstack dashboard.
Ignazio



Il giorno mar 29 gen 2019 alle ore 16:08 Alfredo De Luca <
alfredo.deluca at gmail.com> ha scritto:

> Hi Ignazio and Clemens. I haven\t configure the proxy  and all the logs on
> the kube master keep saying the following
>
> + '[' ok = '[-]poststarthook/bootstrap-controller failed: not finished
> [+]poststarthook/extensions/third-party-resources ok
> [-]poststarthook/rbac/bootstrap-roles failed: not finished
> healthz check failed' ']'
> + sleep 5
> ++ curl --silent http://127.0.0.1:8080/healthz
> + '[' ok = '' ']'
> + sleep 5
> ++ curl --silent http://127.0.0.1:8080/healthz
> + '[' ok = '[-]poststarthook/bootstrap-controller failed: not finished
> [+]poststarthook/extensions/third-party-resources ok
> [-]poststarthook/rbac/bootstrap-roles failed: not finished
> healthz check failed' ']'
> + sleep 5
>
> Not sure what to do.
> My configuration is ...
> eth0 - 10.1.8.113
>
> But the openstack configration in terms of networkin is the default from
> ansible-openstack which is 172.29.236.100/22
>
> Maybe that's the problem?
>
>
>
>
>
>
> On Tue, Jan 29, 2019 at 2:26 PM Ignazio Cassano <ignaziocassano at gmail.com>
> wrote:
>
>> Hello Alfredo,
>> your external network is using proxy ?
>> If you using a proxy, and yuo configured it in cluster template, you must
>> setup no proxy for 127.0.0.1
>> Ignazio
>>
>> Il giorno mar 29 gen 2019 alle ore 12:26 Clemens Hardewig <
>> clemens.hardewig at crandale.de> ha scritto:
>>
>>> At least on fedora there is a second cloud Init log as far as I
>>> remember-Look into both
>>>
>>> Br c
>>>
>>> Von meinem iPhone gesendet
>>>
>>> Am 29.01.2019 um 12:08 schrieb Alfredo De Luca <alfredo.deluca at gmail.com
>>> >:
>>>
>>> thanks Clemens.
>>> I looked at the cloud-init-output.log  on the master... and at the
>>> moment is doing the following....
>>>
>>> ++ curl --silent http://127.0.0.1:8080/healthz
>>> + '[' ok = '' ']'
>>> + sleep 5
>>> ++ curl --silent http://127.0.0.1:8080/healthz
>>> + '[' ok = '' ']'
>>> + sleep 5
>>> ++ curl --silent http://127.0.0.1:8080/healthz
>>> + '[' ok = '' ']'
>>> + sleep 5
>>>
>>> Network ....could be but not sure where to look at
>>>
>>>
>>> On Tue, Jan 29, 2019 at 11:34 AM Clemens Hardewig <
>>> clemens.hardewig at crandale.de> wrote:
>>>
>>>> Yes, you should check the cloud-init logs of your master. Without
>>>> having seen them, I would guess a network issue or you have selected for
>>>> your minion nodes a flavor using swap perhaps ...
>>>> So, log files are the first step you could dig into...
>>>> Br c
>>>> Von meinem iPhone gesendet
>>>>
>>>> Am 28.01.2019 um 15:34 schrieb Alfredo De Luca <
>>>> alfredo.deluca at gmail.com>:
>>>>
>>>> Hi all.
>>>> I finally instaledl successufully openstack ansible (queens) but, after
>>>> creating a cluster template I create k8s cluster, it stuck on
>>>>
>>>>
>>>> kube_masters
>>>> <https://10.1.8.113/project/stacks/stack/6221608c-e7f1-4d76-b694-cdd7ec22c386/kube_masters/>
>>>> b7204f0c-b9d8-4ef2-8f0b-afe4c077d039
>>>> <https://10.1.8.113/project/stacks/stack/b7204f0c-b9d8-4ef2-8f0b-afe4c077d039/>
>>>> OS::Heat::ResourceGroup 16 minutes Create In Progress state changedcreate
>>>> in progress....and after around an hour it says...time out. k8s master
>>>> seems to be up.....at least as VM.
>>>>
>>>> any idea?
>>>>
>>>>
>>>>
>>>>
>>>> *Alfredo*
>>>>
>>>>
>>>
>>> --
>>> *Alfredo*
>>>
>>>
>
> --
> *Alfredo*
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.openstack.org/pipermail/openstack-discuss/attachments/20190129/48b4a285/attachment.html>


More information about the openstack-discuss mailing list