Wallaby Magnum Issue

Karera Tony tonykarera at gmail.com
Thu Aug 26 05:14:07 UTC 2021


Hello Guys,

Attached are the two logs from the /var/log/heat-config/heat-config-script
directory
Regards

Tony Karera




On Thu, Aug 26, 2021 at 5:59 AM Karera Tony <tonykarera at gmail.com> wrote:

> Dear Sir,
>
> You are right.
>
> I am getting this error
>
> kubectl get --raw=/healthz
> The connection to the server localhost:8080 was refused - did you specify
> the right host or port?
>
>
> Regards
>
> Tony Karera
>
>
>
>
> On Wed, Aug 25, 2021 at 10:55 PM Bharat Kunwar <bharat at stackhpc.com>
> wrote:
>
>> I’d check the logs under /var/log/heat-config.
>>
>> Sent from my iPhone
>>
>> On 25 Aug 2021, at 19:39, Karera Tony <tonykarera at gmail.com> wrote:
>>
>> 
>> DeaR Ammad,
>>
>> I was able to make the communication work and the Worker nodes were
>> created as well but the cluster failed.
>>
>> I logged in to the master node and there was no error but below are the
>> error when I run systemctl status heat-container-agent on the worker noed.
>>
>> Aug 25 17:52:24 cluster1-fmkpva3nozf7-node-0 podman[2268]:
>> /var/lib/os-collect-config/local-data not found. Skipping
>> Aug 25 17:52:55 cluster1-fmkpva3nozf7-node-0 podman[2268]:
>> /var/lib/os-collect-config/local-data not found. Skipping
>> Aug 25 17:53:26 cluster1-fmkpva3nozf7-node-0 podman[2268]:
>> /var/lib/os-collect-config/local-data not found. Skipping
>> Aug 25 17:53:57 cluster1-fmkpva3nozf7-node-0 podman[2268]:
>> /var/lib/os-collect-config/local-data not found. Skipping
>> Aug 25 17:54:28 cluster1-fmkpva3nozf7-node-0 podman[2268]:
>> /var/lib/os-collect-config/local-data not found. Skipping
>> Aug 25 17:54:59 cluster1-fmkpva3nozf7-node-0 podman[2268]:
>> /var/lib/os-collect-config/local-data not found. Skipping
>> Aug 25 17:55:29 cluster1-fmkpva3nozf7-node-0 podman[2268]:
>> /var/lib/os-collect-config/local-data not found. Skipping
>> Aug 25 17:56:00 cluster1-fmkpva3nozf7-node-0 podman[2268]:
>> /var/lib/os-collect-config/local-data not found. Skipping
>> Aug 25 17:56:31 cluster1-fmkpva3nozf7-node-0 podman[2268]:
>> /var/lib/os-collect-config/local-data not found. Skipping
>> Aug 25 17:57:02 cluster1-fmkpva3nozf7-node-0 podman[2268]:
>> /var/lib/os-collect-config/local-data not found. Skipping
>> Regards
>>
>> Tony Karera
>>
>>
>>
>>
>> On Wed, Aug 25, 2021 at 10:38 AM Ammad Syed <syedammad83 at gmail.com>
>> wrote:
>>
>>> Yes, keystone, Heat, Barbicane and magnum public endpoints must be
>>> reachable from master and worker nodes.
>>>
>>> You can use below guide for the reference as well.
>>>
>>>
>>> https://www.server-world.info/en/note?os=Ubuntu_20.04&p=openstack_victoria4&f=11
>>>
>>> Ammad
>>>
>>> On Wed, Aug 25, 2021 at 12:08 PM Karera Tony <tonykarera at gmail.com>
>>> wrote:
>>>
>>>> Hello Ammad,
>>>>
>>>> I have deployed using the given image but I think there is an issue
>>>> with keystone as per the screen shot below when I checked the master node's
>>>> heat-container-agent status
>>>>
>>>> <image.png>
>>>>
>>>> Regards
>>>>
>>>> Tony Karera
>>>>
>>>>
>>>>
>>>>
>>>> On Wed, Aug 25, 2021 at 8:28 AM Karera Tony <tonykarera at gmail.com>
>>>> wrote:
>>>>
>>>>> Hello Ammad,
>>>>>
>>>>> I actually first used that one and it was also getting stuck.
>>>>>
>>>>> I will try this one again and update you with the Logs though.
>>>>>
>>>>>
>>>>> Regards
>>>>>
>>>>> Tony Karera
>>>>>
>>>>>
>>>>>
>>>>>
>>>>> On Wed, Aug 25, 2021 at 8:25 AM Ammad Syed <syedammad83 at gmail.com>
>>>>> wrote:
>>>>>
>>>>>> It seems from the logs that you are using fedora atomic. Can you try
>>>>>> with FCOS 32 image.
>>>>>>
>>>>>>
>>>>>> https://builds.coreos.fedoraproject.org/prod/streams/stable/builds/32.20201004.3.0/x86_64/fedora-coreos-32.20201004.3.0-openstack.x86_64.qcow2.xz
>>>>>>
>>>>>> Ammad
>>>>>>
>>>>>> On Wed, Aug 25, 2021 at 11:20 AM Karera Tony <tonykarera at gmail.com>
>>>>>> wrote:
>>>>>>
>>>>>>> Hello Sir,
>>>>>>>
>>>>>>> Attached is the Log file
>>>>>>>
>>>>>>> Regards
>>>>>>>
>>>>>>> Tony Karera
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> On Wed, Aug 25, 2021 at 7:31 AM Ammad Syed <syedammad83 at gmail.com>
>>>>>>> wrote:
>>>>>>>
>>>>>>>> Hi Karera,
>>>>>>>>
>>>>>>>> Can you share us the full log file.
>>>>>>>>
>>>>>>>> Ammad
>>>>>>>>
>>>>>>>> On Wed, Aug 25, 2021 at 9:42 AM Karera Tony <tonykarera at gmail.com>
>>>>>>>> wrote:
>>>>>>>>
>>>>>>>>> Hello Guys,
>>>>>>>>>
>>>>>>>>> Thanks a lot for the help but unfortunately I dont see much
>>>>>>>>> information in the log file indicating a failure apart from the log that
>>>>>>>>> keeps appearing.
>>>>>>>>>
>>>>>>>>> <image.png>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> Regards
>>>>>>>>>
>>>>>>>>> Tony Karera
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> On Tue, Aug 24, 2021 at 8:12 PM Mohammed Naser <
>>>>>>>>> mnaser at vexxhost.com> wrote:
>>>>>>>>>
>>>>>>>>>> Also check out /var/log/cloud-init.log :)
>>>>>>>>>>
>>>>>>>>>> On Tue, Aug 24, 2021 at 1:39 PM Ammad Syed <syedammad83 at gmail.com>
>>>>>>>>>> wrote:
>>>>>>>>>> >
>>>>>>>>>> > Then check journalctl -xe or status of heat agent service
>>>>>>>>>> status.
>>>>>>>>>> >
>>>>>>>>>> >
>>>>>>>>>> > Ammad
>>>>>>>>>> > On Tue, Aug 24, 2021 at 10:36 PM Karera Tony <
>>>>>>>>>> tonykarera at gmail.com> wrote:
>>>>>>>>>> >>
>>>>>>>>>> >> Hello Ammad,
>>>>>>>>>> >>
>>>>>>>>>> >> There is no directory or log relevant to heat in the /var/log
>>>>>>>>>> directory
>>>>>>>>>> >>
>>>>>>>>>> >> Regards
>>>>>>>>>> >>
>>>>>>>>>> >> Tony Karera
>>>>>>>>>> >>
>>>>>>>>>> >>
>>>>>>>>>> >>
>>>>>>>>>> >>
>>>>>>>>>> >> On Tue, Aug 24, 2021 at 12:43 PM Ammad Syed <
>>>>>>>>>> syedammad83 at gmail.com> wrote:
>>>>>>>>>> >>>
>>>>>>>>>> >>> Hi Karera,
>>>>>>>>>> >>>
>>>>>>>>>> >>> Login to master node and check the logs of heat agent in var
>>>>>>>>>> log. There must be something the cluster is stucking somewhere in creating.
>>>>>>>>>> >>>
>>>>>>>>>> >>> Ammad
>>>>>>>>>> >>>
>>>>>>>>>> >>> On Tue, Aug 24, 2021 at 3:41 PM Karera Tony <
>>>>>>>>>> tonykarera at gmail.com> wrote:
>>>>>>>>>> >>>>
>>>>>>>>>> >>>> Hello Ammad,
>>>>>>>>>> >>>>
>>>>>>>>>> >>>> I had done as explained and it worked upto a certain point.
>>>>>>>>>> The master node was created but the cluster remained in Creation in
>>>>>>>>>> progress for over an hour and failed with error below
>>>>>>>>>> >>>>
>>>>>>>>>> >>>> Stack Faults
>>>>>>>>>> >>>> as follows:
>>>>>>>>>> >>>> default-master
>>>>>>>>>> >>>> Timed out
>>>>>>>>>> >>>> default-worker
>>>>>>>>>> >>>> Timed out
>>>>>>>>>> >>>>
>>>>>>>>>> >>>>
>>>>>>>>>> >>>> Regards
>>>>>>>>>> >>>>
>>>>>>>>>> >>>> Tony Karera
>>>>>>>>>> >>>>
>>>>>>>>>> >>>>
>>>>>>>>>> >>>>
>>>>>>>>>> >>>>
>>>>>>>>>> >>>> On Tue, Aug 24, 2021 at 9:25 AM Ammad Syed <
>>>>>>>>>> syedammad83 at gmail.com> wrote:
>>>>>>>>>> >>>>>
>>>>>>>>>> >>>>> Hi Tony,
>>>>>>>>>> >>>>>
>>>>>>>>>> >>>>> You can try by creating your private vxlan network prior to
>>>>>>>>>> deployment of cluster and explicitly create your cluster in vxlan network.
>>>>>>>>>> >>>>>
>>>>>>>>>> >>>>> --fixed-network private --fixed-subnet private-subnet
>>>>>>>>>> >>>>>
>>>>>>>>>> >>>>> You can specify above while creating a cluster.
>>>>>>>>>> >>>>>
>>>>>>>>>> >>>>> Ammad
>>>>>>>>>> >>>>>
>>>>>>>>>> >>>>> On Tue, Aug 24, 2021 at 11:59 AM Karera Tony <
>>>>>>>>>> tonykarera at gmail.com> wrote:
>>>>>>>>>> >>>>>>
>>>>>>>>>> >>>>>> Hello MOhamed,
>>>>>>>>>> >>>>>>
>>>>>>>>>> >>>>>> I think the Kubernetes cluster is ok but it when I deploy
>>>>>>>>>> it, It creates a fixed network using vlan which I am not using for internal
>>>>>>>>>> networks.
>>>>>>>>>> >>>>>>
>>>>>>>>>> >>>>>> When I create a a vxlan Network and use it in the cluster
>>>>>>>>>> creation, It fails. Is there a trick around this ?
>>>>>>>>>> >>>>>> Regards
>>>>>>>>>> >>>>>>
>>>>>>>>>> >>>>>> Tony Karera
>>>>>>>>>> >>>>>>
>>>>>>>>>> >>>>>>
>>>>>>>>>> >>>>>>
>>>>>>>>>> >>>>>>
>>>>>>>>>> >>>>>> On Fri, Aug 20, 2021 at 9:00 AM feilong <
>>>>>>>>>> feilong at catalyst.net.nz> wrote:
>>>>>>>>>> >>>>>>>
>>>>>>>>>> >>>>>>> Oooh, are you using Swarm? I don't think that driver is
>>>>>>>>>> well maintained. I didn't see any interest in the last 4 years since I
>>>>>>>>>> involved in the Magnum project. If there is no specific reason, I would
>>>>>>>>>> suggest go for k8s.
>>>>>>>>>> >>>>>>>
>>>>>>>>>> >>>>>>>
>>>>>>>>>> >>>>>>> On 20/08/21 5:08 pm, Mohammed Naser wrote:
>>>>>>>>>> >>>>>>>
>>>>>>>>>> >>>>>>> Please keep replies on list so others can help too.
>>>>>>>>>> >>>>>>>
>>>>>>>>>> >>>>>>> I don’t know how well tested the Swarm driver is at this
>>>>>>>>>> point. I believe most Magnum users are using it for Kubernetes only.
>>>>>>>>>> >>>>>>>
>>>>>>>>>> >>>>>>> On Fri, Aug 20, 2021 at 1:05 AM Karera Tony <
>>>>>>>>>> tonykarera at gmail.com> wrote:
>>>>>>>>>> >>>>>>>>
>>>>>>>>>> >>>>>>>> Hello Naser,
>>>>>>>>>> >>>>>>>>
>>>>>>>>>> >>>>>>>> Please check below.
>>>>>>>>>> >>>>>>>>
>>>>>>>>>> >>>>>>>> openstack coe cluster template create
>>>>>>>>>> swarm-cluster-template1 \
>>>>>>>>>> >>>>>>>>                      --image fedora-atomic-latest \
>>>>>>>>>> >>>>>>>>                      --external-network External_1700\
>>>>>>>>>> >>>>>>>>                      --dns-nameserver 8.8.8.8 \
>>>>>>>>>> >>>>>>>>                      --master-flavor m1.small \
>>>>>>>>>> >>>>>>>>                      --flavor m1.small \
>>>>>>>>>> >>>>>>>>                      --coe swarm
>>>>>>>>>> >>>>>>>> openstack coe cluster create swarm-cluster \
>>>>>>>>>> >>>>>>>>                         --cluster-template
>>>>>>>>>> swarm-cluster-template \
>>>>>>>>>> >>>>>>>>                         --master-count 1 \
>>>>>>>>>> >>>>>>>>                         --node-count 2 \
>>>>>>>>>> >>>>>>>>                         --keypair Newkey
>>>>>>>>>> >>>>>>>>
>>>>>>>>>> >>>>>>>> Regards
>>>>>>>>>> >>>>>>>>
>>>>>>>>>> >>>>>>>> Tony Karera
>>>>>>>>>> >>>>>>>>
>>>>>>>>>> >>>>>>>>
>>>>>>>>>> >>>>>>>>
>>>>>>>>>> >>>>>>>>
>>>>>>>>>> >>>>>>>> On Fri, Aug 20, 2021 at 7:03 AM Mohammed Naser <
>>>>>>>>>> mnaser at vexxhost.com> wrote:
>>>>>>>>>> >>>>>>>>>
>>>>>>>>>> >>>>>>>>> What does your cluster template and cluster create
>>>>>>>>>> command look like?
>>>>>>>>>> >>>>>>>>>
>>>>>>>>>> >>>>>>>>> On Fri, Aug 20, 2021 at 12:59 AM Karera Tony <
>>>>>>>>>> tonykarera at gmail.com> wrote:
>>>>>>>>>> >>>>>>>>>>
>>>>>>>>>> >>>>>>>>>> Hello Wang,
>>>>>>>>>> >>>>>>>>>>
>>>>>>>>>> >>>>>>>>>> Thanks for the feedback.
>>>>>>>>>> >>>>>>>>>>
>>>>>>>>>> >>>>>>>>>> Unfortunately Octavia is not deployed in my
>>>>>>>>>> environment (at least not yet) and LB is not enabled on either the cluster
>>>>>>>>>> template or the cluster itself.
>>>>>>>>>> >>>>>>>>>>
>>>>>>>>>> >>>>>>>>>> Regards
>>>>>>>>>> >>>>>>>>>>
>>>>>>>>>> >>>>>>>>>> Tony Karera
>>>>>>>>>> >>>>>>>>>>
>>>>>>>>>> >>>>>>>>>>
>>>>>>>>>> >>>>>>>>>>
>>>>>>>>>> >>>>>>>>>>
>>>>>>>>>> >>>>>>>>>> On Fri, Aug 20, 2021 at 6:52 AM feilong <
>>>>>>>>>> feilong at catalyst.net.nz> wrote:
>>>>>>>>>> >>>>>>>>>>>
>>>>>>>>>> >>>>>>>>>>> Hi Karera,
>>>>>>>>>> >>>>>>>>>>>
>>>>>>>>>> >>>>>>>>>>> It's probably a bug. If you do have Octavia deployed,
>>>>>>>>>> can you try to not disable the LB and see how it goes?
>>>>>>>>>> >>>>>>>>>>>
>>>>>>>>>> >>>>>>>>>>>
>>>>>>>>>> >>>>>>>>>>> On 20/08/21 4:18 pm, Karera Tony wrote:
>>>>>>>>>> >>>>>>>>>>>
>>>>>>>>>> >>>>>>>>>>> Hello Team,
>>>>>>>>>> >>>>>>>>>>>
>>>>>>>>>> >>>>>>>>>>> I deployed Openstack Wallby on Ubuntu20 and enabled
>>>>>>>>>> Magum, however when I create a cluster I get the error below.
>>>>>>>>>> >>>>>>>>>>>
>>>>>>>>>> >>>>>>>>>>> Status Reason
>>>>>>>>>> >>>>>>>>>>> ERROR: Property error: : resources.api_lb.properties:
>>>>>>>>>> : Property allowed_cidrs not assigned
>>>>>>>>>> >>>>>>>>>>> Can someone advise on where I could be wrong. Btw, I
>>>>>>>>>> disabled load balancer while creating the cluster.
>>>>>>>>>> >>>>>>>>>>>
>>>>>>>>>> >>>>>>>>>>> Regards
>>>>>>>>>> >>>>>>>>>>>
>>>>>>>>>> >>>>>>>>>>> Tony Karera
>>>>>>>>>> >>>>>>>>>>>
>>>>>>>>>> >>>>>>>>>>>
>>>>>>>>>> >>>>>>>>>>> --
>>>>>>>>>> >>>>>>>>>>> Cheers & Best regards,
>>>>>>>>>> >>>>>>>>>>>
>>>>>>>>>> ------------------------------------------------------------------------------
>>>>>>>>>> >>>>>>>>>>> Feilong Wang (王飞龙) (he/him)
>>>>>>>>>> >>>>>>>>>>> Head of Research & Development
>>>>>>>>>> >>>>>>>>>>>
>>>>>>>>>> >>>>>>>>>>> Catalyst Cloud
>>>>>>>>>> >>>>>>>>>>> Aotearoa's own
>>>>>>>>>> >>>>>>>>>>>
>>>>>>>>>> >>>>>>>>>>> Mob: +64 21 0832 6348 | www.catalystcloud.nz
>>>>>>>>>> >>>>>>>>>>> Level 6, 150 Willis Street, Wellington 6011, New
>>>>>>>>>> Zealand
>>>>>>>>>> >>>>>>>>>>>
>>>>>>>>>> >>>>>>>>>>> CONFIDENTIALITY NOTICE: This email is intended for
>>>>>>>>>> the named recipients only.
>>>>>>>>>> >>>>>>>>>>> It may contain privileged, confidential or copyright
>>>>>>>>>> information. If you are
>>>>>>>>>> >>>>>>>>>>> not the named recipient, any use, reliance upon,
>>>>>>>>>> disclosure or copying of this
>>>>>>>>>> >>>>>>>>>>> email or its attachments is unauthorised. If you have
>>>>>>>>>> received this email in
>>>>>>>>>> >>>>>>>>>>> error, please reply via email or call +64 21 0832
>>>>>>>>>> 6348.
>>>>>>>>>> >>>>>>>>>>>
>>>>>>>>>> ------------------------------------------------------------------------------
>>>>>>>>>> >>>>>>>>>
>>>>>>>>>> >>>>>>>>> --
>>>>>>>>>> >>>>>>>>> Mohammed Naser
>>>>>>>>>> >>>>>>>>> VEXXHOST, Inc.
>>>>>>>>>> >>>>>>>
>>>>>>>>>> >>>>>>> --
>>>>>>>>>> >>>>>>> Mohammed Naser
>>>>>>>>>> >>>>>>> VEXXHOST, Inc.
>>>>>>>>>> >>>>>>>
>>>>>>>>>> >>>>>>> --
>>>>>>>>>> >>>>>>> Cheers & Best regards,
>>>>>>>>>> >>>>>>>
>>>>>>>>>> ------------------------------------------------------------------------------
>>>>>>>>>> >>>>>>> Feilong Wang (王飞龙) (he/him)
>>>>>>>>>> >>>>>>> Head of Research & Development
>>>>>>>>>> >>>>>>>
>>>>>>>>>> >>>>>>> Catalyst Cloud
>>>>>>>>>> >>>>>>> Aotearoa's own
>>>>>>>>>> >>>>>>>
>>>>>>>>>> >>>>>>> Mob: +64 21 0832 6348 | www.catalystcloud.nz
>>>>>>>>>> >>>>>>> Level 6, 150 Willis Street, Wellington 6011, New Zealand
>>>>>>>>>> >>>>>>>
>>>>>>>>>> >>>>>>> CONFIDENTIALITY NOTICE: This email is intended for the
>>>>>>>>>> named recipients only.
>>>>>>>>>> >>>>>>> It may contain privileged, confidential or copyright
>>>>>>>>>> information. If you are
>>>>>>>>>> >>>>>>> not the named recipient, any use, reliance upon,
>>>>>>>>>> disclosure or copying of this
>>>>>>>>>> >>>>>>> email or its attachments is unauthorised. If you have
>>>>>>>>>> received this email in
>>>>>>>>>> >>>>>>> error, please reply via email or call +64 21 0832 6348.
>>>>>>>>>> >>>>>>>
>>>>>>>>>> ------------------------------------------------------------------------------
>>>>>>>>>> >>>>>
>>>>>>>>>> >>>>>
>>>>>>>>>> >>>>>
>>>>>>>>>> >>>>> --
>>>>>>>>>> >>>>> Regards,
>>>>>>>>>> >>>>>
>>>>>>>>>> >>>>>
>>>>>>>>>> >>>>> Syed Ammad Ali
>>>>>>>>>> >>>
>>>>>>>>>> >>>
>>>>>>>>>> >>>
>>>>>>>>>> >>> --
>>>>>>>>>> >>> Regards,
>>>>>>>>>> >>>
>>>>>>>>>> >>>
>>>>>>>>>> >>> Syed Ammad Ali
>>>>>>>>>> >
>>>>>>>>>> > --
>>>>>>>>>> > Regards,
>>>>>>>>>> >
>>>>>>>>>> >
>>>>>>>>>> > Syed Ammad Ali
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> --
>>>>>>>>>> Mohammed Naser
>>>>>>>>>> VEXXHOST, Inc.
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>
>>>>>>>> --
>>>>>>>> Regards,
>>>>>>>>
>>>>>>>>
>>>>>>>> Syed Ammad Ali
>>>>>>>>
>>>>>>>
>>>>>>
>>>>>> --
>>>>>> Regards,
>>>>>>
>>>>>>
>>>>>> Syed Ammad Ali
>>>>>>
>>>>>
>>>
>>> --
>>> Regards,
>>>
>>>
>>> Syed Ammad Ali
>>>
>>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.openstack.org/pipermail/openstack-discuss/attachments/20210826/ca9ef16e/attachment-0001.html>
-------------- next part --------------
A non-text attachment was scrubbed...
Name: 29a37aff-f1f6-46b3-8541-887491c6cfe8-k8s-cluster3-dcu52bgzpbuu-kube_masters-ocfrn2ikpcgd-0-32tmkqgdq7wl-master_config-gihyfv3wlyzd.log
Type: application/octet-stream
Size: 11011 bytes
Desc: not available
URL: <http://lists.openstack.org/pipermail/openstack-discuss/attachments/20210826/ca9ef16e/attachment-0002.obj>
-------------- next part --------------
A non-text attachment was scrubbed...
Name: 6fca39b1-8eda-4786-8424-e5b04434cce7-k8s-cluster3-dcu52bgzpbuu-kube_cluster_config-aht4it6p7wfk.log
Type: application/octet-stream
Size: 1614278 bytes
Desc: not available
URL: <http://lists.openstack.org/pipermail/openstack-discuss/attachments/20210826/ca9ef16e/attachment-0003.obj>


More information about the openstack-discuss mailing list