Wallaby Magnum Issue

Bharat Kunwar bharat at stackhpc.com
Thu Aug 26 05:53:04 UTC 2021


I assume these are from the master nodes? Can you share the logs shortly after creation rather than when it times out? I think there is some missing logs from the top.

Sent from my iPhone

> On 26 Aug 2021, at 06:14, Karera Tony <tonykarera at gmail.com> wrote:
> 
> 
> Hello Guys,
> 
> Attached are the two logs from the /var/log/heat-config/heat-config-script directory
> Regards
> 
> Tony Karera
> 
> 
> 
> 
>> On Thu, Aug 26, 2021 at 5:59 AM Karera Tony <tonykarera at gmail.com> wrote:
>> Dear Sir,
>> 
>> You are right.
>> 
>> I am getting this error 
>> 
>> kubectl get --raw=/healthz
>> The connection to the server localhost:8080 was refused - did you specify the right host or port?
>> 
>> 
>> Regards
>> 
>> Tony Karera
>> 
>> 
>> 
>> 
>>> On Wed, Aug 25, 2021 at 10:55 PM Bharat Kunwar <bharat at stackhpc.com> wrote:
>>> I’d check the logs under /var/log/heat-config.
>>> 
>>> Sent from my iPhone
>>> 
>>>>> On 25 Aug 2021, at 19:39, Karera Tony <tonykarera at gmail.com> wrote:
>>>>> 
>>>> 
>>>> DeaR Ammad,
>>>> 
>>>> I was able to make the communication work and the Worker nodes were created as well but the cluster failed.
>>>> 
>>>> I logged in to the master node and there was no error but below are the error when I run systemctl status heat-container-agent on the worker noed.
>>>> 
>>>> Aug 25 17:52:24 cluster1-fmkpva3nozf7-node-0 podman[2268]: /var/lib/os-collect-config/local-data not found. Skipping
>>>> Aug 25 17:52:55 cluster1-fmkpva3nozf7-node-0 podman[2268]: /var/lib/os-collect-config/local-data not found. Skipping
>>>> Aug 25 17:53:26 cluster1-fmkpva3nozf7-node-0 podman[2268]: /var/lib/os-collect-config/local-data not found. Skipping
>>>> Aug 25 17:53:57 cluster1-fmkpva3nozf7-node-0 podman[2268]: /var/lib/os-collect-config/local-data not found. Skipping
>>>> Aug 25 17:54:28 cluster1-fmkpva3nozf7-node-0 podman[2268]: /var/lib/os-collect-config/local-data not found. Skipping
>>>> Aug 25 17:54:59 cluster1-fmkpva3nozf7-node-0 podman[2268]: /var/lib/os-collect-config/local-data not found. Skipping
>>>> Aug 25 17:55:29 cluster1-fmkpva3nozf7-node-0 podman[2268]: /var/lib/os-collect-config/local-data not found. Skipping
>>>> Aug 25 17:56:00 cluster1-fmkpva3nozf7-node-0 podman[2268]: /var/lib/os-collect-config/local-data not found. Skipping
>>>> Aug 25 17:56:31 cluster1-fmkpva3nozf7-node-0 podman[2268]: /var/lib/os-collect-config/local-data not found. Skipping
>>>> Aug 25 17:57:02 cluster1-fmkpva3nozf7-node-0 podman[2268]: /var/lib/os-collect-config/local-data not found. Skipping
>>>> Regards
>>>> 
>>>> Tony Karera
>>>> 
>>>> 
>>>> 
>>>> 
>>>>> On Wed, Aug 25, 2021 at 10:38 AM Ammad Syed <syedammad83 at gmail.com> wrote:
>>>>> Yes, keystone, Heat, Barbicane and magnum public endpoints must be reachable from master and worker nodes.
>>>>> 
>>>>> You can use below guide for the reference as well.
>>>>> 
>>>>> https://www.server-world.info/en/note?os=Ubuntu_20.04&p=openstack_victoria4&f=11
>>>>> 
>>>>> Ammad
>>>>> 
>>>>>> On Wed, Aug 25, 2021 at 12:08 PM Karera Tony <tonykarera at gmail.com> wrote:
>>>>>> Hello Ammad,
>>>>>> 
>>>>>> I have deployed using the given image but I think there is an issue with keystone as per the screen shot below when I checked the master node's heat-container-agent status
>>>>>> 
>>>>>> <image.png>
>>>>>> 
>>>>>> Regards
>>>>>> 
>>>>>> Tony Karera
>>>>>> 
>>>>>> 
>>>>>> 
>>>>>> 
>>>>>>> On Wed, Aug 25, 2021 at 8:28 AM Karera Tony <tonykarera at gmail.com> wrote:
>>>>>>> Hello Ammad,
>>>>>>> 
>>>>>>> I actually first used that one and it was also getting stuck.
>>>>>>> 
>>>>>>> I will try this one again and update you with the Logs though.
>>>>>>> 
>>>>>>> 
>>>>>>> Regards
>>>>>>> 
>>>>>>> Tony Karera
>>>>>>> 
>>>>>>> 
>>>>>>> 
>>>>>>> 
>>>>>>>> On Wed, Aug 25, 2021 at 8:25 AM Ammad Syed <syedammad83 at gmail.com> wrote:
>>>>>>>> It seems from the logs that you are using fedora atomic. Can you try with FCOS 32 image.
>>>>>>>> 
>>>>>>>> https://builds.coreos.fedoraproject.org/prod/streams/stable/builds/32.20201004.3.0/x86_64/fedora-coreos-32.20201004.3.0-openstack.x86_64.qcow2.xz
>>>>>>>> 
>>>>>>>> Ammad
>>>>>>>> 
>>>>>>>>> On Wed, Aug 25, 2021 at 11:20 AM Karera Tony <tonykarera at gmail.com> wrote:
>>>>>>>>> Hello Sir,
>>>>>>>>> 
>>>>>>>>> Attached is the Log file
>>>>>>>>> 
>>>>>>>>> Regards
>>>>>>>>> 
>>>>>>>>> Tony Karera
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>>> On Wed, Aug 25, 2021 at 7:31 AM Ammad Syed <syedammad83 at gmail.com> wrote:
>>>>>>>>>> Hi Karera,
>>>>>>>>>> 
>>>>>>>>>> Can you share us the full log file.
>>>>>>>>>> 
>>>>>>>>>> Ammad
>>>>>>>>>> 
>>>>>>>>>>> On Wed, Aug 25, 2021 at 9:42 AM Karera Tony <tonykarera at gmail.com> wrote:
>>>>>>>>>>> Hello Guys,
>>>>>>>>>>> 
>>>>>>>>>>> Thanks a lot for the help but unfortunately I dont see much information in the log file indicating a failure apart from the log that keeps appearing.
>>>>>>>>>>> 
>>>>>>>>>>> <image.png>
>>>>>>>>>>> 
>>>>>>>>>>> 
>>>>>>>>>>> Regards
>>>>>>>>>>> 
>>>>>>>>>>> Tony Karera
>>>>>>>>>>> 
>>>>>>>>>>> 
>>>>>>>>>>> 
>>>>>>>>>>> 
>>>>>>>>>>>> On Tue, Aug 24, 2021 at 8:12 PM Mohammed Naser <mnaser at vexxhost.com> wrote:
>>>>>>>>>>>> Also check out /var/log/cloud-init.log :)
>>>>>>>>>>>> 
>>>>>>>>>>>> On Tue, Aug 24, 2021 at 1:39 PM Ammad Syed <syedammad83 at gmail.com> wrote:
>>>>>>>>>>>> >
>>>>>>>>>>>> > Then check journalctl -xe or status of heat agent service status.
>>>>>>>>>>>> >
>>>>>>>>>>>> >
>>>>>>>>>>>> > Ammad
>>>>>>>>>>>> > On Tue, Aug 24, 2021 at 10:36 PM Karera Tony <tonykarera at gmail.com> wrote:
>>>>>>>>>>>> >>
>>>>>>>>>>>> >> Hello Ammad,
>>>>>>>>>>>> >>
>>>>>>>>>>>> >> There is no directory or log relevant to heat in the /var/log directory
>>>>>>>>>>>> >>
>>>>>>>>>>>> >> Regards
>>>>>>>>>>>> >>
>>>>>>>>>>>> >> Tony Karera
>>>>>>>>>>>> >>
>>>>>>>>>>>> >>
>>>>>>>>>>>> >>
>>>>>>>>>>>> >>
>>>>>>>>>>>> >> On Tue, Aug 24, 2021 at 12:43 PM Ammad Syed <syedammad83 at gmail.com> wrote:
>>>>>>>>>>>> >>>
>>>>>>>>>>>> >>> Hi Karera,
>>>>>>>>>>>> >>>
>>>>>>>>>>>> >>> Login to master node and check the logs of heat agent in var log. There must be something the cluster is stucking somewhere in creating.
>>>>>>>>>>>> >>>
>>>>>>>>>>>> >>> Ammad
>>>>>>>>>>>> >>>
>>>>>>>>>>>> >>> On Tue, Aug 24, 2021 at 3:41 PM Karera Tony <tonykarera at gmail.com> wrote:
>>>>>>>>>>>> >>>>
>>>>>>>>>>>> >>>> Hello Ammad,
>>>>>>>>>>>> >>>>
>>>>>>>>>>>> >>>> I had done as explained and it worked upto a certain point. The master node was created but the cluster remained in Creation in progress for over an hour and failed with error below
>>>>>>>>>>>> >>>>
>>>>>>>>>>>> >>>> Stack Faults
>>>>>>>>>>>> >>>> as follows:
>>>>>>>>>>>> >>>> default-master
>>>>>>>>>>>> >>>> Timed out
>>>>>>>>>>>> >>>> default-worker
>>>>>>>>>>>> >>>> Timed out
>>>>>>>>>>>> >>>>
>>>>>>>>>>>> >>>>
>>>>>>>>>>>> >>>> Regards
>>>>>>>>>>>> >>>>
>>>>>>>>>>>> >>>> Tony Karera
>>>>>>>>>>>> >>>>
>>>>>>>>>>>> >>>>
>>>>>>>>>>>> >>>>
>>>>>>>>>>>> >>>>
>>>>>>>>>>>> >>>> On Tue, Aug 24, 2021 at 9:25 AM Ammad Syed <syedammad83 at gmail.com> wrote:
>>>>>>>>>>>> >>>>>
>>>>>>>>>>>> >>>>> Hi Tony,
>>>>>>>>>>>> >>>>>
>>>>>>>>>>>> >>>>> You can try by creating your private vxlan network prior to deployment of cluster and explicitly create your cluster in vxlan network.
>>>>>>>>>>>> >>>>>
>>>>>>>>>>>> >>>>> --fixed-network private --fixed-subnet private-subnet
>>>>>>>>>>>> >>>>>
>>>>>>>>>>>> >>>>> You can specify above while creating a cluster.
>>>>>>>>>>>> >>>>>
>>>>>>>>>>>> >>>>> Ammad
>>>>>>>>>>>> >>>>>
>>>>>>>>>>>> >>>>> On Tue, Aug 24, 2021 at 11:59 AM Karera Tony <tonykarera at gmail.com> wrote:
>>>>>>>>>>>> >>>>>>
>>>>>>>>>>>> >>>>>> Hello MOhamed,
>>>>>>>>>>>> >>>>>>
>>>>>>>>>>>> >>>>>> I think the Kubernetes cluster is ok but it when I deploy it, It creates a fixed network using vlan which I am not using for internal networks.
>>>>>>>>>>>> >>>>>>
>>>>>>>>>>>> >>>>>> When I create a a vxlan Network and use it in the cluster creation, It fails. Is there a trick around this ?
>>>>>>>>>>>> >>>>>> Regards
>>>>>>>>>>>> >>>>>>
>>>>>>>>>>>> >>>>>> Tony Karera
>>>>>>>>>>>> >>>>>>
>>>>>>>>>>>> >>>>>>
>>>>>>>>>>>> >>>>>>
>>>>>>>>>>>> >>>>>>
>>>>>>>>>>>> >>>>>> On Fri, Aug 20, 2021 at 9:00 AM feilong <feilong at catalyst.net.nz> wrote:
>>>>>>>>>>>> >>>>>>>
>>>>>>>>>>>> >>>>>>> Oooh, are you using Swarm? I don't think that driver is well maintained. I didn't see any interest in the last 4 years since I involved in the Magnum project. If there is no specific reason, I would suggest go for k8s.
>>>>>>>>>>>> >>>>>>>
>>>>>>>>>>>> >>>>>>>
>>>>>>>>>>>> >>>>>>> On 20/08/21 5:08 pm, Mohammed Naser wrote:
>>>>>>>>>>>> >>>>>>>
>>>>>>>>>>>> >>>>>>> Please keep replies on list so others can help too.
>>>>>>>>>>>> >>>>>>>
>>>>>>>>>>>> >>>>>>> I don’t know how well tested the Swarm driver is at this point. I believe most Magnum users are using it for Kubernetes only.
>>>>>>>>>>>> >>>>>>>
>>>>>>>>>>>> >>>>>>> On Fri, Aug 20, 2021 at 1:05 AM Karera Tony <tonykarera at gmail.com> wrote:
>>>>>>>>>>>> >>>>>>>>
>>>>>>>>>>>> >>>>>>>> Hello Naser,
>>>>>>>>>>>> >>>>>>>>
>>>>>>>>>>>> >>>>>>>> Please check below.
>>>>>>>>>>>> >>>>>>>>
>>>>>>>>>>>> >>>>>>>> openstack coe cluster template create swarm-cluster-template1 \
>>>>>>>>>>>> >>>>>>>>                      --image fedora-atomic-latest \
>>>>>>>>>>>> >>>>>>>>                      --external-network External_1700\
>>>>>>>>>>>> >>>>>>>>                      --dns-nameserver 8.8.8.8 \
>>>>>>>>>>>> >>>>>>>>                      --master-flavor m1.small \
>>>>>>>>>>>> >>>>>>>>                      --flavor m1.small \
>>>>>>>>>>>> >>>>>>>>                      --coe swarm
>>>>>>>>>>>> >>>>>>>> openstack coe cluster create swarm-cluster \
>>>>>>>>>>>> >>>>>>>>                         --cluster-template swarm-cluster-template \
>>>>>>>>>>>> >>>>>>>>                         --master-count 1 \
>>>>>>>>>>>> >>>>>>>>                         --node-count 2 \
>>>>>>>>>>>> >>>>>>>>                         --keypair Newkey
>>>>>>>>>>>> >>>>>>>>
>>>>>>>>>>>> >>>>>>>> Regards
>>>>>>>>>>>> >>>>>>>>
>>>>>>>>>>>> >>>>>>>> Tony Karera
>>>>>>>>>>>> >>>>>>>>
>>>>>>>>>>>> >>>>>>>>
>>>>>>>>>>>> >>>>>>>>
>>>>>>>>>>>> >>>>>>>>
>>>>>>>>>>>> >>>>>>>> On Fri, Aug 20, 2021 at 7:03 AM Mohammed Naser <mnaser at vexxhost.com> wrote:
>>>>>>>>>>>> >>>>>>>>>
>>>>>>>>>>>> >>>>>>>>> What does your cluster template and cluster create command look like?
>>>>>>>>>>>> >>>>>>>>>
>>>>>>>>>>>> >>>>>>>>> On Fri, Aug 20, 2021 at 12:59 AM Karera Tony <tonykarera at gmail.com> wrote:
>>>>>>>>>>>> >>>>>>>>>>
>>>>>>>>>>>> >>>>>>>>>> Hello Wang,
>>>>>>>>>>>> >>>>>>>>>>
>>>>>>>>>>>> >>>>>>>>>> Thanks for the feedback.
>>>>>>>>>>>> >>>>>>>>>>
>>>>>>>>>>>> >>>>>>>>>> Unfortunately Octavia is not deployed in my environment (at least not yet) and LB is not enabled on either the cluster template or the cluster itself.
>>>>>>>>>>>> >>>>>>>>>>
>>>>>>>>>>>> >>>>>>>>>> Regards
>>>>>>>>>>>> >>>>>>>>>>
>>>>>>>>>>>> >>>>>>>>>> Tony Karera
>>>>>>>>>>>> >>>>>>>>>>
>>>>>>>>>>>> >>>>>>>>>>
>>>>>>>>>>>> >>>>>>>>>>
>>>>>>>>>>>> >>>>>>>>>>
>>>>>>>>>>>> >>>>>>>>>> On Fri, Aug 20, 2021 at 6:52 AM feilong <feilong at catalyst.net.nz> wrote:
>>>>>>>>>>>> >>>>>>>>>>>
>>>>>>>>>>>> >>>>>>>>>>> Hi Karera,
>>>>>>>>>>>> >>>>>>>>>>>
>>>>>>>>>>>> >>>>>>>>>>> It's probably a bug. If you do have Octavia deployed, can you try to not disable the LB and see how it goes?
>>>>>>>>>>>> >>>>>>>>>>>
>>>>>>>>>>>> >>>>>>>>>>>
>>>>>>>>>>>> >>>>>>>>>>> On 20/08/21 4:18 pm, Karera Tony wrote:
>>>>>>>>>>>> >>>>>>>>>>>
>>>>>>>>>>>> >>>>>>>>>>> Hello Team,
>>>>>>>>>>>> >>>>>>>>>>>
>>>>>>>>>>>> >>>>>>>>>>> I deployed Openstack Wallby on Ubuntu20 and enabled Magum, however when I create a cluster I get the error below.
>>>>>>>>>>>> >>>>>>>>>>>
>>>>>>>>>>>> >>>>>>>>>>> Status Reason
>>>>>>>>>>>> >>>>>>>>>>> ERROR: Property error: : resources.api_lb.properties: : Property allowed_cidrs not assigned
>>>>>>>>>>>> >>>>>>>>>>> Can someone advise on where I could be wrong. Btw, I disabled load balancer while creating the cluster.
>>>>>>>>>>>> >>>>>>>>>>>
>>>>>>>>>>>> >>>>>>>>>>> Regards
>>>>>>>>>>>> >>>>>>>>>>>
>>>>>>>>>>>> >>>>>>>>>>> Tony Karera
>>>>>>>>>>>> >>>>>>>>>>>
>>>>>>>>>>>> >>>>>>>>>>>
>>>>>>>>>>>> >>>>>>>>>>> --
>>>>>>>>>>>> >>>>>>>>>>> Cheers & Best regards,
>>>>>>>>>>>> >>>>>>>>>>> ------------------------------------------------------------------------------
>>>>>>>>>>>> >>>>>>>>>>> Feilong Wang (王飞龙) (he/him)
>>>>>>>>>>>> >>>>>>>>>>> Head of Research & Development
>>>>>>>>>>>> >>>>>>>>>>>
>>>>>>>>>>>> >>>>>>>>>>> Catalyst Cloud
>>>>>>>>>>>> >>>>>>>>>>> Aotearoa's own
>>>>>>>>>>>> >>>>>>>>>>>
>>>>>>>>>>>> >>>>>>>>>>> Mob: +64 21 0832 6348 | www.catalystcloud.nz
>>>>>>>>>>>> >>>>>>>>>>> Level 6, 150 Willis Street, Wellington 6011, New Zealand
>>>>>>>>>>>> >>>>>>>>>>>
>>>>>>>>>>>> >>>>>>>>>>> CONFIDENTIALITY NOTICE: This email is intended for the named recipients only.
>>>>>>>>>>>> >>>>>>>>>>> It may contain privileged, confidential or copyright information. If you are
>>>>>>>>>>>> >>>>>>>>>>> not the named recipient, any use, reliance upon, disclosure or copying of this
>>>>>>>>>>>> >>>>>>>>>>> email or its attachments is unauthorised. If you have received this email in
>>>>>>>>>>>> >>>>>>>>>>> error, please reply via email or call +64 21 0832 6348.
>>>>>>>>>>>> >>>>>>>>>>> ------------------------------------------------------------------------------
>>>>>>>>>>>> >>>>>>>>>
>>>>>>>>>>>> >>>>>>>>> --
>>>>>>>>>>>> >>>>>>>>> Mohammed Naser
>>>>>>>>>>>> >>>>>>>>> VEXXHOST, Inc.
>>>>>>>>>>>> >>>>>>>
>>>>>>>>>>>> >>>>>>> --
>>>>>>>>>>>> >>>>>>> Mohammed Naser
>>>>>>>>>>>> >>>>>>> VEXXHOST, Inc.
>>>>>>>>>>>> >>>>>>>
>>>>>>>>>>>> >>>>>>> --
>>>>>>>>>>>> >>>>>>> Cheers & Best regards,
>>>>>>>>>>>> >>>>>>> ------------------------------------------------------------------------------
>>>>>>>>>>>> >>>>>>> Feilong Wang (王飞龙) (he/him)
>>>>>>>>>>>> >>>>>>> Head of Research & Development
>>>>>>>>>>>> >>>>>>>
>>>>>>>>>>>> >>>>>>> Catalyst Cloud
>>>>>>>>>>>> >>>>>>> Aotearoa's own
>>>>>>>>>>>> >>>>>>>
>>>>>>>>>>>> >>>>>>> Mob: +64 21 0832 6348 | www.catalystcloud.nz
>>>>>>>>>>>> >>>>>>> Level 6, 150 Willis Street, Wellington 6011, New Zealand
>>>>>>>>>>>> >>>>>>>
>>>>>>>>>>>> >>>>>>> CONFIDENTIALITY NOTICE: This email is intended for the named recipients only.
>>>>>>>>>>>> >>>>>>> It may contain privileged, confidential or copyright information. If you are
>>>>>>>>>>>> >>>>>>> not the named recipient, any use, reliance upon, disclosure or copying of this
>>>>>>>>>>>> >>>>>>> email or its attachments is unauthorised. If you have received this email in
>>>>>>>>>>>> >>>>>>> error, please reply via email or call +64 21 0832 6348.
>>>>>>>>>>>> >>>>>>> ------------------------------------------------------------------------------
>>>>>>>>>>>> >>>>>
>>>>>>>>>>>> >>>>>
>>>>>>>>>>>> >>>>>
>>>>>>>>>>>> >>>>> --
>>>>>>>>>>>> >>>>> Regards,
>>>>>>>>>>>> >>>>>
>>>>>>>>>>>> >>>>>
>>>>>>>>>>>> >>>>> Syed Ammad Ali
>>>>>>>>>>>> >>>
>>>>>>>>>>>> >>>
>>>>>>>>>>>> >>>
>>>>>>>>>>>> >>> --
>>>>>>>>>>>> >>> Regards,
>>>>>>>>>>>> >>>
>>>>>>>>>>>> >>>
>>>>>>>>>>>> >>> Syed Ammad Ali
>>>>>>>>>>>> >
>>>>>>>>>>>> > --
>>>>>>>>>>>> > Regards,
>>>>>>>>>>>> >
>>>>>>>>>>>> >
>>>>>>>>>>>> > Syed Ammad Ali
>>>>>>>>>>>> 
>>>>>>>>>>>> 
>>>>>>>>>>>> 
>>>>>>>>>>>> -- 
>>>>>>>>>>>> Mohammed Naser
>>>>>>>>>>>> VEXXHOST, Inc.
>>>>>>>>>> 
>>>>>>>>>> 
>>>>>>>>>> -- 
>>>>>>>>>> Regards,
>>>>>>>>>> 
>>>>>>>>>> 
>>>>>>>>>> Syed Ammad Ali
>>>>>>>> 
>>>>>>>> 
>>>>>>>> -- 
>>>>>>>> Regards,
>>>>>>>> 
>>>>>>>> 
>>>>>>>> Syed Ammad Ali
>>>>> 
>>>>> 
>>>>> -- 
>>>>> Regards,
>>>>> 
>>>>> 
>>>>> Syed Ammad Ali
> 
> <29a37aff-f1f6-46b3-8541-887491c6cfe8-k8s-cluster3-dcu52bgzpbuu-kube_masters-ocfrn2ikpcgd-0-32tmkqgdq7wl-master_config-gihyfv3wlyzd.log>
> <6fca39b1-8eda-4786-8424-e5b04434cce7-k8s-cluster3-dcu52bgzpbuu-kube_cluster_config-aht4it6p7wfk.log>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.openstack.org/pipermail/openstack-discuss/attachments/20210826/216e9f35/attachment-0001.html>


More information about the openstack-discuss mailing list