Tenant vlan Network Issue

Karera Tony tonykarera at gmail.com
Wed Sep 8 16:17:42 UTC 2021


Hello Team,

I  tried to create another cluster with tls enabled and below are the
results for podman ps . I have also attached the heat logs

 podman ps
CONTAINER ID  IMAGE
   COMMAND               CREATED         STATUS             PORTS
NAMES
e9c028723166
docker.io/openstackmagnum/heat-container-agent:wallaby-stable-1
 /usr/bin/start-he...  34 minutes ago  Up 34 minutes ago
 heat-container-agent
4128a9e951de  k8s.gcr.io/hyperkube:v1.18.16
   kube-apiserver --...  30 minutes ago  Up 30 minutes ago
 kube-apiserver
5430aa72e7f0  k8s.gcr.io/hyperkube:v1.18.16
   kube-controller-m...  30 minutes ago  Up 30 minutes ago
 kube-controller-manager
8f2c76272916  k8s.gcr.io/hyperkube:v1.18.16
   kube-scheduler --...  30 minutes ago  Up 30 minutes ago
 kube-scheduler
af5f1a661c65  quay.io/coreos/etcd:v3.4.6
    /usr/local/bin/et...  30 minutes ago  Up 30 minutes ago
 etcd

Regards

Tony Karera




On Wed, Sep 8, 2021 at 5:37 PM Ammad Syed <syedammad83 at gmail.com> wrote:

> Hi karera,
>
> I don’t think your the error is related to network side. Its because
> kublet is not installed properly in your kubernetes master node. You need
> to check logs that why hyperkube services are not started.
>
> You can check it in heat logs on mater node and you can check the status
> of kubelet container via podman ps commad.
>
> Ammad
> On Wed, Sep 8, 2021 at 8:28 PM Karera Tony <tonykarera at gmail.com> wrote:
>
>> Hello Slawek,
>>
>> Unfortunately I am not sure because all I did is configure the cluster.
>> Nothing else.
>>
>>
>> Regards
>>
>> Tony Karera
>>
>>
>>
>>
>> On Wed, Sep 8, 2021 at 3:31 PM Slawek Kaplonski <skaplons at redhat.com>
>> wrote:
>>
>>> Hi,
>>>
>>> On środa, 8 września 2021 15:16:57 CEST Karera Tony wrote:
>>> > Hello Slawek,
>>> >
>>> > I really need your help.
>>> >
>>> > I have tried all I can but still the k8s-cluster gets stuck with this
>>> log
>>> > here
>>> >
>>> > root at k8s-cluster-1-vo5s5jvtjoa7-master-0 core]# cd
>>> > /var/log/heat-config/heat-config-script/
>>> > [root at k8s-cluster-1-vo5s5jvtjoa7-master-0 heat-config-script]# ls
>>> > a70c65af-34e2-4071-be18-2e4d66f85cf6-k8s-cluster-1-vo5s5jvtjoa7-
>>> kube_cluster_c
>>> > onfig-asfqq352zosg.log
>>> > c42dcbf1-7812-49a2-b312-96c24225b407-k8s-cluster-1-vo5s5jvtjoa7-
>>> kube_masters-
>>> > waugw6vksw5p-0-vrt2gkfugedg-master_config-xv3vlmr27r46.log
>>> > [root at k8s-cluster-1-vo5s5jvtjoa7-master-0 heat-config-script]# cat
>>> > a70c65af-34e2-4071-be18-2e4d66f85cf6-k8s-cluster-1-vo5s5jvtjoa7-
>>> kube_cluster_c
>>> > onfig-asfqq352zosg.log [root at k8s-cluster-1-vo5s5jvtjoa7-master-0
>>> > heat-config-script]# cat
>>> > a70c65af-34e2-4071-be18-2e4d66f85cf6-k8s-cluster-1-vo5s5jvtjoa7-
>>> kube_cluster_c
>>> > onfig-asfqq352zosg.log [root at k8s-cluster-1-vo5s5jvtjoa7-master-0
>>> > heat-config-script]# cat
>>> > a70c65af-34e2-4071-be18-2e4d66f85cf6-k8s-cluster-1-vo5s5jvtjoa7-
>>> kube_cluster_c
>>> > onfig-asfqq352zosg.log Starting to run kube-apiserver-to-kubelet-role
>>> > + echo 'Waiting for Kubernetes API...'
>>> > Waiting for Kubernetes API...
>>> > ++ kubectl get --raw=/healthz
>>> > The connection to the server localhost:8080 was refused - did you
>>> specify
>>> > the right host or port?
>>> > + '[' ok = '' ']'
>>> > + sleep 5
>>> > ++ kubectl get --raw=/healthz
>>> > The connection to the server localhost:8080 was refused - did you
>>> specify
>>> > the right host or port?
>>>
>>> But why Your script is trying to connect to localhost? It's not the VM
>>> address
>>> I guess :)
>>>
>>> > + '[' ok = '' ']'
>>> > + sleep 5
>>> >
>>> >
>>> > Regards
>>> >
>>> > Tony Karera
>>> >
>>> >
>>> >
>>> >
>>> > On Mon, Aug 30, 2021 at 8:50 AM Slawek Kaplonski <skaplons at redhat.com>
>>> >
>>> > wrote:
>>> > > Hi,
>>> > >
>>> > > You can use vlan as tenant network too. But You have to remember
>>> that vlan
>>> > > is
>>> > > provider network, which means that You need to have configured vlans
>>> on
>>> > > Your
>>> > > network infrastructure too. And in case of tenant network, Neutron
>>> will
>>> > > automatically allocate one of the vlan ids to the network as user
>>> don't
>>> > > have
>>> > > possibility to choose vlan id for network (at least that's default
>>> policy
>>> > > in
>>> > > Neutron).
>>> > > So, You should check what vlan id is used by such tenant network,
>>> check if
>>> > > this vlan is properly confiugred on Your switches and then, if all is
>>> > > good,
>>> > > check where exactly such dhcp requests are dropped (are they going
>>> from
>>> > > compute node? are packets comming to the node with dhcp agent?).
>>> Base on
>>> > > that
>>> > > You can narrow down where the issue can be.
>>> > >
>>> > > On sobota, 28 sierpnia 2021 18:07:07 CEST Karera Tony wrote:
>>> > > > How can i make them both vxlan and vlan
>>> > > >
>>> > > > On Sat, 28 Aug 2021, 17:43 Mohammed Naser, <mnaser at vexxhost.com>
>>> wrote:
>>> > > > > Your issue is that tenant_network_types should be vxlan.
>>> > > > >
>>> > > > > On Sat, Aug 28, 2021 at 4:29 AM Karera Tony <
>>> tonykarera at gmail.com>
>>> > >
>>> > > wrote:
>>> > > > >> Hello Team,
>>> > > > >>
>>> > > > >> Whenever I create an internal network  with type vlan.
>>> > > > >>
>>> > > > >> The Instances don't get IPs but for external Network it's fine.
>>> > > > >>
>>> > > > >>
>>> > > > >> Below is the etc/kolla/config/neutron/ml2_conf.ini file
>>> > > > >>
>>> > > > >>
>>> > > > >> [ml2]
>>> > > > >> type_drivers = flat,vlan,vxlan
>>> > > > >> tenant_network_types = vlan,flat
>>> > > > >> mechanism_drivers = openvswitch,l2population
>>> > > > >>
>>> > > > >> [ml2_type_vlan]
>>> > > > >> network_vlan_ranges = physnet1
>>> > > > >> [ml2_type_flat]
>>> > > > >> flat_networks = physnet1
>>> > > > >>
>>> > > > >> Is there something I need to change or I need to configure the
>>> > >
>>> > > interface
>>> > >
>>> > > > >> differently
>>> > > > >> Regards
>>> > > > >>
>>> > > > >> Tony Karera
>>> > > > >>
>>> > > > >>
>>> > > > >> --
>>> > > > >
>>> > > > > Mohammed Naser
>>> > > > > VEXXHOST, Inc.
>>> > >
>>> > > --
>>> > > Slawek Kaplonski
>>> > > Principal Software Engineer
>>> > > Red Hat
>>>
>>>
>>> --
>>> Slawek Kaplonski
>>> Principal Software Engineer
>>> Red Hat
>>
>> --
> Regards,
>
>
> Syed Ammad Ali
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.openstack.org/pipermail/openstack-discuss/attachments/20210908/ad7ddf1b/attachment-0001.html>
-------------- next part --------------
A non-text attachment was scrubbed...
Name: 965d744e-2a63-4aec-9a65-ab83c92a47b4-k8s-cluster1-vnfx5pwx57gt-kube_masters-ikh4or2pa3o6-0-vawscctajrms-master_config-dchrxrgxwlou.log
Type: application/octet-stream
Size: 238556 bytes
Desc: not available
URL: <http://lists.openstack.org/pipermail/openstack-discuss/attachments/20210908/ad7ddf1b/attachment-0001.obj>


More information about the openstack-discuss mailing list