[openstack-dev] [magnum] [Rocky] K8 deployment on fedora-atomic is failed
Vikrant Aggarwal
ervikrant06 at gmail.com
Sat Dec 8 11:51:46 UTC 2018
Hello Team,
Any help on this issue?
Thanks & Regards,
Vikrant Aggarwal
On Tue, Dec 4, 2018 at 6:49 PM Vikrant Aggarwal <ervikrant06 at gmail.com>
wrote:
> Hello Team,
>
> Any help on this issue?
>
> Thanks & Regards,
> Vikrant Aggarwal
>
>
> On Fri, Nov 30, 2018 at 9:13 AM Vikrant Aggarwal <ervikrant06 at gmail.com>
> wrote:
>
>> Hi Feilong,
>>
>> Thanks for your reply.
>>
>> Kindly find the below outputs.
>>
>> [root at packstack1 ~]# rpm -qa | grep -i magnum
>> python-magnum-7.0.1-1.el7.noarch
>> openstack-magnum-conductor-7.0.1-1.el7.noarch
>> openstack-magnum-ui-5.0.1-1.el7.noarch
>> openstack-magnum-api-7.0.1-1.el7.noarch
>> puppet-magnum-13.3.1-1.el7.noarch
>> python2-magnumclient-2.10.0-1.el7.noarch
>> openstack-magnum-common-7.0.1-1.el7.noarch
>>
>> [root at packstack1 ~]# rpm -qa | grep -i heat
>> openstack-heat-ui-1.4.0-1.el7.noarch
>> openstack-heat-api-cfn-11.0.0-1.el7.noarch
>> openstack-heat-engine-11.0.0-1.el7.noarch
>> puppet-heat-13.3.1-1.el7.noarch
>> python2-heatclient-1.16.1-1.el7.noarch
>> openstack-heat-api-11.0.0-1.el7.noarch
>> openstack-heat-common-11.0.0-1.el7.noarch
>>
>> Thanks & Regards,
>> Vikrant Aggarwal
>>
>>
>> On Fri, Nov 30, 2018 at 2:44 AM Feilong Wang <feilong at catalyst.net.nz>
>> wrote:
>>
>>> Hi Vikrant,
>>>
>>> Before we dig more, it would be nice if you can let us know the version
>>> of your Magnum and Heat. Cheers.
>>>
>>>
>>> On 30/11/18 12:12 AM, Vikrant Aggarwal wrote:
>>>
>>> Hello Team,
>>>
>>> Trying to deploy on K8 on fedora atomic.
>>>
>>> Here is the output of cluster template:
>>> ~~~
>>> [root at packstack1 k8s_fedora_atomic_v1(keystone_admin)]# magnum
>>> cluster-template-show 16eb91f7-18fe-4ce3-98db-c732603f2e57
>>> WARNING: The magnum client is deprecated and will be removed in a future
>>> release.
>>> Use the OpenStack client to avoid seeing this message.
>>> +-----------------------+--------------------------------------+
>>> | Property | Value |
>>> +-----------------------+--------------------------------------+
>>> | insecure_registry | - |
>>> | labels | {} |
>>> | updated_at | - |
>>> | floating_ip_enabled | True |
>>> | fixed_subnet | - |
>>> | master_flavor_id | - |
>>> | user_id | 203617849df9490084dde1897b28eb53 |
>>> | uuid | 16eb91f7-18fe-4ce3-98db-c732603f2e57 |
>>> | no_proxy | - |
>>> | https_proxy | - |
>>> | tls_disabled | False |
>>> | keypair_id | kubernetes |
>>> | project_id | 45a6706c831c42d5bf2da928573382b1 |
>>> | public | False |
>>> | http_proxy | - |
>>> | docker_volume_size | 10 |
>>> | server_type | vm |
>>> | external_network_id | external1 |
>>> | cluster_distro | fedora-atomic |
>>> | image_id | f5954340-f042-4de3-819e-a3b359591770 |
>>> | volume_driver | - |
>>> | registry_enabled | False |
>>> | docker_storage_driver | devicemapper |
>>> | apiserver_port | - |
>>> | name | coe-k8s-template |
>>> | created_at | 2018-11-28T12:58:21+00:00 |
>>> | network_driver | flannel |
>>> | fixed_network | - |
>>> | coe | kubernetes |
>>> | flavor_id | m1.small |
>>> | master_lb_enabled | False |
>>> | dns_nameserver | 8.8.8.8 |
>>> +-----------------------+--------------------------------------+
>>> ~~~
>>> Found couple of issues in the logs of VM started by magnum.
>>>
>>> - etcd was not getting started because of incorrect permission on file
>>> "/etc/etcd/certs/server.key". This file is owned by root by default have
>>> 0440 as permission. Changed the permission to 0444 so that etcd can read
>>> the file. After that etcd started successfully.
>>>
>>> - etcd DB doesn't contain anything:
>>>
>>> [root at kube-cluster1-qobaagdob75g-master-0 ~]# etcdctl ls / -r
>>> [root at kube-cluster1-qobaagdob75g-master-0 ~]#
>>>
>>> - Flanneld is stuck in activating status.
>>> ~~~
>>> [root at kube-cluster1-qobaagdob75g-master-0 ~]# systemctl status flanneld
>>> ● flanneld.service - Flanneld overlay address etcd agent
>>> Loaded: loaded (/usr/lib/systemd/system/flanneld.service; enabled;
>>> vendor preset: disabled)
>>> Active: activating (start) since Thu 2018-11-29 11:05:39 UTC; 14s ago
>>> Main PID: 6491 (flanneld)
>>> Tasks: 6 (limit: 4915)
>>> Memory: 4.7M
>>> CPU: 53ms
>>> CGroup: /system.slice/flanneld.service
>>> └─6491 /usr/bin/flanneld -etcd-endpoints=
>>> http://127.0.0.1:2379 -etcd-prefix=/atomic.io/network
>>>
>>> Nov 29 11:05:44 kube-cluster1-qobaagdob75g-master-0.novalocal
>>> flanneld[6491]: E1129 11:05:44.569376 6491 network.go:102] failed to
>>> retrieve network config: 100: Key not found (/atomic.io) [3]
>>> Nov 29 11:05:45 kube-cluster1-qobaagdob75g-master-0.novalocal
>>> flanneld[6491]: E1129 11:05:45.584532 6491 network.go:102] failed to
>>> retrieve network config: 100: Key not found (/atomic.io) [3]
>>> Nov 29 11:05:46 kube-cluster1-qobaagdob75g-master-0.novalocal
>>> flanneld[6491]: E1129 11:05:46.646255 6491 network.go:102] failed to
>>> retrieve network config: 100: Key not found (/atomic.io) [3]
>>> Nov 29 11:05:47 kube-cluster1-qobaagdob75g-master-0.novalocal
>>> flanneld[6491]: E1129 11:05:47.673062 6491 network.go:102] failed to
>>> retrieve network config: 100: Key not found (/atomic.io) [3]
>>> Nov 29 11:05:48 kube-cluster1-qobaagdob75g-master-0.novalocal
>>> flanneld[6491]: E1129 11:05:48.686919 6491 network.go:102] failed to
>>> retrieve network config: 100: Key not found (/atomic.io) [3]
>>> Nov 29 11:05:49 kube-cluster1-qobaagdob75g-master-0.novalocal
>>> flanneld[6491]: E1129 11:05:49.709136 6491 network.go:102] failed to
>>> retrieve network config: 100: Key not found (/atomic.io) [3]
>>> Nov 29 11:05:50 kube-cluster1-qobaagdob75g-master-0.novalocal
>>> flanneld[6491]: E1129 11:05:50.729548 6491 network.go:102] failed to
>>> retrieve network config: 100: Key not found (/atomic.io) [3]
>>> Nov 29 11:05:51 kube-cluster1-qobaagdob75g-master-0.novalocal
>>> flanneld[6491]: E1129 11:05:51.749425 6491 network.go:102] failed to
>>> retrieve network config: 100: Key not found (/atomic.io) [3]
>>> Nov 29 11:05:52 kube-cluster1-qobaagdob75g-master-0.novalocal
>>> flanneld[6491]: E1129 11:05:52.776612 6491 network.go:102] failed to
>>> retrieve network config: 100: Key not found (/atomic.io) [3]
>>> Nov 29 11:05:53 kube-cluster1-qobaagdob75g-master-0.novalocal
>>> flanneld[6491]: E1129 11:05:53.790418 6491 network.go:102] failed to
>>> retrieve network config: 100: Key not found (/atomic.io) [3]
>>> ~~~
>>>
>>> - Continuously in the jouralctl logs following messages are printed.
>>>
>>> ~~~
>>> Nov 29 11:06:39 kube-cluster1-qobaagdob75g-master-0.novalocal
>>> kube-apiserver[6888]: F1129 11:06:39.338416 6888 server.go:269] Invalid
>>> Authorization Config: Unknown authorization mode Node specified
>>> Nov 29 11:06:39 kube-cluster1-qobaagdob75g-master-0.novalocal
>>> systemd[1]: kube-apiserver.service: Main process exited, code=exited,
>>> status=255/n/a
>>> Nov 29 11:06:39 kube-cluster1-qobaagdob75g-master-0.novalocal
>>> kube-scheduler[2540]: E1129 11:06:39.408272 2540 reflector.go:199]
>>> k8s.io/kubernetes/plugin/pkg/scheduler/factory/factory.go:463: Failed
>>> to list *api.Node: Get
>>> http://127.0.0.1:8080/api/v1/nodes?resourceVersion=0: dial tcp
>>> 127.0.0.1:8080: getsockopt: connection refused
>>> Nov 29 11:06:39 kube-cluster1-qobaagdob75g-master-0.novalocal
>>> kube-scheduler[2540]: E1129 11:06:39.444737 2540 reflector.go:199]
>>> k8s.io/kubernetes/plugin/pkg/scheduler/factory/factory.go:460: Failed
>>> to list *api.Pod: Get
>>> http://127.0.0.1:8080/api/v1/pods?fieldSelector=spec.nodeName%21%3D%2Cstatus.phase%21%3DFailed%2Cstatus.phase%21%3DSucceeded&resourceVersion=0:
>>> dial tcp 127.0.0.1:8080: getsockopt: connection refused
>>> Nov 29 11:06:39 kube-cluster1-qobaagdob75g-master-0.novalocal
>>> kube-scheduler[2540]: E1129 11:06:39.445793 2540 reflector.go:199]
>>> k8s.io/kubernetes/plugin/pkg/scheduler/factory/factory.go:466: Failed
>>> to list *api.PersistentVolume: Get
>>> http://127.0.0.1:8080/api/v1/persistentvolumes?resourceVersion=0: dial
>>> tcp 127.0.0.1:8080: getsockopt: connection refused
>>> Nov 29 11:06:39 kube-cluster1-qobaagdob75g-master-0.novalocal audit[1]:
>>> SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295
>>> subj=system_u:system_r:init_t:s0 msg='unit=kube-apiserver comm="systemd"
>>> exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed'
>>> Nov 29 11:06:39 kube-cluster1-qobaagdob75g-master-0.novalocal
>>> systemd[1]: Failed to start Kubernetes API Server.
>>> Nov 29 11:06:39 kube-cluster1-qobaagdob75g-master-0.novalocal
>>> systemd[1]: kube-apiserver.service: Unit entered failed state.
>>> Nov 29 11:06:39 kube-cluster1-qobaagdob75g-master-0.novalocal
>>> systemd[1]: kube-apiserver.service: Failed with result 'exit-code'.
>>> Nov 29 11:06:39 kube-cluster1-qobaagdob75g-master-0.novalocal
>>> kube-scheduler[2540]: E1129 11:06:39.611699 2540 reflector.go:199]
>>> k8s.io/kubernetes/plugin/pkg/scheduler/factory/factory.go:481: Failed
>>> to list *extensions.ReplicaSet: Get
>>> http://127.0.0.1:8080/apis/extensions/v1beta1/replicasets?resourceVersion=0:
>>> dial tcp 127.0.0.1:8080: getsockopt: connection refused
>>> ~~~
>>>
>>> Any help on above issue is highly appreciated.
>>>
>>> Thanks & Regards,
>>> Vikrant Aggarwal
>>>
>>> __________________________________________________________________________
>>> OpenStack Development Mailing List (not for usage questions)
>>> Unsubscribe: OpenStack-dev-request at lists.openstack.org?subject:unsubscribehttp://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
>>>
>>> --
>>> Cheers & Best regards,
>>> Feilong Wang (王飞龙)
>>> --------------------------------------------------------------------------
>>> Senior Cloud Software Engineer
>>> Tel: +64-48032246
>>> Email: flwang at catalyst.net.nz
>>> Catalyst IT Limited
>>> Level 6, Catalyst House, 150 Willis Street, Wellington
>>> --------------------------------------------------------------------------
>>>
>>>
>>> __________________________________________________________________________
>>> OpenStack Development Mailing List (not for usage questions)
>>> Unsubscribe:
>>> OpenStack-dev-request at lists.openstack.org?subject:unsubscribe
>>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
>>
>>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.openstack.org/pipermail/openstack-discuss/attachments/20181208/2abf2fb3/attachment-0001.html>
More information about the openstack-discuss
mailing list