Need help deploying Openstack

wodel youchi wodel.youchi at gmail.com
Tue Aug 24 21:25:59 UTC 2021


Hello,

After digging after grafana, it seems it needed to download something from
the internet, and i didn't really configure a proper gateway on the
external network.
So I started by configuring a proper gateway and I tested it with the half
deployed nodes, the I redid the deployment, and again I got this error :

2021-08-24 21:29:29.616805 | 525400e8-92c8-d397-6f7e-000000006133 |
>  FATAL | Clean up legacy Cinder keystone catalog entries | undercloud |
> error={"changed": false, "module_stderr": "Fa
> iled to discover available identity versions when contacting
> http://10.0.2.40:5000. Attempting to parse version from URL.\nTraceback
> (most recent call last):\n  File \"/usr/lib/python3.6/si
> te-packages/urllib3/connection.py\", line 162, in _new_conn\n
>  (self._dns_host, self.port), self.timeout, **extra_kw)\n  File
> \"/usr/lib/python3.6/site-packages/urllib3/util/connection.py
> \", line 80, in create_connection\n    raise err\n  File
> \"/usr/lib/python3.6/site-packages/urllib3/util/connection.py\", line 70,
> in create_connection\n    sock.connect(sa)\nTimeoutError:
> [Errno 110] Connection timed out\n\nDuring handling of the above
> exception, another exception occurred:\n\nTraceback (most recent call
> last):\n  File \"/usr/lib/python3.6/site-packages/urll
> ib3/connectionpool.py\", line 600, in urlopen\n    chunked=chunked)\n
>  File \"/usr/lib/python3.6/site-packages/urllib3/connectionpool.py\", line
> 354, in _make_request\n    conn.request(meth
> od, url, **httplib_request_kw)\n  File
> \"/usr/lib64/python3.6/http/client.py\", line 1269, in request\n
>  self._send_request(method, url, body, headers, encode_chunked)\n  File
> \"/usr/lib6
> 4/python3.6/http/client.py\", line 1315, in _send_request\n
>  self.endheaders(body, encode_chunked=encode_chunked)\n  File
> \"/usr/lib64/python3.6/http/client.py\", line 1264, in endheaders
> \n    self._send_output(message_body, encode_chunked=encode_chunked)\n
>  File \"/usr/lib64/python3.6/http/client.py\", line 1040, in _send_output\n
>    self.send(msg)\n  File \"/usr/lib64/pyt
> hon3.6/http/client.py\", line 978, in send\n    self.connect()\n  File
> \"/usr/lib/python3.6/site-packages/urllib3/connection.py\", line 184, in
> connect\n    conn = self._new_conn()\n  File
> \"/usr/lib/python3.6/site-packages/urllib3/connection.py\", line 171, in
> _new_conn\n    self, \"Failed to establish a new connection: %s\" %
> e)\nurllib3.exceptions.NewConnectionError: <urll
> ib3.connection.HTTPConnection object at 0x7f96f7b10cc0>: Failed to
> establish a new connection: [Errno 110] Connection timed out\n\nDuring
> handling of the above exception, another exception
> occurred:\n\nTraceback (most recent call last):\n  File
> \"/usr/lib/python3.6/site-packages/requests/adapters.py\", line 449, in
> send\n    timeout=timeout\n  File \"/usr/lib/python3.6/site-p
> ackages/urllib3/connectionpool.py\", line 638, in urlopen\n
>  _stacktrace=sys.exc_info()[2])\n  File
> \"/usr/lib/python3.6/site-packages/urllib3/util/retry.py\", line 399, in
> increment\n
>  raise MaxRetryError(_pool, url, error or
> ResponseError(cause))\nurllib3.exceptions.MaxRetryError:
> HTTPConnectionPool(host='10.0.2.40', port=5000): Max retries exceeded with
> url: / (Caused
> by NewConnectionError('<urllib3.connection.HTTPConnection object at
> 0x7f96f7b10cc0>: Failed to establish a new connection: [Errno 110]
> Connection timed out',))\n\nDuring handling of the ab$
> ve exception, another exception occurred:\n\nTraceback (most recent call
> last):\n  File
> \"/usr/lib/python3.6/site-packages/keystoneauth1/session.py\", line 997, in
> _send_request\n    resp $
>  self.session.request(method, url, **kwargs)\n  File
> \"/usr/lib/python3.6/site-packages/requests/sessions.py\", line 533, in
> request\n    resp = self.send(prep, **send_kwargs)\n  File \"/u$
> r/lib/python3.6/site-packages/requests/sessions.py\", line 646, in send\n
>    r = adapter.send(request, **kwargs)\n  File
> \"/usr/lib/python3.6/site-packages/requests/adapters.py\", line 516$
>  in send\n    raise ConnectionError(e,
> request=request)\nrequests.exceptions.ConnectionError:
> HTTPConnectionPool(host='10.0.2.40', port=5000): Max retries exceeded with
> url: / (Caused by N$wConnectionError('<urllib3.connection.HTTPConnection
> object at 0x7f96f7b10cc0>: Failed to establish a new connection: [Errno
> 110] Connection timed out',))\n\nDuring handling of the above e$
> ception, another exception occurred:\n\nTraceback (most recent call
> last):\n  File
> \"/usr/lib/python3.6/site-packages/keystoneauth1/identity/generic/base.py\",
> line 138, in _do_create_plug$
> n\n    authenticated=False)\n  File
> \"/usr/lib/python3.6/site-packages/keystoneauth1/identity/base.py\", line
> 610, in get_discovery\n    authenticated=authenticated)\n  File
> \"/usr/lib/pyt$
> on3.6/site-packages/keystoneauth1/discover.py\", line 1442, in
> get_discovery\n    disc = Discover(session, url,
> authenticated=authenticated)\n  File
> \"/usr/lib/python3.6/site-packages/keys$
> oneauth1/discover.py\", line 526, in __init__\n
>  authenticated=authenticated)\n  File
> \"/usr/lib/python3.6/site-packages/keystoneauth1/discover.py\", line 101,
> in get_version_data\n    r$
> sp = session.get(url, headers=headers, authenticated=authenticated)\n
>  File \"/usr/lib/python3.6/site-packages/keystoneauth1/session.py\", line
> 1116, in get\n    return self.request(url, '$
> ET', **kwargs)\n  File
> \"/usr/lib/python3.6/site-packages/keystoneauth1/session.py\", line 906, in
> request\n    resp = send(**kwargs)\n  File
> \"/usr/lib/python3.6/site-packages/keystoneaut$
> 1/session.py\", line 1013, in _send_request\n    raise
> exceptions.ConnectFailure(msg)\nkeystoneauth1.exceptions.connection.ConnectFailure:
> Unable to establish connection to http://10.0.2.4$
> :5000: HTTPConnectionPool(host='10.0.2.40', port=5000): Max retries
> exceeded with url: / (Caused by
> NewConnectionError('<urllib3.connection.HTTPConnection object at
> 0x7f96f7b10cc0>: Failed
> to establish a new connection: [Errno 110] Connection timed
> out',))\n\nDuring handling of the above exception, another exception
> occurred:\n\nTraceback (most recent call last):\n  File \"<$
> tdin>\", line 102, in <module>\n  File \"<stdin>\", line 94, in
> _ansiballz_main\n  File \"<stdin>\", line 40, in invoke_module\n  File
> \"/usr/lib64/python3.6/runpy.py\", line 205, in run_m$
> dule\n    return _run_module_code(code, init_globals, run_name,
> mod_spec)\n  File \"/usr/lib64/python3.6/runpy.py\", line 96, in
> _run_module_code\n    mod_name, mod_spec, pkg_name, script_$
> ame)\n  File \"/usr/lib64/python3.6/runpy.py\", line 85, in _run_code\n
>  exec(code, run_globals)\n  File
> \"/tmp/ansible_os_keystone_service_payload_wcyk6h37/ansible_os_keystone_service_p$
> yload.zip/ansible/modules/cloud/openstack/os_keystone_service.py\", line
> 194, in <module>\n  File
> \"/tmp/ansible_os_keystone_service_payload_wcyk6h37/ansible_os_keystone_service_payload.zi$
> /ansible/modules/cloud/openstack/os_keystone_service.py\", line 153, in
> main\n  File
> \"/usr/lib/python3.6/site-packages/openstack/cloud/_identity.py\", line
> 510, in search_services\n    se$
> vices = self.list_services()\n  File
> \"/usr/lib/python3.6/site-packages/openstack/cloud/_identity.py\", line
> 485, in list_services\n    if self._is_client_version('identity', 2):\n
>  File \$
> /usr/lib/python3.6/site-packages/openstack/cloud/openstackcloud.py\", line
> 459, in _is_client_version\n    client = getattr(self, client_name)\n  File
> \"/usr/lib/python3.6/site-packages/op$
> nstack/cloud/_identity.py\", line 32, in _identity_client\n    'identity',
> min_version=2, max_version='3.latest')\n  File
> \"/usr/lib/python3.6/site-packages/openstack/cloud/openstackcloud.$
> y\", line 406, in _get_versioned_client\n    if adapter.get_endpoint():\n
>  File \"/usr/lib/python3.6/site-packages/keystoneauth1/adapter.py\", line
> 282, in get_endpoint\n    return self.se$
> sion.get_endpoint(auth or self.auth, **kwargs)\n  File
> \"/usr/lib/python3.6/site-packages/keystoneauth1/session.py\", line 1218,
> in get_endpoint\n    return auth.get_endpoint(self, **kwarg$
> )\n  File
> \"/usr/lib/python3.6/site-packages/keystoneauth1/identity/base.py\", line
> 380, in get_endpoint\n    allow_version_hack=allow_version_hack,
> **kwargs)\n  File \"/usr/lib/python3.6/$
> ite-packages/keystoneauth1/identity/base.py\", line 271, in
> get_endpoint_data\n    service_catalog =
> self.get_access(session).service_catalog\n  File
> \"/usr/lib/python3.6/site-packages/key$
> toneauth1/identity/base.py\", line 134, in get_access\n    self.auth_ref =
> self.get_auth_ref(session)\n  File
> \"/usr/lib/python3.6/site-packages/keystoneauth1/identity/generic/base.py\",
> l$
> ne 206, in get_auth_ref\n    self._plugin =
> self._do_create_plugin(session)\n  File
> \"/usr/lib/python3.6/site-packages/keystoneauth1/identity/generic/base.py\",
> line 161, in _do_create_plu$
> in\n    'auth_url is correct. %s' %
> e)\nkeystoneauth1.exceptions.discovery.DiscoveryFailure: Could not find
> versioned identity endpoints when attempting to authenticate. Please check
> that $our auth_url is correct.
>
> *Unable to establish connection to http://10.0.2.40:5000
> <http://10.0.2.40:5000>: HTTPConnectionPool(host='10.0.2.40', port=5000):
> Max retries exceeded with url: / (Caused by
> NewConnectionError('<urllib3.connection.HTTPConnection object at
> 0x7f96f7b10cc0>: Failed to establish a new connection: [Errno 110]
> Connection timed out',))\n", "module_stdout": "", "msg": "MODULE
> FAILURE\nSee stdout/stderr for the exact error", "rc": 1} *
>
>
> 2021-08-24 21:29:29.617697 | 525400e8-92c8-d397-6f7e-000000006133 |
> TIMING | Clean up legacy Cinder keystone catalog entries | undercloud |
> 1:07:40.666419 | 130.85s
>
>
>
> PLAY RECAP
> *********************************************************************
>
>
> overcloud-computehci-0     : ok=260  changed=145  unreachable=0
>  failed=0    skipped=140  rescued=0    ignored=0
>
> overcloud-computehci-1     : ok=258  changed=145  unreachable=0
>  failed=0    skipped=140  rescued=0    ignored=0
>
> overcloud-computehci-2     : ok=255  changed=145  unreachable=0
>  failed=0    skipped=140  rescued=0    ignored=0
>
> overcloud-controller-0     : ok=295  changed=181  unreachable=0
>  failed=0    skipped=151  rescued=0    ignored=0
>
> overcloud-controller-1     : ok=289  changed=177  unreachable=0
>  failed=0    skipped=152  rescued=0    ignored=0
>
> overcloud-controller-2     : ok=288  changed=177  unreachable=0
>  failed=0    skipped=152  rescued=0    ignored=0
>
> undercloud                 : ok=105  changed=21   unreachable=0
>  failed=1    skipped=45   rescued=0    ignored=0
>
>
>
>
> 2021-08-24 21:29:29.730778 | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ Summary
> Information ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
>
> 2021-08-24 21:29:29.731007 | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ Total Tasks:
> 1723       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
>
> 2021-08-24 21:29:29.731098 | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ Elapsed Time:
> 1:07:40.779840 ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
>
> 2021-08-24 21:29:29.731172 |                                 UUID |
> Info |       Host |   Task Name |   Run Time
>
> 2021-08-24 21:29:29.731251 | 525400e8-92c8-d397-6f7e-000000003b9a |
>  SUMMARY | undercloud | Run tripleo-container-image-prepare logged to:
> /var/log/tripleo-container-image-prepare.log | 1762.93s
>
>
>
> 2021-08-24 21:29:29.731349 | 525400e8-92c8-d397-6f7e-0000000057aa |
>  SUMMARY | undercloud | tripleo-ceph-run-ansible : run ceph-ansible |
> 990.24s
> 2021-08-24 21:29:29.731433 | 525400e8-92c8-d397-6f7e-000000005951 |
>  SUMMARY | overcloud-controller-0 | tripleo_ha_wrapper : Run init bundle
> puppet on the host for haproxy | 133.22s
> 2021-08-24 21:29:29.731503 | 525400e8-92c8-d397-6f7e-000000006133 |
>  SUMMARY | undercloud | Clean up legacy Cinder keystone catalog entries |
> 130.85s
> 2021-08-24 21:29:29.731569 | 525400e8-92c8-d397-6f7e-000000006012 |
>  SUMMARY | overcloud-controller-0 | Wait for containers to start for step 3
> using paunch | 103.45s
> 2021-08-24 21:29:29.731643 | 525400e8-92c8-d397-6f7e-000000004337 |
>  SUMMARY | overcloud-computehci-0 | Pre-fetch all the containers | 94.00s
>
> 2021-08-24 21:29:29.731729 | 525400e8-92c8-d397-6f7e-000000004378 |
>  SUMMARY | overcloud-computehci-2 | Pre-fetch all the containers | 92.64s
>
> 2021-08-24 21:29:29.731795 | 525400e8-92c8-d397-6f7e-000000004337 |
>  SUMMARY | overcloud-computehci-1 | Pre-fetch all the containers | 86.38s
>
> 2021-08-24 21:29:29.731867 | 525400e8-92c8-d397-6f7e-000000004d68 |
>  SUMMARY | overcloud-controller-0 | Wait for container-puppet tasks
> (generate config) to finish | 84.13s
> 2021-08-24 21:29:29.731946 | 525400e8-92c8-d397-6f7e-000000004d99 |
>  SUMMARY | overcloud-controller-2 | Wait for container-puppet tasks
> (generate config) to finish | 80.76s
> 2021-08-24 21:29:29.732012 | 525400e8-92c8-d397-6f7e-00000000427c |
>  SUMMARY | overcloud-controller-1 | Pre-fetch all the containers | 80.21s
>
> 2021-08-24 21:29:29.732073 | 525400e8-92c8-d397-6f7e-00000000427c |
>  SUMMARY | overcloud-controller-0 | Pre-fetch all the containers | 77.03s
>
> 2021-08-24 21:29:29.732138 | 525400e8-92c8-d397-6f7e-0000000042f5 |
>  SUMMARY | overcloud-controller-2 | Pre-fetch all the containers | 76.32s
>
> 2021-08-24 21:29:29.732202 | 525400e8-92c8-d397-6f7e-000000004dd3 |
>  SUMMARY | overcloud-controller-1 | Wait for container-puppet tasks
> (generate config) to finish | 74.36s
> 2021-08-24 21:29:29.732266 | 525400e8-92c8-d397-6f7e-000000005da7 |
>  SUMMARY | overcloud-controller-0 | tripleo_ha_wrapper : Run init bundle
> puppet on the host for ovn_dbs | 68.39s
> 2021-08-24 21:29:29.732329 | 525400e8-92c8-d397-6f7e-000000005ce2 |
>  SUMMARY | overcloud-controller-0 | Wait for containers to start for step 2
> using paunch | 64.55s
> 2021-08-24 21:29:29.732398 | 525400e8-92c8-d397-6f7e-000000004b97 |
>  SUMMARY | overcloud-controller-2 | Wait for puppet host configuration to
> finish | 58.13s
> 2021-08-24 21:29:29.732463 | 525400e8-92c8-d397-6f7e-000000004c1a |
>  SUMMARY | overcloud-controller-1 | Wait for puppet host configuration to
> finish | 58.11s
> 2021-08-24 21:29:29.732526 | 525400e8-92c8-d397-6f7e-000000005bd3 |
>  SUMMARY | overcloud-controller-1 | Wait for containers to start for step 2
> using paunch | 58.09s
> 2021-08-24 21:29:29.732589 | 525400e8-92c8-d397-6f7e-000000005b9b |
>  SUMMARY | overcloud-controller-2 | Wait for containers to start for step 2
> using paunch | 58.09s
>


Thank you again for your assistance.

Regards.

Le mar. 24 août 2021 à 08:59, wodel youchi <wodel.youchi at gmail.com> a
écrit :

> Hi, and thanks for your help
>
> As for Ceph, here is  container prepare
> parameter_defaults:
>  ContainerImagePrepare:
>  - push_destination: true
>    set:
>      ceph_alertmanager_image: alertmanager
>      ceph_alertmanager_namespace: quay.ceph.io/prometheus
>      ceph_alertmanager_tag: v0.16.2
>      ceph_grafana_image: grafana
>      ceph_grafana_namespace: quay.ceph.io/app-sre
>      *ceph_grafana_tag: 5.4.3*
>      ceph_image: daemon
>      ceph_namespace: quay.ceph.io/ceph-ci
>      ceph_node_exporter_image: node-exporter
>      ceph_node_exporter_namespace: quay.ceph.io/prometheus
>      ceph_node_exporter_tag: v0.17.0
>      ceph_prometheus_image: prometheus
>      ceph_prometheus_namespace: quay.ceph.io/prometheus
>      ceph_prometheus_tag: v2.7.2
>      *ceph_tag: v4.0.19-stable-4.0-nautilus-centos-7-x86_64*
>      name_prefix: centos-binary-
>      name_suffix: ''
>      namespace: quay.io/tripleotraincentos8
>      neutron_driver: ovn
>      rhel_containers: false
>      tag: current-tripleo
>    tag_from_label: rdo_version
>
> And yes, the 10.200.7.0/24 network is my storage network
> Here is a snippet from my network_data.yaml
>
> - name: Storage
>  vip: true
>  vlan: 1107
>  name_lower: storage
>  ip_subnet: '10.200.7.0/24'
>  allocation_pools: [{'start': '10.200.7.150', 'end': '10.200.7.169'}]
>
> I will look into the grafana service to see why it's not booting and get
> back to you.
>
> Regards.
>
> Le lun. 23 août 2021 à 17:28, Francesco Pantano <fpantano at redhat.com> a
> écrit :
>
>> Hello,
>> thanks John for your reply here.
>> A few more comments inline:
>>
>> On Mon, Aug 23, 2021 at 6:16 PM John Fulton <johfulto at redhat.com> wrote:
>>
>>> On Mon, Aug 23, 2021 at 10:52 AM wodel youchi <wodel.youchi at gmail.com>
>>> wrote:
>>> >
>>> > Hi,
>>> >
>>> > I redid the undercloud deployment for the Train version for now. And I
>>> verified the download URL for the images.
>>> > My overcloud deployment has moved forward but I still get errors.
>>> >
>>> > This is what I got this time :
>>> >>
>>> >>        "TASK [ceph-grafana : wait for grafana to start]
>>> ********************************",
>>> >>        "Monday 23 August 2021  14:55:21 +0100 (0:00:00.961)
>>>  0:12:59.319 ********* ",
>>> >>        "fatal: [overcloud-controller-0]: FAILED! => {\"changed\":
>>> false, \"elapsed\": 300, \"msg\": \"Timeout when waiting for 10.20
>>> >> 0.7.151:3100\"}",
>>> >>        "fatal: [overcloud-controller-1]: FAILED! => {\"changed\":
>>> false, \"elapsed\": 300, \"msg\": \"Timeout when waiting for 10.20
>>> >> 0.7.155:3100\"}",
>>> >>        "fatal: [overcloud-controller-2]: FAILED! => {\"changed\":
>>> false, \"elapsed\": 300, \"msg\": \"Timeout when waiting for 10.20
>>> >> 0.7.165:3100\"}",
>>>
>>> I'm not certain of the ceph-ansible version you're using but it should
>>> be a version 4 with train. ceph-ansible should already be installed on
>>> your undercloud judging by this error and in the latest version 4 this
>>> task is where it failed:
>>>
>>>
>>> https://github.com/ceph/ceph-ansible/blob/v4.0.64/roles/ceph-grafana/tasks/configure_grafana.yml#L112-L115
>>>
>>> You can check the status of this service on your three controllers and
>>> then debug it directly.
>>
>> As John pointed out, ceph-ansible is able to configure, render and start
>> the associated
>> systemd unit for all the ceph monitoring stack components (node-exported,
>> prometheus, alertmanager and
>> grafana).
>> You can ssh to your controllers, and check the systemd unit associated,
>> checking the journal to see why
>> they failed to start (I saw there's a timeout waiting for the container
>> to start).
>> A potential plan, in this case, could be:
>>
>> 1. check the systemd unit (I guess you can start with grafana which is
>> the failed service)
>> 2. look at the journal logs (feel free to attach here the relevant part
>> of the output)
>> 3. double check the network where the service is bound (can you attach
>> the /var/lib/mistral/<stack>/ceph-ansible/group_vars/all.yaml)
>>     * The grafana process should be run on the storage network, but I see
>> a "Timeout when waiting for 10.200.7.165:3100": is that network the
>> right one?
>>
>>>
>>
>>
>>>   John
>>>
>>> >>        "RUNNING HANDLER [ceph-prometheus : service handler]
>>> ****************************",
>>> >>        "Monday 23 August 2021  15:00:22 +0100 (0:05:00.767)
>>>  0:18:00.087 ********* ",
>>> >>        "PLAY RECAP
>>> *********************************************************************",
>>> >>        "overcloud-computehci-0     : ok=224  changed=23
>>>  unreachable=0    failed=0    skipped=415  rescued=0    ignored=0   ",
>>> >>        "overcloud-computehci-1     : ok=199  changed=18
>>>  unreachable=0    failed=0    skipped=392  rescued=0    ignored=0   ",
>>> >>        "overcloud-computehci-2     : ok=212  changed=23
>>>  unreachable=0    failed=0    skipped=390  rescued=0    ignored=0   ",
>>> >>        "overcloud-controller-0     : ok=370  changed=52
>>>  unreachable=0    failed=1    skipped=539  rescued=0    ignored=0   ",
>>> >>        "overcloud-controller-1     : ok=308  changed=43
>>>  unreachable=0    failed=1    skipped=495  rescued=0    ignored=0   ",
>>> >>        "overcloud-controller-2     : ok=317  changed=45
>>>  unreachable=0    failed=1    skipped=493  rescued=0    ignored=0   ",
>>> >>
>>> >>        "INSTALLER STATUS
>>> ***************************************************************",
>>> >>        "Install Ceph Monitor           : Complete (0:00:52)",
>>> >>        "Install Ceph Manager           : Complete (0:05:49)",
>>> >>        "Install Ceph OSD               : Complete (0:02:28)",
>>> >>        "Install Ceph RGW               : Complete (0:00:27)",
>>> >>        "Install Ceph Client            : Complete (0:00:33)",
>>> >>        "Install Ceph Grafana           : In Progress (0:05:54)",
>>> >>        "\tThis phase can be restarted by running:
>>> roles/ceph-grafana/tasks/main.yml",
>>> >>        "Install Ceph Node Exporter     : Complete (0:00:28)",
>>> >>        "Monday 23 August 2021  15:00:22 +0100 (0:00:00.006)
>>>  0:18:00.094 ********* ",
>>> >>
>>> "===============================================================================
>>> ",
>>> >>        "ceph-grafana : wait for grafana to start
>>> ------------------------------ 300.77s",
>>> >>        "ceph-facts : get ceph current status
>>> ---------------------------------- 300.27s",
>>> >>        "ceph-container-common : pulling
>>> udtrain.ctlplane.umaitek.dz:8787/ceph-ci/daemon:v4.0.19-stable-4.0-nautilus-centos-7-x86_64
>>> >> image -- 19.04s",
>>> >>        "ceph-mon : waiting for the monitor(s) to form the quorum...
>>> ------------ 12.83s",
>>> >>        "ceph-osd : use ceph-volume lvm batch to create bluestore osds
>>> ---------- 12.13s",
>>> >>        "ceph-osd : wait for all osd to be up
>>> ----------------------------------- 11.88s",
>>> >>        "ceph-osd : set pg_autoscale_mode value on pool(s)
>>> ---------------------- 11.00s",
>>> >>        "ceph-osd : create openstack pool(s)
>>> ------------------------------------ 10.80s",
>>> >>        "ceph-grafana : make sure grafana is down
>>> ------------------------------- 10.66s",
>>> >>        "ceph-osd : customize pool crush_rule
>>> ----------------------------------- 10.15s",
>>> >>        "ceph-osd : customize pool size
>>> ----------------------------------------- 10.15s",
>>> >>        "ceph-osd : customize pool min_size
>>> ------------------------------------- 10.14s",
>>> >>        "ceph-osd : assign application to pool(s)
>>> ------------------------------- 10.13s",
>>> >>        "ceph-osd : list existing pool(s)
>>> ---------------------------------------- 8.59s",
>>> >>
>>> >>        "ceph-mon : fetch ceph initial keys
>>> -------------------------------------- 7.01s",
>>> >>        "ceph-container-common : get ceph version
>>> -------------------------------- 6.75s",
>>> >>        "ceph-prometheus : start prometheus services
>>> ----------------------------- 6.67s",
>>> >>        "ceph-mgr : wait for all mgr to be up
>>> ------------------------------------ 6.66s",
>>> >>        "ceph-grafana : start the grafana-server service
>>> ------------------------- 6.33s",
>>> >>        "ceph-mgr : create ceph mgr keyring(s) on a mon node
>>> --------------------- 6.26s"
>>> >>    ],
>>> >>    "failed_when_result": true
>>> >> }
>>> >> 2021-08-23 15:00:24.427687 | 525400e8-92c8-47b1-e162-00000000597d |
>>>    TIMING | tripleo-ceph-run-ansible : print ceph-ansible outpu$
>>> >> in case of failure | undercloud | 0:37:30.226345 | 0.25s
>>> >>
>>> >> PLAY RECAP
>>> *********************************************************************
>>> >> overcloud-computehci-0     : ok=213  changed=117  unreachable=0
>>> failed=0    skipped=120  rescued=0    ignored=0
>>> >> overcloud-computehci-1     : ok=207  changed=117  unreachable=0
>>> failed=0    skipped=120  rescued=0    ignored=0
>>> >> overcloud-computehci-2     : ok=207  changed=117  unreachable=0
>>> failed=0    skipped=120  rescued=0    ignored=0
>>> >> overcloud-controller-0     : ok=237  changed=145  unreachable=0
>>> failed=0    skipped=128  rescued=0    ignored=0
>>> >> overcloud-controller-1     : ok=232  changed=145  unreachable=0
>>> failed=0    skipped=128  rescued=0    ignored=0
>>> >> overcloud-controller-2     : ok=232  changed=145  unreachable=0
>>> failed=0    skipped=128  rescued=0    ignored=0
>>> >> undercloud                 : ok=100  changed=18   unreachable=0
>>> failed=1    skipped=37   rescued=0    ignored=0
>>> >>
>>> >> 2021-08-23 15:00:24.559997 | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
>>> Summary Information ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
>>> >> 2021-08-23 15:00:24.560328 | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ Total
>>> Tasks: 1366       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
>>> >> 2021-08-23 15:00:24.560419 | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ Elapsed
>>> Time: 0:37:30.359090 ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
>>> >> 2021-08-23 15:00:24.560490 |                                 UUID |
>>>      Info |       Host |   Task Name |   Run Time
>>> >> 2021-08-23 15:00:24.560589 | 525400e8-92c8-47b1-e162-00000000597b |
>>>   SUMMARY | undercloud | tripleo-ceph-run-ansible : run ceph-ans
>>> >> ible | 1082.71s
>>> >> 2021-08-23 15:00:24.560675 | 525400e8-92c8-47b1-e162-000000004d9a |
>>>   SUMMARY | overcloud-controller-1 | Wait for container-puppet t
>>> >> asks (generate config) to finish | 356.02s
>>> >> 2021-08-23 15:00:24.560763 | 525400e8-92c8-47b1-e162-000000004d6a |
>>>   SUMMARY | overcloud-controller-0 | Wait for container-puppet t
>>> >> asks (generate config) to finish | 355.74s
>>> >> 2021-08-23 15:00:24.560839 | 525400e8-92c8-47b1-e162-000000004dd0 |
>>>   SUMMARY | overcloud-controller-2 | Wait for container-puppet t
>>> >> asks (generate config) to finish | 355.68s
>>> >> 2021-08-23 15:00:24.560912 | 525400e8-92c8-47b1-e162-000000003bb1 |
>>>   SUMMARY | undercloud | Run tripleo-container-image-prepare log
>>> >> ged to: /var/log/tripleo-container-image-prepare.log | 143.03s
>>> >> 2021-08-23 15:00:24.560986 | 525400e8-92c8-47b1-e162-000000004b13 |
>>>   SUMMARY | overcloud-controller-0 | Wait for puppet host config
>>> >> uration to finish | 125.36s
>>> >> 2021-08-23 15:00:24.561057 | 525400e8-92c8-47b1-e162-000000004b88 |
>>>   SUMMARY | overcloud-controller-2 | Wait for puppet host config
>>> >> uration to finish | 125.33s
>>> >> 2021-08-23 15:00:24.561128 | 525400e8-92c8-47b1-e162-000000004b4b |
>>>   SUMMARY | overcloud-controller-1 | Wait for puppet host config
>>> >> uration to finish | 125.25s
>>> >> 2021-08-23 15:00:24.561300 | 525400e8-92c8-47b1-e162-000000001dc4 |
>>>   SUMMARY | overcloud-controller-2 | Run puppet on the host to a
>>> >> pply IPtables rules | 108.08s
>>> >> 2021-08-23 15:00:24.561374 | 525400e8-92c8-47b1-e162-000000001e4f |
>>>   SUMMARY | overcloud-controller-0 | Run puppet on the host to a
>>> >> pply IPtables rules | 107.34s
>>> >> 2021-08-23 15:00:24.561444 | 525400e8-92c8-47b1-e162-000000004c8d |
>>>   SUMMARY | overcloud-computehci-2 | Wait for container-puppet t
>>> >> asks (generate config) to finish | 96.56s
>>> >> 2021-08-23 15:00:24.561514 | 525400e8-92c8-47b1-e162-000000004c33 |
>>>   SUMMARY | overcloud-computehci-0 | Wait for container-puppet t
>>> >> asks (generate config) to finish | 96.38s
>>> >> 2021-08-23 15:00:24.561580 | 525400e8-92c8-47b1-e162-000000004c60 |
>>>   SUMMARY | overcloud-computehci-1 | Wait for container-puppet t
>>> >> asks (generate config) to finish | 93.41s
>>> >> 2021-08-23 15:00:24.561645 | 525400e8-92c8-47b1-e162-00000000434d |
>>>   SUMMARY | overcloud-computehci-0 | Pre-fetch all the container
>>> >> s | 92.70s
>>> >> 2021-08-23 15:00:24.561712 | 525400e8-92c8-47b1-e162-0000000043ed |
>>>   SUMMARY | overcloud-computehci-2 | Pre-fetch all the container
>>> >> s | 91.90s
>>> >> 2021-08-23 15:00:24.561782 | 525400e8-92c8-47b1-e162-000000004385 |
>>>   SUMMARY | overcloud-computehci-1 | Pre-fetch all the container
>>> >> s | 91.88s
>>> >> 2021-08-23 15:00:24.561876 | 525400e8-92c8-47b1-e162-00000000491c |
>>>   SUMMARY | overcloud-computehci-1 | Wait for puppet host config
>>> >> uration to finish | 90.37s
>>> >> 2021-08-23 15:00:24.561947 | 525400e8-92c8-47b1-e162-000000004951 |
>>>   SUMMARY | overcloud-computehci-2 | Wait for puppet host config
>>> >> uration to finish | 90.37s
>>> >> 2021-08-23 15:00:24.562016 | 525400e8-92c8-47b1-e162-0000000048e6 |
>>>   SUMMARY | overcloud-computehci-0 | Wait for puppet host config
>>> >> uration to finish | 90.35s
>>> >> 2021-08-23 15:00:24.562080 | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ End
>>> Summary Information ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
>>> >> 2021-08-23 15:00:24.562196 | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
>>> State Information ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
>>> >> 2021-08-23 15:00:24.562311 | ~~~~~~~~~~~~~~~~~~ Number of nodes which
>>> did not deploy successfully: 1 ~~~~~~~~~~~~~~~~~
>>> >> 2021-08-23 15:00:24.562379 |  The following node(s) had failures:
>>> undercloud
>>> >> 2021-08-23 15:00:24.562456 |
>>> ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
>>> >> Host 10.0.2.40 not found in /home/stack/.ssh/known_hosts
>>> >> Ansible failed, check log at
>>> /var/lib/mistral/overcloud/ansible.log.Overcloud Endpoint:
>>> http://10.0.2.40:5000
>>> >> Overcloud Horizon Dashboard URL: http://10.0.2.40:80/dashboard
>>> >> Overcloud rc file: /home/stack/overcloudrc
>>> >> Overcloud Deployed with error
>>> >> Overcloud configuration failed.
>>> >>
>>> >
>>> >
>>> > Could someone help debug this, the ansible.log is huge, I can't see
>>> what's the origin of the problem, if someone can point me to the right
>>> direction it will aprecciated.
>>> > Thanks in advance.
>>> >
>>> > Regards.
>>> >
>>> > Le mer. 18 août 2021 à 18:02, Wesley Hayutin <whayutin at redhat.com> a
>>> écrit :
>>> >>
>>> >>
>>> >>
>>> >> On Wed, Aug 18, 2021 at 10:10 AM Dmitry Tantsur <dtantsur at redhat.com>
>>> wrote:
>>> >>>
>>> >>> Hi,
>>> >>>
>>> >>> On Wed, Aug 18, 2021 at 4:39 PM wodel youchi <wodel.youchi at gmail.com>
>>> wrote:
>>> >>>>
>>> >>>> Hi,
>>> >>>> I am trying to deploy openstack with tripleO using VMs and
>>> nested-KVM for the compute node. This is for test and learning purposes.
>>> >>>>
>>> >>>> I am using the Train version and following some tutorials.
>>> >>>> I prepared my different template files and started the deployment,
>>> but I got these errors :
>>> >>>>
>>> >>>> Failed to provision instance fc40457e-4b3c-4402-ae9d-c528f2c2ad30:
>>> Asynchronous exception: Node failed to deploy. Exception: Agent API for
>>> node 6d3724fc-6f13-4588-bbe5-56bc4f9a4f87 returned HTTP status code 404
>>> with error: Not found: Extension with id iscsi not found. for node
>>> >>>>
>>> >>>
>>> >>> You somehow ended up using master (Xena release) deploy ramdisk with
>>> Train TripleO. You need to make sure to download Train images. I hope
>>> TripleO people can point you at the right place.
>>> >>>
>>> >>> Dmitry
>>> >>
>>> >>
>>> >> http://images.rdoproject.org/centos8/
>>> >> http://images.rdoproject.org/centos8/train/rdo_trunk/current-tripleo/
>>> >>
>>> >>>
>>> >>>
>>> >>>>
>>> >>>> and
>>> >>>>
>>> >>>> Got HTTP 409: {"errors": [{"status": 409, "title": "Conflict",
>>> "detail": "There was a conflict when trying to complete your request.\n\n
>>> Unable to allocate inventory: Unable to create allocation for
>>> 'CUSTOM_BAREMETAL' on resource provider
>>> '6d3724fc-6f13-4588-bbe5-56bc4f9a4f87'. The requested amount would exceed
>>> the capacity. ",
>>> >>>>
>>> >>>> Could you help understand what those errors mean? I couldn't find
>>> anything similar on the net.
>>> >>>>
>>> >>>> Thanks in advance.
>>> >>>>
>>> >>>> Regards.
>>> >>>
>>> >>>
>>> >>>
>>> >>> --
>>> >>> Red Hat GmbH, https://de.redhat.com/ , Registered seat: Grasbrunn,
>>> >>> Commercial register: Amtsgericht Muenchen, HRB 153243,
>>> >>> Managing Directors: Charles Cachera, Brian Klemm, Laurie Krebs,
>>> Michael O'Neill
>>>
>>>
>>>
>>
>> --
>> Francesco Pantano
>> GPG KEY: F41BD75C
>>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.openstack.org/pipermail/openstack-discuss/attachments/20210824/f2261e98/attachment-0001.html>


More information about the openstack-discuss mailing list