Need help deploying Openstack

wodel youchi wodel.youchi at gmail.com
Wed Aug 25 20:45:40 UTC 2021


Hi,

And thank you all for your help, I've managed to deploy my first overcloud.

But, again I have another problem. I am using HCI deployment and I did
include ceph dashboard in my deployment script, but I didn't find the
dashboard, after reviewing the RedHat documentation, it seems that I have
to use this role "ControllerStorageDashboard". This is what I did, but I
got this :


>
> *RESP BODY: {"resources": [{"updated_time": "2021-08-25T20:14:20Z",
> "creation_time": "2021-08-25T20:14:20Z", "logical_resource_id": "0",
> "resource_name": "0", "physical_resource_id":
> "a21b3498-fbdb-4a19-8e23-9dd71232b473", "resource_status": "CREATE_FAILED",
> "resource_status_reason": "BadRequest:
> resources[0].resources.OVNMacAddressPort: Invalid input for operation:
> 'tripleo_ovn_mac_port_name=ControllerStorageDashboard-ovn-mac-0' exceeds
> maximum length of 60.\nNeutron server returns request_ids:
> ['req-467b58ef-dfd7-42c5-bb07-4f0f99b77332']", "resource_type": "OS*
> ::TripleO::OVNMacAddressPort", "links": [{"href": "
> https://10.200.24.2:13004/v1/4f94deb9a28549c0a78f232756c7599a/stacks/overcloud-ControllerStorageDashboard-vtmxtvxpzggi-1-ue2d2riknvna-Cont
> rollerStorageDashboardOVNChassisMacPorts-ui4dsb2tnkbk/ae81eb26-2f4b-4ae0-8826-af32be18ce14/resources/0",
> "rel": "self"}, {"href": "
> https://10.200.24.2:13004/v1/4f94deb9a28549c0a78f232756c75
> 99a/stacks/overcloud-ControllerStorageDashboard-vtmxtvxpzggi-1-ue2d2riknvna-ControllerStorageDashboardOVNChassisMacPorts-ui4dsb2tnkbk/ae81eb26-2f4b-4ae0-8826-af32be18ce14",
> "rel": "stack"},
>  {"href": "
> https://10.200.24.2:13004/v1/4f94deb9a28549c0a78f232756c7599a/stacks/overcloud-ControllerStorageDashboard-vtmxtvxpzggi-1-ue2d2riknvna-ControllerStorageDashboardOVNChassisMacPorts
> -ui4dsb2tnkbk-0-yfuxj4ahxviu/a21b3498-fbdb-4a19-8e23-9dd71232b473", "rel":
> "nested"}], "required_by": [], "parent_resource":
> "ControllerStorageDashboardOVNChassisMacPorts"}]}
> GET call to orchestration for
> https://10.200.24.2:13004/v1/4f94deb9a28549c0a78f232756c7599a/stacks/overcloud-ControllerStorageDashboard-vtmxtvxpzggi-1-ue2d2riknvna-ControllerStorageDashboar
> dOVNChassisMacPorts-ui4dsb2tnkbk/ae81eb26-2f4b-4ae0-8826-af32be18ce14/resources
> used request id req-9609844f-f173-4e80-a3bd-bc287e88b00f
> REQ: curl -g -i --cacert
> "/etc/pki/ca-trust/source/anchors/cm-local-ca.pem" -X GET
> https://10.200.24.2:13004/v1/4f94deb9a28549c0a78f232756c7599a/stacks/a21b3498-fbdb-4a19-8e23-9dd71232b473/
> resources -H "Accept: application/json" -H "Content-Type:
> application/json" -H "User-Agent: python-heatclient" -H "X-Auth-Token:
> {SHA256}d296097c7cdf0beb50127e0a1d03cb8a702e18d543600f51b16d
> ab4987811a6a" -H "X-Region-Name: "
> https://10.200.24.2:13004 "GET
> /v1/4f94deb9a28549c0a78f232756c7599a/stacks/a21b3498-fbdb-4a19-8e23-9dd71232b473/resources
> HTTP/1.1" 302 649
> RESP: [302] Content-Length: 649 Content-Type: application/json Date: Wed,
> 25 Aug 2021 20:15:11 GMT Location:
> https://10.200.24.2:13004/v1/4f94deb9a28549c0a78f232756c7599a/stacks/overcloud-C
> ontrollerStorageDashboard-vtmxtvxpzggi-1-ue2d2rik
>
...
>
...
>
overcloud.ControllerStorageDashboard.0.ControllerStorageDashboardOVNChassisMacPorts.0.OVNMacAddressPort:
>   resource_type: OS::Neutron::Port
>   physical_resource_id: 259e39f8-9e7b-4494-bb2d-ff7b2cf0ad40
>   status: CREATE_FAILED
>   status_reason: |
>
>
>
> *BadRequest: resources.OVNMacAddressPort: Invalid input for operation:
> 'tripleo_ovn_mac_port_name=ControllerStorageDashboard-ovn-mac-0' exceeds
> maximum length of 60.    Neutron server returns request_ids:
> ['req-322ab0aa-0e1c-416f-be81-b48230d3dab1']overcloud.ControllerStorageDashboard.2.ControllerStorageDashboardOVNChassisMacPorts.0.OVNMacAddressPort:
> resource_type: OS::Neutron::Port*
>   physical_resource_id: c7daf26b-7f96-43cf-8678-11d456b5cdfe
>   status: CREATE_FAILED
>   status_reason: |
>     BadRequest: resources.OVNMacAddressPort: Invalid input for operation:
> 'tripleo_ovn_mac_port_name=ControllerStorageDashboard-ovn-mac-0' exceeds
> maximum length of 60.
>     Neutron server returns request_ids:
> ['req-9e3e19dd-4974-4007-9df0-ee9774369495']
>
> overcloud.ControllerStorageDashboard.1.ControllerStorageDashboardOVNChassisMacPorts.0.OVNMacAddressPort:
>   resource_type: OS::Neutron::Port
>   physical_resource_id: c902e259-f299-457f-8b0d-c37fb40e0d32
>   status: CREATE_FAILED
>   status_reason: |
>     BadRequest: resources.OVNMacAddressPort: Invalid input for operation:
> 'tripleo_ovn_mac_port_name=ControllerStorageDashboard-ovn-mac-0' exceeds
> maximum length of 60.
>     Neutron server returns request_ids:
> ['req-467b58ef-dfd7-42c5-bb07-4f0f99b77332']
> clean_up ListStackFailures:
> END return value: 0
> Instantiating messaging websocket client: wss://10.200.24.2:3000
>

I couldn't find anything on the web about this error.


Regards

Le mar. 24 août 2021 à 22:25, wodel youchi <wodel.youchi at gmail.com> a
écrit :

> Hello,
>
> After digging after grafana, it seems it needed to download something from
> the internet, and i didn't really configure a proper gateway on the
> external network.
> So I started by configuring a proper gateway and I tested it with the half
> deployed nodes, the I redid the deployment, and again I got this error :
>
> 2021-08-24 21:29:29.616805 | 525400e8-92c8-d397-6f7e-000000006133 |
>>  FATAL | Clean up legacy Cinder keystone catalog entries | undercloud |
>> error={"changed": false, "module_stderr": "Fa
>> iled to discover available identity versions when contacting
>> http://10.0.2.40:5000. Attempting to parse version from URL.\nTraceback
>> (most recent call last):\n  File \"/usr/lib/python3.6/si
>> te-packages/urllib3/connection.py\", line 162, in _new_conn\n
>>  (self._dns_host, self.port), self.timeout, **extra_kw)\n  File
>> \"/usr/lib/python3.6/site-packages/urllib3/util/connection.py
>> \", line 80, in create_connection\n    raise err\n  File
>> \"/usr/lib/python3.6/site-packages/urllib3/util/connection.py\", line 70,
>> in create_connection\n    sock.connect(sa)\nTimeoutError:
>> [Errno 110] Connection timed out\n\nDuring handling of the above
>> exception, another exception occurred:\n\nTraceback (most recent call
>> last):\n  File \"/usr/lib/python3.6/site-packages/urll
>> ib3/connectionpool.py\", line 600, in urlopen\n    chunked=chunked)\n
>>  File \"/usr/lib/python3.6/site-packages/urllib3/connectionpool.py\", line
>> 354, in _make_request\n    conn.request(meth
>> od, url, **httplib_request_kw)\n  File
>> \"/usr/lib64/python3.6/http/client.py\", line 1269, in request\n
>>  self._send_request(method, url, body, headers, encode_chunked)\n  File
>> \"/usr/lib6
>> 4/python3.6/http/client.py\", line 1315, in _send_request\n
>>  self.endheaders(body, encode_chunked=encode_chunked)\n  File
>> \"/usr/lib64/python3.6/http/client.py\", line 1264, in endheaders
>> \n    self._send_output(message_body, encode_chunked=encode_chunked)\n
>>  File \"/usr/lib64/python3.6/http/client.py\", line 1040, in _send_output\n
>>    self.send(msg)\n  File \"/usr/lib64/pyt
>> hon3.6/http/client.py\", line 978, in send\n    self.connect()\n  File
>> \"/usr/lib/python3.6/site-packages/urllib3/connection.py\", line 184, in
>> connect\n    conn = self._new_conn()\n  File
>> \"/usr/lib/python3.6/site-packages/urllib3/connection.py\", line 171, in
>> _new_conn\n    self, \"Failed to establish a new connection: %s\" %
>> e)\nurllib3.exceptions.NewConnectionError: <urll
>> ib3.connection.HTTPConnection object at 0x7f96f7b10cc0>: Failed to
>> establish a new connection: [Errno 110] Connection timed out\n\nDuring
>> handling of the above exception, another exception
>> occurred:\n\nTraceback (most recent call last):\n  File
>> \"/usr/lib/python3.6/site-packages/requests/adapters.py\", line 449, in
>> send\n    timeout=timeout\n  File \"/usr/lib/python3.6/site-p
>> ackages/urllib3/connectionpool.py\", line 638, in urlopen\n
>>  _stacktrace=sys.exc_info()[2])\n  File
>> \"/usr/lib/python3.6/site-packages/urllib3/util/retry.py\", line 399, in
>> increment\n
>>  raise MaxRetryError(_pool, url, error or
>> ResponseError(cause))\nurllib3.exceptions.MaxRetryError:
>> HTTPConnectionPool(host='10.0.2.40', port=5000): Max retries exceeded with
>> url: / (Caused
>> by NewConnectionError('<urllib3.connection.HTTPConnection object at
>> 0x7f96f7b10cc0>: Failed to establish a new connection: [Errno 110]
>> Connection timed out',))\n\nDuring handling of the ab$
>> ve exception, another exception occurred:\n\nTraceback (most recent call
>> last):\n  File
>> \"/usr/lib/python3.6/site-packages/keystoneauth1/session.py\", line 997, in
>> _send_request\n    resp $
>>  self.session.request(method, url, **kwargs)\n  File
>> \"/usr/lib/python3.6/site-packages/requests/sessions.py\", line 533, in
>> request\n    resp = self.send(prep, **send_kwargs)\n  File \"/u$
>> r/lib/python3.6/site-packages/requests/sessions.py\", line 646, in send\n
>>    r = adapter.send(request, **kwargs)\n  File
>> \"/usr/lib/python3.6/site-packages/requests/adapters.py\", line 516$
>>  in send\n    raise ConnectionError(e,
>> request=request)\nrequests.exceptions.ConnectionError:
>> HTTPConnectionPool(host='10.0.2.40', port=5000): Max retries exceeded with
>> url: / (Caused by N$wConnectionError('<urllib3.connection.HTTPConnection
>> object at 0x7f96f7b10cc0>: Failed to establish a new connection: [Errno
>> 110] Connection timed out',))\n\nDuring handling of the above e$
>> ception, another exception occurred:\n\nTraceback (most recent call
>> last):\n  File
>> \"/usr/lib/python3.6/site-packages/keystoneauth1/identity/generic/base.py\",
>> line 138, in _do_create_plug$
>> n\n    authenticated=False)\n  File
>> \"/usr/lib/python3.6/site-packages/keystoneauth1/identity/base.py\", line
>> 610, in get_discovery\n    authenticated=authenticated)\n  File
>> \"/usr/lib/pyt$
>> on3.6/site-packages/keystoneauth1/discover.py\", line 1442, in
>> get_discovery\n    disc = Discover(session, url,
>> authenticated=authenticated)\n  File
>> \"/usr/lib/python3.6/site-packages/keys$
>> oneauth1/discover.py\", line 526, in __init__\n
>>  authenticated=authenticated)\n  File
>> \"/usr/lib/python3.6/site-packages/keystoneauth1/discover.py\", line 101,
>> in get_version_data\n    r$
>> sp = session.get(url, headers=headers, authenticated=authenticated)\n
>>  File \"/usr/lib/python3.6/site-packages/keystoneauth1/session.py\", line
>> 1116, in get\n    return self.request(url, '$
>> ET', **kwargs)\n  File
>> \"/usr/lib/python3.6/site-packages/keystoneauth1/session.py\", line 906, in
>> request\n    resp = send(**kwargs)\n  File
>> \"/usr/lib/python3.6/site-packages/keystoneaut$
>> 1/session.py\", line 1013, in _send_request\n    raise
>> exceptions.ConnectFailure(msg)\nkeystoneauth1.exceptions.connection.ConnectFailure:
>> Unable to establish connection to http://10.0.2.4$
>> :5000: HTTPConnectionPool(host='10.0.2.40', port=5000): Max retries
>> exceeded with url: / (Caused by
>> NewConnectionError('<urllib3.connection.HTTPConnection object at
>> 0x7f96f7b10cc0>: Failed
>> to establish a new connection: [Errno 110] Connection timed
>> out',))\n\nDuring handling of the above exception, another exception
>> occurred:\n\nTraceback (most recent call last):\n  File \"<$
>> tdin>\", line 102, in <module>\n  File \"<stdin>\", line 94, in
>> _ansiballz_main\n  File \"<stdin>\", line 40, in invoke_module\n  File
>> \"/usr/lib64/python3.6/runpy.py\", line 205, in run_m$
>> dule\n    return _run_module_code(code, init_globals, run_name,
>> mod_spec)\n  File \"/usr/lib64/python3.6/runpy.py\", line 96, in
>> _run_module_code\n    mod_name, mod_spec, pkg_name, script_$
>> ame)\n  File \"/usr/lib64/python3.6/runpy.py\", line 85, in _run_code\n
>>  exec(code, run_globals)\n  File
>> \"/tmp/ansible_os_keystone_service_payload_wcyk6h37/ansible_os_keystone_service_p$
>> yload.zip/ansible/modules/cloud/openstack/os_keystone_service.py\", line
>> 194, in <module>\n  File
>> \"/tmp/ansible_os_keystone_service_payload_wcyk6h37/ansible_os_keystone_service_payload.zi$
>> /ansible/modules/cloud/openstack/os_keystone_service.py\", line 153, in
>> main\n  File
>> \"/usr/lib/python3.6/site-packages/openstack/cloud/_identity.py\", line
>> 510, in search_services\n    se$
>> vices = self.list_services()\n  File
>> \"/usr/lib/python3.6/site-packages/openstack/cloud/_identity.py\", line
>> 485, in list_services\n    if self._is_client_version('identity', 2):\n
>>  File \$
>> /usr/lib/python3.6/site-packages/openstack/cloud/openstackcloud.py\",
>> line 459, in _is_client_version\n    client = getattr(self, client_name)\n
>>  File \"/usr/lib/python3.6/site-packages/op$
>> nstack/cloud/_identity.py\", line 32, in _identity_client\n
>>  'identity', min_version=2, max_version='3.latest')\n  File
>> \"/usr/lib/python3.6/site-packages/openstack/cloud/openstackcloud.$
>> y\", line 406, in _get_versioned_client\n    if adapter.get_endpoint():\n
>>  File \"/usr/lib/python3.6/site-packages/keystoneauth1/adapter.py\", line
>> 282, in get_endpoint\n    return self.se$
>> sion.get_endpoint(auth or self.auth, **kwargs)\n  File
>> \"/usr/lib/python3.6/site-packages/keystoneauth1/session.py\", line 1218,
>> in get_endpoint\n    return auth.get_endpoint(self, **kwarg$
>> )\n  File
>> \"/usr/lib/python3.6/site-packages/keystoneauth1/identity/base.py\", line
>> 380, in get_endpoint\n    allow_version_hack=allow_version_hack,
>> **kwargs)\n  File \"/usr/lib/python3.6/$
>> ite-packages/keystoneauth1/identity/base.py\", line 271, in
>> get_endpoint_data\n    service_catalog =
>> self.get_access(session).service_catalog\n  File
>> \"/usr/lib/python3.6/site-packages/key$
>> toneauth1/identity/base.py\", line 134, in get_access\n    self.auth_ref
>> = self.get_auth_ref(session)\n  File
>> \"/usr/lib/python3.6/site-packages/keystoneauth1/identity/generic/base.py\",
>> l$
>> ne 206, in get_auth_ref\n    self._plugin =
>> self._do_create_plugin(session)\n  File
>> \"/usr/lib/python3.6/site-packages/keystoneauth1/identity/generic/base.py\",
>> line 161, in _do_create_plu$
>> in\n    'auth_url is correct. %s' %
>> e)\nkeystoneauth1.exceptions.discovery.DiscoveryFailure: Could not find
>> versioned identity endpoints when attempting to authenticate. Please check
>> that $our auth_url is correct.
>>
>> *Unable to establish connection to http://10.0.2.40:5000
>> <http://10.0.2.40:5000>: HTTPConnectionPool(host='10.0.2.40', port=5000):
>> Max retries exceeded with url: / (Caused by
>> NewConnectionError('<urllib3.connection.HTTPConnection object at
>> 0x7f96f7b10cc0>: Failed to establish a new connection: [Errno 110]
>> Connection timed out',))\n", "module_stdout": "", "msg": "MODULE
>> FAILURE\nSee stdout/stderr for the exact error", "rc": 1} *
>>
>>
>> 2021-08-24 21:29:29.617697 | 525400e8-92c8-d397-6f7e-000000006133 |
>> TIMING | Clean up legacy Cinder keystone catalog entries | undercloud |
>> 1:07:40.666419 | 130.85s
>>
>>
>>
>> PLAY RECAP
>> *********************************************************************
>>
>>
>> overcloud-computehci-0     : ok=260  changed=145  unreachable=0
>>  failed=0    skipped=140  rescued=0    ignored=0
>>
>> overcloud-computehci-1     : ok=258  changed=145  unreachable=0
>>  failed=0    skipped=140  rescued=0    ignored=0
>>
>> overcloud-computehci-2     : ok=255  changed=145  unreachable=0
>>  failed=0    skipped=140  rescued=0    ignored=0
>>
>> overcloud-controller-0     : ok=295  changed=181  unreachable=0
>>  failed=0    skipped=151  rescued=0    ignored=0
>>
>> overcloud-controller-1     : ok=289  changed=177  unreachable=0
>>  failed=0    skipped=152  rescued=0    ignored=0
>>
>> overcloud-controller-2     : ok=288  changed=177  unreachable=0
>>  failed=0    skipped=152  rescued=0    ignored=0
>>
>> undercloud                 : ok=105  changed=21   unreachable=0
>>  failed=1    skipped=45   rescued=0    ignored=0
>>
>>
>>
>>
>> 2021-08-24 21:29:29.730778 | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ Summary
>> Information ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
>>
>> 2021-08-24 21:29:29.731007 | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ Total
>> Tasks: 1723       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
>>
>> 2021-08-24 21:29:29.731098 | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ Elapsed Time:
>> 1:07:40.779840 ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
>>
>> 2021-08-24 21:29:29.731172 |                                 UUID |
>> Info |       Host |   Task Name |   Run Time
>>
>> 2021-08-24 21:29:29.731251 | 525400e8-92c8-d397-6f7e-000000003b9a |
>>  SUMMARY | undercloud | Run tripleo-container-image-prepare logged to:
>> /var/log/tripleo-container-image-prepare.log | 1762.93s
>>
>>
>>
>> 2021-08-24 21:29:29.731349 | 525400e8-92c8-d397-6f7e-0000000057aa |
>>  SUMMARY | undercloud | tripleo-ceph-run-ansible : run ceph-ansible |
>> 990.24s
>> 2021-08-24 21:29:29.731433 | 525400e8-92c8-d397-6f7e-000000005951 |
>>  SUMMARY | overcloud-controller-0 | tripleo_ha_wrapper : Run init bundle
>> puppet on the host for haproxy | 133.22s
>> 2021-08-24 21:29:29.731503 | 525400e8-92c8-d397-6f7e-000000006133 |
>>  SUMMARY | undercloud | Clean up legacy Cinder keystone catalog entries |
>> 130.85s
>> 2021-08-24 21:29:29.731569 | 525400e8-92c8-d397-6f7e-000000006012 |
>>  SUMMARY | overcloud-controller-0 | Wait for containers to start for step 3
>> using paunch | 103.45s
>> 2021-08-24 21:29:29.731643 | 525400e8-92c8-d397-6f7e-000000004337 |
>>  SUMMARY | overcloud-computehci-0 | Pre-fetch all the containers | 94.00s
>>
>> 2021-08-24 21:29:29.731729 | 525400e8-92c8-d397-6f7e-000000004378 |
>>  SUMMARY | overcloud-computehci-2 | Pre-fetch all the containers | 92.64s
>>
>> 2021-08-24 21:29:29.731795 | 525400e8-92c8-d397-6f7e-000000004337 |
>>  SUMMARY | overcloud-computehci-1 | Pre-fetch all the containers | 86.38s
>>
>> 2021-08-24 21:29:29.731867 | 525400e8-92c8-d397-6f7e-000000004d68 |
>>  SUMMARY | overcloud-controller-0 | Wait for container-puppet tasks
>> (generate config) to finish | 84.13s
>> 2021-08-24 21:29:29.731946 | 525400e8-92c8-d397-6f7e-000000004d99 |
>>  SUMMARY | overcloud-controller-2 | Wait for container-puppet tasks
>> (generate config) to finish | 80.76s
>> 2021-08-24 21:29:29.732012 | 525400e8-92c8-d397-6f7e-00000000427c |
>>  SUMMARY | overcloud-controller-1 | Pre-fetch all the containers | 80.21s
>>
>> 2021-08-24 21:29:29.732073 | 525400e8-92c8-d397-6f7e-00000000427c |
>>  SUMMARY | overcloud-controller-0 | Pre-fetch all the containers | 77.03s
>>
>> 2021-08-24 21:29:29.732138 | 525400e8-92c8-d397-6f7e-0000000042f5 |
>>  SUMMARY | overcloud-controller-2 | Pre-fetch all the containers | 76.32s
>>
>> 2021-08-24 21:29:29.732202 | 525400e8-92c8-d397-6f7e-000000004dd3 |
>>  SUMMARY | overcloud-controller-1 | Wait for container-puppet tasks
>> (generate config) to finish | 74.36s
>> 2021-08-24 21:29:29.732266 | 525400e8-92c8-d397-6f7e-000000005da7 |
>>  SUMMARY | overcloud-controller-0 | tripleo_ha_wrapper : Run init bundle
>> puppet on the host for ovn_dbs | 68.39s
>> 2021-08-24 21:29:29.732329 | 525400e8-92c8-d397-6f7e-000000005ce2 |
>>  SUMMARY | overcloud-controller-0 | Wait for containers to start for step 2
>> using paunch | 64.55s
>> 2021-08-24 21:29:29.732398 | 525400e8-92c8-d397-6f7e-000000004b97 |
>>  SUMMARY | overcloud-controller-2 | Wait for puppet host configuration to
>> finish | 58.13s
>> 2021-08-24 21:29:29.732463 | 525400e8-92c8-d397-6f7e-000000004c1a |
>>  SUMMARY | overcloud-controller-1 | Wait for puppet host configuration to
>> finish | 58.11s
>> 2021-08-24 21:29:29.732526 | 525400e8-92c8-d397-6f7e-000000005bd3 |
>>  SUMMARY | overcloud-controller-1 | Wait for containers to start for step 2
>> using paunch | 58.09s
>> 2021-08-24 21:29:29.732589 | 525400e8-92c8-d397-6f7e-000000005b9b |
>>  SUMMARY | overcloud-controller-2 | Wait for containers to start for step 2
>> using paunch | 58.09s
>>
>
>
> Thank you again for your assistance.
>
> Regards.
>
> Le mar. 24 août 2021 à 08:59, wodel youchi <wodel.youchi at gmail.com> a
> écrit :
>
>> Hi, and thanks for your help
>>
>> As for Ceph, here is  container prepare
>> parameter_defaults:
>>  ContainerImagePrepare:
>>  - push_destination: true
>>    set:
>>      ceph_alertmanager_image: alertmanager
>>      ceph_alertmanager_namespace: quay.ceph.io/prometheus
>>      ceph_alertmanager_tag: v0.16.2
>>      ceph_grafana_image: grafana
>>      ceph_grafana_namespace: quay.ceph.io/app-sre
>>      *ceph_grafana_tag: 5.4.3*
>>      ceph_image: daemon
>>      ceph_namespace: quay.ceph.io/ceph-ci
>>      ceph_node_exporter_image: node-exporter
>>      ceph_node_exporter_namespace: quay.ceph.io/prometheus
>>      ceph_node_exporter_tag: v0.17.0
>>      ceph_prometheus_image: prometheus
>>      ceph_prometheus_namespace: quay.ceph.io/prometheus
>>      ceph_prometheus_tag: v2.7.2
>>      *ceph_tag: v4.0.19-stable-4.0-nautilus-centos-7-x86_64*
>>      name_prefix: centos-binary-
>>      name_suffix: ''
>>      namespace: quay.io/tripleotraincentos8
>>      neutron_driver: ovn
>>      rhel_containers: false
>>      tag: current-tripleo
>>    tag_from_label: rdo_version
>>
>> And yes, the 10.200.7.0/24 network is my storage network
>> Here is a snippet from my network_data.yaml
>>
>> - name: Storage
>>  vip: true
>>  vlan: 1107
>>  name_lower: storage
>>  ip_subnet: '10.200.7.0/24'
>>  allocation_pools: [{'start': '10.200.7.150', 'end': '10.200.7.169'}]
>>
>> I will look into the grafana service to see why it's not booting and get
>> back to you.
>>
>> Regards.
>>
>> Le lun. 23 août 2021 à 17:28, Francesco Pantano <fpantano at redhat.com> a
>> écrit :
>>
>>> Hello,
>>> thanks John for your reply here.
>>> A few more comments inline:
>>>
>>> On Mon, Aug 23, 2021 at 6:16 PM John Fulton <johfulto at redhat.com> wrote:
>>>
>>>> On Mon, Aug 23, 2021 at 10:52 AM wodel youchi <wodel.youchi at gmail.com>
>>>> wrote:
>>>> >
>>>> > Hi,
>>>> >
>>>> > I redid the undercloud deployment for the Train version for now. And
>>>> I verified the download URL for the images.
>>>> > My overcloud deployment has moved forward but I still get errors.
>>>> >
>>>> > This is what I got this time :
>>>> >>
>>>> >>        "TASK [ceph-grafana : wait for grafana to start]
>>>> ********************************",
>>>> >>        "Monday 23 August 2021  14:55:21 +0100 (0:00:00.961)
>>>>  0:12:59.319 ********* ",
>>>> >>        "fatal: [overcloud-controller-0]: FAILED! => {\"changed\":
>>>> false, \"elapsed\": 300, \"msg\": \"Timeout when waiting for 10.20
>>>> >> 0.7.151:3100\"}",
>>>> >>        "fatal: [overcloud-controller-1]: FAILED! => {\"changed\":
>>>> false, \"elapsed\": 300, \"msg\": \"Timeout when waiting for 10.20
>>>> >> 0.7.155:3100\"}",
>>>> >>        "fatal: [overcloud-controller-2]: FAILED! => {\"changed\":
>>>> false, \"elapsed\": 300, \"msg\": \"Timeout when waiting for 10.20
>>>> >> 0.7.165:3100\"}",
>>>>
>>>> I'm not certain of the ceph-ansible version you're using but it should
>>>> be a version 4 with train. ceph-ansible should already be installed on
>>>> your undercloud judging by this error and in the latest version 4 this
>>>> task is where it failed:
>>>>
>>>>
>>>> https://github.com/ceph/ceph-ansible/blob/v4.0.64/roles/ceph-grafana/tasks/configure_grafana.yml#L112-L115
>>>>
>>>> You can check the status of this service on your three controllers and
>>>> then debug it directly.
>>>
>>> As John pointed out, ceph-ansible is able to configure, render and start
>>> the associated
>>> systemd unit for all the ceph monitoring stack components
>>> (node-exported, prometheus, alertmanager and
>>> grafana).
>>> You can ssh to your controllers, and check the systemd unit associated,
>>> checking the journal to see why
>>> they failed to start (I saw there's a timeout waiting for the container
>>> to start).
>>> A potential plan, in this case, could be:
>>>
>>> 1. check the systemd unit (I guess you can start with grafana which is
>>> the failed service)
>>> 2. look at the journal logs (feel free to attach here the relevant part
>>> of the output)
>>> 3. double check the network where the service is bound (can you attach
>>> the /var/lib/mistral/<stack>/ceph-ansible/group_vars/all.yaml)
>>>     * The grafana process should be run on the storage network, but I
>>> see a "Timeout when waiting for 10.200.7.165:3100": is that network the
>>> right one?
>>>
>>>>
>>>
>>>
>>>>   John
>>>>
>>>> >>        "RUNNING HANDLER [ceph-prometheus : service handler]
>>>> ****************************",
>>>> >>        "Monday 23 August 2021  15:00:22 +0100 (0:05:00.767)
>>>>  0:18:00.087 ********* ",
>>>> >>        "PLAY RECAP
>>>> *********************************************************************",
>>>> >>        "overcloud-computehci-0     : ok=224  changed=23
>>>>  unreachable=0    failed=0    skipped=415  rescued=0    ignored=0   ",
>>>> >>        "overcloud-computehci-1     : ok=199  changed=18
>>>>  unreachable=0    failed=0    skipped=392  rescued=0    ignored=0   ",
>>>> >>        "overcloud-computehci-2     : ok=212  changed=23
>>>>  unreachable=0    failed=0    skipped=390  rescued=0    ignored=0   ",
>>>> >>        "overcloud-controller-0     : ok=370  changed=52
>>>>  unreachable=0    failed=1    skipped=539  rescued=0    ignored=0   ",
>>>> >>        "overcloud-controller-1     : ok=308  changed=43
>>>>  unreachable=0    failed=1    skipped=495  rescued=0    ignored=0   ",
>>>> >>        "overcloud-controller-2     : ok=317  changed=45
>>>>  unreachable=0    failed=1    skipped=493  rescued=0    ignored=0   ",
>>>> >>
>>>> >>        "INSTALLER STATUS
>>>> ***************************************************************",
>>>> >>        "Install Ceph Monitor           : Complete (0:00:52)",
>>>> >>        "Install Ceph Manager           : Complete (0:05:49)",
>>>> >>        "Install Ceph OSD               : Complete (0:02:28)",
>>>> >>        "Install Ceph RGW               : Complete (0:00:27)",
>>>> >>        "Install Ceph Client            : Complete (0:00:33)",
>>>> >>        "Install Ceph Grafana           : In Progress (0:05:54)",
>>>> >>        "\tThis phase can be restarted by running:
>>>> roles/ceph-grafana/tasks/main.yml",
>>>> >>        "Install Ceph Node Exporter     : Complete (0:00:28)",
>>>> >>        "Monday 23 August 2021  15:00:22 +0100 (0:00:00.006)
>>>>  0:18:00.094 ********* ",
>>>> >>
>>>> "===============================================================================
>>>> ",
>>>> >>        "ceph-grafana : wait for grafana to start
>>>> ------------------------------ 300.77s",
>>>> >>        "ceph-facts : get ceph current status
>>>> ---------------------------------- 300.27s",
>>>> >>        "ceph-container-common : pulling
>>>> udtrain.ctlplane.umaitek.dz:8787/ceph-ci/daemon:v4.0.19-stable-4.0-nautilus-centos-7-x86_64
>>>> >> image -- 19.04s",
>>>> >>        "ceph-mon : waiting for the monitor(s) to form the quorum...
>>>> ------------ 12.83s",
>>>> >>        "ceph-osd : use ceph-volume lvm batch to create bluestore
>>>> osds ---------- 12.13s",
>>>> >>        "ceph-osd : wait for all osd to be up
>>>> ----------------------------------- 11.88s",
>>>> >>        "ceph-osd : set pg_autoscale_mode value on pool(s)
>>>> ---------------------- 11.00s",
>>>> >>        "ceph-osd : create openstack pool(s)
>>>> ------------------------------------ 10.80s",
>>>> >>        "ceph-grafana : make sure grafana is down
>>>> ------------------------------- 10.66s",
>>>> >>        "ceph-osd : customize pool crush_rule
>>>> ----------------------------------- 10.15s",
>>>> >>        "ceph-osd : customize pool size
>>>> ----------------------------------------- 10.15s",
>>>> >>        "ceph-osd : customize pool min_size
>>>> ------------------------------------- 10.14s",
>>>> >>        "ceph-osd : assign application to pool(s)
>>>> ------------------------------- 10.13s",
>>>> >>        "ceph-osd : list existing pool(s)
>>>> ---------------------------------------- 8.59s",
>>>> >>
>>>> >>        "ceph-mon : fetch ceph initial keys
>>>> -------------------------------------- 7.01s",
>>>> >>        "ceph-container-common : get ceph version
>>>> -------------------------------- 6.75s",
>>>> >>        "ceph-prometheus : start prometheus services
>>>> ----------------------------- 6.67s",
>>>> >>        "ceph-mgr : wait for all mgr to be up
>>>> ------------------------------------ 6.66s",
>>>> >>        "ceph-grafana : start the grafana-server service
>>>> ------------------------- 6.33s",
>>>> >>        "ceph-mgr : create ceph mgr keyring(s) on a mon node
>>>> --------------------- 6.26s"
>>>> >>    ],
>>>> >>    "failed_when_result": true
>>>> >> }
>>>> >> 2021-08-23 15:00:24.427687 | 525400e8-92c8-47b1-e162-00000000597d |
>>>>    TIMING | tripleo-ceph-run-ansible : print ceph-ansible outpu$
>>>> >> in case of failure | undercloud | 0:37:30.226345 | 0.25s
>>>> >>
>>>> >> PLAY RECAP
>>>> *********************************************************************
>>>> >> overcloud-computehci-0     : ok=213  changed=117  unreachable=0
>>>> failed=0    skipped=120  rescued=0    ignored=0
>>>> >> overcloud-computehci-1     : ok=207  changed=117  unreachable=0
>>>> failed=0    skipped=120  rescued=0    ignored=0
>>>> >> overcloud-computehci-2     : ok=207  changed=117  unreachable=0
>>>> failed=0    skipped=120  rescued=0    ignored=0
>>>> >> overcloud-controller-0     : ok=237  changed=145  unreachable=0
>>>> failed=0    skipped=128  rescued=0    ignored=0
>>>> >> overcloud-controller-1     : ok=232  changed=145  unreachable=0
>>>> failed=0    skipped=128  rescued=0    ignored=0
>>>> >> overcloud-controller-2     : ok=232  changed=145  unreachable=0
>>>> failed=0    skipped=128  rescued=0    ignored=0
>>>> >> undercloud                 : ok=100  changed=18   unreachable=0
>>>> failed=1    skipped=37   rescued=0    ignored=0
>>>> >>
>>>> >> 2021-08-23 15:00:24.559997 | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
>>>> Summary Information ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
>>>> >> 2021-08-23 15:00:24.560328 | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ Total
>>>> Tasks: 1366       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
>>>> >> 2021-08-23 15:00:24.560419 | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ Elapsed
>>>> Time: 0:37:30.359090 ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
>>>> >> 2021-08-23 15:00:24.560490 |                                 UUID |
>>>>      Info |       Host |   Task Name |   Run Time
>>>> >> 2021-08-23 15:00:24.560589 | 525400e8-92c8-47b1-e162-00000000597b |
>>>>   SUMMARY | undercloud | tripleo-ceph-run-ansible : run ceph-ans
>>>> >> ible | 1082.71s
>>>> >> 2021-08-23 15:00:24.560675 | 525400e8-92c8-47b1-e162-000000004d9a |
>>>>   SUMMARY | overcloud-controller-1 | Wait for container-puppet t
>>>> >> asks (generate config) to finish | 356.02s
>>>> >> 2021-08-23 15:00:24.560763 | 525400e8-92c8-47b1-e162-000000004d6a |
>>>>   SUMMARY | overcloud-controller-0 | Wait for container-puppet t
>>>> >> asks (generate config) to finish | 355.74s
>>>> >> 2021-08-23 15:00:24.560839 | 525400e8-92c8-47b1-e162-000000004dd0 |
>>>>   SUMMARY | overcloud-controller-2 | Wait for container-puppet t
>>>> >> asks (generate config) to finish | 355.68s
>>>> >> 2021-08-23 15:00:24.560912 | 525400e8-92c8-47b1-e162-000000003bb1 |
>>>>   SUMMARY | undercloud | Run tripleo-container-image-prepare log
>>>> >> ged to: /var/log/tripleo-container-image-prepare.log | 143.03s
>>>> >> 2021-08-23 15:00:24.560986 | 525400e8-92c8-47b1-e162-000000004b13 |
>>>>   SUMMARY | overcloud-controller-0 | Wait for puppet host config
>>>> >> uration to finish | 125.36s
>>>> >> 2021-08-23 15:00:24.561057 | 525400e8-92c8-47b1-e162-000000004b88 |
>>>>   SUMMARY | overcloud-controller-2 | Wait for puppet host config
>>>> >> uration to finish | 125.33s
>>>> >> 2021-08-23 15:00:24.561128 | 525400e8-92c8-47b1-e162-000000004b4b |
>>>>   SUMMARY | overcloud-controller-1 | Wait for puppet host config
>>>> >> uration to finish | 125.25s
>>>> >> 2021-08-23 15:00:24.561300 | 525400e8-92c8-47b1-e162-000000001dc4 |
>>>>   SUMMARY | overcloud-controller-2 | Run puppet on the host to a
>>>> >> pply IPtables rules | 108.08s
>>>> >> 2021-08-23 15:00:24.561374 | 525400e8-92c8-47b1-e162-000000001e4f |
>>>>   SUMMARY | overcloud-controller-0 | Run puppet on the host to a
>>>> >> pply IPtables rules | 107.34s
>>>> >> 2021-08-23 15:00:24.561444 | 525400e8-92c8-47b1-e162-000000004c8d |
>>>>   SUMMARY | overcloud-computehci-2 | Wait for container-puppet t
>>>> >> asks (generate config) to finish | 96.56s
>>>> >> 2021-08-23 15:00:24.561514 | 525400e8-92c8-47b1-e162-000000004c33 |
>>>>   SUMMARY | overcloud-computehci-0 | Wait for container-puppet t
>>>> >> asks (generate config) to finish | 96.38s
>>>> >> 2021-08-23 15:00:24.561580 | 525400e8-92c8-47b1-e162-000000004c60 |
>>>>   SUMMARY | overcloud-computehci-1 | Wait for container-puppet t
>>>> >> asks (generate config) to finish | 93.41s
>>>> >> 2021-08-23 15:00:24.561645 | 525400e8-92c8-47b1-e162-00000000434d |
>>>>   SUMMARY | overcloud-computehci-0 | Pre-fetch all the container
>>>> >> s | 92.70s
>>>> >> 2021-08-23 15:00:24.561712 | 525400e8-92c8-47b1-e162-0000000043ed |
>>>>   SUMMARY | overcloud-computehci-2 | Pre-fetch all the container
>>>> >> s | 91.90s
>>>> >> 2021-08-23 15:00:24.561782 | 525400e8-92c8-47b1-e162-000000004385 |
>>>>   SUMMARY | overcloud-computehci-1 | Pre-fetch all the container
>>>> >> s | 91.88s
>>>> >> 2021-08-23 15:00:24.561876 | 525400e8-92c8-47b1-e162-00000000491c |
>>>>   SUMMARY | overcloud-computehci-1 | Wait for puppet host config
>>>> >> uration to finish | 90.37s
>>>> >> 2021-08-23 15:00:24.561947 | 525400e8-92c8-47b1-e162-000000004951 |
>>>>   SUMMARY | overcloud-computehci-2 | Wait for puppet host config
>>>> >> uration to finish | 90.37s
>>>> >> 2021-08-23 15:00:24.562016 | 525400e8-92c8-47b1-e162-0000000048e6 |
>>>>   SUMMARY | overcloud-computehci-0 | Wait for puppet host config
>>>> >> uration to finish | 90.35s
>>>> >> 2021-08-23 15:00:24.562080 | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ End
>>>> Summary Information ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
>>>> >> 2021-08-23 15:00:24.562196 | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
>>>> State Information ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
>>>> >> 2021-08-23 15:00:24.562311 | ~~~~~~~~~~~~~~~~~~ Number of nodes
>>>> which did not deploy successfully: 1 ~~~~~~~~~~~~~~~~~
>>>> >> 2021-08-23 15:00:24.562379 |  The following node(s) had failures:
>>>> undercloud
>>>> >> 2021-08-23 15:00:24.562456 |
>>>> ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
>>>> >> Host 10.0.2.40 not found in /home/stack/.ssh/known_hosts
>>>> >> Ansible failed, check log at
>>>> /var/lib/mistral/overcloud/ansible.log.Overcloud Endpoint:
>>>> http://10.0.2.40:5000
>>>> >> Overcloud Horizon Dashboard URL: http://10.0.2.40:80/dashboard
>>>> >> Overcloud rc file: /home/stack/overcloudrc
>>>> >> Overcloud Deployed with error
>>>> >> Overcloud configuration failed.
>>>> >>
>>>> >
>>>> >
>>>> > Could someone help debug this, the ansible.log is huge, I can't see
>>>> what's the origin of the problem, if someone can point me to the right
>>>> direction it will aprecciated.
>>>> > Thanks in advance.
>>>> >
>>>> > Regards.
>>>> >
>>>> > Le mer. 18 août 2021 à 18:02, Wesley Hayutin <whayutin at redhat.com> a
>>>> écrit :
>>>> >>
>>>> >>
>>>> >>
>>>> >> On Wed, Aug 18, 2021 at 10:10 AM Dmitry Tantsur <dtantsur at redhat.com>
>>>> wrote:
>>>> >>>
>>>> >>> Hi,
>>>> >>>
>>>> >>> On Wed, Aug 18, 2021 at 4:39 PM wodel youchi <
>>>> wodel.youchi at gmail.com> wrote:
>>>> >>>>
>>>> >>>> Hi,
>>>> >>>> I am trying to deploy openstack with tripleO using VMs and
>>>> nested-KVM for the compute node. This is for test and learning purposes.
>>>> >>>>
>>>> >>>> I am using the Train version and following some tutorials.
>>>> >>>> I prepared my different template files and started the deployment,
>>>> but I got these errors :
>>>> >>>>
>>>> >>>> Failed to provision instance fc40457e-4b3c-4402-ae9d-c528f2c2ad30:
>>>> Asynchronous exception: Node failed to deploy. Exception: Agent API for
>>>> node 6d3724fc-6f13-4588-bbe5-56bc4f9a4f87 returned HTTP status code 404
>>>> with error: Not found: Extension with id iscsi not found. for node
>>>> >>>>
>>>> >>>
>>>> >>> You somehow ended up using master (Xena release) deploy ramdisk
>>>> with Train TripleO. You need to make sure to download Train images. I hope
>>>> TripleO people can point you at the right place.
>>>> >>>
>>>> >>> Dmitry
>>>> >>
>>>> >>
>>>> >> http://images.rdoproject.org/centos8/
>>>> >>
>>>> http://images.rdoproject.org/centos8/train/rdo_trunk/current-tripleo/
>>>> >>
>>>> >>>
>>>> >>>
>>>> >>>>
>>>> >>>> and
>>>> >>>>
>>>> >>>> Got HTTP 409: {"errors": [{"status": 409, "title": "Conflict",
>>>> "detail": "There was a conflict when trying to complete your request.\n\n
>>>> Unable to allocate inventory: Unable to create allocation for
>>>> 'CUSTOM_BAREMETAL' on resource provider
>>>> '6d3724fc-6f13-4588-bbe5-56bc4f9a4f87'. The requested amount would exceed
>>>> the capacity. ",
>>>> >>>>
>>>> >>>> Could you help understand what those errors mean? I couldn't find
>>>> anything similar on the net.
>>>> >>>>
>>>> >>>> Thanks in advance.
>>>> >>>>
>>>> >>>> Regards.
>>>> >>>
>>>> >>>
>>>> >>>
>>>> >>> --
>>>> >>> Red Hat GmbH, https://de.redhat.com/ , Registered seat: Grasbrunn,
>>>> >>> Commercial register: Amtsgericht Muenchen, HRB 153243,
>>>> >>> Managing Directors: Charles Cachera, Brian Klemm, Laurie Krebs,
>>>> Michael O'Neill
>>>>
>>>>
>>>>
>>>
>>> --
>>> Francesco Pantano
>>> GPG KEY: F41BD75C
>>>
>>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.openstack.org/pipermail/openstack-discuss/attachments/20210825/0b2da70c/attachment-0001.html>


More information about the openstack-discuss mailing list