DCN compute service goes down when a instance is scheduled to launch | wallaby | tripleo

Alan Bishop abishop at redhat.com
Thu Mar 23 17:05:30 UTC 2023


On Thu, Mar 23, 2023 at 9:01 AM Swogat Pradhan <swogatpradhan22 at gmail.com>
wrote:

> Hi,
> Can someone please help me identify the issue here?
> Latest cinder-volume logs from dcn02:
> (ATTACHED)
>

It's really not possible to analyze what's happening with just one or two
log entries. Do you have
debug logs enabled? One thing I noticed is the glance image's disk_format
is qcow2. You should
use "raw" images with ceph RBD.

Alan


>
> The volume is stuck in creating state.
>
> With regards,
> Swogat Pradhan
>
> On Thu, Mar 23, 2023 at 6:12 PM Swogat Pradhan <swogatpradhan22 at gmail.com>
> wrote:
>
>> Hi Jhon,
>> Thank you for clarifying that.
>> Right now the cinder volume is stuck in *creating *state when adding
>> image as volume source.
>> But when creating an empty volume the volumes are getting created
>> successfully without any errors.
>>
>> We are getting volume creation request in cinder-volume.log as such:
>> 2023-03-23 12:34:40.152 108 INFO
>> cinder.volume.flows.manager.create_volume
>> [req-18556796-a61c-4097-8fa8-b136ce9814f7 b240e3e89d99489284cd731e75f2a5db
>> 4160ce999a31485fa643aed0936dfef0 - - -] Volume
>> 872a2ae6-c75b-4fc0-8172-17a29d07a66c: being created as image with
>> specification: {'status': 'creating', 'volume_name':
>> 'volume-872a2ae6-c75b-4fc0-8172-17a29d07a66c', 'volume_size': 1,
>> 'image_id': '131ed4e0-0474-45be-b74a-43b599a7d6c5', 'image_location':
>> ('rbd://a5ae877c-bcba-53fe-8336-450e63014757/images/131ed4e0-0474-45be-b74a-43b599a7d6c5/snap',
>> [{'url':
>> 'rbd://a5ae877c-bcba-53fe-8336-450e63014757/images/131ed4e0-0474-45be-b74a-43b599a7d6c5/snap',
>> 'metadata': {'store': 'ceph'}}, {'url':
>> 'rbd://a8d5f1f5-48e7-5ede-89ab-8aca59b6397b/images/131ed4e0-0474-45be-b74a-43b599a7d6c5/snap',
>> 'metadata': {'store': 'dcn02'}}]), 'image_meta': {'name': 'cirros',
>> 'disk_format': 'qcow2', 'container_format': 'bare', 'visibility': 'public',
>> 'size': 16338944, 'virtual_size': 117440512, 'status': 'active',
>> 'checksum': '1d3062cd89af34e419f7100277f38b2b', 'protected': False,
>> 'min_ram': 0, 'min_disk': 0, 'owner': '4160ce999a31485fa643aed0936dfef0',
>> 'os_hidden': False, 'os_hash_algo': 'sha512', 'os_hash_value':
>> '553d220ed58cfee7dafe003c446a9f197ab5edf8ffc09396c74187cf83873c877e7ae041cb80f3b91489acf687183adcd689b53b38e3ddd22e627e7f98a09c46',
>> 'id': '131ed4e0-0474-45be-b74a-43b599a7d6c5', 'created_at':
>> datetime.datetime(2023, 3, 23, 11, 41, 51, tzinfo=datetime.timezone.utc),
>> 'updated_at': datetime.datetime(2023, 3, 23, 11, 46, 37,
>> tzinfo=datetime.timezone.utc), 'locations': [{'url':
>> 'rbd://a5ae877c-bcba-53fe-8336-450e63014757/images/131ed4e0-0474-45be-b74a-43b599a7d6c5/snap',
>> 'metadata': {'store': 'ceph'}}, {'url':
>> 'rbd://a8d5f1f5-48e7-5ede-89ab-8aca59b6397b/images/131ed4e0-0474-45be-b74a-43b599a7d6c5/snap',
>> 'metadata': {'store': 'dcn02'}}], 'direct_url':
>> 'rbd://a5ae877c-bcba-53fe-8336-450e63014757/images/131ed4e0-0474-45be-b74a-43b599a7d6c5/snap',
>> 'tags': [], 'file': '/v2/images/131ed4e0-0474-45be-b74a-43b599a7d6c5/file',
>> 'stores': 'ceph,dcn02', 'properties': {'os_glance_failed_import': '',
>> 'os_glance_importing_to_stores': '', 'owner_specified.openstack.md5': '',
>> 'owner_specified.openstack.object': 'images/cirros',
>> 'owner_specified.openstack.sha256': ''}}, 'image_service':
>> <cinder.image.glance.GlanceImageService object at 0x7f98d869ed68>}
>>
>> But there is nothing else after that and the volume doesn't even timeout,
>> it just gets stuck in creating state.
>> Can you advise what might be the issue here?
>> All the containers are in a healthy state now.
>>
>> With regards,
>> Swogat Pradhan
>>
>>
>> On Thu, Mar 23, 2023 at 6:06 PM Alan Bishop <abishop at redhat.com> wrote:
>>
>>>
>>>
>>> On Thu, Mar 23, 2023 at 5:20 AM Swogat Pradhan <
>>> swogatpradhan22 at gmail.com> wrote:
>>>
>>>> Hi,
>>>> Is this bind not required for cinder_scheduler container?
>>>>
>>>> "/var/lib/tripleo-config/ceph:/var/lib/kolla/config_files/src-ceph:ro,rprivate,rbind",
>>>> I do not see this particular bind on the cinder scheduler containers on
>>>> my controller nodes.
>>>>
>>>
>>> That is correct, because the scheduler does not access the ceph cluster.
>>>
>>> Alan
>>>
>>>
>>>> With regards,
>>>> Swogat Pradhan
>>>>
>>>> On Thu, Mar 23, 2023 at 2:46 AM Swogat Pradhan <
>>>> swogatpradhan22 at gmail.com> wrote:
>>>>
>>>>> Cinder volume config:
>>>>>
>>>>> [tripleo_ceph]
>>>>> volume_backend_name=tripleo_ceph
>>>>> volume_driver=cinder.volume.drivers.rbd.RBDDriver
>>>>> rbd_user=openstack
>>>>> rbd_pool=volumes
>>>>> rbd_flatten_volume_from_snapshot=False
>>>>> rbd_secret_uuid=a8d5f1f5-48e7-5ede-89ab-8aca59b6397b
>>>>> report_discard_supported=True
>>>>> rbd_ceph_conf=/etc/ceph/dcn02.conf
>>>>> rbd_cluster_name=dcn02
>>>>>
>>>>> Glance api config:
>>>>>
>>>>> [dcn02]
>>>>> rbd_store_ceph_conf=/etc/ceph/dcn02.conf
>>>>> rbd_store_user=openstack
>>>>> rbd_store_pool=images
>>>>> rbd_thin_provisioning=False
>>>>> store_description=dcn02 rbd glance store
>>>>> [ceph]
>>>>> rbd_store_ceph_conf=/etc/ceph/ceph.conf
>>>>> rbd_store_user=openstack
>>>>> rbd_store_pool=images
>>>>> rbd_thin_provisioning=False
>>>>> store_description=Default glance store backend.
>>>>>
>>>>> On Thu, Mar 23, 2023 at 2:29 AM Swogat Pradhan <
>>>>> swogatpradhan22 at gmail.com> wrote:
>>>>>
>>>>>> I still have the same issue, I'm not sure what's left to try.
>>>>>> All the pods are now in a healthy state, I am getting log entries 3
>>>>>> mins after I hit the create volume button in cinder-volume when I try to
>>>>>> create a volume with an image.
>>>>>> And the volumes are just stuck in creating state for more than 20
>>>>>> mins now.
>>>>>>
>>>>>> Cinder logs:
>>>>>> 2023-03-22 20:32:44.010 108 INFO cinder.rpc
>>>>>> [req-0d2093a0-efbd-45a5-bd7d-cce25ddc200e b240e3e89d99489284cd731e75f2a5db
>>>>>> 4160ce999a31485fa643aed0936dfef0 - - -] Automatically selected
>>>>>> cinder-volume RPC version 3.17 as minimum service version.
>>>>>> 2023-03-22 20:34:59.166 108 INFO
>>>>>> cinder.volume.flows.manager.create_volume
>>>>>> [req-0d2093a0-efbd-45a5-bd7d-cce25ddc200e b240e3e89d99489284cd731e75f2a5db
>>>>>> 4160ce999a31485fa643aed0936dfef0 - - -] Volume
>>>>>> 5743a879-090d-46db-bc7c-1c0b0669a112: being created as image with
>>>>>> specification: {'status': 'creating', 'volume_name':
>>>>>> 'volume-5743a879-090d-46db-bc7c-1c0b0669a112', 'volume_size': 2,
>>>>>> 'image_id': 'acfd0a14-69e0-44d6-a6a1-aa9dc83e9d5b', 'image_location':
>>>>>> ('rbd://a5ae877c-bcba-53fe-8336-450e63014757/images/acfd0a14-69e0-44d6-a6a1-aa9dc83e9d5b/snap',
>>>>>> [{'url':
>>>>>> 'rbd://a5ae877c-bcba-53fe-8336-450e63014757/images/acfd0a14-69e0-44d6-a6a1-aa9dc83e9d5b/snap',
>>>>>> 'metadata': {'store': 'ceph'}}, {'url':
>>>>>> 'rbd://a8d5f1f5-48e7-5ede-89ab-8aca59b6397b/images/acfd0a14-69e0-44d6-a6a1-aa9dc83e9d5b/snap',
>>>>>> 'metadata': {'store': 'dcn02'}}]), 'image_meta': {'name': 'cirros',
>>>>>> 'disk_format': 'qcow2', 'container_format': 'bare', 'visibility': 'public',
>>>>>> 'size': 16338944, 'virtual_size': 117440512, 'status': 'active',
>>>>>> 'checksum': '1d3062cd89af34e419f7100277f38b2b', 'protected': False,
>>>>>> 'min_ram': 0, 'min_disk': 0, 'owner': '4160ce999a31485fa643aed0936dfef0',
>>>>>> 'os_hidden': False, 'os_hash_algo': 'sha512', 'os_hash_value':
>>>>>> '553d220ed58cfee7dafe003c446a9f197ab5edf8ffc09396c74187cf83873c877e7ae041cb80f3b91489acf687183adcd689b53b38e3ddd22e627e7f98a09c46',
>>>>>> 'id': 'acfd0a14-69e0-44d6-a6a1-aa9dc83e9d5b', 'created_at':
>>>>>> datetime.datetime(2023, 3, 22, 18, 50, 5, tzinfo=datetime.timezone.utc),
>>>>>> 'updated_at': datetime.datetime(2023, 3, 22, 20, 3, 54,
>>>>>> tzinfo=datetime.timezone.utc), 'locations': [{'url':
>>>>>> 'rbd://a5ae877c-bcba-53fe-8336-450e63014757/images/acfd0a14-69e0-44d6-a6a1-aa9dc83e9d5b/snap',
>>>>>> 'metadata': {'store': 'ceph'}}, {'url':
>>>>>> 'rbd://a8d5f1f5-48e7-5ede-89ab-8aca59b6397b/images/acfd0a14-69e0-44d6-a6a1-aa9dc83e9d5b/snap',
>>>>>> 'metadata': {'store': 'dcn02'}}], 'direct_url':
>>>>>> 'rbd://a5ae877c-bcba-53fe-8336-450e63014757/images/acfd0a14-69e0-44d6-a6a1-aa9dc83e9d5b/snap',
>>>>>> 'tags': [], 'file': '/v2/images/acfd0a14-69e0-44d6-a6a1-aa9dc83e9d5b/file',
>>>>>> 'stores': 'ceph,dcn02', 'properties': {'os_glance_failed_import': '',
>>>>>> 'os_glance_importing_to_stores': '', 'owner_specified.openstack.md5': '',
>>>>>> 'owner_specified.openstack.object': 'images/cirros',
>>>>>> 'owner_specified.openstack.sha256': ''}}, 'image_service':
>>>>>> <cinder.image.glance.GlanceImageService object at 0x7f8147973438>}
>>>>>>
>>>>>> With regards,
>>>>>> Swogat Pradhan
>>>>>>
>>>>>> On Wed, Mar 22, 2023 at 9:19 PM Alan Bishop <abishop at redhat.com>
>>>>>> wrote:
>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> On Wed, Mar 22, 2023 at 8:38 AM Swogat Pradhan <
>>>>>>> swogatpradhan22 at gmail.com> wrote:
>>>>>>>
>>>>>>>> Hi Adam,
>>>>>>>> The systems are in same LAN, in this case it seemed like the image
>>>>>>>> was getting pulled from the central site which was caused due to an
>>>>>>>> misconfiguration in ceph.conf file in /var/lib/tripleo-config/ceph/
>>>>>>>> directory, which seems to have been resolved after the changes i made to
>>>>>>>> fix it.
>>>>>>>>
>>>>>>>> Right now the glance api podman is running in unhealthy state and
>>>>>>>> the podman logs don't show any error whatsoever and when issued the command
>>>>>>>> netstat -nultp i do not see any entry for glance port i.e. 9292 in the dcn
>>>>>>>> site, which is why cinder is throwing an error stating:
>>>>>>>>
>>>>>>>> 2023-03-22 13:32:29.786 108 ERROR oslo_messaging.rpc.server
>>>>>>>> cinder.exception.GlanceConnectionFailed: Connection to glance failed: Error
>>>>>>>> finding address for
>>>>>>>> http://172.25.228.253:9292/v2/images/736d8779-07cd-4510-bab2-adcb653cc538:
>>>>>>>> Unable to establish connection to
>>>>>>>> http://172.25.228.253:9292/v2/images/736d8779-07cd-4510-bab2-adcb653cc538:
>>>>>>>> HTTPConnectionPool(host='172.25.228.253', port=9292): Max retries exceeded
>>>>>>>> with url: /v2/images/736d8779-07cd-4510-bab2-adcb653cc538 (Caused by
>>>>>>>> NewConnectionError('<urllib3.connection.HTTPConnection object at
>>>>>>>> 0x7f7682d2cd30>: Failed to establish a new connection: [Errno 111]
>>>>>>>> ECONNREFUSED',))
>>>>>>>>
>>>>>>>> Now i need to find out why the port is not listed as the glance
>>>>>>>> service is running, which i am not sure how to find out.
>>>>>>>>
>>>>>>>
>>>>>>> One other thing to investigate is whether your deployment includes
>>>>>>> this patch [1]. If it does, then bear in mind
>>>>>>> the glance-api service running at the edge site will be an
>>>>>>> "internal" (non public facing) instance that uses port 9293
>>>>>>> instead of 9292. You should familiarize yourself with the release
>>>>>>> note [2].
>>>>>>>
>>>>>>> [1]
>>>>>>> https://opendev.org/openstack/tripleo-heat-templates/commit/3605d45e417a77a1d0f153fbeffcbb283ec85fe6
>>>>>>> [2]
>>>>>>> https://opendev.org/openstack/tripleo-heat-templates/src/branch/stable/wallaby/releasenotes/notes/glance-internal-service-86274f56712ffaac.yaml
>>>>>>>
>>>>>>> Alan
>>>>>>>
>>>>>>>
>>>>>>>> With regards,
>>>>>>>> Swogat Pradhan
>>>>>>>>
>>>>>>>> On Wed, Mar 22, 2023 at 8:11 PM Alan Bishop <abishop at redhat.com>
>>>>>>>> wrote:
>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> On Wed, Mar 22, 2023 at 6:37 AM Swogat Pradhan <
>>>>>>>>> swogatpradhan22 at gmail.com> wrote:
>>>>>>>>>
>>>>>>>>>> Update:
>>>>>>>>>> Here is the log when creating a volume using cirros image:
>>>>>>>>>>
>>>>>>>>>> 2023-03-22 11:04:38.449 109 INFO
>>>>>>>>>> cinder.volume.flows.manager.create_volume
>>>>>>>>>> [req-646b9ac8-a5a7-45ac-a96d-8dd6bb45da17 b240e3e89d99489284cd731e75f2a5db
>>>>>>>>>> 4160ce999a31485fa643aed0936dfef0 - - -] Volume
>>>>>>>>>> bf341343-6609-4b8c-b9e0-93e2a89c8c8f: being created as image with
>>>>>>>>>> specification: {'status': 'creating', 'volume_name':
>>>>>>>>>> 'volume-bf341343-6609-4b8c-b9e0-93e2a89c8c8f', 'volume_size': 4,
>>>>>>>>>> 'image_id': '736d8779-07cd-4510-bab2-adcb653cc538', 'image_location':
>>>>>>>>>> ('rbd://a5ae877c-bcba-53fe-8336-450e63014757/images/736d8779-07cd-4510-bab2-adcb653cc538/snap',
>>>>>>>>>> [{'url':
>>>>>>>>>> 'rbd://a5ae877c-bcba-53fe-8336-450e63014757/images/736d8779-07cd-4510-bab2-adcb653cc538/snap',
>>>>>>>>>> 'metadata': {'store': 'ceph'}}, {'url':
>>>>>>>>>> 'rbd://a8d5f1f5-48e7-5ede-89ab-8aca59b6397b/images/736d8779-07cd-4510-bab2-adcb653cc538/snap',
>>>>>>>>>> 'metadata': {'store': 'dcn02'}}]), 'image_meta': {'name': 'cirros',
>>>>>>>>>> 'disk_format': 'qcow2', 'container_format': 'bare', 'visibility': 'public',
>>>>>>>>>> 'size': 16338944, 'virtual_size': 117440512, 'status': 'active',
>>>>>>>>>> 'checksum': '1d3062cd89af34e419f7100277f38b2b', 'protected': False,
>>>>>>>>>> 'min_ram': 0, 'min_disk': 0, 'owner': '4160ce999a31485fa643aed0936dfef0',
>>>>>>>>>> 'os_hidden': False, 'os_hash_algo': 'sha512', 'os_hash_value':
>>>>>>>>>> '553d220ed58cfee7dafe003c446a9f197ab5edf8ffc09396c74187cf83873c877e7ae041cb80f3b91489acf687183adcd689b53b38e3ddd22e627e7f98a09c46',
>>>>>>>>>> 'id': '736d8779-07cd-4510-bab2-adcb653cc538', 'created_at':
>>>>>>>>>> datetime.datetime(2023, 3, 22, 10, 44, 12, tzinfo=datetime.timezone.utc),
>>>>>>>>>> 'updated_at': datetime.datetime(2023, 3, 22, 10, 54, 1,
>>>>>>>>>> tzinfo=datetime.timezone.utc), 'locations': [{'url':
>>>>>>>>>> 'rbd://a5ae877c-bcba-53fe-8336-450e63014757/images/736d8779-07cd-4510-bab2-adcb653cc538/snap',
>>>>>>>>>> 'metadata': {'store': 'ceph'}}, {'url':
>>>>>>>>>> 'rbd://a8d5f1f5-48e7-5ede-89ab-8aca59b6397b/images/736d8779-07cd-4510-bab2-adcb653cc538/snap',
>>>>>>>>>> 'metadata': {'store': 'dcn02'}}], 'direct_url':
>>>>>>>>>> 'rbd://a5ae877c-bcba-53fe-8336-450e63014757/images/736d8779-07cd-4510-bab2-adcb653cc538/snap',
>>>>>>>>>> 'tags': [], 'file': '/v2/images/736d8779-07cd-4510-bab2-adcb653cc538/file',
>>>>>>>>>> 'stores': 'ceph,dcn02', 'properties': {'os_glance_failed_import': '',
>>>>>>>>>> 'os_glance_importing_to_stores': '', 'owner_specified.openstack.md5': '',
>>>>>>>>>> 'owner_specified.openstack.object': 'images/cirros',
>>>>>>>>>> 'owner_specified.openstack.sha256': ''}}, 'image_service':
>>>>>>>>>> <cinder.image.glance.GlanceImageService object at 0x7f449ded1198>}
>>>>>>>>>> 2023-03-22 11:06:16.570 109 INFO cinder.image.image_utils
>>>>>>>>>> [req-646b9ac8-a5a7-45ac-a96d-8dd6bb45da17 b240e3e89d99489284cd731e75f2a5db
>>>>>>>>>> 4160ce999a31485fa643aed0936dfef0 - - -] Image download 15.58 MB at 0.16 MB/s
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>> As Adam Savage would say, well there's your problem ^^ (Image
>>>>>>>>> download 15.58 MB at 0.16 MB/s). Downloading the image takes too long, and
>>>>>>>>> 0.16 MB/s suggests you have a network issue.
>>>>>>>>>
>>>>>>>>> John Fulton previously stated your cinder-volume service at the
>>>>>>>>> edge site is not using the local ceph image store. Assuming you are
>>>>>>>>> deploying GlanceApiEdge service [1], then the cinder-volume service should
>>>>>>>>> be configured to use the local glance service [2]. You should check
>>>>>>>>> cinder's glance_api_servers to confirm it's the edge site's glance service.
>>>>>>>>>
>>>>>>>>> [1]
>>>>>>>>> https://github.com/openstack/tripleo-heat-templates/blob/stable/wallaby/environments/dcn.yaml#L29
>>>>>>>>> [2]
>>>>>>>>> https://github.com/openstack/tripleo-heat-templates/blob/stable/wallaby/deployment/glance/glance-api-edge-container-puppet.yaml#L80
>>>>>>>>>
>>>>>>>>> Alan
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>> 2023-03-22 11:07:54.023 109 WARNING py.warnings
>>>>>>>>>> [req-646b9ac8-a5a7-45ac-a96d-8dd6bb45da17 b240e3e89d99489284cd731e75f2a5db
>>>>>>>>>> 4160ce999a31485fa643aed0936dfef0 - - -]
>>>>>>>>>> /usr/lib/python3.6/site-packages/oslo_utils/imageutils.py:75:
>>>>>>>>>> FutureWarning: The human format is deprecated and the format parameter will
>>>>>>>>>> be removed. Use explicitly json instead in version 'xena'
>>>>>>>>>>   category=FutureWarning)
>>>>>>>>>>
>>>>>>>>>> 2023-03-22 11:11:12.161 109 WARNING py.warnings
>>>>>>>>>> [req-646b9ac8-a5a7-45ac-a96d-8dd6bb45da17 b240e3e89d99489284cd731e75f2a5db
>>>>>>>>>> 4160ce999a31485fa643aed0936dfef0 - - -]
>>>>>>>>>> /usr/lib/python3.6/site-packages/oslo_utils/imageutils.py:75:
>>>>>>>>>> FutureWarning: The human format is deprecated and the format parameter will
>>>>>>>>>> be removed. Use explicitly json instead in version 'xena'
>>>>>>>>>>   category=FutureWarning)
>>>>>>>>>>
>>>>>>>>>> 2023-03-22 11:11:12.163 109 INFO cinder.image.image_utils
>>>>>>>>>> [req-646b9ac8-a5a7-45ac-a96d-8dd6bb45da17 b240e3e89d99489284cd731e75f2a5db
>>>>>>>>>> 4160ce999a31485fa643aed0936dfef0 - - -] Converted 112.00 MB image at 112.00
>>>>>>>>>> MB/s
>>>>>>>>>> 2023-03-22 11:11:14.998 109 INFO
>>>>>>>>>> cinder.volume.flows.manager.create_volume
>>>>>>>>>> [req-646b9ac8-a5a7-45ac-a96d-8dd6bb45da17 b240e3e89d99489284cd731e75f2a5db
>>>>>>>>>> 4160ce999a31485fa643aed0936dfef0 - - -] Volume
>>>>>>>>>> volume-bf341343-6609-4b8c-b9e0-93e2a89c8c8f
>>>>>>>>>> (bf341343-6609-4b8c-b9e0-93e2a89c8c8f): created successfully
>>>>>>>>>> 2023-03-22 11:11:15.195 109 INFO cinder.volume.manager
>>>>>>>>>> [req-646b9ac8-a5a7-45ac-a96d-8dd6bb45da17 b240e3e89d99489284cd731e75f2a5db
>>>>>>>>>> 4160ce999a31485fa643aed0936dfef0 - - -] Created volume successfully.
>>>>>>>>>>
>>>>>>>>>> The image is present in dcn02 store but still it downloaded the
>>>>>>>>>> image in 0.16 MB/s and then created the volume.
>>>>>>>>>>
>>>>>>>>>> With regards,
>>>>>>>>>> Swogat Pradhan
>>>>>>>>>>
>>>>>>>>>> On Tue, Mar 21, 2023 at 6:10 PM Swogat Pradhan <
>>>>>>>>>> swogatpradhan22 at gmail.com> wrote:
>>>>>>>>>>
>>>>>>>>>>> Hi Jhon,
>>>>>>>>>>> This seems to be an issue.
>>>>>>>>>>> When i deployed the dcn ceph in both dcn01 and dcn02 the
>>>>>>>>>>> --cluster parameter was specified to the respective cluster names but the
>>>>>>>>>>> config files were created in the name of ceph.conf and keyring was
>>>>>>>>>>> ceph.client.openstack.keyring.
>>>>>>>>>>>
>>>>>>>>>>> Which created issues in glance as well as the naming convention
>>>>>>>>>>> of the files didn't match the cluster names, so i had to manually rename
>>>>>>>>>>> the central ceph conf file as such:
>>>>>>>>>>>
>>>>>>>>>>> [root at dcn02-compute-0 ~]# cd /var/lib/tripleo-config/ceph/
>>>>>>>>>>> [root at dcn02-compute-0 ceph]# ll
>>>>>>>>>>> total 16
>>>>>>>>>>> -rw-------. 1 root root 257 Mar 13 13:56
>>>>>>>>>>> ceph_central.client.openstack.keyring
>>>>>>>>>>> -rw-r--r--. 1 root root 428 Mar 13 13:56 ceph_central.conf
>>>>>>>>>>> -rw-------. 1 root root 205 Mar 15 18:45
>>>>>>>>>>> ceph.client.openstack.keyring
>>>>>>>>>>> -rw-r--r--. 1 root root 362 Mar 15 18:45 ceph.conf
>>>>>>>>>>> [root at dcn02-compute-0 ceph]#
>>>>>>>>>>>
>>>>>>>>>>> ceph.conf and ceph.client.openstack.keyring contain the fsid of
>>>>>>>>>>> the respective clusters in both dcn01 and dcn02.
>>>>>>>>>>> In the above cli output, the ceph.conf and ceph.client... are
>>>>>>>>>>> the files used to access dcn02 ceph cluster and ceph_central* files are
>>>>>>>>>>> used in for accessing central ceph cluster.
>>>>>>>>>>>
>>>>>>>>>>> glance multistore config:
>>>>>>>>>>> [dcn02]
>>>>>>>>>>> rbd_store_ceph_conf=/etc/ceph/ceph.conf
>>>>>>>>>>> rbd_store_user=openstack
>>>>>>>>>>> rbd_store_pool=images
>>>>>>>>>>> rbd_thin_provisioning=False
>>>>>>>>>>> store_description=dcn02 rbd glance store
>>>>>>>>>>>
>>>>>>>>>>> [ceph_central]
>>>>>>>>>>> rbd_store_ceph_conf=/etc/ceph/ceph_central.conf
>>>>>>>>>>> rbd_store_user=openstack
>>>>>>>>>>> rbd_store_pool=images
>>>>>>>>>>> rbd_thin_provisioning=False
>>>>>>>>>>> store_description=Default glance store backend.
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> With regards,
>>>>>>>>>>> Swogat Pradhan
>>>>>>>>>>>
>>>>>>>>>>> On Tue, Mar 21, 2023 at 5:52 PM John Fulton <johfulto at redhat.com>
>>>>>>>>>>> wrote:
>>>>>>>>>>>
>>>>>>>>>>>> On Tue, Mar 21, 2023 at 8:03 AM Swogat Pradhan
>>>>>>>>>>>> <swogatpradhan22 at gmail.com> wrote:
>>>>>>>>>>>> >
>>>>>>>>>>>> > Hi,
>>>>>>>>>>>> > Seems like cinder is not using the local ceph.
>>>>>>>>>>>>
>>>>>>>>>>>> That explains the issue. It's a misconfiguration.
>>>>>>>>>>>>
>>>>>>>>>>>> I hope this is not a production system since the mailing list
>>>>>>>>>>>> now has
>>>>>>>>>>>> the cinder.conf which contains passwords.
>>>>>>>>>>>>
>>>>>>>>>>>> The section that looks like this:
>>>>>>>>>>>>
>>>>>>>>>>>> [tripleo_ceph]
>>>>>>>>>>>> volume_backend_name=tripleo_ceph
>>>>>>>>>>>> volume_driver=cinder.volume.drivers.rbd.RBDDriver
>>>>>>>>>>>> rbd_ceph_conf=/etc/ceph/ceph.conf
>>>>>>>>>>>> rbd_user=openstack
>>>>>>>>>>>> rbd_pool=volumes
>>>>>>>>>>>> rbd_flatten_volume_from_snapshot=False
>>>>>>>>>>>> rbd_secret_uuid=<redacted>
>>>>>>>>>>>> report_discard_supported=True
>>>>>>>>>>>>
>>>>>>>>>>>> Should be updated to refer to the local DCN ceph cluster and
>>>>>>>>>>>> not the
>>>>>>>>>>>> central one. Use the ceph conf file for that cluster and ensure
>>>>>>>>>>>> the
>>>>>>>>>>>> rbd_secret_uuid corresponds to that one.
>>>>>>>>>>>>
>>>>>>>>>>>> TripleO’s convention is to set the rbd_secret_uuid to the FSID
>>>>>>>>>>>> of the
>>>>>>>>>>>> Ceph cluster. The FSID should be in the ceph.conf file. The
>>>>>>>>>>>> tripleo_nova_libvirt role will use virsh secret-* commands so
>>>>>>>>>>>> that
>>>>>>>>>>>> libvirt can retrieve the cephx secret using the FSID as a key.
>>>>>>>>>>>> This
>>>>>>>>>>>> can be confirmed with `podman exec nova_virtsecretd virsh
>>>>>>>>>>>> secret-get-value $FSID`.
>>>>>>>>>>>>
>>>>>>>>>>>> The documentation describes how to configure the central and
>>>>>>>>>>>> DCN sites
>>>>>>>>>>>> correctly but an error seems to have occurred while you were
>>>>>>>>>>>> following
>>>>>>>>>>>> it.
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> https://docs.openstack.org/project-deploy-guide/tripleo-docs/latest/features/distributed_multibackend_storage.html
>>>>>>>>>>>>
>>>>>>>>>>>>   John
>>>>>>>>>>>>
>>>>>>>>>>>> >
>>>>>>>>>>>> > Ceph Output:
>>>>>>>>>>>> > [ceph: root at dcn02-ceph-all-0 /]# rbd -p images ls -l
>>>>>>>>>>>> > NAME                                       SIZE     PARENT
>>>>>>>>>>>> FMT  PROT  LOCK
>>>>>>>>>>>> > 2abfafaa-eff4-4c2e-a538-dc2e1249ab65         8 MiB
>>>>>>>>>>>> 2        excl
>>>>>>>>>>>> > 55f40c8a-8f79-48c5-a52a-9b679b762f19        16 MiB
>>>>>>>>>>>> 2
>>>>>>>>>>>> > 55f40c8a-8f79-48c5-a52a-9b679b762f19 at snap   16 MiB
>>>>>>>>>>>>   2  yes
>>>>>>>>>>>> > 59f6a9cd-721c-45b5-a15f-fd021b08160d       321 MiB
>>>>>>>>>>>> 2
>>>>>>>>>>>> > 59f6a9cd-721c-45b5-a15f-fd021b08160d at snap  321 MiB
>>>>>>>>>>>>   2  yes
>>>>>>>>>>>> > 5f5ddd77-35f3-45e8-9dd3-8c1cbb1f39f0       386 MiB
>>>>>>>>>>>> 2
>>>>>>>>>>>> > 5f5ddd77-35f3-45e8-9dd3-8c1cbb1f39f0 at snap  386 MiB
>>>>>>>>>>>>   2  yes
>>>>>>>>>>>> > 9b27248e-a8cf-4f00-a039-d3e3066cd26a        15 GiB
>>>>>>>>>>>> 2
>>>>>>>>>>>> > 9b27248e-a8cf-4f00-a039-d3e3066cd26a at snap   15 GiB
>>>>>>>>>>>>   2  yes
>>>>>>>>>>>> > b7356adc-bb47-4c05-968b-6d3c9ca0079b        15 GiB
>>>>>>>>>>>> 2
>>>>>>>>>>>> > b7356adc-bb47-4c05-968b-6d3c9ca0079b at snap   15 GiB
>>>>>>>>>>>>   2  yes
>>>>>>>>>>>> > e77e78ad-d369-4a1d-b758-8113621269a3        15 GiB
>>>>>>>>>>>> 2
>>>>>>>>>>>> > e77e78ad-d369-4a1d-b758-8113621269a3 at snap   15 GiB
>>>>>>>>>>>>   2  yes
>>>>>>>>>>>> >
>>>>>>>>>>>> > [ceph: root at dcn02-ceph-all-0 /]# rbd -p volumes ls -l
>>>>>>>>>>>> > NAME                                         SIZE     PARENT
>>>>>>>>>>>> FMT  PROT  LOCK
>>>>>>>>>>>> > volume-c644086f-d3cf-406d-b0f1-7691bde5981d  100 GiB
>>>>>>>>>>>>   2
>>>>>>>>>>>> > volume-f0969935-a742-4744-9375-80bf323e4d63   10 GiB
>>>>>>>>>>>>   2
>>>>>>>>>>>> > [ceph: root at dcn02-ceph-all-0 /]#
>>>>>>>>>>>> >
>>>>>>>>>>>> > Attached the cinder config.
>>>>>>>>>>>> > Please let me know how I can solve this issue.
>>>>>>>>>>>> >
>>>>>>>>>>>> > With regards,
>>>>>>>>>>>> > Swogat Pradhan
>>>>>>>>>>>> >
>>>>>>>>>>>> > On Tue, Mar 21, 2023 at 3:53 PM John Fulton <
>>>>>>>>>>>> johfulto at redhat.com> wrote:
>>>>>>>>>>>> >>
>>>>>>>>>>>> >> in my last message under the line "On a DCN site if you run
>>>>>>>>>>>> a command like this:" I suggested some steps you could try to confirm the
>>>>>>>>>>>> image is a COW from the local glance as well as how to look at your cinder
>>>>>>>>>>>> config.
>>>>>>>>>>>> >>
>>>>>>>>>>>> >> On Tue, Mar 21, 2023, 12:06 AM Swogat Pradhan <
>>>>>>>>>>>> swogatpradhan22 at gmail.com> wrote:
>>>>>>>>>>>> >>>
>>>>>>>>>>>> >>> Update:
>>>>>>>>>>>> >>> I uploaded an image directly to the dcn02 store, and it
>>>>>>>>>>>> takes around 10,15 minutes to create a volume with image in dcn02.
>>>>>>>>>>>> >>> The image size is 389 MB.
>>>>>>>>>>>> >>>
>>>>>>>>>>>> >>> On Mon, Mar 20, 2023 at 10:26 PM Swogat Pradhan <
>>>>>>>>>>>> swogatpradhan22 at gmail.com> wrote:
>>>>>>>>>>>> >>>>
>>>>>>>>>>>> >>>> Hi Jhon,
>>>>>>>>>>>> >>>> I checked in the ceph od dcn02, I can see the images
>>>>>>>>>>>> created after importing from the central site.
>>>>>>>>>>>> >>>> But launching an instance normally fails as it takes a
>>>>>>>>>>>> long time for the volume to get created.
>>>>>>>>>>>> >>>>
>>>>>>>>>>>> >>>> When launching an instance from volume the instance is
>>>>>>>>>>>> getting created properly without any errors.
>>>>>>>>>>>> >>>>
>>>>>>>>>>>> >>>> I tried to cache images in nova using
>>>>>>>>>>>> https://docs.openstack.org/project-deploy-guide/tripleo-docs/latest/post_deployment/pre_cache_images.html
>>>>>>>>>>>> but getting checksum failed error.
>>>>>>>>>>>> >>>>
>>>>>>>>>>>> >>>> With regards,
>>>>>>>>>>>> >>>> Swogat Pradhan
>>>>>>>>>>>> >>>>
>>>>>>>>>>>> >>>> On Thu, Mar 16, 2023 at 5:24 PM John Fulton <
>>>>>>>>>>>> johfulto at redhat.com> wrote:
>>>>>>>>>>>> >>>>>
>>>>>>>>>>>> >>>>> On Wed, Mar 15, 2023 at 8:05 PM Swogat Pradhan
>>>>>>>>>>>> >>>>> <swogatpradhan22 at gmail.com> wrote:
>>>>>>>>>>>> >>>>> >
>>>>>>>>>>>> >>>>> > Update: After restarting the nova services on the
>>>>>>>>>>>> controller and running the deploy script on the edge site, I was able to
>>>>>>>>>>>> launch the VM from volume.
>>>>>>>>>>>> >>>>> >
>>>>>>>>>>>> >>>>> > Right now the instance creation is failing as the block
>>>>>>>>>>>> device creation is stuck in creating state, it is taking more than 10 mins
>>>>>>>>>>>> for the volume to be created, whereas the image has already been imported
>>>>>>>>>>>> to the edge glance.
>>>>>>>>>>>> >>>>>
>>>>>>>>>>>> >>>>> Try following this document and making the same
>>>>>>>>>>>> observations in your
>>>>>>>>>>>> >>>>> environment for AZs and their local ceph cluster.
>>>>>>>>>>>> >>>>>
>>>>>>>>>>>> >>>>>
>>>>>>>>>>>> https://docs.openstack.org/project-deploy-guide/tripleo-docs/latest/features/distributed_multibackend_storage.html#confirm-images-may-be-copied-between-sites
>>>>>>>>>>>> >>>>>
>>>>>>>>>>>> >>>>> On a DCN site if you run a command like this:
>>>>>>>>>>>> >>>>>
>>>>>>>>>>>> >>>>> $ sudo cephadm shell --config /etc/ceph/dcn0.conf
>>>>>>>>>>>> --keyring
>>>>>>>>>>>> >>>>> /etc/ceph/dcn0.client.admin.keyring
>>>>>>>>>>>> >>>>> $ rbd --cluster dcn0 -p volumes ls -l
>>>>>>>>>>>> >>>>> NAME                                      SIZE  PARENT
>>>>>>>>>>>> >>>>>                           FMT PROT LOCK
>>>>>>>>>>>> >>>>> volume-28c6fc32-047b-4306-ad2d-de2be02716b7 8 GiB
>>>>>>>>>>>> >>>>> images/8083c7e7-32d8-4f7a-b1da-0ed7884f1076 at snap   2
>>>>>>>>>>>>   excl
>>>>>>>>>>>> >>>>> $
>>>>>>>>>>>> >>>>>
>>>>>>>>>>>> >>>>> Then, you should see the parent of the volume is the
>>>>>>>>>>>> image which is on
>>>>>>>>>>>> >>>>> the same local ceph cluster.
>>>>>>>>>>>> >>>>>
>>>>>>>>>>>> >>>>> I wonder if something is misconfigured and thus you're
>>>>>>>>>>>> encountering
>>>>>>>>>>>> >>>>> the streaming behavior described here:
>>>>>>>>>>>> >>>>>
>>>>>>>>>>>> >>>>> Ideally all images should reside in the central Glance
>>>>>>>>>>>> and be copied
>>>>>>>>>>>> >>>>> to DCN sites before instances of those images are booted
>>>>>>>>>>>> on DCN sites.
>>>>>>>>>>>> >>>>> If an image is not copied to a DCN site before it is
>>>>>>>>>>>> booted, then the
>>>>>>>>>>>> >>>>> image will be streamed to the DCN site and then the image
>>>>>>>>>>>> will boot as
>>>>>>>>>>>> >>>>> an instance. This happens because Glance at the DCN site
>>>>>>>>>>>> has access to
>>>>>>>>>>>> >>>>> the images store at the Central ceph cluster. Though the
>>>>>>>>>>>> booting of
>>>>>>>>>>>> >>>>> the image will take time because it has not been copied
>>>>>>>>>>>> in advance,
>>>>>>>>>>>> >>>>> this is still preferable to failing to boot the image.
>>>>>>>>>>>> >>>>>
>>>>>>>>>>>> >>>>> You can also exec into the cinder container at the DCN
>>>>>>>>>>>> site and
>>>>>>>>>>>> >>>>> confirm it's using it's local ceph cluster.
>>>>>>>>>>>> >>>>>
>>>>>>>>>>>> >>>>>   John
>>>>>>>>>>>> >>>>>
>>>>>>>>>>>> >>>>> >
>>>>>>>>>>>> >>>>> > I will try and create a new fresh image and test again
>>>>>>>>>>>> then update.
>>>>>>>>>>>> >>>>> >
>>>>>>>>>>>> >>>>> > With regards,
>>>>>>>>>>>> >>>>> > Swogat Pradhan
>>>>>>>>>>>> >>>>> >
>>>>>>>>>>>> >>>>> > On Wed, Mar 15, 2023 at 11:13 PM Swogat Pradhan <
>>>>>>>>>>>> swogatpradhan22 at gmail.com> wrote:
>>>>>>>>>>>> >>>>> >>
>>>>>>>>>>>> >>>>> >> Update:
>>>>>>>>>>>> >>>>> >> In the hypervisor list the compute node state is
>>>>>>>>>>>> showing down.
>>>>>>>>>>>> >>>>> >>
>>>>>>>>>>>> >>>>> >>
>>>>>>>>>>>> >>>>> >> On Wed, Mar 15, 2023 at 11:11 PM Swogat Pradhan <
>>>>>>>>>>>> swogatpradhan22 at gmail.com> wrote:
>>>>>>>>>>>> >>>>> >>>
>>>>>>>>>>>> >>>>> >>> Hi Brendan,
>>>>>>>>>>>> >>>>> >>> Now i have deployed another site where i have used 2
>>>>>>>>>>>> linux bonds network template for both 3 compute nodes and 3 ceph nodes.
>>>>>>>>>>>> >>>>> >>> The bonding options is set to mode=802.3ad
>>>>>>>>>>>> (lacp=active).
>>>>>>>>>>>> >>>>> >>> I used a cirros image to launch instance but the
>>>>>>>>>>>> instance timed out so i waited for the volume to be created.
>>>>>>>>>>>> >>>>> >>> Once the volume was created i tried launching the
>>>>>>>>>>>> instance from the volume and still the instance is stuck in spawning state.
>>>>>>>>>>>> >>>>> >>>
>>>>>>>>>>>> >>>>> >>> Here is the nova-compute log:
>>>>>>>>>>>> >>>>> >>>
>>>>>>>>>>>> >>>>> >>> 2023-03-15 17:35:47.739 185437 INFO
>>>>>>>>>>>> oslo.privsep.daemon [-] privsep daemon starting
>>>>>>>>>>>> >>>>> >>> 2023-03-15 17:35:47.744 185437 INFO
>>>>>>>>>>>> oslo.privsep.daemon [-] privsep process running with uid/gid: 0/0
>>>>>>>>>>>> >>>>> >>> 2023-03-15 17:35:47.749 185437 INFO
>>>>>>>>>>>> oslo.privsep.daemon [-] privsep process running with capabilities
>>>>>>>>>>>> (eff/prm/inh): CAP_SYS_ADMIN/CAP_SYS_ADMIN/none
>>>>>>>>>>>> >>>>> >>> 2023-03-15 17:35:47.749 185437 INFO
>>>>>>>>>>>> oslo.privsep.daemon [-] privsep daemon running as pid 185437
>>>>>>>>>>>> >>>>> >>> 2023-03-15 17:35:47.974 8 WARNING
>>>>>>>>>>>> os_brick.initiator.connectors.nvmeof
>>>>>>>>>>>> [req-dbb11a9b-317e-4957-b141-f9e0bdf6a266 b240e3e89d99489284cd731e75f2a5db
>>>>>>>>>>>> 4160ce999a31485fa643aed0936dfef0 - default default] Process execution error
>>>>>>>>>>>> in _get_host_uuid: Unexpected error while running command.
>>>>>>>>>>>> >>>>> >>> Command: blkid overlay -s UUID -o value
>>>>>>>>>>>> >>>>> >>> Exit code: 2
>>>>>>>>>>>> >>>>> >>> Stdout: ''
>>>>>>>>>>>> >>>>> >>> Stderr: '':
>>>>>>>>>>>> oslo_concurrency.processutils.ProcessExecutionError: Unexpected error while
>>>>>>>>>>>> running command.
>>>>>>>>>>>> >>>>> >>> 2023-03-15 17:35:51.616 8 INFO
>>>>>>>>>>>> nova.virt.libvirt.driver [req-dbb11a9b-317e-4957-b141-f9e0bdf6a266
>>>>>>>>>>>> b240e3e89d99489284cd731e75f2a5db 4160ce999a31485fa643aed0936dfef0 - default
>>>>>>>>>>>> default] [instance: 450b749c-a10a-4308-80a9-3b8020fee758] Creating image
>>>>>>>>>>>> >>>>> >>>
>>>>>>>>>>>> >>>>> >>> It is stuck in creating image, do i need to run the
>>>>>>>>>>>> template mentioned here ?:
>>>>>>>>>>>> https://docs.openstack.org/project-deploy-guide/tripleo-docs/latest/post_deployment/pre_cache_images.html
>>>>>>>>>>>> >>>>> >>>
>>>>>>>>>>>> >>>>> >>> The volume is already created and i do not understand
>>>>>>>>>>>> why the instance is stuck in spawning state.
>>>>>>>>>>>> >>>>> >>>
>>>>>>>>>>>> >>>>> >>> With regards,
>>>>>>>>>>>> >>>>> >>> Swogat Pradhan
>>>>>>>>>>>> >>>>> >>>
>>>>>>>>>>>> >>>>> >>>
>>>>>>>>>>>> >>>>> >>> On Sun, Mar 5, 2023 at 4:02 PM Brendan Shephard <
>>>>>>>>>>>> bshephar at redhat.com> wrote:
>>>>>>>>>>>> >>>>> >>>>
>>>>>>>>>>>> >>>>> >>>> Does your environment use different network
>>>>>>>>>>>> interfaces for each of the networks? Or does it have a bond with everything
>>>>>>>>>>>> on it?
>>>>>>>>>>>> >>>>> >>>>
>>>>>>>>>>>> >>>>> >>>> One issue I have seen before is that when launching
>>>>>>>>>>>> instances, there is a lot of network traffic between nodes as the
>>>>>>>>>>>> hypervisor needs to download the image from Glance. Along with various
>>>>>>>>>>>> other services sending normal network traffic, it can be enough to cause
>>>>>>>>>>>> issues if everything is running over a single 1Gbe interface.
>>>>>>>>>>>> >>>>> >>>>
>>>>>>>>>>>> >>>>> >>>> I have seen the same situation in fact when using a
>>>>>>>>>>>> single active/backup bond on 1Gbe nics. It’s worth checking the network
>>>>>>>>>>>> traffic while you try to spawn the instance to see if you’re dropping
>>>>>>>>>>>> packets. In the situation I described, there were dropped packets which
>>>>>>>>>>>> resulted in a loss of communication between nova_compute and RMQ, so the
>>>>>>>>>>>> node appeared offline. You should also confirm that nova_compute is being
>>>>>>>>>>>> disconnected in the nova_compute logs if you tail them on the Hypervisor
>>>>>>>>>>>> while spawning the instance.
>>>>>>>>>>>> >>>>> >>>>
>>>>>>>>>>>> >>>>> >>>> In my case, changing from active/backup to LACP
>>>>>>>>>>>> helped. So, based on that experience, from my perspective, is certainly
>>>>>>>>>>>> sounds like some kind of network issue.
>>>>>>>>>>>> >>>>> >>>>
>>>>>>>>>>>> >>>>> >>>> Regards,
>>>>>>>>>>>> >>>>> >>>>
>>>>>>>>>>>> >>>>> >>>> Brendan Shephard
>>>>>>>>>>>> >>>>> >>>> Senior Software Engineer
>>>>>>>>>>>> >>>>> >>>> Red Hat Australia
>>>>>>>>>>>> >>>>> >>>>
>>>>>>>>>>>> >>>>> >>>>
>>>>>>>>>>>> >>>>> >>>>
>>>>>>>>>>>> >>>>> >>>> On 5 Mar 2023, at 6:47 am, Eugen Block <
>>>>>>>>>>>> eblock at nde.ag> wrote:
>>>>>>>>>>>> >>>>> >>>>
>>>>>>>>>>>> >>>>> >>>> Hi,
>>>>>>>>>>>> >>>>> >>>>
>>>>>>>>>>>> >>>>> >>>> I tried to help someone with a similar issue some
>>>>>>>>>>>> time ago in this thread:
>>>>>>>>>>>> >>>>> >>>>
>>>>>>>>>>>> https://serverfault.com/questions/1116771/openstack-oslo-messaging-exception-in-nova-conductor
>>>>>>>>>>>> >>>>> >>>>
>>>>>>>>>>>> >>>>> >>>> But apparently a neutron reinstallation fixed it for
>>>>>>>>>>>> that user, not sure if that could apply here. But is it possible that your
>>>>>>>>>>>> nova and neutron versions are different between central and edge site? Have
>>>>>>>>>>>> you restarted nova and neutron services on the compute nodes after
>>>>>>>>>>>> installation? Have you debug logs of nova-conductor and maybe nova-compute?
>>>>>>>>>>>> Maybe they can help narrow down the issue.
>>>>>>>>>>>> >>>>> >>>> If there isn't any additional information in the
>>>>>>>>>>>> debug logs I probably would start "tearing down" rabbitmq. I didn't have to
>>>>>>>>>>>> do that in a production system yet so be careful. I can think of two routes:
>>>>>>>>>>>> >>>>> >>>>
>>>>>>>>>>>> >>>>> >>>> - Either remove queues, exchanges etc. while rabbit
>>>>>>>>>>>> is running, this will most likely impact client IO depending on your load.
>>>>>>>>>>>> Check out the rabbitmqctl commands.
>>>>>>>>>>>> >>>>> >>>> - Or stop the rabbitmq cluster, remove the mnesia
>>>>>>>>>>>> tables from all nodes and restart rabbitmq so the exchanges, queues etc.
>>>>>>>>>>>> rebuild.
>>>>>>>>>>>> >>>>> >>>>
>>>>>>>>>>>> >>>>> >>>> I can imagine that the failed reply "survives" while
>>>>>>>>>>>> being replicated across the rabbit nodes. But I don't really know the
>>>>>>>>>>>> rabbit internals too well, so maybe someone else can chime in here and give
>>>>>>>>>>>> a better advice.
>>>>>>>>>>>> >>>>> >>>>
>>>>>>>>>>>> >>>>> >>>> Regards,
>>>>>>>>>>>> >>>>> >>>> Eugen
>>>>>>>>>>>> >>>>> >>>>
>>>>>>>>>>>> >>>>> >>>> Zitat von Swogat Pradhan <swogatpradhan22 at gmail.com
>>>>>>>>>>>> >:
>>>>>>>>>>>> >>>>> >>>>
>>>>>>>>>>>> >>>>> >>>> Hi,
>>>>>>>>>>>> >>>>> >>>> Can someone please help me out on this issue?
>>>>>>>>>>>> >>>>> >>>>
>>>>>>>>>>>> >>>>> >>>> With regards,
>>>>>>>>>>>> >>>>> >>>> Swogat Pradhan
>>>>>>>>>>>> >>>>> >>>>
>>>>>>>>>>>> >>>>> >>>> On Thu, Mar 2, 2023 at 1:24 PM Swogat Pradhan <
>>>>>>>>>>>> swogatpradhan22 at gmail.com>
>>>>>>>>>>>> >>>>> >>>> wrote:
>>>>>>>>>>>> >>>>> >>>>
>>>>>>>>>>>> >>>>> >>>> Hi
>>>>>>>>>>>> >>>>> >>>> I don't see any major packet loss.
>>>>>>>>>>>> >>>>> >>>> It seems the problem is somewhere in rabbitmq maybe
>>>>>>>>>>>> but not due to packet
>>>>>>>>>>>> >>>>> >>>> loss.
>>>>>>>>>>>> >>>>> >>>>
>>>>>>>>>>>> >>>>> >>>> with regards,
>>>>>>>>>>>> >>>>> >>>> Swogat Pradhan
>>>>>>>>>>>> >>>>> >>>>
>>>>>>>>>>>> >>>>> >>>> On Wed, Mar 1, 2023 at 3:34 PM Swogat Pradhan <
>>>>>>>>>>>> swogatpradhan22 at gmail.com>
>>>>>>>>>>>> >>>>> >>>> wrote:
>>>>>>>>>>>> >>>>> >>>>
>>>>>>>>>>>> >>>>> >>>> Hi,
>>>>>>>>>>>> >>>>> >>>> Yes the MTU is the same as the default '1500'.
>>>>>>>>>>>> >>>>> >>>> Generally I haven't seen any packet loss, but never
>>>>>>>>>>>> checked when
>>>>>>>>>>>> >>>>> >>>> launching the instance.
>>>>>>>>>>>> >>>>> >>>> I will check that and come back.
>>>>>>>>>>>> >>>>> >>>> But everytime i launch an instance the instance gets
>>>>>>>>>>>> stuck at spawning
>>>>>>>>>>>> >>>>> >>>> state and there the hypervisor becomes down, so not
>>>>>>>>>>>> sure if packet loss
>>>>>>>>>>>> >>>>> >>>> causes this.
>>>>>>>>>>>> >>>>> >>>>
>>>>>>>>>>>> >>>>> >>>> With regards,
>>>>>>>>>>>> >>>>> >>>> Swogat pradhan
>>>>>>>>>>>> >>>>> >>>>
>>>>>>>>>>>> >>>>> >>>> On Wed, Mar 1, 2023 at 3:30 PM Eugen Block <
>>>>>>>>>>>> eblock at nde.ag> wrote:
>>>>>>>>>>>> >>>>> >>>>
>>>>>>>>>>>> >>>>> >>>> One more thing coming to mind is MTU size. Are they
>>>>>>>>>>>> identical between
>>>>>>>>>>>> >>>>> >>>> central and edge site? Do you see packet loss
>>>>>>>>>>>> through the tunnel?
>>>>>>>>>>>> >>>>> >>>>
>>>>>>>>>>>> >>>>> >>>> Zitat von Swogat Pradhan <swogatpradhan22 at gmail.com
>>>>>>>>>>>> >:
>>>>>>>>>>>> >>>>> >>>>
>>>>>>>>>>>> >>>>> >>>> > Hi Eugen,
>>>>>>>>>>>> >>>>> >>>> > Request you to please add my email either on 'to'
>>>>>>>>>>>> or 'cc' as i am not
>>>>>>>>>>>> >>>>> >>>> > getting email's from you.
>>>>>>>>>>>> >>>>> >>>> > Coming to the issue:
>>>>>>>>>>>> >>>>> >>>> >
>>>>>>>>>>>> >>>>> >>>> > [root at overcloud-controller-no-ceph-3 /]#
>>>>>>>>>>>> rabbitmqctl list_policies -p
>>>>>>>>>>>> >>>>> >>>> /
>>>>>>>>>>>> >>>>> >>>> > Listing policies for vhost "/" ...
>>>>>>>>>>>> >>>>> >>>> > vhost   name    pattern apply-to
>>>>>>>>>>>> definition      priority
>>>>>>>>>>>> >>>>> >>>> > /       ha-all  ^(?!amq\.).*    queues
>>>>>>>>>>>> >>>>> >>>> >
>>>>>>>>>>>> >>>>> >>>>
>>>>>>>>>>>> {"ha-mode":"exactly","ha-params":2,"ha-promote-on-shutdown":"always"}   0
>>>>>>>>>>>> >>>>> >>>> >
>>>>>>>>>>>> >>>>> >>>> > I have the edge site compute nodes up, it only
>>>>>>>>>>>> goes down when i am
>>>>>>>>>>>> >>>>> >>>> trying
>>>>>>>>>>>> >>>>> >>>> > to launch an instance and the instance comes to a
>>>>>>>>>>>> spawning state and
>>>>>>>>>>>> >>>>> >>>> then
>>>>>>>>>>>> >>>>> >>>> > gets stuck.
>>>>>>>>>>>> >>>>> >>>> >
>>>>>>>>>>>> >>>>> >>>> > I have a tunnel setup between the central and the
>>>>>>>>>>>> edge sites.
>>>>>>>>>>>> >>>>> >>>> >
>>>>>>>>>>>> >>>>> >>>> > With regards,
>>>>>>>>>>>> >>>>> >>>> > Swogat Pradhan
>>>>>>>>>>>> >>>>> >>>> >
>>>>>>>>>>>> >>>>> >>>> > On Tue, Feb 28, 2023 at 9:11 PM Swogat Pradhan <
>>>>>>>>>>>> >>>>> >>>> swogatpradhan22 at gmail.com>
>>>>>>>>>>>> >>>>> >>>> > wrote:
>>>>>>>>>>>> >>>>> >>>> >
>>>>>>>>>>>> >>>>> >>>> >> Hi Eugen,
>>>>>>>>>>>> >>>>> >>>> >> For some reason i am not getting your email to me
>>>>>>>>>>>> directly, i am
>>>>>>>>>>>> >>>>> >>>> checking
>>>>>>>>>>>> >>>>> >>>> >> the email digest and there i am able to find your
>>>>>>>>>>>> reply.
>>>>>>>>>>>> >>>>> >>>> >> Here is the log for download:
>>>>>>>>>>>> https://we.tl/t-L8FEkGZFSq
>>>>>>>>>>>> >>>>> >>>> >> Yes, these logs are from the time when the issue
>>>>>>>>>>>> occurred.
>>>>>>>>>>>> >>>>> >>>> >>
>>>>>>>>>>>> >>>>> >>>> >> *Note: i am able to create vm's and perform other
>>>>>>>>>>>> activities in the
>>>>>>>>>>>> >>>>> >>>> >> central site, only facing this issue in the edge
>>>>>>>>>>>> site.*
>>>>>>>>>>>> >>>>> >>>> >>
>>>>>>>>>>>> >>>>> >>>> >> With regards,
>>>>>>>>>>>> >>>>> >>>> >> Swogat Pradhan
>>>>>>>>>>>> >>>>> >>>> >>
>>>>>>>>>>>> >>>>> >>>> >> On Mon, Feb 27, 2023 at 5:12 PM Swogat Pradhan <
>>>>>>>>>>>> >>>>> >>>> swogatpradhan22 at gmail.com>
>>>>>>>>>>>> >>>>> >>>> >> wrote:
>>>>>>>>>>>> >>>>> >>>> >>
>>>>>>>>>>>> >>>>> >>>> >>> Hi Eugen,
>>>>>>>>>>>> >>>>> >>>> >>> Thanks for your response.
>>>>>>>>>>>> >>>>> >>>> >>> I have actually a 4 controller setup so here are
>>>>>>>>>>>> the details:
>>>>>>>>>>>> >>>>> >>>> >>>
>>>>>>>>>>>> >>>>> >>>> >>> *PCS Status:*
>>>>>>>>>>>> >>>>> >>>> >>>   * Container bundle set: rabbitmq-bundle [
>>>>>>>>>>>> >>>>> >>>> >>>
>>>>>>>>>>>> 172.25.201.68:8787/tripleomaster/openstack-rabbitmq:pcmklatest
>>>>>>>>>>>> ]:
>>>>>>>>>>>> >>>>> >>>> >>>     * rabbitmq-bundle-0
>>>>>>>>>>>> (ocf::heartbeat:rabbitmq-cluster):
>>>>>>>>>>>> >>>>> >>>> Started
>>>>>>>>>>>> >>>>> >>>> >>> overcloud-controller-no-ceph-3
>>>>>>>>>>>> >>>>> >>>> >>>     * rabbitmq-bundle-1
>>>>>>>>>>>> (ocf::heartbeat:rabbitmq-cluster):
>>>>>>>>>>>> >>>>> >>>> Started
>>>>>>>>>>>> >>>>> >>>> >>> overcloud-controller-2
>>>>>>>>>>>> >>>>> >>>> >>>     * rabbitmq-bundle-2
>>>>>>>>>>>> (ocf::heartbeat:rabbitmq-cluster):
>>>>>>>>>>>> >>>>> >>>> Started
>>>>>>>>>>>> >>>>> >>>> >>> overcloud-controller-1
>>>>>>>>>>>> >>>>> >>>> >>>     * rabbitmq-bundle-3
>>>>>>>>>>>> (ocf::heartbeat:rabbitmq-cluster):
>>>>>>>>>>>> >>>>> >>>> Started
>>>>>>>>>>>> >>>>> >>>> >>> overcloud-controller-0
>>>>>>>>>>>> >>>>> >>>> >>>
>>>>>>>>>>>> >>>>> >>>> >>> I have tried restarting the bundle multiple
>>>>>>>>>>>> times but the issue is
>>>>>>>>>>>> >>>>> >>>> still
>>>>>>>>>>>> >>>>> >>>> >>> present.
>>>>>>>>>>>> >>>>> >>>> >>>
>>>>>>>>>>>> >>>>> >>>> >>> *Cluster status:*
>>>>>>>>>>>> >>>>> >>>> >>> [root at overcloud-controller-0 /]# rabbitmqctl
>>>>>>>>>>>> cluster_status
>>>>>>>>>>>> >>>>> >>>> >>> Cluster status of node
>>>>>>>>>>>> >>>>> >>>> >>>
>>>>>>>>>>>> rabbit at overcloud-controller-0.internalapi.bdxworld.com ...
>>>>>>>>>>>> >>>>> >>>> >>> Basics
>>>>>>>>>>>> >>>>> >>>> >>>
>>>>>>>>>>>> >>>>> >>>> >>> Cluster name:
>>>>>>>>>>>> rabbit at overcloud-controller-no-ceph-3.bdxworld.com
>>>>>>>>>>>> >>>>> >>>> >>>
>>>>>>>>>>>> >>>>> >>>> >>> Disk Nodes
>>>>>>>>>>>> >>>>> >>>> >>>
>>>>>>>>>>>> >>>>> >>>> >>>
>>>>>>>>>>>> rabbit at overcloud-controller-0.internalapi.bdxworld.com
>>>>>>>>>>>> >>>>> >>>> >>>
>>>>>>>>>>>> rabbit at overcloud-controller-1.internalapi.bdxworld.com
>>>>>>>>>>>> >>>>> >>>> >>>
>>>>>>>>>>>> rabbit at overcloud-controller-2.internalapi.bdxworld.com
>>>>>>>>>>>> >>>>> >>>> >>>
>>>>>>>>>>>> rabbit at overcloud-controller-no-ceph-3.internalapi.bdxworld.com
>>>>>>>>>>>> >>>>> >>>> >>>
>>>>>>>>>>>> >>>>> >>>> >>> Running Nodes
>>>>>>>>>>>> >>>>> >>>> >>>
>>>>>>>>>>>> >>>>> >>>> >>>
>>>>>>>>>>>> rabbit at overcloud-controller-0.internalapi.bdxworld.com
>>>>>>>>>>>> >>>>> >>>> >>>
>>>>>>>>>>>> rabbit at overcloud-controller-1.internalapi.bdxworld.com
>>>>>>>>>>>> >>>>> >>>> >>>
>>>>>>>>>>>> rabbit at overcloud-controller-2.internalapi.bdxworld.com
>>>>>>>>>>>> >>>>> >>>> >>>
>>>>>>>>>>>> rabbit at overcloud-controller-no-ceph-3.internalapi.bdxworld.com
>>>>>>>>>>>> >>>>> >>>> >>>
>>>>>>>>>>>> >>>>> >>>> >>> Versions
>>>>>>>>>>>> >>>>> >>>> >>>
>>>>>>>>>>>> >>>>> >>>> >>>
>>>>>>>>>>>> rabbit at overcloud-controller-0.internalapi.bdxworld.com:
>>>>>>>>>>>> RabbitMQ
>>>>>>>>>>>> >>>>> >>>> 3.8.3
>>>>>>>>>>>> >>>>> >>>> >>> on Erlang 22.3.4.1
>>>>>>>>>>>> >>>>> >>>> >>>
>>>>>>>>>>>> rabbit at overcloud-controller-1.internalapi.bdxworld.com:
>>>>>>>>>>>> RabbitMQ
>>>>>>>>>>>> >>>>> >>>> 3.8.3
>>>>>>>>>>>> >>>>> >>>> >>> on Erlang 22.3.4.1
>>>>>>>>>>>> >>>>> >>>> >>>
>>>>>>>>>>>> rabbit at overcloud-controller-2.internalapi.bdxworld.com:
>>>>>>>>>>>> RabbitMQ
>>>>>>>>>>>> >>>>> >>>> 3.8.3
>>>>>>>>>>>> >>>>> >>>> >>> on Erlang 22.3.4.1
>>>>>>>>>>>> >>>>> >>>> >>>
>>>>>>>>>>>> rabbit at overcloud-controller-no-ceph-3.internalapi.bdxworld.com:
>>>>>>>>>>>> >>>>> >>>> RabbitMQ
>>>>>>>>>>>> >>>>> >>>> >>> 3.8.3 on Erlang 22.3.4.1
>>>>>>>>>>>> >>>>> >>>> >>>
>>>>>>>>>>>> >>>>> >>>> >>> Alarms
>>>>>>>>>>>> >>>>> >>>> >>>
>>>>>>>>>>>> >>>>> >>>> >>> (none)
>>>>>>>>>>>> >>>>> >>>> >>>
>>>>>>>>>>>> >>>>> >>>> >>> Network Partitions
>>>>>>>>>>>> >>>>> >>>> >>>
>>>>>>>>>>>> >>>>> >>>> >>> (none)
>>>>>>>>>>>> >>>>> >>>> >>>
>>>>>>>>>>>> >>>>> >>>> >>> Listeners
>>>>>>>>>>>> >>>>> >>>> >>>
>>>>>>>>>>>> >>>>> >>>> >>> Node:
>>>>>>>>>>>> rabbit at overcloud-controller-0.internalapi.bdxworld.com,
>>>>>>>>>>>> >>>>> >>>> interface:
>>>>>>>>>>>> >>>>> >>>> >>> [::], port: 25672, protocol: clustering,
>>>>>>>>>>>> purpose: inter-node and CLI
>>>>>>>>>>>> >>>>> >>>> tool
>>>>>>>>>>>> >>>>> >>>> >>> communication
>>>>>>>>>>>> >>>>> >>>> >>> Node:
>>>>>>>>>>>> rabbit at overcloud-controller-0.internalapi.bdxworld.com,
>>>>>>>>>>>> >>>>> >>>> interface:
>>>>>>>>>>>> >>>>> >>>> >>> 172.25.201.212, port: 5672, protocol: amqp,
>>>>>>>>>>>> purpose: AMQP 0-9-1
>>>>>>>>>>>> >>>>> >>>> >>> and AMQP 1.0
>>>>>>>>>>>> >>>>> >>>> >>> Node:
>>>>>>>>>>>> rabbit at overcloud-controller-0.internalapi.bdxworld.com,
>>>>>>>>>>>> >>>>> >>>> interface:
>>>>>>>>>>>> >>>>> >>>> >>> [::], port: 15672, protocol: http, purpose: HTTP
>>>>>>>>>>>> API
>>>>>>>>>>>> >>>>> >>>> >>> Node:
>>>>>>>>>>>> rabbit at overcloud-controller-1.internalapi.bdxworld.com,
>>>>>>>>>>>> >>>>> >>>> interface:
>>>>>>>>>>>> >>>>> >>>> >>> [::], port: 25672, protocol: clustering,
>>>>>>>>>>>> purpose: inter-node and CLI
>>>>>>>>>>>> >>>>> >>>> tool
>>>>>>>>>>>> >>>>> >>>> >>> communication
>>>>>>>>>>>> >>>>> >>>> >>> Node:
>>>>>>>>>>>> rabbit at overcloud-controller-1.internalapi.bdxworld.com,
>>>>>>>>>>>> >>>>> >>>> interface:
>>>>>>>>>>>> >>>>> >>>> >>> 172.25.201.205, port: 5672, protocol: amqp,
>>>>>>>>>>>> purpose: AMQP 0-9-1
>>>>>>>>>>>> >>>>> >>>> >>> and AMQP 1.0
>>>>>>>>>>>> >>>>> >>>> >>> Node:
>>>>>>>>>>>> rabbit at overcloud-controller-1.internalapi.bdxworld.com,
>>>>>>>>>>>> >>>>> >>>> interface:
>>>>>>>>>>>> >>>>> >>>> >>> [::], port: 15672, protocol: http, purpose: HTTP
>>>>>>>>>>>> API
>>>>>>>>>>>> >>>>> >>>> >>> Node:
>>>>>>>>>>>> rabbit at overcloud-controller-2.internalapi.bdxworld.com,
>>>>>>>>>>>> >>>>> >>>> interface:
>>>>>>>>>>>> >>>>> >>>> >>> [::], port: 25672, protocol: clustering,
>>>>>>>>>>>> purpose: inter-node and CLI
>>>>>>>>>>>> >>>>> >>>> tool
>>>>>>>>>>>> >>>>> >>>> >>> communication
>>>>>>>>>>>> >>>>> >>>> >>> Node:
>>>>>>>>>>>> rabbit at overcloud-controller-2.internalapi.bdxworld.com,
>>>>>>>>>>>> >>>>> >>>> interface:
>>>>>>>>>>>> >>>>> >>>> >>> 172.25.201.201, port: 5672, protocol: amqp,
>>>>>>>>>>>> purpose: AMQP 0-9-1
>>>>>>>>>>>> >>>>> >>>> >>> and AMQP 1.0
>>>>>>>>>>>> >>>>> >>>> >>> Node:
>>>>>>>>>>>> rabbit at overcloud-controller-2.internalapi.bdxworld.com,
>>>>>>>>>>>> >>>>> >>>> interface:
>>>>>>>>>>>> >>>>> >>>> >>> [::], port: 15672, protocol: http, purpose: HTTP
>>>>>>>>>>>> API
>>>>>>>>>>>> >>>>> >>>> >>> Node:
>>>>>>>>>>>> rabbit at overcloud-controller-no-ceph-3.internalapi.bdxworld.com
>>>>>>>>>>>> >>>>> >>>> ,
>>>>>>>>>>>> >>>>> >>>> >>> interface: [::], port: 25672, protocol:
>>>>>>>>>>>> clustering, purpose:
>>>>>>>>>>>> >>>>> >>>> inter-node and
>>>>>>>>>>>> >>>>> >>>> >>> CLI tool communication
>>>>>>>>>>>> >>>>> >>>> >>> Node:
>>>>>>>>>>>> rabbit at overcloud-controller-no-ceph-3.internalapi.bdxworld.com
>>>>>>>>>>>> >>>>> >>>> ,
>>>>>>>>>>>> >>>>> >>>> >>> interface: 172.25.201.209, port: 5672, protocol:
>>>>>>>>>>>> amqp, purpose: AMQP
>>>>>>>>>>>> >>>>> >>>> 0-9-1
>>>>>>>>>>>> >>>>> >>>> >>> and AMQP 1.0
>>>>>>>>>>>> >>>>> >>>> >>> Node:
>>>>>>>>>>>> rabbit at overcloud-controller-no-ceph-3.internalapi.bdxworld.com
>>>>>>>>>>>> >>>>> >>>> ,
>>>>>>>>>>>> >>>>> >>>> >>> interface: [::], port: 15672, protocol: http,
>>>>>>>>>>>> purpose: HTTP API
>>>>>>>>>>>> >>>>> >>>> >>>
>>>>>>>>>>>> >>>>> >>>> >>> Feature flags
>>>>>>>>>>>> >>>>> >>>> >>>
>>>>>>>>>>>> >>>>> >>>> >>> Flag: drop_unroutable_metric, state: enabled
>>>>>>>>>>>> >>>>> >>>> >>> Flag: empty_basic_get_metric, state: enabled
>>>>>>>>>>>> >>>>> >>>> >>> Flag: implicit_default_bindings, state: enabled
>>>>>>>>>>>> >>>>> >>>> >>> Flag: quorum_queue, state: enabled
>>>>>>>>>>>> >>>>> >>>> >>> Flag: virtual_host_metadata, state: enabled
>>>>>>>>>>>> >>>>> >>>> >>>
>>>>>>>>>>>> >>>>> >>>> >>> *Logs:*
>>>>>>>>>>>> >>>>> >>>> >>> *(Attached)*
>>>>>>>>>>>> >>>>> >>>> >>>
>>>>>>>>>>>> >>>>> >>>> >>> With regards,
>>>>>>>>>>>> >>>>> >>>> >>> Swogat Pradhan
>>>>>>>>>>>> >>>>> >>>> >>>
>>>>>>>>>>>> >>>>> >>>> >>> On Sun, Feb 26, 2023 at 2:34 PM Swogat Pradhan <
>>>>>>>>>>>> >>>>> >>>> swogatpradhan22 at gmail.com>
>>>>>>>>>>>> >>>>> >>>> >>> wrote:
>>>>>>>>>>>> >>>>> >>>> >>>
>>>>>>>>>>>> >>>>> >>>> >>>> Hi,
>>>>>>>>>>>> >>>>> >>>> >>>> Please find the nova conductor as well as nova
>>>>>>>>>>>> api log.
>>>>>>>>>>>> >>>>> >>>> >>>>
>>>>>>>>>>>> >>>>> >>>> >>>> nova-conuctor:
>>>>>>>>>>>> >>>>> >>>> >>>>
>>>>>>>>>>>> >>>>> >>>> >>>> 2023-02-26 08:45:01.108 31 WARNING
>>>>>>>>>>>> >>>>> >>>> oslo_messaging._drivers.amqpdriver
>>>>>>>>>>>> >>>>> >>>> >>>> [req-caefe26d-153a-4dfd-9ea6-bc5ca0d46679 - - -
>>>>>>>>>>>> - -]
>>>>>>>>>>>> >>>>> >>>> >>>> reply_349bcb075f8c49329435a0f884b33066 doesn't
>>>>>>>>>>>> exist, drop reply to
>>>>>>>>>>>> >>>>> >>>> >>>> 16152921c1eb45c2b1f562087140168b
>>>>>>>>>>>> >>>>> >>>> >>>> 2023-02-26 08:45:02.144 26 WARNING
>>>>>>>>>>>> >>>>> >>>> oslo_messaging._drivers.amqpdriver
>>>>>>>>>>>> >>>>> >>>> >>>> [req-7b43c4e5-0475-4598-92c0-fcacb51d9813 - - -
>>>>>>>>>>>> - -]
>>>>>>>>>>>> >>>>> >>>> >>>> reply_276049ec36a84486a8a406911d9802f4 doesn't
>>>>>>>>>>>> exist, drop reply to
>>>>>>>>>>>> >>>>> >>>> >>>> 83dbe5f567a940b698acfe986f6194fa
>>>>>>>>>>>> >>>>> >>>> >>>> 2023-02-26 08:45:02.314 32 WARNING
>>>>>>>>>>>> >>>>> >>>> oslo_messaging._drivers.amqpdriver
>>>>>>>>>>>> >>>>> >>>> >>>> [req-7b43c4e5-0475-4598-92c0-fcacb51d9813 - - -
>>>>>>>>>>>> - -]
>>>>>>>>>>>> >>>>> >>>> >>>> reply_276049ec36a84486a8a406911d9802f4 doesn't
>>>>>>>>>>>> exist, drop reply to
>>>>>>>>>>>> >>>>> >>>> >>>> f3bfd7f65bd542b18d84cea3033abb43:
>>>>>>>>>>>> >>>>> >>>> >>>> oslo_messaging.exceptions.MessageUndeliverable
>>>>>>>>>>>> >>>>> >>>> >>>> 2023-02-26 08:45:02.316 32 ERROR
>>>>>>>>>>>> oslo_messaging._drivers.amqpdriver
>>>>>>>>>>>> >>>>> >>>> >>>> [req-7b43c4e5-0475-4598-92c0-fcacb51d9813 - - -
>>>>>>>>>>>> - -] The reply
>>>>>>>>>>>> >>>>> >>>> >>>> f3bfd7f65bd542b18d84cea3033abb43 failed to send
>>>>>>>>>>>> after 60 seconds
>>>>>>>>>>>> >>>>> >>>> due to a
>>>>>>>>>>>> >>>>> >>>> >>>> missing queue
>>>>>>>>>>>> (reply_276049ec36a84486a8a406911d9802f4).
>>>>>>>>>>>> >>>>> >>>> Abandoning...:
>>>>>>>>>>>> >>>>> >>>> >>>> oslo_messaging.exceptions.MessageUndeliverable
>>>>>>>>>>>> >>>>> >>>> >>>> 2023-02-26 08:48:01.282 35 WARNING
>>>>>>>>>>>> >>>>> >>>> oslo_messaging._drivers.amqpdriver
>>>>>>>>>>>> >>>>> >>>> >>>> [req-caefe26d-153a-4dfd-9ea6-bc5ca0d46679 - - -
>>>>>>>>>>>> - -]
>>>>>>>>>>>> >>>>> >>>> >>>> reply_349bcb075f8c49329435a0f884b33066 doesn't
>>>>>>>>>>>> exist, drop reply to
>>>>>>>>>>>> >>>>> >>>> >>>> d4b9180f91a94f9a82c3c9c4b7595566:
>>>>>>>>>>>> >>>>> >>>> >>>> oslo_messaging.exceptions.MessageUndeliverable
>>>>>>>>>>>> >>>>> >>>> >>>> 2023-02-26 08:48:01.284 35 ERROR
>>>>>>>>>>>> oslo_messaging._drivers.amqpdriver
>>>>>>>>>>>> >>>>> >>>> >>>> [req-caefe26d-153a-4dfd-9ea6-bc5ca0d46679 - - -
>>>>>>>>>>>> - -] The reply
>>>>>>>>>>>> >>>>> >>>> >>>> d4b9180f91a94f9a82c3c9c4b7595566 failed to send
>>>>>>>>>>>> after 60 seconds
>>>>>>>>>>>> >>>>> >>>> due to a
>>>>>>>>>>>> >>>>> >>>> >>>> missing queue
>>>>>>>>>>>> (reply_349bcb075f8c49329435a0f884b33066).
>>>>>>>>>>>> >>>>> >>>> Abandoning...:
>>>>>>>>>>>> >>>>> >>>> >>>> oslo_messaging.exceptions.MessageUndeliverable
>>>>>>>>>>>> >>>>> >>>> >>>> 2023-02-26 08:49:01.303 33 WARNING
>>>>>>>>>>>> >>>>> >>>> oslo_messaging._drivers.amqpdriver
>>>>>>>>>>>> >>>>> >>>> >>>> [req-caefe26d-153a-4dfd-9ea6-bc5ca0d46679 - - -
>>>>>>>>>>>> - -]
>>>>>>>>>>>> >>>>> >>>> >>>> reply_349bcb075f8c49329435a0f884b33066 doesn't
>>>>>>>>>>>> exist, drop reply to
>>>>>>>>>>>> >>>>> >>>> >>>> 897911a234a445d8a0d8af02ece40f6f:
>>>>>>>>>>>> >>>>> >>>> >>>> oslo_messaging.exceptions.MessageUndeliverable
>>>>>>>>>>>> >>>>> >>>> >>>> 2023-02-26 08:49:01.304 33 ERROR
>>>>>>>>>>>> oslo_messaging._drivers.amqpdriver
>>>>>>>>>>>> >>>>> >>>> >>>> [req-caefe26d-153a-4dfd-9ea6-bc5ca0d46679 - - -
>>>>>>>>>>>> - -] The reply
>>>>>>>>>>>> >>>>> >>>> >>>> 897911a234a445d8a0d8af02ece40f6f failed to send
>>>>>>>>>>>> after 60 seconds
>>>>>>>>>>>> >>>>> >>>> due to a
>>>>>>>>>>>> >>>>> >>>> >>>> missing queue
>>>>>>>>>>>> (reply_349bcb075f8c49329435a0f884b33066).
>>>>>>>>>>>> >>>>> >>>> Abandoning...:
>>>>>>>>>>>> >>>>> >>>> >>>> oslo_messaging.exceptions.MessageUndeliverable
>>>>>>>>>>>> >>>>> >>>> >>>> 2023-02-26 08:49:52.254 31 WARNING
>>>>>>>>>>>> nova.cache_utils
>>>>>>>>>>>> >>>>> >>>> >>>> [req-3a1547ea-326f-4dd0-9127-7f4a4bdf1e45
>>>>>>>>>>>> >>>>> >>>> b240e3e89d99489284cd731e75f2a5db
>>>>>>>>>>>> >>>>> >>>> >>>> 4160ce999a31485fa643aed0936dfef0 - default
>>>>>>>>>>>> default] Cache enabled
>>>>>>>>>>>> >>>>> >>>> with
>>>>>>>>>>>> >>>>> >>>> >>>> backend dogpile.cache.null.
>>>>>>>>>>>> >>>>> >>>> >>>> 2023-02-26 08:50:01.264 27 WARNING
>>>>>>>>>>>> >>>>> >>>> oslo_messaging._drivers.amqpdriver
>>>>>>>>>>>> >>>>> >>>> >>>> [req-caefe26d-153a-4dfd-9ea6-bc5ca0d46679 - - -
>>>>>>>>>>>> - -]
>>>>>>>>>>>> >>>>> >>>> >>>> reply_349bcb075f8c49329435a0f884b33066 doesn't
>>>>>>>>>>>> exist, drop reply to
>>>>>>>>>>>> >>>>> >>>> >>>> 8f723ceb10c3472db9a9f324861df2bb:
>>>>>>>>>>>> >>>>> >>>> >>>> oslo_messaging.exceptions.MessageUndeliverable
>>>>>>>>>>>> >>>>> >>>> >>>> 2023-02-26 08:50:01.266 27 ERROR
>>>>>>>>>>>> oslo_messaging._drivers.amqpdriver
>>>>>>>>>>>> >>>>> >>>> >>>> [req-caefe26d-153a-4dfd-9ea6-bc5ca0d46679 - - -
>>>>>>>>>>>> - -] The reply
>>>>>>>>>>>> >>>>> >>>> >>>> 8f723ceb10c3472db9a9f324861df2bb failed to send
>>>>>>>>>>>> after 60 seconds
>>>>>>>>>>>> >>>>> >>>> due to a
>>>>>>>>>>>> >>>>> >>>> >>>> missing queue
>>>>>>>>>>>> (reply_349bcb075f8c49329435a0f884b33066).
>>>>>>>>>>>> >>>>> >>>> Abandoning...:
>>>>>>>>>>>> >>>>> >>>> >>>> oslo_messaging.exceptions.MessageUndeliverable
>>>>>>>>>>>> >>>>> >>>> >>>>
>>>>>>>>>>>> >>>>> >>>> >>>> With regards,
>>>>>>>>>>>> >>>>> >>>> >>>> Swogat Pradhan
>>>>>>>>>>>> >>>>> >>>> >>>>
>>>>>>>>>>>> >>>>> >>>> >>>> On Sun, Feb 26, 2023 at 2:26 PM Swogat Pradhan <
>>>>>>>>>>>> >>>>> >>>> >>>> swogatpradhan22 at gmail.com> wrote:
>>>>>>>>>>>> >>>>> >>>> >>>>
>>>>>>>>>>>> >>>>> >>>> >>>>> Hi,
>>>>>>>>>>>> >>>>> >>>> >>>>> I currently have 3 compute nodes on edge site1
>>>>>>>>>>>> where i am trying to
>>>>>>>>>>>> >>>>> >>>> >>>>> launch vm's.
>>>>>>>>>>>> >>>>> >>>> >>>>> When the VM is in spawning state the node goes
>>>>>>>>>>>> down (openstack
>>>>>>>>>>>> >>>>> >>>> compute
>>>>>>>>>>>> >>>>> >>>> >>>>> service list), the node comes backup when i
>>>>>>>>>>>> restart the nova
>>>>>>>>>>>> >>>>> >>>> compute
>>>>>>>>>>>> >>>>> >>>> >>>>> service but then the launch of the vm fails.
>>>>>>>>>>>> >>>>> >>>> >>>>>
>>>>>>>>>>>> >>>>> >>>> >>>>> nova-compute.log
>>>>>>>>>>>> >>>>> >>>> >>>>>
>>>>>>>>>>>> >>>>> >>>> >>>>> 2023-02-26 08:15:51.808 7 INFO
>>>>>>>>>>>> nova.compute.manager
>>>>>>>>>>>> >>>>> >>>> >>>>> [req-bc0f5f2e-53fc-4dae-b1da-82f1f972d617 - -
>>>>>>>>>>>> - - -] Running
>>>>>>>>>>>> >>>>> >>>> >>>>> instance usage
>>>>>>>>>>>> >>>>> >>>> >>>>> audit for host dcn01-hci-0.bdxworld.com from
>>>>>>>>>>>> 2023-02-26 07:00:00
>>>>>>>>>>>> >>>>> >>>> to
>>>>>>>>>>>> >>>>> >>>> >>>>> 2023-02-26 08:00:00. 0 instances.
>>>>>>>>>>>> >>>>> >>>> >>>>> 2023-02-26 08:49:52.813 7 INFO
>>>>>>>>>>>> nova.compute.claims
>>>>>>>>>>>> >>>>> >>>> >>>>> [req-3a1547ea-326f-4dd0-9127-7f4a4bdf1e45
>>>>>>>>>>>> >>>>> >>>> >>>>> b240e3e89d99489284cd731e75f2a5db
>>>>>>>>>>>> >>>>> >>>> >>>>> 4160ce999a31485fa643aed0936dfef0 - default
>>>>>>>>>>>> default] [instance:
>>>>>>>>>>>> >>>>> >>>> >>>>> 0c62c1ef-9010-417d-a05f-4db77e901600] Claim
>>>>>>>>>>>> successful on node
>>>>>>>>>>>> >>>>> >>>> >>>>> dcn01-hci-0.bdxworld.com
>>>>>>>>>>>> >>>>> >>>> >>>>> 2023-02-26 08:49:54.225 7 INFO
>>>>>>>>>>>> nova.virt.libvirt.driver
>>>>>>>>>>>> >>>>> >>>> >>>>> [req-3a1547ea-326f-4dd0-9127-7f4a4bdf1e45
>>>>>>>>>>>> >>>>> >>>> >>>>> b240e3e89d99489284cd731e75f2a5db
>>>>>>>>>>>> >>>>> >>>> >>>>> 4160ce999a31485fa643aed0936dfef0 - default
>>>>>>>>>>>> default] [instance:
>>>>>>>>>>>> >>>>> >>>> >>>>> 0c62c1ef-9010-417d-a05f-4db77e901600] Ignoring
>>>>>>>>>>>> supplied device
>>>>>>>>>>>> >>>>> >>>> name:
>>>>>>>>>>>> >>>>> >>>> >>>>> /dev/vda. Libvirt can't honour user-supplied
>>>>>>>>>>>> dev names
>>>>>>>>>>>> >>>>> >>>> >>>>> 2023-02-26 08:49:54.398 7 INFO
>>>>>>>>>>>> nova.virt.block_device
>>>>>>>>>>>> >>>>> >>>> >>>>> [req-3a1547ea-326f-4dd0-9127-7f4a4bdf1e45
>>>>>>>>>>>> >>>>> >>>> >>>>> b240e3e89d99489284cd731e75f2a5db
>>>>>>>>>>>> >>>>> >>>> >>>>> 4160ce999a31485fa643aed0936dfef0 - default
>>>>>>>>>>>> default] [instance:
>>>>>>>>>>>> >>>>> >>>> >>>>> 0c62c1ef-9010-417d-a05f-4db77e901600] Booting
>>>>>>>>>>>> with volume
>>>>>>>>>>>> >>>>> >>>> >>>>> c4bd7885-5973-4860-bbe6-7a2f726baeee at
>>>>>>>>>>>> /dev/vda
>>>>>>>>>>>> >>>>> >>>> >>>>> 2023-02-26 08:49:55.216 7 WARNING
>>>>>>>>>>>> nova.cache_utils
>>>>>>>>>>>> >>>>> >>>> >>>>> [req-3a1547ea-326f-4dd0-9127-7f4a4bdf1e45
>>>>>>>>>>>> >>>>> >>>> >>>>> b240e3e89d99489284cd731e75f2a5db
>>>>>>>>>>>> >>>>> >>>> >>>>> 4160ce999a31485fa643aed0936dfef0 - default
>>>>>>>>>>>> default] Cache enabled
>>>>>>>>>>>> >>>>> >>>> with
>>>>>>>>>>>> >>>>> >>>> >>>>> backend dogpile.cache.null.
>>>>>>>>>>>> >>>>> >>>> >>>>> 2023-02-26 08:49:55.283 7 INFO
>>>>>>>>>>>> oslo.privsep.daemon
>>>>>>>>>>>> >>>>> >>>> >>>>> [req-3a1547ea-326f-4dd0-9127-7f4a4bdf1e45
>>>>>>>>>>>> >>>>> >>>> >>>>> b240e3e89d99489284cd731e75f2a5db
>>>>>>>>>>>> >>>>> >>>> >>>>> 4160ce999a31485fa643aed0936dfef0 - default
>>>>>>>>>>>> default] Running
>>>>>>>>>>>> >>>>> >>>> >>>>> privsep helper:
>>>>>>>>>>>> >>>>> >>>> >>>>> ['sudo', 'nova-rootwrap',
>>>>>>>>>>>> '/etc/nova/rootwrap.conf',
>>>>>>>>>>>> >>>>> >>>> 'privsep-helper',
>>>>>>>>>>>> >>>>> >>>> >>>>> '--config-file', '/etc/nova/nova.conf',
>>>>>>>>>>>> '--config-file',
>>>>>>>>>>>> >>>>> >>>> >>>>> '/etc/nova/nova-compute.conf',
>>>>>>>>>>>> '--privsep_context',
>>>>>>>>>>>> >>>>> >>>> >>>>> 'os_brick.privileged.default',
>>>>>>>>>>>> '--privsep_sock_path',
>>>>>>>>>>>> >>>>> >>>> >>>>> '/tmp/tmpin40tah6/privsep.sock']
>>>>>>>>>>>> >>>>> >>>> >>>>> 2023-02-26 08:49:55.791 7 INFO
>>>>>>>>>>>> oslo.privsep.daemon
>>>>>>>>>>>> >>>>> >>>> >>>>> [req-3a1547ea-326f-4dd0-9127-7f4a4bdf1e45
>>>>>>>>>>>> >>>>> >>>> >>>>> b240e3e89d99489284cd731e75f2a5db
>>>>>>>>>>>> >>>>> >>>> >>>>> 4160ce999a31485fa643aed0936dfef0 - default
>>>>>>>>>>>> default] Spawned new
>>>>>>>>>>>> >>>>> >>>> privsep
>>>>>>>>>>>> >>>>> >>>> >>>>> daemon via rootwrap
>>>>>>>>>>>> >>>>> >>>> >>>>> 2023-02-26 08:49:55.717 2647 INFO
>>>>>>>>>>>> oslo.privsep.daemon [-] privsep
>>>>>>>>>>>> >>>>> >>>> >>>>> daemon starting
>>>>>>>>>>>> >>>>> >>>> >>>>> 2023-02-26 08:49:55.722 2647 INFO
>>>>>>>>>>>> oslo.privsep.daemon [-] privsep
>>>>>>>>>>>> >>>>> >>>> >>>>> process running with uid/gid: 0/0
>>>>>>>>>>>> >>>>> >>>> >>>>> 2023-02-26 08:49:55.726 2647 INFO
>>>>>>>>>>>> oslo.privsep.daemon [-] privsep
>>>>>>>>>>>> >>>>> >>>> >>>>> process running with capabilities
>>>>>>>>>>>> (eff/prm/inh):
>>>>>>>>>>>> >>>>> >>>> >>>>> CAP_SYS_ADMIN/CAP_SYS_ADMIN/none
>>>>>>>>>>>> >>>>> >>>> >>>>> 2023-02-26 08:49:55.726 2647 INFO
>>>>>>>>>>>> oslo.privsep.daemon [-] privsep
>>>>>>>>>>>> >>>>> >>>> >>>>> daemon running as pid 2647
>>>>>>>>>>>> >>>>> >>>> >>>>> 2023-02-26 08:49:55.956 7 WARNING
>>>>>>>>>>>> >>>>> >>>> os_brick.initiator.connectors.nvmeof
>>>>>>>>>>>> >>>>> >>>> >>>>> [req-3a1547ea-326f-4dd0-9127-7f4a4bdf1e45
>>>>>>>>>>>> >>>>> >>>> >>>>> b240e3e89d99489284cd731e75f2a5db
>>>>>>>>>>>> >>>>> >>>> >>>>> 4160ce999a31485fa643aed0936dfef0 - default
>>>>>>>>>>>> default] Process
>>>>>>>>>>>> >>>>> >>>> >>>>> execution error
>>>>>>>>>>>> >>>>> >>>> >>>>> in _get_host_uuid: Unexpected error while
>>>>>>>>>>>> running command.
>>>>>>>>>>>> >>>>> >>>> >>>>> Command: blkid overlay -s UUID -o value
>>>>>>>>>>>> >>>>> >>>> >>>>> Exit code: 2
>>>>>>>>>>>> >>>>> >>>> >>>>> Stdout: ''
>>>>>>>>>>>> >>>>> >>>> >>>>> Stderr: '':
>>>>>>>>>>>> oslo_concurrency.processutils.ProcessExecutionError:
>>>>>>>>>>>> >>>>> >>>> >>>>> Unexpected error while running command.
>>>>>>>>>>>> >>>>> >>>> >>>>> 2023-02-26 08:49:58.247 7 INFO
>>>>>>>>>>>> nova.virt.libvirt.driver
>>>>>>>>>>>> >>>>> >>>> >>>>> [req-3a1547ea-326f-4dd0-9127-7f4a4bdf1e45
>>>>>>>>>>>> >>>>> >>>> >>>>> b240e3e89d99489284cd731e75f2a5db
>>>>>>>>>>>> >>>>> >>>> >>>>> 4160ce999a31485fa643aed0936dfef0 - default
>>>>>>>>>>>> default] [instance:
>>>>>>>>>>>> >>>>> >>>> >>>>> 0c62c1ef-9010-417d-a05f-4db77e901600] Creating
>>>>>>>>>>>> image
>>>>>>>>>>>> >>>>> >>>> >>>>>
>>>>>>>>>>>> >>>>> >>>> >>>>> Is there a way to solve this issue?
>>>>>>>>>>>> >>>>> >>>> >>>>>
>>>>>>>>>>>> >>>>> >>>> >>>>>
>>>>>>>>>>>> >>>>> >>>> >>>>> With regards,
>>>>>>>>>>>> >>>>> >>>> >>>>>
>>>>>>>>>>>> >>>>> >>>> >>>>> Swogat Pradhan
>>>>>>>>>>>> >>>>> >>>> >>>>>
>>>>>>>>>>>> >>>>> >>>> >>>>
>>>>>>>>>>>> >>>>> >>>>
>>>>>>>>>>>> >>>>> >>>>
>>>>>>>>>>>> >>>>> >>>>
>>>>>>>>>>>> >>>>> >>>>
>>>>>>>>>>>> >>>>> >>>>
>>>>>>>>>>>> >>>>> >>>>
>>>>>>>>>>>> >>>>> >>>>
>>>>>>>>>>>> >>>>> >>>>
>>>>>>>>>>>> >>>>> >>>>
>>>>>>>>>>>> >>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <https://lists.openstack.org/pipermail/openstack-discuss/attachments/20230323/1d3ac151/attachment-0001.htm>


More information about the openstack-discuss mailing list