yes. I too suspect this not the problem with image. I have seen the logs: it throwing Errors which pasted here. Its waiting for nova-conductor service as in the logs. Is this the issue? Do I need to explictly statrt conductor service or some other service ? nova-compute logs pasted below the lines. share your thoughts what may the issue. On Fri, Aug 20, 2021 at 8:11 AM Mohammed Naser <mnaser@vexxhost.com> wrote:
I suggest that you have a look at the compute node logs when it fails to spawn. I suspect the problem is not Your images but something inside openstack
tail: no files remaining cloud@Meghctrl1:~$ sudo tail -f /var/log/nova/nova-compute.log [sudo] password for cloud: 2021-08-19 11:59:59.477 2847961 ERROR oslo_service.periodic_task transport_options=transport_options) 2021-08-19 11:59:59.477 2847961 ERROR oslo_service.periodic_task File "/usr/local/lib/python3.7/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 642, in _send 2021-08-19 11:59:59.477 2847961 ERROR oslo_service.periodic_task call_monitor_timeout) 2021-08-19 11:59:59.477 2847961 ERROR oslo_service.periodic_task File "/usr/local/lib/python3.7/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 531, in wait 2021-08-19 11:59:59.477 2847961 ERROR oslo_service.periodic_task message = self.waiters.get(msg_id, timeout=timeout) 2021-08-19 11:59:59.477 2847961 ERROR oslo_service.periodic_task File "/usr/local/lib/python3.7/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 409, in get 2021-08-19 11:59:59.477 2847961 ERROR oslo_service.periodic_task 'to messageID %s' % msg_id) 2021-08-19 11:59:59.477 2847961 ERROR oslo_service.periodic_task oslo_messaging.exceptions.MessagingTimeout: Timed out waiting for a reply to message ID 7a4e3a4acad3403a9966570cc259f7b7 2021-08-19 11:59:59.477 2847961 ERROR oslo_service.periodic_task 2021-08-19 12:00:01.467 2847961 WARNING oslo.service.loopingcall [-] Function 'nova.servicegroup.drivers.db.DbDriver._report_state' run outlasted interval by 50.03 sec 2021-08-19 12:00:34.012 2847812 WARNING nova.conductor.api [req-afca4d09-bf9c-4bbc-ae88-22008d12b978 - - - - -] Timed out waiting for nova-conductor. Is it running? Or did this service start before nova-conductor? Reattempting establishment of nova-conductor connection...: oslo_messaging.exceptions.MessagingTimeout: Timed out waiting for a reply to message ID 9f5b593361b34c8d91585c00e0514047 2021-08-19 12:00:59.480 2847961 DEBUG oslo_service.periodic_task [req-58a0e182-6e56-4695-b4db-75624ca69c7b - - - - -] Running periodic task ComputeManager._check_instance_build_time run_periodic_tasks /usr/local/lib/python3.7/dist-packages/oslo_service/periodic_task.py:211 2021-08-19 12:00:59.481 2847961 DEBUG oslo_service.periodic_task [req-58a0e182-6e56-4695-b4db-75624ca69c7b - - - - -] Running periodic task ComputeManager._sync_scheduler_instance_info run_periodic_tasks /usr/local/lib/python3.7/dist-packages/oslo_service/periodic_task.py:211 2021-08-19 12:01:01.499 2847961 WARNING oslo.service.loopingcall [-] Function 'nova.servicegroup.drivers.db.DbDriver._report_state' run outlasted interval by 50.03 sec 2021-08-19 12:01:34.040 2847812 WARNING nova.conductor.api [req-afca4d09-bf9c-4bbc-ae88-22008d12b978 - - - - -] Timed out waiting for nova-conductor. Is it running? Or did this service start before nova-conductor? Reattempting establishment of nova-conductor connection...: oslo_messaging.exceptions.MessagingTimeout: Timed out waiting for a reply to message ID e8e478be640e4ecab8a3d27c01960440 essage ID d620b1b3938e4428a0500c00df8d68ee 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task [req-58a0e182-6e56-4695-b4db-75624ca69c7b - - - - -] Error during ComputeManager._cleanup_expired_console_auth_tokens: oslo_messaging.exceptions.MessagingTimeout: Timed out waiting for a reply to message ID a493182a404242b79c8f11f8ec350e36 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task Traceback (most recent call last): 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task File "/usr/local/lib/python3.7/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 405, in get 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task return self._queues[msg_id].get(block=True, timeout=timeout) 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task File "/usr/local/lib/python3.7/dist-packages/eventlet/queue.py", line 322, in get 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task return waiter.wait() 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task File "/usr/local/lib/python3.7/dist-packages/eventlet/queue.py", line 141, in wait 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task return get_hub().switch() 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task File "/usr/local/lib/python3.7/dist-packages/eventlet/hubs/hub.py", line 298, in switch 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task return self.greenlet.switch() 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task _queue.Empty 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task During handling of the above exception, another exception occurred: 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task Traceback (most recent call last): 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task File "/usr/local/lib/python3.7/dist-packages/oslo_service/periodic_task.py", line 216, in run_periodic_tasks 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task task(self, context) 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task File "/usr/local/lib/python3.7/dist-packages/nova/compute/manager.py", line 10450, in _cleanup_expired_console_auth_tokens 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task objects.ConsoleAuthToken.clean_expired_console_auths(context) 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task File "/usr/local/lib/python3.7/dist-packages/oslo_versionedobjects/base.py", line 177, in wrapper 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task args, kwargs) 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task File "/usr/local/lib/python3.7/dist-packages/nova/conductor/rpcapi.py", line 243, in object_class_action_versions 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task args=args, kwargs=kwargs) 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task File "/usr/local/lib/python3.7/dist-packages/oslo_messaging/rpc/client.py", line 179, in call 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task transport_options=self.transport_options) 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task File "/usr/local/lib/python3.7/dist-packages/oslo_messaging/transport.py", line 128, in _send 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task transport_options=transport_options) 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task File "/usr/local/lib/python3.7/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 654, in send 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task transport_options=transport_options) 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task File "/usr/local/lib/python3.7/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 642, in _send 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task call_monitor_timeout) 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task File "/usr/local/lib/python3.7/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 531, in wait 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task message = self.waiters.get(msg_id, timeout=timeout) 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task File "/usr/local/lib/python3.7/dist-packages/oslo_messaging/_drivers/amqpdriver.py", line 409, in get 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task 'to message ID %s' % msg_id) 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task oslo_messaging.exceptions.MessagingTimeout: Timed out waiting for a reply to message ID a493182a404242b79c8f11f8ec350e36 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task 2021-08-19 12:27:01.681 2847961 DEBUG oslo_service.periodic_task [req-58a0e182-6e56-4695-b4db-75624ca69c7b - - - - -] Running periodic task ComputeManager._check_instance_build_time run_periodic_tasks /usr/local/lib/python3.7/dist-packages/oslo_service/periodic_task.py:211 2021-08-19 12:27:01.687 2847961 DEBUG oslo_service.periodic_task [req-58a0e182-6e56-4695-b4db-75624ca69c7b - - - - -] Running periodic task ComputeManager._sync_scheduler_instance_info run_periodic_tasks /usr/local/lib/python3.7/dist-packages/oslo_service/periodic_task.py:211 2021-08-19 12:27:02.368 2847961 WARNING oslo.service.loopingcall [-] Function 'nova.servicegroup.drivers.db.DbDriver._report_state' run outlasted interval by 50.03 sec Error-nov-log-file_while_sch ... e_in _horizaon_dashboard.txt Open with Displaying Error-nov-log-file_while_scheduling_Windows_VM_instance_in _horizaon_dashboard.txt.
If I was to guess it’s probably missing UEFI firmware packages :)
On Wed, Aug 18, 2021 at 9:17 AM KK CHN <kkchn.in@gmail.com> wrote:
Error : failed to perform requested operation on instance "WindowsVM "the instance has error status. Please try again later [Error exceeded maximum number of retries. Exhausted all hosts available for retrying build failures for instance e3d5c095-7d26-4b1e-89d1-d1a6e20a45041
I am trying to import a WIndows2012 Single disk VM, to OpenStack Ussuri, with glance and Qemu KVM.
In bare machine KVM I am able to import and boot this Windows VM which exported from rhevm hypervisor as vhdx image. what I have done is
1. converted this windows image from vhdx to qcow2 2. root@MeghCtrol1:/home/cloud/CMOBB_APP#cirt-install --name WINDOWS --ram=1048 --vcups=1 --cpu host --hvm --dick path=BackUP2_CMAPP_disk_1_Windows_qcow2_imagefile,device=disk, format=qcow2,bus=virtio --graphics vnc --boot uefi
This uploaded the qcow2 image of WindowsVM to KVM hypervisor and its working.
But when I do importing to openstack unable to launch instance from the image .
These are the steps I performed..
1. openstack image create "WindowsVM" --file CMAPP_disk_1.qcow2 --disk-format qcow2 --container-format bare --public
4.openstack image set --property hw_firmware_type=uefi --property os_secure_boot=required WindowsVM
5.openstack image set --property hw_firmware_type=uefi --property hw_disk_bus=ide WindowsVM
6.openstack image show WindowsVM
7. root@dmzcloud:/home/cloud# openstack image show WindowsVM|grep "properties" | properties | hw_disk_bus='ide', hw_firmware_type='uefi', os_hash_algo='sha512',
os_hash_value='753ee596980409e1e72d6d020c8219c56a6ada8b43f634fb575c594a245725a398e45982c0a1ad72b3fc3451cde62cceb9ff22be044863b31ecdd7893b049349', os_hidden='False', os_secure_boot='required', owner_specified.openstack.md5='', owner_specified.openstack.object='images/WindowsVM', owner_specified.openstack.sha256='' | root@dmzcloud:/home/cloud#
Then I logged into horizon dashboard, from the images selected the imported image and try to launch the instance. With a Flavour of 550 GB disk, 4 vcpus and 8GB Ram ..
The instance spawning ran for 30 minutes and throws the error which I pasted first in the right top corner of horizon dashboard.
How to solve this error and boot the Windows machine successfully ..
""" Error : failed to perform requested operation on instance "WindowsVM "the instance has error status. Please try again later [Error exceeded maximum number of retries. Exhausted all hosts available for retrying build failures for instance e3d5c095-7d26-4b1e-89d1-d1a6e20a45041
"""
Any help highly appreciated.
Kris
-- Mohammed Naser VEXXHOST, Inc.