Windows2012 VM image importing and Instance Launch Failure

Eugen Block eblock at nde.ag
Tue Aug 24 08:08:49 UTC 2021


Of course you need a running nova-conductor, how did you launch VMs before?


Zitat von KK CHN <kkchn.in at gmail.com>:

> yes. I too suspect this not the problem with image.   I have seen the logs:
> it throwing Errors which pasted here.
>
> Its waiting for nova-conductor service as in the logs.   Is this the issue?
> Do I need to explictly statrt conductor service or some other service ?
> nova-compute logs pasted below the lines.   share your thoughts what may
> the issue.
>
>
> On Fri, Aug 20, 2021 at 8:11 AM Mohammed Naser <mnaser at vexxhost.com> wrote:
>
>> I suggest that you have a look at the compute node logs when it fails to
>> spawn. I suspect the problem is not Your images but something inside
>> openstack
>>
>>
> tail: no files remaining
> cloud at Meghctrl1:~$ sudo tail -f /var/log/nova/nova-compute.log
> [sudo] password for cloud:
> 2021-08-19 11:59:59.477 2847961 ERROR oslo_service.periodic_task
> transport_options=transport_options)
> 2021-08-19 11:59:59.477 2847961 ERROR oslo_service.periodic_task
> File  
> "/usr/local/lib/python3.7/dist-packages/oslo_messaging/_drivers/amqpdriver.py",
> line 642, in _send
> 2021-08-19 11:59:59.477 2847961 ERROR oslo_service.periodic_task
> call_monitor_timeout)
> 2021-08-19 11:59:59.477 2847961 ERROR oslo_service.periodic_task
> File  
> "/usr/local/lib/python3.7/dist-packages/oslo_messaging/_drivers/amqpdriver.py",
> line 531, in wait
> 2021-08-19 11:59:59.477 2847961 ERROR oslo_service.periodic_task
> message = self.waiters.get(msg_id, timeout=timeout)
> 2021-08-19 11:59:59.477 2847961 ERROR oslo_service.periodic_task
> File  
> "/usr/local/lib/python3.7/dist-packages/oslo_messaging/_drivers/amqpdriver.py",
> line 409, in get
> 2021-08-19 11:59:59.477 2847961 ERROR oslo_service.periodic_task
> 'to messageID %s' % msg_id)
> 2021-08-19 11:59:59.477 2847961 ERROR oslo_service.periodic_task
> oslo_messaging.exceptions.MessagingTimeout: Timed out waiting for a
> reply to message ID 7a4e3a4acad3403a9966570cc259f7b7
> 2021-08-19 11:59:59.477 2847961 ERROR oslo_service.periodic_task
> 2021-08-19 12:00:01.467 2847961 WARNING oslo.service.loopingcall [-]
> Function 'nova.servicegroup.drivers.db.DbDriver._report_state' run
> outlasted interval by 50.03 sec
> 2021-08-19 12:00:34.012 2847812 WARNING nova.conductor.api
> [req-afca4d09-bf9c-4bbc-ae88-22008d12b978 - - - - -] Timed out waiting
> for nova-conductor.  Is it running? Or did this service start before
> nova-conductor?  Reattempting establishment of nova-conductor
> connection...: oslo_messaging.exceptions.MessagingTimeout: Timed out
> waiting for a reply to message ID 9f5b593361b34c8d91585c00e0514047
> 2021-08-19 12:00:59.480 2847961 DEBUG oslo_service.periodic_task
> [req-58a0e182-6e56-4695-b4db-75624ca69c7b - - - - -] Running periodic
> task ComputeManager._check_instance_build_time run_periodic_tasks
> /usr/local/lib/python3.7/dist-packages/oslo_service/periodic_task.py:211
> 2021-08-19 12:00:59.481 2847961 DEBUG oslo_service.periodic_task
> [req-58a0e182-6e56-4695-b4db-75624ca69c7b - - - - -] Running periodic
> task ComputeManager._sync_scheduler_instance_info run_periodic_tasks
> /usr/local/lib/python3.7/dist-packages/oslo_service/periodic_task.py:211
> 2021-08-19 12:01:01.499 2847961 WARNING oslo.service.loopingcall [-]
> Function 'nova.servicegroup.drivers.db.DbDriver._report_state' run
> outlasted interval by 50.03 sec
> 2021-08-19 12:01:34.040 2847812 WARNING nova.conductor.api
> [req-afca4d09-bf9c-4bbc-ae88-22008d12b978 - - - - -] Timed out waiting
> for nova-conductor.  Is it running? Or did this service start before
> nova-conductor?  Reattempting establishment of nova-conductor
> connection...: oslo_messaging.exceptions.MessagingTimeout: Timed out
> waiting for a reply to message ID e8e478be640e4ecab8a3d27c01960440
>
>
> essage ID d620b1b3938e4428a0500c00df8d68ee
> 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task
> [req-58a0e182-6e56-4695-b4db-75624ca69c7b - - - - -] Error during
> ComputeManager._cleanup_expired_console_auth_tokens:
> oslo_messaging.exceptions.MessagingTimeout: Timed out waiting for a
> reply to message ID a493182a404242b79c8f11f8ec350e36
> 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task
> Traceback (most recent call last):
> 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task
> File  
> "/usr/local/lib/python3.7/dist-packages/oslo_messaging/_drivers/amqpdriver.py",
> line 405, in get
> 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task
> return self._queues[msg_id].get(block=True, timeout=timeout)
> 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task
> File "/usr/local/lib/python3.7/dist-packages/eventlet/queue.py", line
> 322, in get
> 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task
> return waiter.wait()
> 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task
> File "/usr/local/lib/python3.7/dist-packages/eventlet/queue.py", line
> 141, in wait
> 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task
> return get_hub().switch()
> 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task
> File "/usr/local/lib/python3.7/dist-packages/eventlet/hubs/hub.py",
> line 298, in switch
> 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task
> return self.greenlet.switch()
> 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task _queue.Empty
> 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task
> 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task
> During handling of the above exception, another exception occurred:
> 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task
> 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task
> Traceback (most recent call last):
> 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task
> File "/usr/local/lib/python3.7/dist-packages/oslo_service/periodic_task.py",
> line 216, in run_periodic_tasks
> 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task
> task(self, context)
> 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task
> File "/usr/local/lib/python3.7/dist-packages/nova/compute/manager.py",
> line 10450, in _cleanup_expired_console_auth_tokens
> 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task
> objects.ConsoleAuthToken.clean_expired_console_auths(context)
> 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task
> File "/usr/local/lib/python3.7/dist-packages/oslo_versionedobjects/base.py",
> line 177, in wrapper
> 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task
> args, kwargs)
> 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task
> File "/usr/local/lib/python3.7/dist-packages/nova/conductor/rpcapi.py",
> line 243, in object_class_action_versions
> 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task
> args=args, kwargs=kwargs)
> 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task
> File "/usr/local/lib/python3.7/dist-packages/oslo_messaging/rpc/client.py",
> line 179, in call
> 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task
> transport_options=self.transport_options)
> 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task
> File "/usr/local/lib/python3.7/dist-packages/oslo_messaging/transport.py",
> line 128, in _send
> 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task
> transport_options=transport_options)
> 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task
> File  
> "/usr/local/lib/python3.7/dist-packages/oslo_messaging/_drivers/amqpdriver.py",
> line 654, in send
> 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task
> transport_options=transport_options)
> 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task
> File  
> "/usr/local/lib/python3.7/dist-packages/oslo_messaging/_drivers/amqpdriver.py",
> line 642, in _send
> 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task
> call_monitor_timeout)
> 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task
> File  
> "/usr/local/lib/python3.7/dist-packages/oslo_messaging/_drivers/amqpdriver.py",
> line 531, in wait
> 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task
> message = self.waiters.get(msg_id, timeout=timeout)
> 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task
> File  
> "/usr/local/lib/python3.7/dist-packages/oslo_messaging/_drivers/amqpdriver.py",
> line 409, in get
> 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task
> 'to message ID %s' % msg_id)
> 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task
> oslo_messaging.exceptions.MessagingTimeout: Timed out waiting for a
> reply to message ID a493182a404242b79c8f11f8ec350e36
> 2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task
> 2021-08-19 12:27:01.681 2847961 DEBUG oslo_service.periodic_task
> [req-58a0e182-6e56-4695-b4db-75624ca69c7b - - - - -] Running periodic
> task ComputeManager._check_instance_build_time run_periodic_tasks
> /usr/local/lib/python3.7/dist-packages/oslo_service/periodic_task.py:211
> 2021-08-19 12:27:01.687 2847961 DEBUG oslo_service.periodic_task
> [req-58a0e182-6e56-4695-b4db-75624ca69c7b - - - - -] Running periodic
> task ComputeManager._sync_scheduler_instance_info run_periodic_tasks
> /usr/local/lib/python3.7/dist-packages/oslo_service/periodic_task.py:211
> 2021-08-19 12:27:02.368 2847961 WARNING oslo.service.loopingcall [-]
> Function 'nova.servicegroup.drivers.db.DbDriver._report_state' run
> outlasted interval by 50.03 sec
>
>
>
> Error-nov-log-file_while_sch ... e_in _horizaon_dashboard.txt
> Open with
> Displaying Error-nov-log-file_while_scheduling_Windows_VM_instance_in
> _horizaon_dashboard.txt.
>
>
>
>
>
>> If I was to guess it’s probably missing UEFI firmware packages :)
>>
>> On Wed, Aug 18, 2021 at 9:17 AM KK CHN <kkchn.in at gmail.com> wrote:
>>
>>> Error : failed to perform requested operation on instance "WindowsVM "the
>>> instance has error status. Please try again later [Error exceeded maximum
>>> number of retries. Exhausted all hosts available for retrying build
>>> failures for instance e3d5c095-7d26-4b1e-89d1-d1a6e20a45041
>>>
>>> I am trying to import a WIndows2012 Single disk VM, to OpenStack Ussuri,
>>> with glance and Qemu KVM.
>>>
>>> In bare machine KVM I am able to import and boot this Windows VM which
>>> exported from rhevm hypervisor as  vhdx image.
>>> what I have done is
>>>
>>> 1. converted this windows  image from vhdx to qcow2
>>> 2.  root at MeghCtrol1:/home/cloud/CMOBB_APP#cirt-install --name WINDOWS
>>> --ram=1048 --vcups=1 --cpu host --hvm --dick
>>> path=BackUP2_CMAPP_disk_1_Windows_qcow2_imagefile,device=disk,
>>> format=qcow2,bus=virtio --graphics vnc --boot uefi
>>>
>>> This uploaded the qcow2 image of WindowsVM to  KVM  hypervisor and its
>>> working.
>>>
>>> But when I do importing to openstack   unable to launch  instance from
>>> the image .
>>>
>>> These are the steps I performed..
>>>
>>> 1. openstack image create "WindowsVM" --file CMAPP_disk_1.qcow2
>>> --disk-format qcow2 --container-format bare --public
>>>
>>> 4.openstack image set --property hw_firmware_type=uefi --property
>>> os_secure_boot=required WindowsVM
>>>
>>> 5.openstack image set --property hw_firmware_type=uefi --property
>>> hw_disk_bus=ide WindowsVM
>>>
>>> 6.openstack image show WindowsVM
>>>
>>> 7. root at dmzcloud:/home/cloud# openstack image show WindowsVM|grep
>>> "properties"
>>> | properties       | hw_disk_bus='ide', hw_firmware_type='uefi',
>>> os_hash_algo='sha512',
>>>
>>> os_hash_value='753ee596980409e1e72d6d020c8219c56a6ada8b43f634fb575c594a245725a398e45982c0a1ad72b3fc3451cde62cceb9ff22be044863b31ecdd7893b049349',
>>> os_hidden='False', os_secure_boot='required',
>>> owner_specified.openstack.md5='',
>>> owner_specified.openstack.object='images/WindowsVM',
>>> owner_specified.openstack.sha256='' |
>>> root at dmzcloud:/home/cloud#
>>>
>>>
>>> Then  I logged into horizon dashboard,  from the images  selected the
>>> imported image and try to launch the instance.  With  a Flavour of 550 GB
>>> disk, 4 vcpus and 8GB Ram ..
>>>
>>> The instance spawning ran for 30 minutes and throws the error which I
>>> pasted first in the right top corner of horizon dashboard.
>>>
>>> How to solve this error and boot the Windows machine successfully ..
>>>
>>>
>>> """
>>> Error : failed to perform requested operation on instance "WindowsVM "the
>>> instance has error status. Please try again later [Error exceeded maximum
>>> number of retries. Exhausted all hosts available for retrying build
>>> failures for instance e3d5c095-7d26-4b1e-89d1-d1a6e20a45041
>>>
>>>
>>> """
>>>
>>> Any help highly appreciated.
>>>
>>> Kris
>>>
>>> --
>> Mohammed Naser
>> VEXXHOST, Inc.
>>






More information about the openstack-discuss mailing list