Windows2012 VM image importing and Instance Launch Failure

KK CHN kkchn.in at gmail.com
Sun Aug 22 06:38:03 UTC 2021


yes. I too suspect this not the problem with image.   I have seen the logs:
it throwing Errors which pasted here.

Its waiting for nova-conductor service as in the logs.   Is this the issue?
Do I need to explictly statrt conductor service or some other service ?
nova-compute logs pasted below the lines.   share your thoughts what may
the issue.


On Fri, Aug 20, 2021 at 8:11 AM Mohammed Naser <mnaser at vexxhost.com> wrote:

> I suggest that you have a look at the compute node logs when it fails to
> spawn. I suspect the problem is not Your images but something inside
> openstack
>
>
tail: no files remaining
cloud at Meghctrl1:~$ sudo tail -f /var/log/nova/nova-compute.log
[sudo] password for cloud:
2021-08-19 11:59:59.477 2847961 ERROR oslo_service.periodic_task
transport_options=transport_options)
2021-08-19 11:59:59.477 2847961 ERROR oslo_service.periodic_task
File "/usr/local/lib/python3.7/dist-packages/oslo_messaging/_drivers/amqpdriver.py",
line 642, in _send
2021-08-19 11:59:59.477 2847961 ERROR oslo_service.periodic_task
call_monitor_timeout)
2021-08-19 11:59:59.477 2847961 ERROR oslo_service.periodic_task
File "/usr/local/lib/python3.7/dist-packages/oslo_messaging/_drivers/amqpdriver.py",
line 531, in wait
2021-08-19 11:59:59.477 2847961 ERROR oslo_service.periodic_task
message = self.waiters.get(msg_id, timeout=timeout)
2021-08-19 11:59:59.477 2847961 ERROR oslo_service.periodic_task
File "/usr/local/lib/python3.7/dist-packages/oslo_messaging/_drivers/amqpdriver.py",
line 409, in get
2021-08-19 11:59:59.477 2847961 ERROR oslo_service.periodic_task
'to messageID %s' % msg_id)
2021-08-19 11:59:59.477 2847961 ERROR oslo_service.periodic_task
oslo_messaging.exceptions.MessagingTimeout: Timed out waiting for a
reply to message ID 7a4e3a4acad3403a9966570cc259f7b7
2021-08-19 11:59:59.477 2847961 ERROR oslo_service.periodic_task
2021-08-19 12:00:01.467 2847961 WARNING oslo.service.loopingcall [-]
Function 'nova.servicegroup.drivers.db.DbDriver._report_state' run
outlasted interval by 50.03 sec
2021-08-19 12:00:34.012 2847812 WARNING nova.conductor.api
[req-afca4d09-bf9c-4bbc-ae88-22008d12b978 - - - - -] Timed out waiting
for nova-conductor.  Is it running? Or did this service start before
nova-conductor?  Reattempting establishment of nova-conductor
connection...: oslo_messaging.exceptions.MessagingTimeout: Timed out
waiting for a reply to message ID 9f5b593361b34c8d91585c00e0514047
2021-08-19 12:00:59.480 2847961 DEBUG oslo_service.periodic_task
[req-58a0e182-6e56-4695-b4db-75624ca69c7b - - - - -] Running periodic
task ComputeManager._check_instance_build_time run_periodic_tasks
/usr/local/lib/python3.7/dist-packages/oslo_service/periodic_task.py:211
2021-08-19 12:00:59.481 2847961 DEBUG oslo_service.periodic_task
[req-58a0e182-6e56-4695-b4db-75624ca69c7b - - - - -] Running periodic
task ComputeManager._sync_scheduler_instance_info run_periodic_tasks
/usr/local/lib/python3.7/dist-packages/oslo_service/periodic_task.py:211
2021-08-19 12:01:01.499 2847961 WARNING oslo.service.loopingcall [-]
Function 'nova.servicegroup.drivers.db.DbDriver._report_state' run
outlasted interval by 50.03 sec
2021-08-19 12:01:34.040 2847812 WARNING nova.conductor.api
[req-afca4d09-bf9c-4bbc-ae88-22008d12b978 - - - - -] Timed out waiting
for nova-conductor.  Is it running? Or did this service start before
nova-conductor?  Reattempting establishment of nova-conductor
connection...: oslo_messaging.exceptions.MessagingTimeout: Timed out
waiting for a reply to message ID e8e478be640e4ecab8a3d27c01960440


essage ID d620b1b3938e4428a0500c00df8d68ee
2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task
[req-58a0e182-6e56-4695-b4db-75624ca69c7b - - - - -] Error during
ComputeManager._cleanup_expired_console_auth_tokens:
oslo_messaging.exceptions.MessagingTimeout: Timed out waiting for a
reply to message ID a493182a404242b79c8f11f8ec350e36
2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task
Traceback (most recent call last):
2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task
File "/usr/local/lib/python3.7/dist-packages/oslo_messaging/_drivers/amqpdriver.py",
line 405, in get
2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task
return self._queues[msg_id].get(block=True, timeout=timeout)
2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task
File "/usr/local/lib/python3.7/dist-packages/eventlet/queue.py", line
322, in get
2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task
return waiter.wait()
2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task
File "/usr/local/lib/python3.7/dist-packages/eventlet/queue.py", line
141, in wait
2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task
return get_hub().switch()
2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task
File "/usr/local/lib/python3.7/dist-packages/eventlet/hubs/hub.py",
line 298, in switch
2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task
return self.greenlet.switch()
2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task _queue.Empty
2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task
2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task
During handling of the above exception, another exception occurred:
2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task
2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task
Traceback (most recent call last):
2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task
File "/usr/local/lib/python3.7/dist-packages/oslo_service/periodic_task.py",
line 216, in run_periodic_tasks
2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task
task(self, context)
2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task
File "/usr/local/lib/python3.7/dist-packages/nova/compute/manager.py",
line 10450, in _cleanup_expired_console_auth_tokens
2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task
objects.ConsoleAuthToken.clean_expired_console_auths(context)
2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task
File "/usr/local/lib/python3.7/dist-packages/oslo_versionedobjects/base.py",
line 177, in wrapper
2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task
args, kwargs)
2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task
File "/usr/local/lib/python3.7/dist-packages/nova/conductor/rpcapi.py",
line 243, in object_class_action_versions
2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task
args=args, kwargs=kwargs)
2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task
File "/usr/local/lib/python3.7/dist-packages/oslo_messaging/rpc/client.py",
line 179, in call
2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task
transport_options=self.transport_options)
2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task
File "/usr/local/lib/python3.7/dist-packages/oslo_messaging/transport.py",
line 128, in _send
2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task
transport_options=transport_options)
2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task
File "/usr/local/lib/python3.7/dist-packages/oslo_messaging/_drivers/amqpdriver.py",
line 654, in send
2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task
transport_options=transport_options)
2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task
File "/usr/local/lib/python3.7/dist-packages/oslo_messaging/_drivers/amqpdriver.py",
line 642, in _send
2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task
call_monitor_timeout)
2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task
File "/usr/local/lib/python3.7/dist-packages/oslo_messaging/_drivers/amqpdriver.py",
line 531, in wait
2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task
message = self.waiters.get(msg_id, timeout=timeout)
2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task
File "/usr/local/lib/python3.7/dist-packages/oslo_messaging/_drivers/amqpdriver.py",
line 409, in get
2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task
'to message ID %s' % msg_id)
2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task
oslo_messaging.exceptions.MessagingTimeout: Timed out waiting for a
reply to message ID a493182a404242b79c8f11f8ec350e36
2021-08-19 12:27:00.387 2847961 ERROR oslo_service.periodic_task
2021-08-19 12:27:01.681 2847961 DEBUG oslo_service.periodic_task
[req-58a0e182-6e56-4695-b4db-75624ca69c7b - - - - -] Running periodic
task ComputeManager._check_instance_build_time run_periodic_tasks
/usr/local/lib/python3.7/dist-packages/oslo_service/periodic_task.py:211
2021-08-19 12:27:01.687 2847961 DEBUG oslo_service.periodic_task
[req-58a0e182-6e56-4695-b4db-75624ca69c7b - - - - -] Running periodic
task ComputeManager._sync_scheduler_instance_info run_periodic_tasks
/usr/local/lib/python3.7/dist-packages/oslo_service/periodic_task.py:211
2021-08-19 12:27:02.368 2847961 WARNING oslo.service.loopingcall [-]
Function 'nova.servicegroup.drivers.db.DbDriver._report_state' run
outlasted interval by 50.03 sec



Error-nov-log-file_while_sch ... e_in _horizaon_dashboard.txt
Open with
Displaying Error-nov-log-file_while_scheduling_Windows_VM_instance_in
_horizaon_dashboard.txt.





> If I was to guess it’s probably missing UEFI firmware packages :)
>
> On Wed, Aug 18, 2021 at 9:17 AM KK CHN <kkchn.in at gmail.com> wrote:
>
>> Error : failed to perform requested operation on instance "WindowsVM "the
>> instance has error status. Please try again later [Error exceeded maximum
>> number of retries. Exhausted all hosts available for retrying build
>> failures for instance e3d5c095-7d26-4b1e-89d1-d1a6e20a45041
>>
>> I am trying to import a WIndows2012 Single disk VM, to OpenStack Ussuri,
>> with glance and Qemu KVM.
>>
>> In bare machine KVM I am able to import and boot this Windows VM which
>> exported from rhevm hypervisor as  vhdx image.
>> what I have done is
>>
>> 1. converted this windows  image from vhdx to qcow2
>> 2.  root at MeghCtrol1:/home/cloud/CMOBB_APP#cirt-install --name WINDOWS
>> --ram=1048 --vcups=1 --cpu host --hvm --dick
>> path=BackUP2_CMAPP_disk_1_Windows_qcow2_imagefile,device=disk,
>> format=qcow2,bus=virtio --graphics vnc --boot uefi
>>
>> This uploaded the qcow2 image of WindowsVM to  KVM  hypervisor and its
>> working.
>>
>> But when I do importing to openstack   unable to launch  instance from
>> the image .
>>
>> These are the steps I performed..
>>
>> 1. openstack image create "WindowsVM" --file CMAPP_disk_1.qcow2
>> --disk-format qcow2 --container-format bare --public
>>
>> 4.openstack image set --property hw_firmware_type=uefi --property
>> os_secure_boot=required WindowsVM
>>
>> 5.openstack image set --property hw_firmware_type=uefi --property
>> hw_disk_bus=ide WindowsVM
>>
>> 6.openstack image show WindowsVM
>>
>> 7. root at dmzcloud:/home/cloud# openstack image show WindowsVM|grep
>> "properties"
>> | properties       | hw_disk_bus='ide', hw_firmware_type='uefi',
>> os_hash_algo='sha512',
>>
>> os_hash_value='753ee596980409e1e72d6d020c8219c56a6ada8b43f634fb575c594a245725a398e45982c0a1ad72b3fc3451cde62cceb9ff22be044863b31ecdd7893b049349',
>> os_hidden='False', os_secure_boot='required',
>> owner_specified.openstack.md5='',
>> owner_specified.openstack.object='images/WindowsVM',
>> owner_specified.openstack.sha256='' |
>> root at dmzcloud:/home/cloud#
>>
>>
>> Then  I logged into horizon dashboard,  from the images  selected the
>> imported image and try to launch the instance.  With  a Flavour of 550 GB
>> disk, 4 vcpus and 8GB Ram ..
>>
>> The instance spawning ran for 30 minutes and throws the error which I
>> pasted first in the right top corner of horizon dashboard.
>>
>> How to solve this error and boot the Windows machine successfully ..
>>
>>
>> """
>> Error : failed to perform requested operation on instance "WindowsVM "the
>> instance has error status. Please try again later [Error exceeded maximum
>> number of retries. Exhausted all hosts available for retrying build
>> failures for instance e3d5c095-7d26-4b1e-89d1-d1a6e20a45041
>>
>>
>> """
>>
>> Any help highly appreciated.
>>
>> Kris
>>
>> --
> Mohammed Naser
> VEXXHOST, Inc.
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.openstack.org/pipermail/openstack-discuss/attachments/20210822/2537dd92/attachment-0001.html>


More information about the openstack-discuss mailing list