issues creating a second vm with numa affinity

Sean Mooney smooney at redhat.com
Tue Oct 1 23:24:58 UTC 2019


On Tue, 2019-10-01 at 14:39 -0400, Satish Patel wrote:
> did you try to removing "hw:numa_nodes=1"  ?
that will have no effect
the vm implcitly has a numa toplogy of 1 node due to usei cpu pinning.
so hw:numa_nodes=1 is identical to what will be setting hw:cpu_policy=dedicated

openstack flavor create --public xlarge.numa.perf.test --ram 200000 --disk 700 --vcpus 20 --property
hw:cpu_policy=dedicated --property hw:emulator_threads_policy=isolate --property hw:numa_nodes='1' --property
pci_passthrough:alias='nvme:4'

looking at the  numa info that was provided.

node 0 cpus: 0 1 2 3 4 5 6 7 8 9 10 11 12 13 28 29 30 31 32 33 34 35 36 37 38 39 40 41
> > 
> > node 0 size: 262029 MB
> > 
> > node 0 free: 52787 MB
> > 
> > node 1 cpus: 14 15 16 17 18 19 20 21 22 23 24 25 26 27 42 43 44 45 46 47 48 49 50 51 52 53 54 55
> > 

it looks like you have a dual socket host with 14 cores per secket and hyper threading enabled.

looking at the flaovr
hw:cpu_policy=dedicated --property hw:emulator_threads_policy=isolate
enables pinning and allocate 1 addtional pinned for the emulator thread.

since hw:cpu_treads_policy is not defien the behavior will be determined by the numa of cores requrested.

by default if the flavor.vcpu is even it would default to the require policy and try to use hyper tread siblibngs
if flavor.vcpu was odd it would defualt to isolate policy and try to isolate individual cores. this was originally
done to prevent a class of timing based attach that can be executed if two vms were pinned differnet hyperthread on the
same core. i say be defualt as you are also useing hw:emulator_threads_policy=isolate which actully means you are
askign for 21 cores and im not sure of the top of my head which policy will take effect.
strictly speacking the prefer policy is used but it behavior is subtle.

anway form the numa info above reaange the data to show the tread siblings

node 0 cpus: 0  1  2  3  4  5  6  7  8  9  10 11 12 13 
             28 29 30 31 32 33 34 35 36 37 38 39 40 41

node 1 cpus: 14 15 16 17 18 19 20 21 22 23 24 25 26 27
             42 43 44 45 46 47 48 49 50 51 52 53 54 55

if you have look at this long enough you will know after kernel 4.0 enumates in a prediable way that is different
form the predicable way that older kernels used to enuamrete cores in.

if we boot 1 vm say on node0 which is the socket 0 in this case as well with the above flaovr i would expect the free
cores to look like this

node 0 cpus: -  -  -  -  -  -  -  -  -  -  -  11 12 13 
             -  -  -  -  -  -  -  -  -  -  38 39 40 41

node 1 cpus: 14 15 16 17 18 19 20 21 22 23 24 25 26 27
             42 43 44 45 46 47 48 49 50 51 52 53 54 55

looking at the pci white list there is something else that you can see

passthrough_whitelist = [ {"address":"0000:06:00.0"}, {"address":"0000:07:00.0"}, {"address":"0000:08:00.0"},
> > {"address":"0000:09:00.0"}, {"address":"0000:84:00.0"}, {"address":"0000:85:00.0"}, {"address":"0000:86:00.0"},
> > {"address":"0000:87:00.0"} ]

for all the devies the first 2 bytes are 0000 this is the pci domain.

on a multi socket sytems, or at least on any 2 socket system new enouglsht to processor wth 14 cores and hypertreading
you will have 2 different pci roots. 1 pci route complex per phyical processor. in a system with multiple pci root
complex 1 becomes the primary pci root and is assigned the 0000 domain and the second is can be asigned a different
domain adress but that depend on your kernel commandline option and the number of devices.
form my experince when only a single domain is create the second numa node device start with 0000:80:xx.y or higher
so {"address":"0000:06:00.0"}, {"address":"0000:07:00.0"}, {"address":"0000:08:00.0"},{"address":"0000:09:00.0"} shoudl
be on numa node 0 and shuld be assinged to the vm

that leave 
{"address":"0000:84:00.0"}, {"address":"0000:85:00.0"}, {"address":"0000:86:00.0"}, {"address":"0000:87:00.0"}
and 
node 1 cpus: 14 15 16 17 18 19 20 21 22 23 24 25 26 27
             42 43 44 45 46 47 48 49 50 51 52 53 54 55

so from an openstack point of view there are enough core free to pin the second vm and there are devices free on the
same numa node.

node 0 free: 52787 MB
node 1 free: 250624 MB

200G is being used on node 0 by the first vm and there is 250 is free on node 1.

as kashyap pointed out in an earlier reply the most likely cause of the 
"qemu-kvm: kvm_init_vcpu failed: Cannot allocate memory" error is a libvirt interaction with a kvm
kernel bug that was fix in kernel 4.19 (4.19 fixes a lot of kvm bugs and enabled nestexd virt by default so you shoudl
use 4.19+ if you can)
kashyap submitted https://review.opendev.org/#/c/684375/ as a possible way to workaround to the kernel issue by relaxing
the requirement nova places on the memory assgined to a guest that is not used for guest ram.

effectivly we belive that the root case is on the host if you run "grep DMA32 /proc/zoneinfo" the DMA32 zone will only
exist on 1 nuam node.

e.g. sean at workstation:~$ grep DMA32 /proc/zoneinfo  
Node 0, zone    DMA32
Node 1, zone    DMA32

https://review.opendev.org/#/c/684375/ we belive would allow the second vm to booth with numa affined guest ram
but non numa affined DMA memroy however that could have negative performace implication in some cases.
nova connot contol how where DMA memroy is allocated by the kernel so this cannot be fully adress  by nova.

ideally the best way to fix this would be to some how force your kenel to allocate DMA32 zones per numa node
but i am not aware of a way to do that.

so to ansewr the orginal question 
'What "emu-kvm: kvm_init_vcpu failed: Cannot allocate memory" means in this context?'
my understanding is that it mean
qemu could not allcate memory form a DMA32 zone on the
same numa node as the cpus and guest ram for the PCI passthough devices which would be required
when  <memory mode="strict" ...> is defiend. we always require strict mode when we have a vm with a numa
toplogy to ensure that the guest memroy is allocated form the node we requested but if you are using pci
passtouhg and do not have DMA32 zones. it is my understanding that on newewr kernels the kvm modules
allows non local DMA zones to be used. with all that said it is very uncommon to have hardware that
dose not have a DMA and DMA32 zone per numa node so most peopel will never have this problem. 


> On Tue, Oct 1, 2019 at 2:16 PM Manuel Sopena Ballesteros
> <manuel.sb at garvan.org.au> wrote:
> > 
> > Dear Openstack user community,
> > 
> > 
> > 
> > I have a compute node with 2 numa nodes and I would like to create 2 vms, each one using a different numa node
> > through numa affinity with cpu, memory and nvme pci devices.
> > 
> > 
> > 
> > pci passthrough whitelist
> > 
> > [root at zeus-53 ~]# tail /etc/kolla/nova-compute/nova.conf
> > 
> > [notifications]
> > 
> > 
> > 
> > [filter_scheduler]
> > 
> > enabled_filters = enabled_filters = RetryFilter, AvailabilityZoneFilter, ComputeFilter, ComputeCapabilitiesFilter,
> > ImagePropertiesFilter, ServerGroupAntiAffinityFilter, ServerGroupAffinityFilter, PciPassthroughFilter
> > 
> > available_filters = nova.scheduler.filters.all_filters
> > 
> > 
> > 
> > [pci]
> > 
> > passthrough_whitelist = [ {"address":"0000:06:00.0"}, {"address":"0000:07:00.0"}, {"address":"0000:08:00.0"},
> > {"address":"0000:09:00.0"}, {"address":"0000:84:00.0"}, {"address":"0000:85:00.0"}, {"address":"0000:86:00.0"},
> > {"address":"0000:87:00.0"} ]
> > 
> > alias = { "vendor_id":"8086", "product_id":"0953", "device_type":"type-PCI", "name":"nvme"}
> > 
> > 
> > 
> > Openstack flavor
> > 
> > openstack flavor create --public xlarge.numa.perf.test --ram 200000 --disk 700 --vcpus 20 --property
> > hw:cpu_policy=dedicated --property hw:emulator_threads_policy=isolate --property hw:numa_nodes='1' --property
> > pci_passthrough:alias='nvme:4'
> > 
> > 
> > 
> > The first vm is successfully created
> > 
> > openstack server create --network hpc --flavor xlarge.numa.perf.test --image centos7.6-image --availability-zone
> > nova:zeus-53.localdomain --key-name mykey kudu-1
> > 
> > 
> > 
> > However the second vm fails
> > 
> > openstack server create --network hpc --flavor xlarge.numa.perf --image centos7.6-kudu-image --availability-zone
> > nova:zeus-53.localdomain --key-name mykey kudu-4
> > 
> > 
> > 
> > Errors in nova compute node
> > 
> > 2019-09-27 16:45:19.785 7 ERROR nova.compute.manager [req-b5a25c73-8c7d-466c-8128-71f29e7ae8aa
> > 91e83343e9834c8ba0172ff369c8acac b91520cff5bd45c59a8de07c38641582 - default default] [instance: ebe4e78c-501e-4535-
> > ae15-948301cbf1ae] Instance failed to spawn: libvirtError: internal error: qemu unexpectedly closed the monitor:
> > 2019-09-27T06:45:19.118089Z qemu-kvm: kvm_init_vcpu failed: Cannot allocate memory
> > 
> > 2019-09-27 16:45:19.785 7 ERROR nova.compute.manager [instance: ebe4e78c-501e-4535-ae15-948301cbf1ae] Traceback
> > (most recent call last):
> > 
> > 2019-09-27 16:45:19.785 7 ERROR nova.compute.manager [instance: ebe4e78c-501e-4535-ae15-948301cbf1ae]   File
> > "/usr/lib/python2.7/site-packages/nova/compute/manager.py", line 2369, in _build_resources
> > 
> > 2019-09-27 16:45:19.785 7 ERROR nova.compute.manager [instance: ebe4e78c-501e-4535-ae15-948301cbf1ae]     yield
> > resources
> > 
> > 2019-09-27 16:45:19.785 7 ERROR nova.compute.manager [instance: ebe4e78c-501e-4535-ae15-948301cbf1ae]   File
> > "/usr/lib/python2.7/site-packages/nova/compute/manager.py", line 2133, in _build_and_run_instance
> > 
> > 2019-09-27 16:45:19.785 7 ERROR nova.compute.manager [instance: ebe4e78c-501e-4535-ae15-
> > 948301cbf1ae]     block_device_info=block_device_info)
> > 
> > 2019-09-27 16:45:19.785 7 ERROR nova.compute.manager [instance: ebe4e78c-501e-4535-ae15-948301cbf1ae]   File
> > "/usr/lib/python2.7/site-packages/nova/virt/libvirt/driver.py", line 3142, in spawn
> > 
> > 2019-09-27 16:45:19.785 7 ERROR nova.compute.manager [instance: ebe4e78c-501e-4535-ae15-
> > 948301cbf1ae]     destroy_disks_on_failure=True)
> > 
> > 2019-09-27 16:45:19.785 7 ERROR nova.compute.manager [instance: ebe4e78c-501e-4535-ae15-948301cbf1ae]   File
> > "/usr/lib/python2.7/site-packages/nova/virt/libvirt/driver.py", line 5705, in _create_domain_and_network
> > 
> > 2019-09-27 16:45:19.785 7 ERROR nova.compute.manager [instance: ebe4e78c-501e-4535-ae15-
> > 948301cbf1ae]     destroy_disks_on_failure)
> > 
> > 2019-09-27 16:45:19.785 7 ERROR nova.compute.manager [instance: ebe4e78c-501e-4535-ae15-948301cbf1ae]   File
> > "/usr/lib/python2.7/site-packages/oslo_utils/excutils.py", line 220, in __exit__
> > 
> > 2019-09-27 16:45:19.785 7 ERROR nova.compute.manager [instance: ebe4e78c-501e-4535-ae15-
> > 948301cbf1ae]     self.force_reraise()
> > 
> > 2019-09-27 16:45:19.785 7 ERROR nova.compute.manager [instance: ebe4e78c-501e-4535-ae15-948301cbf1ae]   File
> > "/usr/lib/python2.7/site-packages/oslo_utils/excutils.py", line 196, in force_reraise
> > 
> > 2019-09-27 16:45:19.785 7 ERROR nova.compute.manager [instance: ebe4e78c-501e-4535-ae15-
> > 948301cbf1ae]     six.reraise(self.type_, self.value, self.tb)
> > 
> > 2019-09-27 16:45:19.785 7 ERROR nova.compute.manager [instance: ebe4e78c-501e-4535-ae15-948301cbf1ae]   File
> > "/usr/lib/python2.7/site-packages/nova/virt/libvirt/driver.py", line 5674, in _create_domain_and_network
> > 
> > 2019-09-27 16:45:19.785 7 ERROR nova.compute.manager [instance: ebe4e78c-501e-4535-ae15-
> > 948301cbf1ae]     post_xml_callback=post_xml_callback)
> > 
> > 2019-09-27 16:45:19.785 7 ERROR nova.compute.manager [instance: ebe4e78c-501e-4535-ae15-948301cbf1ae]   File
> > "/usr/lib/python2.7/site-packages/nova/virt/libvirt/driver.py", line 5608, in _create_domain
> > 
> > 2019-09-27 16:45:19.785 7 ERROR nova.compute.manager [instance: ebe4e78c-501e-4535-ae15-
> > 948301cbf1ae]     guest.launch(pause=pause)
> > 
> > 2019-09-27 16:45:19.785 7 ERROR nova.compute.manager [instance: ebe4e78c-501e-4535-ae15-948301cbf1ae]   File
> > "/usr/lib/python2.7/site-packages/nova/virt/libvirt/guest.py", line 144, in launch
> > 
> > 2019-09-27 16:45:19.785 7 ERROR nova.compute.manager [instance: ebe4e78c-501e-4535-ae15-
> > 948301cbf1ae]     self._encoded_xml, errors='ignore')
> > 
> > 2019-09-27 16:45:19.785 7 ERROR nova.compute.manager [instance: ebe4e78c-501e-4535-ae15-948301cbf1ae]   File
> > "/usr/lib/python2.7/site-packages/oslo_utils/excutils.py", line 220, in __exit__
> > 
> > 2019-09-27 16:45:19.785 7 ERROR nova.compute.manager [instance: ebe4e78c-501e-4535-ae15-
> > 948301cbf1ae]     self.force_reraise()
> > 
> > 2019-09-27 16:45:19.785 7 ERROR nova.compute.manager [instance: ebe4e78c-501e-4535-ae15-948301cbf1ae]   File
> > "/usr/lib/python2.7/site-packages/oslo_utils/excutils.py", line 196, in force_reraise
> > 
> > 2019-09-27 16:45:19.785 7 ERROR nova.compute.manager [instance: ebe4e78c-501e-4535-ae15-
> > 948301cbf1ae]     six.reraise(self.type_, self.value, self.tb)
> > 
> > 2019-09-27 16:45:19.785 7 ERROR nova.compute.manager [instance: ebe4e78c-501e-4535-ae15-948301cbf1ae]   File
> > "/usr/lib/python2.7/site-packages/nova/virt/libvirt/guest.py", line 139, in launch
> > 
> > 2019-09-27 16:45:19.785 7 ERROR nova.compute.manager [instance: ebe4e78c-501e-4535-ae15-948301cbf1ae]     return
> > self._domain.createWithFlags(flags)
> > 
> > 2019-09-27 16:45:19.785 7 ERROR nova.compute.manager [instance: ebe4e78c-501e-4535-ae15-948301cbf1ae]   File
> > "/usr/lib/python2.7/site-packages/eventlet/tpool.py", line 186, in doit
> > 
> > 2019-09-27 16:45:19.785 7 ERROR nova.compute.manager [instance: ebe4e78c-501e-4535-ae15-948301cbf1ae]     result =
> > proxy_call(self._autowrap, f, *args, **kwargs)
> > 
> > 2019-09-27 16:45:19.785 7 ERROR nova.compute.manager [instance: ebe4e78c-501e-4535-ae15-948301cbf1ae]   File
> > "/usr/lib/python2.7/site-packages/eventlet/tpool.py", line 144, in proxy_call
> > 
> > 2019-09-27 16:45:19.785 7 ERROR nova.compute.manager [instance: ebe4e78c-501e-4535-ae15-948301cbf1ae]     rv =
> > execute(f, *args, **kwargs)
> > 
> > 2019-09-27 16:45:19.785 7 ERROR nova.compute.manager [instance: ebe4e78c-501e-4535-ae15-948301cbf1ae]   File
> > "/usr/lib/python2.7/site-packages/eventlet/tpool.py", line 125, in execute
> > 
> > 2019-09-27 16:45:19.785 7 ERROR nova.compute.manager [instance: ebe4e78c-501e-4535-ae15-
> > 948301cbf1ae]     six.reraise(c, e, tb)
> > 
> > 2019-09-27 16:45:19.785 7 ERROR nova.compute.manager [instance: ebe4e78c-501e-4535-ae15-948301cbf1ae]   File
> > "/usr/lib/python2.7/site-packages/eventlet/tpool.py", line 83, in tworker
> > 
> > 2019-09-27 16:45:19.785 7 ERROR nova.compute.manager [instance: ebe4e78c-501e-4535-ae15-948301cbf1ae]     rv =
> > meth(*args, **kwargs)
> > 
> > 2019-09-27 16:45:19.785 7 ERROR nova.compute.manager [instance: ebe4e78c-501e-4535-ae15-948301cbf1ae]   File
> > "/usr/lib64/python2.7/site-packages/libvirt.py", line 1110, in createWithFlags
> > 
> > 2019-09-27 16:45:19.785 7 ERROR nova.compute.manager [instance: ebe4e78c-501e-4535-ae15-948301cbf1ae]     if ret ==
> > -1: raise libvirtError ('virDomainCreateWithFlags() failed', dom=self)
> > 
> > 2019-09-27 16:45:19.785 7 ERROR nova.compute.manager [instance: ebe4e78c-501e-4535-ae15-948301cbf1ae] libvirtError:
> > internal error: qemu unexpectedly closed the monitor: 2019-09-27T06:45:19.118089Z qemu-kvm: kvm_init_vcpu failed:
> > Cannot allocate memory
> > 
> > 2019-09-27 16:45:19.785 7 ERROR nova.compute.manager [instance: ebe4e78c-501e-4535-ae15-948301cbf1ae]
> > 
> > 
> > 
> > Numa cell/node 1 (the one assigned on kudu-4) has enough cpu, memory, pci devices and disk capacity to fit this vm.
> > NOTE: below is the information relevant I could think of that shows resources available after creating the second
> > vm.
> > 
> > 
> > 
> > [root at zeus-53 ~]# numactl -H
> > 
> > available: 2 nodes (0-1)
> > 
> > node 0 cpus: 0 1 2 3 4 5 6 7 8 9 10 11 12 13 28 29 30 31 32 33 34 35 36 37 38 39 40 41
> > 
> > node 0 size: 262029 MB
> > 
> > node 0 free: 52787 MB
> > 
> > node 1 cpus: 14 15 16 17 18 19 20 21 22 23 24 25 26 27 42 43 44 45 46 47 48 49 50 51 52 53 54 55
> > 
> > node 1 size: 262144 MB
> > 
> > node 1 free: 250624 MB
> > 
> > node distances:
> > 
> > node   0   1
> > 
> >   0:  10  21
> > 
> >   1:  21  10
> > 
> > NOTE: this is to show that numa node/cell 1 has enough resources available (also nova-compute logs shows that kudu-4 
> > is assigned to cell 1)
> > 
> > 
> > 
> > [root at zeus-53 ~]# df -h
> > 
> > Filesystem      Size  Used Avail Use% Mounted on
> > 
> > /dev/md127      3.7T  9.1G  3.7T   1% /
> > 
> > ...
> > 
> > NOTE: vm disk files goes to root (/) partition
> > 
> > 
> > 
> > [root at zeus-53 ~]# lsblk
> > 
> > NAME                                                                                         MAJ:MIN RM  SIZE RO
> > TYPE  MOUNTPOINT
> > 
> > sda                                                                                            8:0    0 59.6G  0
> > disk
> > 
> > ├─sda1                                                                                         8:1    0    1G  0
> > part  /boot
> > 
> > └─sda2                                                                                         8:2    0   16G  0
> > part  [SWAP]
> > 
> > loop0                                                                                          7:0    0  100G  0
> > loop
> > 
> > └─docker-9:127-6979358884-pool                                                               253:0    0  100G  0 dm
> > 
> >   ├─docker-9:127-6979358884-4301cee8d0433729cd6332ca2b6111afc85f14c48d4ce2d888a1da0ef9b5ca01 253:1    0   10G  0 dm
> > 
> >   ├─docker-9:127-6979358884-d59208adcb7cee3418f810f24e6c3a55d39281f713c8e76141fc61a8deba8a2b 253:2    0   10G  0 dm
> > 
> >   ├─docker-9:127-6979358884-106bc0838e37442eca84eb9ab17aa7a45308b7e3a38be3fb21a4fa00366fe306 253:3    0   10G  0 dm
> > 
> >   ├─docker-9:127-6979358884-7e16b5d012ab8744739b671fcdc8e47db5cc64e6c3d5a5fe423bfd68cfb07b20 253:4    0   10G  0 dm
> > 
> >   ├─docker-9:127-6979358884-f1c2545b4edbfd7b42d2a492eda8224fcf7cefc3e3a41e65d307c585acffe6a8 253:5    0   10G  0 dm
> > 
> >   ├─docker-9:127-6979358884-e7fd6c7b3f624f387bdb3746a7944a30c92d8ee5395e75c76288b281bd009d90 253:6    0   10G  0 dm
> > 
> >   ├─docker-9:127-6979358884-95a818cc7afd9867385bb9a9ea750d4cc6e162916c6ae3a157097af74578e1e4 253:7    0   10G  0 dm
> > 
> >   ├─docker-9:127-6979358884-9a7f28d396c149119556f382bf5c19f5925eed5d18b94407649244c7adabb4b3 253:8    0   10G  0 dm
> > 
> >   ├─docker-9:127-6979358884-b25941b6f115300caea977911e2d7fd3541ef187c9aa5736fe10fad638ecd0d1 253:9    0   10G  0 dm
> > 
> >   ├─docker-9:127-6979358884-122b201c6ad24896a205f8db4a64759ba8fbd5bbe245d0f98984268a01e6a0c4 253:10   0   10G  0 dm
> > 
> >   └─docker-9:127-6979358884-bc04120ba59a1b393f338a1cef64b16d920cf4e73400198e4b999bb72a42ff90 253:11   0   10G  0 dm
> > 
> > loop1                                                                                          7:1    0    2G  0
> > loop
> > 
> > └─docker-9:127-6979358884-pool                                                               253:0    0  100G  0 dm
> > 
> >   ├─docker-9:127-6979358884-4301cee8d0433729cd6332ca2b6111afc85f14c48d4ce2d888a1da0ef9b5ca01 253:1    0   10G  0 dm
> > 
> >   ├─docker-9:127-6979358884-d59208adcb7cee3418f810f24e6c3a55d39281f713c8e76141fc61a8deba8a2b 253:2    0   10G  0 dm
> > 
> >   ├─docker-9:127-6979358884-106bc0838e37442eca84eb9ab17aa7a45308b7e3a38be3fb21a4fa00366fe306 253:3    0   10G  0 dm
> > 
> >   ├─docker-9:127-6979358884-7e16b5d012ab8744739b671fcdc8e47db5cc64e6c3d5a5fe423bfd68cfb07b20 253:4    0   10G  0 dm
> > 
> >   ├─docker-9:127-6979358884-f1c2545b4edbfd7b42d2a492eda8224fcf7cefc3e3a41e65d307c585acffe6a8 253:5    0   10G  0 dm
> > 
> >   ├─docker-9:127-6979358884-e7fd6c7b3f624f387bdb3746a7944a30c92d8ee5395e75c76288b281bd009d90 253:6    0   10G  0 dm
> > 
> >   ├─docker-9:127-6979358884-95a818cc7afd9867385bb9a9ea750d4cc6e162916c6ae3a157097af74578e1e4 253:7    0   10G  0 dm
> > 
> >   ├─docker-9:127-6979358884-9a7f28d396c149119556f382bf5c19f5925eed5d18b94407649244c7adabb4b3 253:8    0   10G  0 dm
> > 
> >   ├─docker-9:127-6979358884-b25941b6f115300caea977911e2d7fd3541ef187c9aa5736fe10fad638ecd0d1 253:9    0   10G  0 dm
> > 
> >   ├─docker-9:127-6979358884-122b201c6ad24896a205f8db4a64759ba8fbd5bbe245d0f98984268a01e6a0c4 253:10   0   10G  0 dm
> > 
> >   └─docker-9:127-6979358884-bc04120ba59a1b393f338a1cef64b16d920cf4e73400198e4b999bb72a42ff90 253:11   0   10G  0 dm
> > 
> > nvme0n1                                                                                      259:8    0  1.8T  0
> > disk
> > 
> > └─nvme0n1p1                                                                                  259:9    0  1.8T  0
> > part
> > 
> >   └─md127                                                                                      9:127  0  3.7T  0
> > raid0 /
> > 
> > nvme1n1                                                                                      259:6    0  1.8T  0
> > disk
> > 
> > └─nvme1n1p1                                                                                  259:7    0  1.8T  0
> > part
> > 
> >   └─md127                                                                                      9:127  0  3.7T  0
> > raid0 /
> > 
> > nvme2n1                                                                                      259:2    0  1.8T  0
> > disk
> > 
> > nvme3n1                                                                                      259:1    0  1.8T  0
> > disk
> > 
> > nvme4n1                                                                                      259:0    0  1.8T  0
> > disk
> > 
> > nvme5n1                                                                                      259:3    0  1.8T  0
> > disk
> > 
> > NOTE: this is to show that there are 4 nvme disks (nvme2n1, nvme3n1, nvme4n1, nvme5n1) available for the second vm
> > 
> > 
> > 
> > What "emu-kvm: kvm_init_vcpu failed: Cannot allocate memory" means in this context?
> > 
> > 
> > 
> > Thank you very much
> > 
> > NOTICE
> > Please consider the environment before printing this email. This message and any attachments are intended for the
> > addressee named and may contain legally privileged/confidential/copyright information. If you are not the intended
> > recipient, you should not read, use, disclose, copy or distribute this communication. If you have received this
> > message in error please notify us at once by return email and then delete both messages. We accept no liability for
> > the distribution of viruses or similar in electronic communications. This notice should not be removed.
> 
> 




More information about the openstack-discuss mailing list