[Openstack] multipath - EMC vs NetApp

Takashi Natsume natsume.takashi at lab.ntt.co.jp
Tue Nov 26 08:51:56 UTC 2013


Hello Xing,

Thank you for your reply.

You wrote:
> What OpenStack release are you running?  iSCSI multipath support was added
very late in Grizzly release and there were patches regarding multipath
added even after the Grizzly release.  EMC VNX5300 should work with iSCSI
multipath with Grizzly and beyond.

We tested with 'master' code (nova:1dca95a74788667e52cab664c8a1dd942222d9c8
and cinder:9b599d092ffa168a73ab7fa98ff20cb2cb48fe0b) last month, 
but iSCSI multipath could not be configured.

Our configuration was as follows.
(IP addresses, passwords, prefixes are masked.)

----------------------------------------------------------------
- /etc/nova/nova.conf
volume_api_class = nova.volume.cinder.API
libvirt_iscsi_use_multipath = True

- /etc/cinder/cinder.conf
use_multipath_for_image_xfer = true
iscsi_target_prefix = iqn.1992-04.com.emc:cx.fcn00XXXXXXXX
iscsi_ip_address = XX.XX.XX.11
volume_driver = cinder.volume.drivers.emc.emc_smis_iscsi.EMCSMISISCSIDriver
cinder_emc_config_file = /etc/cinder/cinder_emc_config.xml

- /etc/cinder/cinder_emc_config.xml

<?xml version='1.0' encoding='UTF-8'?>
<EMC>
<StorageType>Pool0</StorageType>
<EcomServerIp>XX.XX.XX.XX</EcomServerIp>
<EcomServerPort>5988</EcomServerPort>
<EcomUserName>admin</EcomUserName>
<EcomPassword>xxxxxxxx</EcomPassword>
</EMC>

- cinder-volume host
A python-pywbem package(Version 0.7.0-4) was installed on the cinder-volume
host.

- SMI-S Provider
EMC SMI-S Provider V4.6.1.1 was set up on RHEL 6.3.
and configured by following the SMI-S Provider Release notes.

- VNX5300
Thin Provisioning enabler was installed.
Thin pool 'Pool0' was created.

And "Register with VNX" by Unisphere
(See
http://docs.openstack.org/grizzly/openstack-block-storage/admin/content/emc-
smis-iscsi-driver.html)
----------------------------------------------------------------

When copy_image_to_volume was executed, the following log was outputed.
(Target IQN and portal IP addresses are masked.)

---
2013-10-31 11:27:43.322 26080 DEBUG cinder.brick.initiator.connector
[req-18a3433e-561b-4f04-b3be-b5d5956780f9 59b7221e2f8041e39f643520b78ba533
964613d2565a4a069da9658e44b544ae] iscsiadm ('--login',): stdout=Logging in
to [iface: default, target: iqn.1992-04.com.emc:cx.fcn00XXXXXXXX.a6, portal:
XX.XX.XX.13,3260]
Login to [iface: default, target: iqn.1992-04.com.emc:cx.fcn00XXXXXXXX.a6,
portal: XX.XX.XX.13,3260]: successful
2013-10-31 11:27:45.680 26080 DEBUG cinder.brick.initiator.connector
[req-18a3433e-561b-4f04-b3be-b5d5956780f9 59b7221e2f8041e39f643520b78ba533
964613d2565a4a069da9658e44b544ae] iscsiadm ('--login',): stdout=Logging in
to [iface: default, target: iqn.1992-04.com.emc:cx.fcn00XXXXXXXX.a6, portal:
XX.XX.XX.17,3260]
 stderr=iscsiadm: Could not login to [iface: default, target:
iqn.1992-04.com.emc:cx.fcn00XXXXXXXX.a6, portal: XX.XX.XX.17,3260]:
2013-10-31 11:27:47.945 26080 DEBUG cinder.brick.initiator.connector
[req-18a3433e-561b-4f04-b3be-b5d5956780f9 59b7221e2f8041e39f643520b78ba533
964613d2565a4a069da9658e44b544ae] iscsiadm ('--login',): stdout=Logging in
to [iface: default, target: iqn.1992-04.com.emc:cx.fcn00XXXXXXXX.a6, portal:
XX.XX.XX.12,3260]
 stderr=iscsiadm: Could not login to [iface: default, target:
iqn.1992-04.com.emc:cx.fcn00XXXXXXXX.a6, portal: XX.XX.XX.12,3260]:
2013-10-31 11:27:50.144 26080 DEBUG cinder.brick.initiator.connector
[req-18a3433e-561b-4f04-b3be-b5d5956780f9 59b7221e2f8041e39f643520b78ba533
964613d2565a4a069da9658e44b544ae] iscsiadm ('--login',): stdout=Logging in
to [iface: default, target: iqn.1992-04.com.emc:cx.fcn00XXXXXXXX.a6, portal:
XX.XX.XX.15,3260]
 stderr=iscsiadm: Could not login to [iface: default, target:
iqn.1992-04.com.emc:cx.fcn00XXXXXXXX.a6, portal: XX.XX.XX.15,3260]:
2013-10-31 11:27:52.374 26080 DEBUG cinder.brick.initiator.connector
[req-18a3433e-561b-4f04-b3be-b5d5956780f9 59b7221e2f8041e39f643520b78ba533
964613d2565a4a069da9658e44b544ae] iscsiadm ('--login',): stdout=Logging in
to [iface: default, target: iqn.1992-04.com.emc:cx.fcn00XXXXXXXX.a6, portal:
XX.XX.XX.11,3260]
 stderr=iscsiadm: Could not login to [iface: default, target:
iqn.1992-04.com.emc:cx.fcn00XXXXXXXX.a6, portal: XX.XX.XX.11,3260]:
2013-10-31 11:27:54.621 26080 DEBUG cinder.brick.initiator.connector
[req-18a3433e-561b-4f04-b3be-b5d5956780f9 59b7221e2f8041e39f643520b78ba533
964613d2565a4a069da9658e44b544ae] iscsiadm ('--login',): stdout=Logging in
to [iface: default, target: iqn.1992-04.com.emc:cx.fcn00XXXXXXXX.a6, portal:
XX.XX.XX.16,3260]
 stderr=iscsiadm: Could not login to [iface: default, target:
iqn.1992-04.com.emc:cx.fcn00XXXXXXXX.a6, portal: XX.XX.XX.16,3260]:
2013-10-31 11:27:56.914 26080 DEBUG cinder.brick.initiator.connector
[req-18a3433e-561b-4f04-b3be-b5d5956780f9 59b7221e2f8041e39f643520b78ba533
964613d2565a4a069da9658e44b544ae] iscsiadm ('--login',): stdout=Logging in
to [iface: default, target: iqn.1992-04.com.emc:cx.fcn00XXXXXXXX.a6, portal:
XX.XX.XX.18,3260]
 stderr=iscsiadm: Could not login to [iface: default, target:
iqn.1992-04.com.emc:cx.fcn00XXXXXXXX.a6, portal: XX.XX.XX.18,3260]:
2013-10-31 11:27:59.174 26080 DEBUG cinder.brick.initiator.connector
[req-18a3433e-561b-4f04-b3be-b5d5956780f9 59b7221e2f8041e39f643520b78ba533
964613d2565a4a069da9658e44b544ae] iscsiadm ('--login',): stdout=Logging in
to [iface: default, target: iqn.1992-04.com.emc:cx.fcn00XXXXXXXX.a6, portal:
XX.XX.XX.14,3260]
 stderr=iscsiadm: Could not login to [iface: default, target:
iqn.1992-04.com.emc:cx.fcn00XXXXXXXX.a6, portal: XX.XX.XX.14,3260]:
---

In cinder.brick.initiator.connector, it is using a single iSCSI target(IQN)
for login to multiple (different) portal IP addresses.
VNX5300 has multiple iSCSI ports with different iSCSI target (IQN)s.
(Normally it should use a different iSCSI target(IQN) for each IP address.)

If our settings was wrong, could you please let us know what are the correct
settings?

Regards,
Takashi Natsume
NTT Software Innovation Center
Tel: +81-422-59-4399
E-mail: natsume.takashi at lab.ntt.co.jp

> -----Original Message-----
> From: yang, xing [mailto:xing.yang at emc.com]
> Sent: Tuesday, November 26, 2013 12:12 AM
> To: Takashi Natsume; openstack at lists.openstack.org
> Subject: RE: [Openstack] multipath - EMC vs NetApp
> 
> Hi Takashi,
> 
> What OpenStack release are you running?  iSCSI multipath support was added
> very late in Grizzly release and there were patches regarding multipath
> added even after the Grizzly release.  EMC VNX5300 should work with iSCSI
> multipath with Grizzly and beyond.
> 
> Thanks,
> Xing
> 
> 
> 
> -----Original Message-----
> From: Takashi Natsume [mailto:natsume.takashi at lab.ntt.co.jp]
> Sent: Monday, November 25, 2013 6:07 AM
> To: yang, xing; openstack at lists.openstack.org
> Subject: RE: [Openstack] multipath - EMC vs NetApp
> 
> Hello Xing and all,
> 
> We tried to set up iSCSI multipath with EMC VNX5300.
> However, we had some problems with setting the iSCSI multipath with EMC
> VNX5300.
> So, we analyzed the logs and source codes, and found out that nova-compute
> (cinder-volume also same) is using a single iSCSI target(IQN) for login
> to multiple (different) portal IP addresses.
> VNX5300 has multiple iSCSI ports with different iSCSI target (IQN)s.
> (Normally it should use a different iSCSI target(IQN) for each IP
address.)
> 
> So, we realized that, iSCSI multipath cannot be configured with VNX5300.
> But we would like to use iSCSI multipath with EMC VNX5300.
> 
> Regards,
> Takashi Natsume
> NTT Software Innovation Center
> Tel: +81-422-59-4399
> E-mail: natsume.takashi at lab.ntt.co.jp
> 
> From: yang, xing [mailto:xing.yang at emc.com]
> Sent: Friday, November 22, 2013 7:25 AM
> To: Carlos Alvarez; openstack at lists.openstack.org
> Subject: Re: [Openstack] multipath - EMC vs NetApp
> 
> Hi Carlos,
> 
> We are working on this issue and will keep you informed.
> 
> Thanks,
> Xing
> 
> 
> From: Carlos Alvarez [mailto:cbalvarez at gmail.com]
> Sent: Monday, November 18, 2013 1:19 PM
> To: openstack at lists.openstack.org
> Subject: [Openstack] multipath - EMC vs NetApp
> 
> Hi All.
> 
> I recently added a EMC V-Max storage system, and realized the multipath
> is not working. The device is called /dev/mapper/XXXX but when I see the
> multipath -ll output, I see just one path. It is working fine with a
NetApp
> 3250.
> 
> Looking into differences, I see the output of the iscsiadm discovery
> differs:
> 
> Netapp output:
> root at grizzly-dev04:~# iscsiadm -m discovery -t sendtargets -p 10.33.5.10
> 10.33.5.10:3260,1030
> iqn.1992-08.com.netapp:sn.0e3c22d9f2ea11e2a2f2123478563412:vs.10
> 10.33.5.11:3260,1031
> iqn.1992-08.com.netapp:sn.0e3c22d9f2ea11e2a2f2123478563412:vs.10
> 
> Emc output:
> root at grizzly-dev04:~# iscsiadm -m discovery -t sendtargets -p 10.33.5.25
> 10.33.5.25:3260,1 iqn.1992-04.com.emc:50000973f00bcd44
> 
> 
> Looking into the code, the former is clearly what the connect_volume
method
> is expecting: a single ip which returns both path. I reported it to EMC
> and the answer is that it works with netapp because the 3250 has a feature
> EMC lacks (called multipath groups).
> 
> Is anybody using multipath with a V-Max?. It should work or EMC is not
> supported?
> 
> Thanks!
> Carlos
> 
> 







More information about the Openstack mailing list