Hi Takashi,

In the logs, you got a login failure.  I'm wondering if the iscsi_ip_address in 
cinder.conf has the correct value.  It should be the iSCSI IP address from SPA 
or SPB.  

The code in cinder.brick.initiator.connector should work for VNX.  It uses one 
iSCSI target IP address to do iscsiadm discovery and that should return 
multiple IP addresses.  If not, we need to look at your configuration.

Let me contact the account manager for NTT to set up a conference call with you 
to get this resolved quickly.

Thanks,
Xing



-----Original Message-----
From: Takashi Natsume [mailto:natsume.taka...@lab.ntt.co.jp] 
Sent: Tuesday, November 26, 2013 3:52 AM
To: yang, xing; openstack@lists.openstack.org
Subject: RE: [Openstack] multipath - EMC vs NetApp

Hello Xing,

Thank you for your reply.

You wrote:
> What OpenStack release are you running?  iSCSI multipath support was 
> added
very late in Grizzly release and there were patches regarding multipath added 
even after the Grizzly release.  EMC VNX5300 should work with iSCSI multipath 
with Grizzly and beyond.

We tested with 'master' code (nova:1dca95a74788667e52cab664c8a1dd942222d9c8
and cinder:9b599d092ffa168a73ab7fa98ff20cb2cb48fe0b) last month, but iSCSI 
multipath could not be configured.

Our configuration was as follows.
(IP addresses, passwords, prefixes are masked.)

----------------------------------------------------------------
- /etc/nova/nova.conf
volume_api_class = nova.volume.cinder.API libvirt_iscsi_use_multipath = True

- /etc/cinder/cinder.conf
use_multipath_for_image_xfer = true
iscsi_target_prefix = iqn.1992-04.com.emc:cx.fcn00XXXXXXXX
iscsi_ip_address = XX.XX.XX.11
volume_driver = cinder.volume.drivers.emc.emc_smis_iscsi.EMCSMISISCSIDriver
cinder_emc_config_file = /etc/cinder/cinder_emc_config.xml

- /etc/cinder/cinder_emc_config.xml

<?xml version='1.0' encoding='UTF-8'?>
<EMC>
<StorageType>Pool0</StorageType>
<EcomServerIp>XX.XX.XX.XX</EcomServerIp>
<EcomServerPort>5988</EcomServerPort>
<EcomUserName>admin</EcomUserName>
<EcomPassword>xxxxxxxx</EcomPassword>
</EMC>

- cinder-volume host
A python-pywbem package(Version 0.7.0-4) was installed on the cinder-volume 
host.

- SMI-S Provider
EMC SMI-S Provider V4.6.1.1 was set up on RHEL 6.3.
and configured by following the SMI-S Provider Release notes.

- VNX5300
Thin Provisioning enabler was installed.
Thin pool 'Pool0' was created.

And "Register with VNX" by Unisphere
(See
http://docs.openstack.org/grizzly/openstack-block-storage/admin/content/emc-
smis-iscsi-driver.html)
----------------------------------------------------------------

When copy_image_to_volume was executed, the following log was outputed.
(Target IQN and portal IP addresses are masked.)

---
2013-10-31 11:27:43.322 26080 DEBUG cinder.brick.initiator.connector
[req-18a3433e-561b-4f04-b3be-b5d5956780f9 59b7221e2f8041e39f643520b78ba533 
964613d2565a4a069da9658e44b544ae] iscsiadm ('--login',): stdout=Logging in to 
[iface: default, target: iqn.1992-04.com.emc:cx.fcn00XXXXXXXX.a6, portal:
XX.XX.XX.13,3260]
Login to [iface: default, target: iqn.1992-04.com.emc:cx.fcn00XXXXXXXX.a6,
portal: XX.XX.XX.13,3260]: successful
2013-10-31 11:27:45.680 26080 DEBUG cinder.brick.initiator.connector
[req-18a3433e-561b-4f04-b3be-b5d5956780f9 59b7221e2f8041e39f643520b78ba533 
964613d2565a4a069da9658e44b544ae] iscsiadm ('--login',): stdout=Logging in to 
[iface: default, target: iqn.1992-04.com.emc:cx.fcn00XXXXXXXX.a6, portal:
XX.XX.XX.17,3260]
 stderr=iscsiadm: Could not login to [iface: default, target:
iqn.1992-04.com.emc:cx.fcn00XXXXXXXX.a6, portal: XX.XX.XX.17,3260]:
2013-10-31 11:27:47.945 26080 DEBUG cinder.brick.initiator.connector
[req-18a3433e-561b-4f04-b3be-b5d5956780f9 59b7221e2f8041e39f643520b78ba533 
964613d2565a4a069da9658e44b544ae] iscsiadm ('--login',): stdout=Logging in to 
[iface: default, target: iqn.1992-04.com.emc:cx.fcn00XXXXXXXX.a6, portal:
XX.XX.XX.12,3260]
 stderr=iscsiadm: Could not login to [iface: default, target:
iqn.1992-04.com.emc:cx.fcn00XXXXXXXX.a6, portal: XX.XX.XX.12,3260]:
2013-10-31 11:27:50.144 26080 DEBUG cinder.brick.initiator.connector
[req-18a3433e-561b-4f04-b3be-b5d5956780f9 59b7221e2f8041e39f643520b78ba533 
964613d2565a4a069da9658e44b544ae] iscsiadm ('--login',): stdout=Logging in to 
[iface: default, target: iqn.1992-04.com.emc:cx.fcn00XXXXXXXX.a6, portal:
XX.XX.XX.15,3260]
 stderr=iscsiadm: Could not login to [iface: default, target:
iqn.1992-04.com.emc:cx.fcn00XXXXXXXX.a6, portal: XX.XX.XX.15,3260]:
2013-10-31 11:27:52.374 26080 DEBUG cinder.brick.initiator.connector
[req-18a3433e-561b-4f04-b3be-b5d5956780f9 59b7221e2f8041e39f643520b78ba533 
964613d2565a4a069da9658e44b544ae] iscsiadm ('--login',): stdout=Logging in to 
[iface: default, target: iqn.1992-04.com.emc:cx.fcn00XXXXXXXX.a6, portal:
XX.XX.XX.11,3260]
 stderr=iscsiadm: Could not login to [iface: default, target:
iqn.1992-04.com.emc:cx.fcn00XXXXXXXX.a6, portal: XX.XX.XX.11,3260]:
2013-10-31 11:27:54.621 26080 DEBUG cinder.brick.initiator.connector
[req-18a3433e-561b-4f04-b3be-b5d5956780f9 59b7221e2f8041e39f643520b78ba533 
964613d2565a4a069da9658e44b544ae] iscsiadm ('--login',): stdout=Logging in to 
[iface: default, target: iqn.1992-04.com.emc:cx.fcn00XXXXXXXX.a6, portal:
XX.XX.XX.16,3260]
 stderr=iscsiadm: Could not login to [iface: default, target:
iqn.1992-04.com.emc:cx.fcn00XXXXXXXX.a6, portal: XX.XX.XX.16,3260]:
2013-10-31 11:27:56.914 26080 DEBUG cinder.brick.initiator.connector
[req-18a3433e-561b-4f04-b3be-b5d5956780f9 59b7221e2f8041e39f643520b78ba533 
964613d2565a4a069da9658e44b544ae] iscsiadm ('--login',): stdout=Logging in to 
[iface: default, target: iqn.1992-04.com.emc:cx.fcn00XXXXXXXX.a6, portal:
XX.XX.XX.18,3260]
 stderr=iscsiadm: Could not login to [iface: default, target:
iqn.1992-04.com.emc:cx.fcn00XXXXXXXX.a6, portal: XX.XX.XX.18,3260]:
2013-10-31 11:27:59.174 26080 DEBUG cinder.brick.initiator.connector
[req-18a3433e-561b-4f04-b3be-b5d5956780f9 59b7221e2f8041e39f643520b78ba533 
964613d2565a4a069da9658e44b544ae] iscsiadm ('--login',): stdout=Logging in to 
[iface: default, target: iqn.1992-04.com.emc:cx.fcn00XXXXXXXX.a6, portal:
XX.XX.XX.14,3260]
 stderr=iscsiadm: Could not login to [iface: default, target:
iqn.1992-04.com.emc:cx.fcn00XXXXXXXX.a6, portal: XX.XX.XX.14,3260]:
---

In cinder.brick.initiator.connector, it is using a single iSCSI target(IQN) for 
login to multiple (different) portal IP addresses.
VNX5300 has multiple iSCSI ports with different iSCSI target (IQN)s.
(Normally it should use a different iSCSI target(IQN) for each IP address.)

If our settings was wrong, could you please let us know what are the correct 
settings?

Regards,
Takashi Natsume
NTT Software Innovation Center
Tel: +81-422-59-4399
E-mail: natsume.taka...@lab.ntt.co.jp

> -----Original Message-----
> From: yang, xing [mailto:xing.y...@emc.com]
> Sent: Tuesday, November 26, 2013 12:12 AM
> To: Takashi Natsume; openstack@lists.openstack.org
> Subject: RE: [Openstack] multipath - EMC vs NetApp
> 
> Hi Takashi,
> 
> What OpenStack release are you running?  iSCSI multipath support was 
> added very late in Grizzly release and there were patches regarding 
> multipath added even after the Grizzly release.  EMC VNX5300 should 
> work with iSCSI multipath with Grizzly and beyond.
> 
> Thanks,
> Xing
> 
> 
> 
> -----Original Message-----
> From: Takashi Natsume [mailto:natsume.taka...@lab.ntt.co.jp]
> Sent: Monday, November 25, 2013 6:07 AM
> To: yang, xing; openstack@lists.openstack.org
> Subject: RE: [Openstack] multipath - EMC vs NetApp
> 
> Hello Xing and all,
> 
> We tried to set up iSCSI multipath with EMC VNX5300.
> However, we had some problems with setting the iSCSI multipath with 
> EMC VNX5300.
> So, we analyzed the logs and source codes, and found out that 
> nova-compute (cinder-volume also same) is using a single iSCSI 
> target(IQN) for login to multiple (different) portal IP addresses.
> VNX5300 has multiple iSCSI ports with different iSCSI target (IQN)s.
> (Normally it should use a different iSCSI target(IQN) for each IP
address.)
> 
> So, we realized that, iSCSI multipath cannot be configured with VNX5300.
> But we would like to use iSCSI multipath with EMC VNX5300.
> 
> Regards,
> Takashi Natsume
> NTT Software Innovation Center
> Tel: +81-422-59-4399
> E-mail: natsume.taka...@lab.ntt.co.jp
> 
> From: yang, xing [mailto:xing.y...@emc.com]
> Sent: Friday, November 22, 2013 7:25 AM
> To: Carlos Alvarez; openstack@lists.openstack.org
> Subject: Re: [Openstack] multipath - EMC vs NetApp
> 
> Hi Carlos,
> 
> We are working on this issue and will keep you informed.
> 
> Thanks,
> Xing
> 
> 
> From: Carlos Alvarez [mailto:cbalva...@gmail.com]
> Sent: Monday, November 18, 2013 1:19 PM
> To: openstack@lists.openstack.org
> Subject: [Openstack] multipath - EMC vs NetApp
> 
> Hi All.
> 
> I recently added a EMC V-Max storage system, and realized the 
> multipath is not working. The device is called /dev/mapper/XXXX but 
> when I see the multipath -ll output, I see just one path. It is 
> working fine with a
NetApp
> 3250.
> 
> Looking into differences, I see the output of the iscsiadm discovery
> differs:
> 
> Netapp output:
> root@grizzly-dev04:~# iscsiadm -m discovery -t sendtargets -p 
> 10.33.5.10
> 10.33.5.10:3260,1030
> iqn.1992-08.com.netapp:sn.0e3c22d9f2ea11e2a2f2123478563412:vs.10
> 10.33.5.11:3260,1031
> iqn.1992-08.com.netapp:sn.0e3c22d9f2ea11e2a2f2123478563412:vs.10
> 
> Emc output:
> root@grizzly-dev04:~# iscsiadm -m discovery -t sendtargets -p 
> 10.33.5.25
> 10.33.5.25:3260,1 iqn.1992-04.com.emc:50000973f00bcd44
> 
> 
> Looking into the code, the former is clearly what the connect_volume
method
> is expecting: a single ip which returns both path. I reported it to 
> EMC and the answer is that it works with netapp because the 3250 has a 
> feature EMC lacks (called multipath groups).
> 
> Is anybody using multipath with a V-Max?. It should work or EMC is not 
> supported?
> 
> Thanks!
> Carlos
> 
> 





_______________________________________________
Mailing list: http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
Post to     : openstack@lists.openstack.org
Unsubscribe : http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack

Reply via email to