On Mon, May 12, 2014 at 10:52:33AM +0100, Andrei Mikhailovsky wrote:
> Leen, 
> 
> thanks for explaining things. I does make sense now. 
> 
> Unfortunately, it does look like this technology would not fulfill my 
> requirements as I do need to have an ability to perform maintenance without 
> shutting down vms. 
> 

Sorry for being cautious. I've seen certain iSCSI-initiators act that way.

I do not know if that is representative for other iSCSI-initiators.

So I don't know if that applies to VMWare.

During failover reads/writes would be stalled of course.

When properly configured, failover of the target could be done in seconds.

> I will open another topic to discuss possible solutions. 
> 
> Thanks for all your help 
> 
> Andrei 
> ----- Original Message -----
> 
> From: "Leen Besselink" <l...@consolejunkie.net> 
> To: ceph-users@lists.ceph.com 
> Cc: "Andrei Mikhailovsky" <and...@arhont.com> 
> Sent: Sunday, 11 May, 2014 11:41:08 PM 
> Subject: Re: [ceph-users] NFS over CEPH - best practice 
> 
> On Sun, May 11, 2014 at 09:24:30PM +0100, Andrei Mikhailovsky wrote: 
> > Sorry if these questions will sound stupid, but I was not able to find an 
> > answer by googling. 
> > 
> 
> As the Astralians say: no worries, mate. 
> 
> It's fine. 
> 
> > 1. Does iSCSI protocol support having multiple target servers to serve the 
> > same disk/block device? 
> > 
> 
> No, I don't think so. What does work is active/standby failover. 
> 
> I suggest to have some kind of clustering, because as far as I can see, you 
> never want to have 2 target servers active if they don't share state 
> (as far as I know there is no Linux iSCSI-target server which can share state 
> between 2 targets). 
> 
> When there is a failure there is time to have all targets offline for a brief 
> moment, before the second target comes online. The initiators should be able 
> to handle short interruptions. 
> 
> > In case of ceph, the same rbd disk image. I was hoping to have multiple 
> > servers to mount the same rbd disk and serve it as an iscsi LUN. This LUN 
> > would be used as a vm image storage on vmware / xenserver. 
> > 
> 
> You'd have one server which handles a LUN, with it goes down, an other should 
> take over the target IP-address and handle requests for that LUN. 
> 
> > 2.Does iscsi multipathing provide failover/HA capability only on the 
> > initiator side? The docs that i came across all mention multipathing on the 
> > client side, like using two different nics. I did not find anything about 
> > having multiple nics on the initiator connecting to multiple iscsi target 
> > servers. 
> > 
> 
> Multipathing for iSCSI, as I see it, only does one thing: it can be used to 
> create multiple network paths between the initiator and the target. They can 
> be used for resiliance (read: failover) or for loadbalancing when you need 
> more bandwidth. 
> 
> The way I would do it is to have 2 switches and connect each initiator and 
> each target to both switches. Also you would have 2 IP-subnets. 
> 
> So both the target and initiator would have 2 IP-addresses, one from each 
> subnet. 
> 
> So for example: the target would have: 10.0.1.1 and 10.0.2.1 and the 
> initiator: 10.0.1.11 and 10.0.2.11 
> 
> Then you run the IP-traffic for 10.0.1.x on switch 1 and the 10.0.2.x traffic 
> on switch 2. 
> 
> Thus, you have created a resilient set up: The target has multiple 
> connections to the network, the initiator has multiple connections to the 
> network and you can also handle a switch failover. 
> 
> > I was hoping to have resilient solution on the storage side so that I can 
> > perform upgrades and maintenance without needing to shutdown vms running on 
> > vmware/xenserver. Is this possible with iscsi? 
> > 
> 
> The failover set up is mostly to handle failures, not really great for 
> maintenance because it does give a short interruption in service. Like 30 
> seconds or so of no writing to the LUN. 
> 
> That might not be a problem for you, I don't know, but it is at least 
> something to be aware of. And also something you should test when you've 
> build the setup. 
> 
> > Cheers 
> > 
> 
> Hope that helps. 
> 
> > Andrei 
> > ----- Original Message ----- 
> > 
> > From: "Leen Besselink" <l...@consolejunkie.net> 
> > To: ceph-users@lists.ceph.com 
> > Sent: Saturday, 10 May, 2014 8:31:02 AM 
> > Subject: Re: [ceph-users] NFS over CEPH - best practice 
> > 
> > On Fri, May 09, 2014 at 12:37:57PM +0100, Andrei Mikhailovsky wrote: 
> > > Ideally I would like to have a setup with 2+ iscsi servers, so that I can 
> > > perform maintenance if necessary without shutting down the vms running on 
> > > the servers. I guess multipathing is what I need. 
> > > 
> > > Also I will need to have more than one xenserver/vmware host servers, so 
> > > the iscsi LUNs will be mounted on several servers. 
> > > 
> > 
> > So you have multiple machines talking to the same LUN at the same time ? 
> > 
> > You'll have to co-ordinate how changes are written to the backing store, 
> > normally you'd have the virtualization servers use some kind of protocol. 
> > 
> > When it's SCSI there are the older Reserve/Release commands and the newer 
> > SCSI-3 Persistent Reservation commands. 
> > 
> > (i)SCSI allows multiple changes to be in-flight, without coordination 
> > things will go wrong. 
> > 
> > Below it was mentioned that you can disable the cache for rbd, if you have 
> > no coordination protocol you'll need to do the same on the iSCSI-side. 
> > 
> > I believe when you do that it will be slower, but it might work. 
> > 
> > > Would the suggested setup not work for my requirements? 
> > > 
> > 
> > It depends on VMWare if they allow such a setup. 
> > 
> > Then there is an other thing. How do the VMWare machines coordinate which 
> > VM they should be running ? 
> > 
> > I don't know VMWare but usually if you have some kind of clustering setup 
> > you'll need to have a 'quorum'. 
> > 
> > A lot of times the quorum is handled by a quorum disk with the SCSI 
> > coordiation protocols mentioned above. 
> > 
> > An other way to have a quorum is to have a majority voting system with an 
> > un-even number of machines talking over the network. This is what Ceph 
> > monitor nodes do. 
> > 
> > As an example of a clustering system that allows it to be used without a 
> > quorum disk with only 2 machines talking over the network is Linux 
> > Pacemaker. When something bad happends, one machine will just turn off the 
> > power of the other machine to prevent things going wrong (this is called 
> > STONITH). 
> > 
> > > Andrei 
> > > ----- Original Message ----- 
> > > 
> > > From: "Leen Besselink" <l...@consolejunkie.net> 
> > > To: ceph-users@lists.ceph.com 
> > > Sent: Thursday, 8 May, 2014 9:35:21 PM 
> > > Subject: Re: [ceph-users] NFS over CEPH - best practice 
> > > 
> > > On Thu, May 08, 2014 at 01:24:17AM +0200, Gilles Mocellin wrote: 
> > > > Le 07/05/2014 15:23, Vlad Gorbunov a écrit : 
> > > > >It's easy to install tgtd with ceph support. ubuntu 12.04 for example: 
> > > > > 
> > > > >Connect ceph-extras repo: 
> > > > >echo deb http://ceph.com/packages/ceph-extras/debian $(lsb_release 
> > > > >-sc) main | sudo tee /etc/apt/sources.list.d/ceph-extras.list 
> > > > > 
> > > > >Install tgtd with rbd support: 
> > > > >apt-get update 
> > > > >apt-get install tgt 
> > > > > 
> > > > >It's important to disable the rbd cache on tgtd host. Set in 
> > > > >/etc/ceph/ceph.conf: 
> > > > >[client] 
> > > > >rbd_cache = false 
> > > > [...] 
> > > > 
> > > > Hello, 
> > > > 
> > > 
> > > Hi, 
> > > 
> > > > Without cache on the tgtd side, it should be possible to have 
> > > > failover and load balancing (active/avtive) multipathing. 
> > > > Have you tested multipath load balancing in this scenario ? 
> > > > 
> > > > If it's reliable, it opens a new way for me to do HA storage with iSCSI 
> > > > ! 
> > > > 
> > > 
> > > I have a question, what is your use case ? 
> > > 
> > > Do you need SCSI-3 persistent reservations so multiple machines can use 
> > > the same LUN at the same time ? 
> > > 
> > > Because in that case I think tgtd won't help you. 
> > > 
> > > Have a good day, 
> > > Leen. 
> > > _______________________________________________ 
> > > ceph-users mailing list 
> > > ceph-users@lists.ceph.com 
> > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com 
> > > 
> > _______________________________________________ 
> > ceph-users mailing list 
> > ceph-users@lists.ceph.com 
> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com 
> > 
> 

> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to