Some answers below,
Mertol Ozyoney Storage Practice - Sales Manager Sun Microsystems, TR Istanbul TR Phone +902123352200 Mobile +905339310752 Fax +902123352222 Email [EMAIL PROTECTED] -----Original Message----- From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of Ross Sent: Tuesday, November 06, 2007 1:50 PM To: zfs-discuss@opensolaris.org Subject: [zfs-discuss] ZFS and clustering - wrong place? I'm just starting to learn about Solaris, ZFS, etc... It's amazing me how much is possible, but it's just shy of what I'd really, really like to see. I can see there's a fair amount of interest in ZFS and clustering, and it seems Sun are actively looking into this, but I'm wondering if that's the right place to do it? Now I may be missing something obvious here, but it seems to me that for really reliable clustering of data you need to be dealing with it at a higher layer, effectively where iSCSI sits. Instead of making ZFS cluster aware, wouldn't it be easier to add support for things like mirroring, striping (even raid) to the iSCSI protocol? You are missing something obvious here. Looking from an aplication layer file system is at a higher level then a storage protokol. Besides iSCSI is a protokol and ZFS is a file system so there is virtualy no reason to compare them. What Sun is doing at the moment is trying to support active active access of cluster nodes to the same ZFS file system. And active active access is managed at FS level. Accessing shared store is an other thing. ZFS defines nothing about how you access raw devices (FCP , iSCSI, Sata etc.) You can access your storage with iSCSI and use ZFS over it. That way you get to use ZFS locally with all the benefits that entails (guaranteed data integrity, etc), and you also have a protocol somewhere in the network layer that guarantees data integrity to the client (confirming writes at multiple locations, painless failover, etc...). Essentially doing for iSCSI what ZFS did for disk. You'd need support for this in the iSCSI target as it would seem make sense to store the configuration of the cluster on every target. That way the client can connect to any target and read the information on how it is to connect. But once that's done, your SAN speed is only limited by the internal speed of your switch. If you need fast performance, add half a dozen devices and stripe data across them. If you need reliability mirror them. If you want both, use a raid approach. Who needs expensive fibre channel when you can just stripe a pile of cheap iSCSI devices? It would make disaster recovery and HA a piece of cake. For any network like ourselves with a couple of offices and a fast link between them (any university campus would fit that model), you just have two completely independent servers and configure the clients to stream data to them both. No messy configuration of clustered servers, and support for multicast on the network means you don't even have to slow your clients down. The iSCSI target would probably need to integrate with the file system to cope with disasters. You'd need an intelligent way to re-synchronise machines when they came back online, but that shouldn't be too difficult with ZFS. I reckon you could turn Solaris & ZFS into the basis for one of the most flexible SAN solutions out there. What do you all think? Am I off my rocker or would an approach like this work? This message posted from opensolaris.org _______________________________________________ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
_______________________________________________ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss