On Thu, 20 Sep 2007, Richard Elling wrote: > 50,000 directories aren't a problem, unless you also need 50,000 quotas > and hence 50,000 file systems. Such a large, single storage pool system > will be an outlier... significantly beyond what we have real world > experience with.
Yes, considering that 45,000 of those users will be students, we definitely need separate quotas for each one :). Hmm, I get a bit of a shiver down my spine at the prospect of deploying a critical central service in a relatively untested configuration 8-/. What is the maximum number of file systems in a given pool that has undergone some reasonable amount of real world deployment? One issue I have is that our previous filesystem, DFS, completely spoiled me with its global namespace and location transparency. We had three fairly large servers, with the content evenly dispersed among them, but from the perspective of the client any user's files were available at /dfs/user/<username>, regardless of which physical server they resided on. We could even move them around between servers transparently. Unfortunately, there aren't really any filesystems available with similar features and enterprise applicability. OpenAFS comes closest, we've been prototyping that but the lack of per file ACLs bites, and as an add-on product we've had issues with kernel compatibility across upgrades. I was hoping to replicate a similar feel by just having one large file server with all the data on it. If I split our user files across multiple servers, we would have to worry about which server contained what files, which would be rather annoying. There are some features in NFSv4 that seem like they might someday help resolve this problem, but I don't think they are readily available in servers and definitely not in the common client. > > I was planning to provide CIFS services via Samba. I noticed a posting a > > while back from a Sun engineer working on integrating NFSv4/ZFS ACL support > > into Samba, but I'm not sure if that was ever completed and shipped either > > in the Sun version or pending inclusion in the official version, does > > anyone happen to have an update on that? Also, I saw a patch proposing a > > different implementation of shadow copies that better supported ZFS > > snapshots, any thoughts on that would also be appreciated. > > This work is done and, AFAIK, has been integrated into S10 8/07. Excellent. I did a little further research myself on the Samba mailing lists, and it looks like ZFS ACL support was merged into the official 3.0.26 release. Unfortunately, the patch to improve shadow copy performance on top of ZFS still appears to be floating around the technical mailing list under discussion. > > Is there any facility for managing ZFS remotely? We have a central identity > > management system that automatically provisions resources as necessary for [...] > This is a loaded question. There is a webconsole interface to ZFS which can > be run from most browsers. But I think you'll find that the CLI is easier > for remote management. Perhaps I should have been more clear -- a remote facility available via programmatic access, not manual user direct access. If I wanted to do something myself, I would absolutely login to the system and use the CLI. However, the question was regarding an automated process. For example, our Perl-based identity management system might create a user in the middle of the night based on the appearance in our authoritative database of that user's identity, and need to create a ZFS filesystem and quota for that user. So, I need to be able to manipulate ZFS remotely via a programmatic API. > Active/passive only. ZFS is not supported over pxfs and ZFS cannot be > mounted simultaneously from two different nodes. That's what I thought, I'll have to get back to that SE. Makes me wonder as to the reliability of his other answers :). > For most large file servers, people will split the file systems across > servers such that under normal circumstances, both nodes are providing > file service. This implies two or more storage pools. Again though, that would imply two different storage locations visible to the clients? I'd really rather avoid that. For example, with our current Samba implementation, a user can just connect to '\\files.csupomona.edu\<username>' to access their home directory or '\\files.csupomona.edu\<groupname>' to access a shared group directory. They don't need to worry on which physical server it resides or determine what server name to connect to. > The SE is mistaken. Sun^H^Holaris Cluster supports a wide variety of > JBOD and RAID array solutions. For ZFS, I recommend a configuration > which allows ZFS to repair corrupted data. That would also be my preference, but if I were forced to use hardware RAID, the additional loss of storage for ZFS redundancy would be painful. Would anyone happen to have any good recommendations for an enterprise scale storage subsystem suitable for ZFS deployment? If I recall correctly, the SE we spoke with recommended the StorageTek 6140 in a hardware raid configuration, and evidently mistakenly claimed that Cluster would not work with JBOD. Thanks... -- Paul B. Henson | (909) 979-6361 | http://www.csupomona.edu/~henson/ Operating Systems and Network Analyst | [EMAIL PROTECTED] California State Polytechnic University | Pomona CA 91768 _______________________________________________ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss