Just my 2c: Is it possible to do an "offline" dedup, kind of like snapshotting?

What I mean in practice, is: we make many Solaris full-root zones. They share a 
lot of data as complete files. This is kind of easy to save space - make one 
zone as a template, snapshot/clone its dataset, make new zones.

However, as projects evolve (software installed, etc.) these zones are filled 
with many similar files, many of which are duplicates.

It seems reasonable to make some dedup process which would create a 
least-common-denominator snapshot for all the datasets involved (zone roots), 
of which all other datasets' current data are to be dubbed "clones with 
modified data". 

For the system (and user) it should be perceived just the same as these 
datasets are currently "clones with modified data" of the original template 
zone-root dataset. Only the "template" becomes different...

Hope this idea makes sense, and perhaps makes its way into code sometime :)
 
 
This message posted from opensolaris.org
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to