On May 9, 2010, at 11:16 AM, Jim Horng wrote: > Okay, so after some test with dedup on snv_134. I decided we can not to use > dedup feature for the time being. > > While unable to destroy a dedupped file system. I decided to migrate the > file system to another pool then destroy the pool. (see below) > > http://opensolaris.org/jive/thread.jspa?threadID=128532&tstart=75 > http://opensolaris.org/jive/thread.jspa?threadID=128620&tstart=60 > > > Now here is my problem. > I did a snapshot of the file system I want to migrate. > I did a send and receive of the file system > > zfs send tank/export/projects/project1...@today | zfs receive -d mpool > > but the file system end up smaller than the original file system without the > dedup turn on. How is this possible?
What you think you are measuring is not what you are measuring. Compare the size of the snapshots. -- richard > Can someone explain. I am not able to trust the data now until I can verify > the data are identical. > > SunOS filearch1 5.11 snv_134 i86pc i386 i86xpv Solaris > r...@filearch1:/var/adm# zpool status > pool: mpool > state: ONLINE > scrub: none requested > config: > > NAME STATE READ WRITE CKSUM > mpool ONLINE 0 0 0 > c7t7d0 ONLINE 0 0 0 > > errors: No known data errors > > pool: rpool > state: ONLINE > scrub: none requested > config: > > NAME STATE READ WRITE CKSUM > rpool ONLINE 0 0 0 > c7t0d0s0 ONLINE 0 0 0 > > errors: No known data errors > > pool: tank > state: ONLINE > scrub: none requested > config: > > NAME STATE READ WRITE CKSUM > tank ONLINE 0 0 0 > raidz1-0 ONLINE 0 0 0 > c7t1d0 ONLINE 0 0 0 > c7t2d0 ONLINE 0 0 0 > c7t3d0 ONLINE 0 0 0 > c7t4d0 ONLINE 0 0 0 > c7t5d0 ONLINE 0 0 0 > c7t6d0 ONLINE 0 0 0 > > errors: No known data errors > > r...@filearch1:/var/adm# zfs list > NAME USED AVAIL REFER MOUNTPOINT > mpool 407G 278G 22K /mpool > mpool/export 407G 278G 22K /mpool/export > mpool/export/projects 407G 278G 23K > /mpool/export/projects > mpool/export/projects/bali_nobackup 407G 278G 407G > /mpool/export/projects/project1_nb > < ...> > tank 520G 4.11T 34.9K /tank > tank/export/projects 515G 4.11T 41.5K /export/projects > tank/export/projects/bali_nobackup 427G 4.11T 424G > /export/projects/project1_nb > > r...@filearch1:/var/adm# zfs get compressratio > NAME PROPERTY VALUE SOURCE > mpool compressratio 2.43x - > mpool/export compressratio 2.43x - > mpool/export/projects compressratio 2.43x - > mpool/export/projects/project1_nb compressratio 2.43x - > mpool/export/projects/project1...@today compressratio 2.43x - > tank compressratio 2.34x - > tank/export compressratio 2.34x - > tank/export/projects compressratio 2.34x - > tank/export/projects/project1_nb compressratio 2.44x - > tank/export/projects/project1...@today compressratio 2.44x - > tank/export/projects/project1_nb_2 compressratio 1.00x - > tank/export/projects/project1_nb_3 compressratio 1.90x - > > r...@filearch1:/var/adm# zpool list > NAME SIZE ALLOC FREE CAP DEDUP HEALTH ALTROOT > mpool 696G 407G 289G 58% 1.00x ONLINE - > rpool 19.9G 9.50G 10.4G 47% 1.00x ONLINE - > tank 5.44T 403G 5.04T 7% 2.53x ONLINE - > -- > This message posted from opensolaris.org > _______________________________________________ > zfs-discuss mailing list > zfs-discuss@opensolaris.org > http://mail.opensolaris.org/mailman/listinfo/zfs-discuss -- ZFS storage and performance consulting at http://www.RichardElling.com _______________________________________________ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss