Just an observation: panic occurs in avl_add when called from find_ds_by_guid that tries to add existing snapshot id to the avl tree (http://src.opensolaris.org/source/xref/onnv/onnv-gate/usr/src/uts/common/fs/zfs/dmu_send.c#find_ds_by_guid).
HTH, Andrey On Tue, Feb 9, 2010 at 1:37 AM, Bruno Damour <br...@ruomad.net> wrote: > On 02/ 8/10 06:38 PM, Lori Alt wrote: > > Can you please send a complete list of the actions taken: The commands you > used to create the send stream, the commands used to receive the stream. > Also the output of `zfs list -t all` on both the sending and receiving > sides. If you were able to collect a core dump (it should be in > /var/crash/<hostname>), it would be good to upload it. > > The panic you're seeing is in the code that is specific to receiving a > dedup'ed stream. It's possible that you could do the migration if you > turned off dedup (i.e. didn't specify -D) when creating the send stream.. > However, then we wouldn't be able to diagnose and fix what appears to be a > bug. > > The best way to get us the crash dump is to upload it here: > > https://supportfiles.sun.com/upload > > We need either both vmcore.X and unix.X OR you can just send us vmdump.X. > > Sometimes big uploads have mixed results, so if there is a problem some > helpful hints are > on > http://wikis.sun.com/display/supportfiles/Sun+Support+Files+-+Help+and+Users+Guide, > specifically in section 7. > > It's best to include your name or your initials or something in the name of > the file you upload. As > you might imagine we get a lot of files uploaded named vmcore.1 > > You might also create a defect report at http://defect.opensolaris.org/bz/ > > Lori > > > On 02/08/10 09:41, Bruno Damour wrote: > > <copied from opensolaris-dicuss as this probably belongs here.> > > I kept on trying to migrate my pool with children (see previous threads) and > had the (bad) idea to try the -d option on the receive part. > The system reboots immediately. > > Here is the log in /var/adm/messages > > Feb 8 16:07:09 amber unix: [ID 836849 kern.notice] > Feb 8 16:07:09 amber ^Mpanic[cpu1]/thread=ffffff014ba86e40: > Feb 8 16:07:09 amber genunix: [ID 169834 kern.notice] avl_find() succeeded > inside avl_add() > Feb 8 16:07:09 amber unix: [ID 100000 kern.notice] > Feb 8 16:07:09 amber genunix: [ID 655072 kern.notice] ffffff00053c4660 > genunix:avl_add+59 () > Feb 8 16:07:09 amber genunix: [ID 655072 kern.notice] ffffff00053c46c0 > zfs:find_ds_by_guid+b9 () > Feb 8 16:07:09 amber genunix: [ID 655072 kern.notice] ffffff00053c46f0 > zfs:findfunc+23 () > Feb 8 16:07:09 amber genunix: [ID 655072 kern.notice] ffffff00053c47d0 > zfs:dmu_objset_find_spa+38c () > Feb 8 16:07:09 amber genunix: [ID 655072 kern.notice] ffffff00053c4810 > zfs:dmu_objset_find+40 () > Feb 8 16:07:09 amber genunix: [ID 655072 kern.notice] ffffff00053c4a70 > zfs:dmu_recv_stream+448 () > Feb 8 16:07:09 amber genunix: [ID 655072 kern.notice] ffffff00053c4c40 > zfs:zfs_ioc_recv+41d () > Feb 8 16:07:09 amber genunix: [ID 655072 kern.notice] ffffff00053c4cc0 > zfs:zfsdev_ioctl+175 () > Feb 8 16:07:09 amber genunix: [ID 655072 kern.notice] ffffff00053c4d00 > genunix:cdev_ioctl+45 () > Feb 8 16:07:09 amber genunix: [ID 655072 kern.notice] ffffff00053c4d40 > specfs:spec_ioctl+5a () > Feb 8 16:07:09 amber genunix: [ID 655072 kern.notice] ffffff00053c4dc0 > genunix:fop_ioctl+7b () > Feb 8 16:07:09 amber genunix: [ID 655072 kern.notice] ffffff00053c4ec0 > genunix:ioctl+18e () > Feb 8 16:07:09 amber genunix: [ID 655072 kern.notice] ffffff00053c4f10 > unix:brand_sys_syscall32+1ca () > Feb 8 16:07:09 amber unix: [ID 100000 kern.notice] > Feb 8 16:07:09 amber genunix: [ID 672855 kern.notice] syncing file > systems... > Feb 8 16:07:09 amber genunix: [ID 904073 kern.notice] done > Feb 8 16:07:10 amber genunix: [ID 111219 kern.notice] dumping to > /dev/zvol/dsk/rpool/dump, offset 65536, content: kernel > Feb 8 16:07:10 amber ahci: [ID 405573 kern.info] NOTICE: ahci0: > ahci_tran_reset_dport port 3 reset port > Feb 8 16:07:35 amber genunix: [ID 100000 kern.notice] > Feb 8 16:07:35 amber genunix: [ID 665016 kern.notice] ^M100% done: 107693 > pages dumped, > Feb 8 16:07:35 amber genunix: [ID 851671 kern.notice] dump succeeded > > > Hello, > I'll try to do my best. > > Here are the commands : > > amber ~ # zfs unmount data > amber ~ # zfs snapshot -r d...@prededup > amber ~ # zpool destroy ezdata > amber ~ # zpool create ezdata c6t1d0 > amber ~ # zfs set dedup=on ezdata > amber ~ # zfs set compress=on ezdata > amber ~ # zfs send -RD d...@prededup |zfs receive ezdata/data > cannot receive new filesystem stream: destination 'ezdata/data' exists > must specify -F to overwrite it > amber ~ # zpool destroy ezdata > amber ~ # zpool create ezdata c6t1d0 > amber ~ # zfs set compression=on ezdata > amber ~ # zfs set dedup=on ezdata > amber ~ # zfs send -RD d...@prededup |zfs receive -F ezdata/data > cannot receive new filesystem stream: destination has snapshots (eg. > ezdata/d...@prededup) > must destroy them to overwrite it > > Each time the send/receive command took some hours and transferred 151G data > before issuing the message > > amber ~ # zfs list > NAME USED AVAIL REFER MOUNTPOINT > data 295G 621G 161G /data > data/archive 134G 621G 67.7G /data/archive > data/archive/scanrisk 9.45G 621G 8.03G /data/archive/scanrisk > data/archive/slbp_g 56.6G 621G 13.5G /data/archive/slbp_g > data/cyrus23 219M 621G 219M /data/cyrus23 > data/postgres84_64 373M 621G 199M /data/postgres84_64 > data/postgres84_64/8k 174M 621G 174M /data/postgres84_64/8k > ezdata 151G 144G 21K /ezdata > ezdata/data 151G 144G 151G /ezdata/data > rpool 16.2G 98.0G 87K /rpool > ... > > And the complete : > > amber ~ # zfs list -t all > NAME USED AVAIL REFER MOUNTPOINT > data 295G 621G 161G /data > data/archive 134G 621G 67.7G /data/archive > data/arch...@20090521 32.9M - 36.4G - > data/archive/scanrisk 9.45G 621G 8.03G /data/archive/scanrisk > data/archive/scanr...@2008pre 1.42G - 9.45G - > data/archive/slbp_g 56.6G 621G 13.5G /data/archive/slbp_g > data/archive/slb...@20081129 2.35G - 9.02G - > data/archive/slb...@20081212 2.33G - 9.70G - > data/archive/slb...@20090110 9.98M - 9.34G - > data/archive/slb...@20090521 15.3M - 9.35G - > data/archive/slb...@20090702 1.58G - 14.1G - > data/archive/slb...@20090809 67K - 18.4G - > data/archive/slb...@20090912 68K - 18.4G - > data/archive/slb...@20090915 4.25G - 22.1G - > data/archive/slb...@20091128 97.7M - 19.0G - > data/archive/slb...@20091130 438M - 19.2G - > data/cyrus23 219M 621G 219M /data/cyrus23 > data/postgres84_64 373M 621G 199M /data/postgres84_64 > data/postgres84_64/8k 174M 621G 174M /data/postgres84_64/8k > ezdata 151G 144G 24K /ezdata > ezd...@now 29K - 31K - > ezdata/data 151G 144G 151G /ezdata/data > ezdata/d...@prededup 46K - 151G - > ezdata/test 31K 144G 31K /ezdata/testamber ~ # > ezdata/t...@now 0 - 31K - > rpool 16.3G 97.8G 87K /rpool > rpool/ROOT 8.29G 97.8G 21K legacy > rpool/ROOT/snv_132 8.29G 97.8G 7.19G / > rpool/ROOT/snv_...@install 1.10G - 3.24G - > rpool/dump 4.00G 97.8G 4.00G - > rpool/export 49.9M 97.8G 23K /export > rpool/export/home 49.8M 97.8G 23K /export/home > rpool/export/home/admin 49.8M 97.8G 49.8M /export/home/admin > rpool/swap 4.00G 102G 109M - > tank 292G 165G 166G /tank > t...@20090517 6.75G - 63.6G - > tank/corwin.raw 15G 166G 14.2G - > tank/dara.raw 15G 168G 12.2G - > tank/deirdre.raw 15G 172G 8.36G - > tank/fiona.raw 20G 179G 6.49G - > tank/oberon.raw 15G 180G 16K - > tank/rebma.raw 22.9G 173G 7.93G - > tank/rebma....@20100202 7.34G - 7.93G - > tank/soas.raw 15G 180G 494M - > tank/test 79K 165G 31K /tank/test > tank/t...@now 18K - 31K - > tank/test/child 30K 165G 30K /tank/test/child > tank/test/ch...@now 0 - 30K - > tank/zones 994M 165G 36K /tank/zones > tank/zones/avalon 994M 165G 24K /tank/zones/avalon > tank/zones/avalon/ROOT 994M 165G 21K legacy > tank/zones/avalon/ROOT/zbe 994M 165G 994M legacy > > I will upload the core as vmdump.amber.0.7z > Good luck > > Bruno > > _______________________________________________ > zfs-discuss mailing list > zfs-discuss@opensolaris.org > http://mail.opensolaris.org/mailman/listinfo/zfs-discuss > > _______________________________________________ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss