I'm trying to add some additional devices to my existing pool, but it's not working. I'm adding a raidz group of 5 300 GB drives, but the command always fails:
r...@kronos:/ # zpool add raid raidz c8t8d0 c8t13d0 c7t8d0 c3t8d0 c5t8d0 Assertion failed: nvlist_lookup_string(cnv, "path", &path) == 0, file zpool_vdev.c, line 631 Abort (core dumped) The disks all work, were labeled easily using 'format' after zfs and other tools refused to look at them. Creating a UFS filesystem with newfs on them runs with no issues, but I can't add them to the existing zpool. I can use the same devices to create a NEW zpool without issue. I fully patched up this system after encountering this problem, no change. The zpool to which I am adding them is fairly large and in a degraded state (three resilvers running, one that never seems to complete and two related to trying to add these new disks), but I didn't think that should prevent me from adding another vdev. For those who suggest waiting 20 minutes for the resilver to finish, it's been estimating less than 30 minutes for the last 12 hours, and we're running out of space, so I wanted to add the new devices sooner rather than later. Can anyone help? extra details below: r...@kronos:/ # uname -a SunOS kronos 5.10 Generic_137137-09 sun4u sparc SUNW,Sun-Fire-480R r...@kronos:/ # smpatch analyze 137276-01 SunOS 5.10: uucico patch 122470-02 Gnome 2.6.0: GNOME Java Help Patch 121430-31 SunOS 5.8 5.9 5.10: Live Upgrade Patch 121428-11 SunOS 5.10: Live Upgrade Zones Support Patch r...@kronos:patch # zpool list NAME SIZE USED AVAIL CAP HEALTH ALTROOT raid 4.32T 4.23T 92.1G 97% DEGRADED - r...@kronos:patch # zpool status pool: raid state: DEGRADED status: One or more devices are faulted in response to persistent errors. Sufficient replicas exist for the pool to continue functioning in a degraded state. action: Replace the faulted device, or use 'zpool clear' to mark the device repaired. scrub: resilver in progress for 12h22m, 97.25% done, 0h20m to go config: NAME STATE READ WRITE CKSUM raid DEGRADED 0 0 0 raidz1 ONLINE 0 0 0 c9t0d0 ONLINE 0 0 0 c6t0d0 ONLINE 0 0 0 c2t0d0 ONLINE 0 0 0 c4t0d0 ONLINE 0 0 0 c10t0d0 ONLINE 0 0 0 raidz1 ONLINE 0 0 0 c9t1d0 ONLINE 0 0 0 c6t1d0 ONLINE 0 0 0 c2t1d0 ONLINE 0 0 0 c4t1d0 ONLINE 0 0 0 c10t1d0 ONLINE 0 0 0 raidz1 ONLINE 0 0 0 c9t3d0 ONLINE 0 0 0 c6t3d0 ONLINE 0 0 0 c2t3d0 ONLINE 0 0 0 c4t3d0 ONLINE 0 0 0 c10t3d0 ONLINE 0 0 0 raidz1 DEGRADED 0 0 0 c9t4d0 ONLINE 0 0 0 spare DEGRADED 0 0 0 c5t13d0 ONLINE 0 0 0 c6t4d0 FAULTED 0 12.3K 0 too many errors c2t4d0 ONLINE 0 0 0 c4t4d0 ONLINE 0 0 0 c10t4d0 ONLINE 0 0 0 raidz1 DEGRADED 0 0 0 c9t5d0 ONLINE 0 0 0 spare DEGRADED 0 0 0 replacing DEGRADED 0 0 0 c6t5d0s0/o UNAVAIL 0 0 0 cannot open c6t5d0 ONLINE 0 0 0 c11t13d0 ONLINE 0 0 0 c2t5d0 ONLINE 0 0 0 c4t5d0 ONLINE 0 0 0 c10t5d0 ONLINE 0 0 0 raidz1 ONLINE 0 0 0 c5t9d0 ONLINE 0 0 0 c7t9d0 ONLINE 0 0 0 c3t9d0 ONLINE 0 0 0 c8t9d0 ONLINE 0 0 0 c11t9d0 ONLINE 0 0 0 raidz1 ONLINE 0 0 0 c5t10d0 ONLINE 0 0 0 c7t10d0 ONLINE 0 0 0 c3t10d0 ONLINE 0 0 0 c8t10d0 ONLINE 0 0 0 c11t10d0 ONLINE 0 0 0 raidz1 ONLINE 0 0 0 c5t11d0 ONLINE 0 0 0 c7t11d0 ONLINE 0 0 0 c3t11d0 ONLINE 0 0 0 c8t11d0 ONLINE 0 0 0 c11t11d0 ONLINE 0 0 0 raidz1 ONLINE 0 0 0 c5t12d0 ONLINE 0 0 0 c7t12d0 ONLINE 0 0 0 c3t12d0 ONLINE 0 0 0 c8t12d0 ONLINE 0 0 0 c11t12d0 ONLINE 0 0 0 raidz1 ONLINE 0 0 0 c9t2d0 ONLINE 0 0 0 c6t2d0 ONLINE 0 0 0 replacing ONLINE 0 0 0 c11t8d0 ONLINE 0 0 0 c2t2d0 ONLINE 0 0 0 c4t2d0 ONLINE 0 0 0 c10t2d0 ONLINE 0 0 0 spares c6t4d0 INUSE currently in use c3t13d0 AVAIL c7t13d0 AVAIL c11t13d0 INUSE currently in use errors: No known data errors -- This message posted from opensolaris.org _______________________________________________ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss