This sounds like a bug I hit - if you have zvols on your pool, and
automatic snapshots enabled, the thousands of resultant snapshots have
to be polled by devfsadm during boot, which take a long time - several
seconds per zvol.

I remove the auto-snapshot property from my zvols and the slow boot stopped.

Blake



On Thu, Jul 23, 2009 at 5:53 AM, Luc De Meyer<no-re...@opensolaris.org> wrote:
> Follow-up : happy end ...
>
> It took quite some thinkering but... i have my data back...
>
> I ended up starting without the troublesome zfs storage array, de-installed 
> the iscsitartget software and re-installed it...just to have solaris boot 
> without complaining about missing modules...
>
> That left me with a system that would boot as long as the storage was 
> disconnected... Reconnecting it made the boot stop at the hostname. Then the 
> disk activity light would flash every second or so forever... I then rebooted 
> using milestone=none. That worked also with the storage attached! So now I 
> was sure that some software process was causing a hangup (or what appeared to 
> be a hangup.) I could now in milestone none verify that the pool was intact: 
> and so it was... fortunately I had not broken the pool itself... all online 
> with no errors to report.
> I then went to milestone-all which again made the system hang with the disk 
> activity every second forever. I think the task doing this was devfsadm. I 
> then "assumed on a gut feeling" that somehow the system was scanning or 
> checking the pool. I left the system overnight in a desperate attempt because 
> I calculated the 500GB checking to take about 8 hrs if the system would 
> *really* scan everything. (I copied a 1 TB drive last week which took nearly 
> 20 hrs, so I learned that sometimes I need to wait... copying these big disks 
> takes a *lot* of time!)
>
> This morning I switched on the monitor and lo and behold : a login screen !!!!
> The store was there!
>
> Lesson for myself and others: you MUST wait at the hostname line: the system 
> WILL eventually come online... but don't ask how long it takes... I hate to 
> think how long it would take if I had a 10TB system. (but then again, a 
> file-system-check on an ext2 disk also takes forever...)
>
> I re-enabled the iscsitgtd and did a list : it saw one of the two targets ! 
> (which was ok because I remembered that I had turned off the shareiscsi flag 
> on the second share.
> I then went ahead and connected the system back into the network and 
> "repaired" the iscsi-target on the virtual mainframe : WORKED ! Copied over 
> the virtual disks to local store so I can at least start up these servers 
> asap again.
> Then set the iscsishare on the second and most important share: OK! Listed 
> the targets: THERE, BOTH! Repaired it's connection too: WORKED...!
>
> I am copying everything away from the ZFS pools now, but my data is 
> recovered... fortunately.
>
> I now have mixed feelings about the ordeal: yes Sun Solaris kept its promise: 
> I did not loose my data. But the time and trouble it took to recover in this 
> case (just to restart a system for example taking an overnight wait!) is 
> something that a few of my customers would *seriously* dislike...
>
> But: a happy end after all... most important data rescued and 2nd important : 
> I learned a lot in the process...
>
> Bye
> Luc De Meyer
> Belgium
> --
> This message posted from opensolaris.org
> _______________________________________________
> zfs-discuss mailing list
> zfs-discuss@opensolaris.org
> http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
>
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to