On Sep 19, 2007, at 4:25 AM, Wilkinson, Alex wrote: > 0n Wed, Sep 19, 2007 at 03:24:25AM -0400, Adam Jacob Muller wrote: > >> I have a server with two ZFS pools, one is an internal raid0 using >> 2 drives >> connected via ahc. The other is an external storage array with 11 >> drives >> also using ahc, using raidz. (This is a dell 1650 and pv220s). >> On reboot, the pools do not come online on their own. Both pools >> consistently show as failed. > > Make sure your hostid doesn't change. If it does. Then ZFS will > fail upon bootstrap. > > -aW > No, The hostid is not changing, just rebooted and replicated the problem. Also it seems like from reading ZFS docs that the symptoms would be that the pool would simply need to be imported again if the host id changed? after another reboot, I see this: # zpool status pool: tank state: UNAVAIL status: One or more devices could not be opened. There are insufficient replicas for the pool to continue functioning. action: Attach the missing device and online it using 'zpool online'. see: http://www.sun.com/msg/ZFS-8000-D3 scrub: none requested config: NAME STATE READ WRITE CKSUM tank UNAVAIL 0 0 0 insufficient replicas da1 ONLINE 0 0 0 da2 UNAVAIL 0 0 0 cannot open ... more output showing the other array with 11 drives is fine # zpool export tank # zpool import tank # zpool status pool: tank state: ONLINE scrub: none requested config: NAME STATE READ WRITE CKSUM tank ONLINE 0 0 0 da1 ONLINE 0 0 0 da2 ONLINE 0 0 0 errors: No known data errors (11-drive raidz is fine still of course)Received on Wed Sep 19 2007 - 20:13:18 UTC
This archive was generated by hypermail 2.4.0 : Wed May 19 2021 - 11:39:17 UTC