[OmniOS-discuss] issue importing zpool on S11.1 from omniOS LUNs

Stephan Budach stephan.budach at jvm.de
Wed Jan 25 13:54:59 UTC 2017


Hi guys,

I have been trying to import a zpool, based on a 3way-mirror provided by 
three omniOS boxes via iSCSI. This zpool had been working flawlessly 
until some random reboot of the S11.1 host. Since then, S11.1 has been 
importing this zpool without success.

This zpool consists of three 108TB LUNs, based on a raidz-2 zvols… yeah 
I know, we shouldn't have done that in the first place, but performance 
was not the primary goal for that, as this one is a backup/archive pool.

When issueing a zpool import, it says this:

root at solaris11atest2:~# zpool import
   pool: vsmPool10
     id: 12653649504720395171
  state: DEGRADED
status: The pool was last accessed by another system.
action: The pool can be imported despite missing or damaged devices.  The
         fault tolerance of the pool may be compromised if imported.
    see: http://support.oracle.com/msg/ZFS-8000-EY
config:

         vsmPool10                                  DEGRADED
           mirror-0                                 DEGRADED
             c0t600144F07A3506580000569398F60001d0 DEGRADED  corrupted data
             c0t600144F07A35066C00005693A0D90001d0 DEGRADED  corrupted data
             c0t600144F07A35001A00005693A2810001d0 DEGRADED  corrupted data

device details:

         c0t600144F07A3506580000569398F60001d0 DEGRADED         
scrub/resilver needed
         status: ZFS detected errors on this device.
                 The device is missing some data that is recoverable.

         c0t600144F07A35066C00005693A0D90001d0 DEGRADED         
scrub/resilver needed
         status: ZFS detected errors on this device.
                 The device is missing some data that is recoverable.

         c0t600144F07A35001A00005693A2810001d0 DEGRADED         
scrub/resilver needed
         status: ZFS detected errors on this device.
                 The device is missing some data that is recoverable.

However, when  actually running zpool import -f vsmPool10, the system 
starts to perform a lot of writes on the LUNs and iostat report an 
alarming increase in h/w errors:

root at solaris11atest2:~# iostat -xeM 5
                          extended device statistics ---- errors ---
device    r/s    w/s   Mr/s   Mw/s wait actv  svc_t  %w  %b s/w h/w trn tot
sd0       0.0    0.0    0.0    0.0  0.0  0.0    0.0   0 0   0   0   0   0
sd1       0.0    0.0    0.0    0.0  0.0  0.0    0.0   0 0   0   0   0   0
sd2       0.0    0.0    0.0    0.0  0.0  0.0    0.0   0 0   0  71   0  71
sd3       0.0    0.0    0.0    0.0  0.0  0.0    0.0   0 0   0   0   0   0
sd4       0.0    0.0    0.0    0.0  0.0  0.0    0.0   0 0   0   0   0   0
sd5       0.0    0.0    0.0    0.0  0.0  0.0    0.0   0 0   0   0   0   0
                          extended device statistics ---- errors ---
device    r/s    w/s   Mr/s   Mw/s wait actv  svc_t  %w  %b s/w h/w trn tot
sd0      14.2  147.3    0.7    0.4  0.2  0.1    2.0   6 9   0   0   0   0
sd1      14.2    8.4    0.4    0.0  0.0  0.0    0.3   0 0   0   0   0   0
sd2       0.0    4.2    0.0    0.0  0.0  0.0    0.0   0 0   0  92   0  92
sd3     157.3   46.2    2.1    0.2  0.0  0.7    3.7   0 14   0  30   0  30
sd4     123.9   29.4    1.6    0.1  0.0  1.7   10.9   0 36   0  40   0  40
sd5     142.5   43.0    2.0    0.1  0.0  1.9   10.2   0 45   0  88   0  88
                          extended device statistics ---- errors ---
device    r/s    w/s   Mr/s   Mw/s wait actv  svc_t  %w  %b s/w h/w trn tot
sd0       0.0  234.5    0.0    0.6  0.2  0.1    1.4   6 10   0   0   0   0
sd1       0.0    0.0    0.0    0.0  0.0  0.0    0.0   0 0   0   0   0   0
sd2       0.0    0.0    0.0    0.0  0.0  0.0    0.0   0 0   0  92   0  92
sd3       3.6   64.0    0.0    0.5  0.0  4.3   63.2   0 63   0 235   0 235
sd4       3.0   67.0    0.0    0.6  0.0  4.2   60.5   0 68   0 298   0 298
sd5       4.2   59.6    0.0    0.4  0.0  5.2   81.0   0 72   0 406   0 406
                          extended device statistics ---- errors ---
device    r/s    w/s   Mr/s   Mw/s wait actv  svc_t  %w  %b s/w h/w trn tot
sd0       0.0  234.8    0.0    0.7  0.4  0.1    2.2  11 10   0   0   0   0
sd1       0.0    0.0    0.0    0.0  0.0  0.0    0.0   0 0   0   0   0   0
sd2       0.0    0.0    0.0    0.0  0.0  0.0    0.0   0 0   0  92   0  92
sd3       5.4   54.4    0.0    0.3  0.0  2.9   48.5   0 67   0 384   0 384
sd4       6.0   53.4    0.0    0.3  0.0  4.6   77.7   0 87   0 519   0 519
sd5       6.0   60.8    0.0    0.3  0.0  4.8   72.5   0 87   0 727   0 727


I have tried pulling data from the LUNs using dd to /dev/null and I 
didn't get any h/w error, this just started, when trying to actually 
import the zpool. As the h/w errors are constantly rising, I am 
wondering what could cause this and if there can something be done about 
this?

Cheers,
Stephan
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <https://omniosce.org/ml-archive/attachments/20170125/24a45e10/attachment-0001.html>
-------------- next part --------------
A non-text attachment was scrubbed...
Name: smime.p7s
Type: application/x-pkcs7-signature
Size: 5546 bytes
Desc: not available
URL: <https://omniosce.org/ml-archive/attachments/20170125/24a45e10/attachment-0001.bin>


More information about the OmniOS-discuss mailing list