[OmniOS-discuss] Kernel Panic - ZFS on iSCSI target and transferring data.

Svavar Örn Eysteinsson svavar at januar.is
Wed May 21 09:53:39 UTC 2014


Hi.

I recently got a kernel panic on my OmniOS ZFS storage server.
I have a 500GB iSCSI target from my ISP that I have created a zpool on, and
one ZFS dataset.
My OmniOS machines uses this zpool to archive some data not used in
production.
I mainly use rsync to move files between.

This morning, which the iSCSI connection had been up since yesterday I was
going to sync about 20GB of files
to the iSCSI target.

My server got into panic mode.

This is what was in my messages.log file :


May 21 09:13:17 media savecore: [ID 570001 auth.error] reboot after panic:
BAD TRAP: type=8 (#df Double fault) rp=ffffff04e3069f10 addr=0
May 21 09:13:17 media savecore: [ID 365739 auth.error] Saving compressed
system crash dump in /var/crash/unknown/vmdump.0
May 21 09:14:24 media savecore: [ID 849871 auth.error] Decompress the crash
dump with
May 21 09:14:24 media 'savecore -vf /var/crash/unknown/vmdump.0'
May 21 09:14:24 media fmd: [ID 377184 daemon.error] SUNW-MSG-ID:
SUNOS-8000-KL, TYPE: Defect, VER: 1, SEVERITY: Major
May 21 09:14:24 media EVENT-TIME: Wed May 21 09:14:24 GMT 2014
May 21 09:14:24 media PLATFORM: X10SAE, CSN: 0123456789, HOSTNAME: media
May 21 09:14:24 media SOURCE: software-diagnosis, REV: 0.1
May 21 09:14:24 media EVENT-ID: f8b8e00b-a409-6f50-aaa6-bc6c3ebadad0
May 21 09:14:24 media DESC: The system has rebooted after a kernel panic.
 Refer to http://illumos.org/msg/SUNOS-8000-KL for more information.
May 21 09:14:24 media AUTO-RESPONSE: The failed system image was dumped to
the dump device.  If savecore is enabled (see dumpadm(1M)) a copy of the
dump will be written to the savecore directory /var/crash/unknown.
May 21 09:14:24 media IMPACT: There may be some performance impact while
the panic is copied to the savecore directory.  Disk space usage by panics
can be substantial.
May 21 09:14:24 media REC-ACTION: If savecore is not enabled then please
take steps to preserve the crash image.
May 21 09:14:24 media Use 'fmdump -Vp -u
f8b8e00b-a409-6f50-aaa6-bc6c3ebadad0' to view more panic detail.  Please
refer to the knowledge article for additional information.


By issuing fmdump -Vp -u f8b8e00b-a409-6f50-aaa6-bc6c3ebadad0
I have this data :


TIME                           UUID
SUNW-MSG-ID
maí 21 2014 09:14:24.861678000 f8b8e00b-a409-6f50-aaa6-bc6c3ebadad0
SUNOS-8000-KL

  TIME                 CLASS                                 ENA
  maí 21 09:14:24.7657 ireport.os.sunos.panic.dump_available
0x0000000000000000
  maí 21 09:13:17.7666 ireport.os.sunos.panic.dump_pending_on_device
0x0000000000000000

nvlist version: 0
        version = 0x0
        class = list.suspect
        uuid = f8b8e00b-a409-6f50-aaa6-bc6c3ebadad0
        code = SUNOS-8000-KL
        diag-time = 1400663664 781451
        de = fmd:///module/software-diagnosis
        fault-list-sz = 0x1
        fault-list = (array of embedded nvlists)
        (start fault-list[0])
        nvlist version: 0
                version = 0x0
                class = defect.sunos.kernel.panic
                certainty = 0x64
                asru =
sw:///:path=/var/crash/unknown/.f8b8e00b-a409-6f50-aaa6-bc6c3ebadad0
                resource =
sw:///:path=/var/crash/unknown/.f8b8e00b-a409-6f50-aaa6-bc6c3ebadad0
                savecore-succcess = 1
                dump-dir = /var/crash/unknown
                dump-files = vmdump.0
                os-instance-uuid = f8b8e00b-a409-6f50-aaa6-bc6c3ebadad0
                panicstr = BAD TRAP: type=8 (#df Double fault)
rp=ffffff04e3069f10 addr=0
                panicstack = unix:real_mode_stop_cpu_stage2_end+9de3 () |
unix:trap+ca5 () | unix:_patch_xrstorq_rbx+196 () |
zfs:zio_vdev_delegated_io+86 () | zfs:vdev_queue_aggregate+298 () |
zfs:vdev_queue_io_to_issue+5e () | zfs:vdev_queue_io_done+88 () |
zfs:zio_vdev_io_done+80 () | zfs:zio_execute+88 () |
zfs:vdev_queue_io_done+78 () | zfs:zio_vdev_io_done+80 () |
zfs:zio_execute+88 () | zfs:vdev_queue_io_done+78 () |
zfs:zio_vdev_io_done+80 () | zfs:zio_execute+88 () |
zfs:vdev_queue_io_done+78 () | zfs:zio_vdev_io_done+80 () |
zfs:zio_execute+88 () | zfs:vdev_queue_io_done+78 () |
zfs:zio_vdev_io_done+80 () | zfs:zio_execute+88 () |
zfs:vdev_queue_io_done+78 () | zfs:zio_vdev_io_done+80 () |
zfs:zio_execute+88 () | zfs:vdev_queue_io_done+78 () |
zfs:zio_vdev_io_done+80 () | zfs:zio_execute+88 () |
zfs:vdev_queue_io_done+78 () | zfs:zio_vdev_io_done+80 () |
zfs:zio_execute+88 () | zfs:vdev_queue_io_done+78 () |
zfs:zio_vdev_io_done+80 () | zfs:zio_execute+88 () |
zfs:vdev_queue_io_done+78 () | zfs:zio_vdev_io_done+80 () |
zfs:zio_execute+88 () | zfs:vdev_queue_io_done+78 () |
zfs:zio_vdev_io_done+80 () | zfs:zio_execute+88 () |
zfs:vdev_queue_io_done+78 () | zfs:zio_vdev_io_done+80 () |
zfs:zio_execute+88 () | zfs:vdev_queue_io_done+78 () |
zfs:zio_vdev_io_done+80 () | zfs:zio_execute+88 () |
zfs:vdev_queue_io_done+78 () | zfs:zio_vdev_io_done+80 () |
zfs:zio_execute+88 () | zfs:vdev_queue_io_done+78 () |
zfs:zio_vdev_io_done+80 () | zfs:zio_execute+88 () |
zfs:vdev_queue_io_done+78 () | zfs:zio_vdev_io_done+80 () |
zfs:zio_execute+88 () | zfs:vdev_queue_io_done+78 () |
zfs:zio_vdev_io_done+80 () | zfs:zio_execute+88 () |
zfs:vdev_queue_io_done+78 () | zfs:zio_vdev_io_done+80 () |
zfs:zio_execute+88 () | zfs:vdev_queue_io_done+78 () |
zfs:zio_vdev_io_done+80 () | zfs:zio_execute+88 () |
zfs:vdev_queue_io_done+78 () | zfs:zio_vdev_io_done+80 () |
zfs:zio_execute+88 () | zfs:vdev_queue_io_done+78 () |
zfs:zio_vdev_io_done+80 () | zfs:zio_execute+88 () |
zfs:vdev_queue_io_done+78 () | zfs:zio_vdev_io_done+80 () |
                crashtime = 1400662215
                panic-time = Wed May 21 08:50:15 2014 GMT
        (end fault-list[0])

        fault-status = 0x1
        severity = Major
        __ttl = 0x1
        __tod = 0x537c6e70 0x335c29b0

Does anyone see anything ? I have no clue or knowledge/experience in
debugging kernel based crashes.
The only feature that I have enabled on this zPool and or ZFS dataset is a
Lz4 compression on the zfs dataset.

Is there any zfs, iSCSI improvements in the latest OmniOS release ?

Any help, and or information would be much appreciated.

Thanks allot people.

Best regards,

Svavar Orn
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <https://omniosce.org/ml-archive/attachments/20140521/d512b932/attachment-0001.html>


More information about the OmniOS-discuss mailing list