![]() | Oracle System Handbook - ISO 7.0 May 2018 Internal/Partner Edition | ||
|
|
![]() |
||||||||||||||||
Solution Type Problem Resolution Sure Solution 2163827.1 : A Fragmented And Near Full Zpool Could Cause RAC Panic
In this Document
Created from <SR 3-12922707506> Applies to:Solaris Operating System - Version 11.1 and laterOracle SuperCluster T5-8 Full Rack - Version All Versions to All Versions [Release All Releases] Information in this document applies to any platform. SymptomsIf the zpool, where the RAC software is installed, becomes fragmented and near full, it could cause RAC panic. Disclaimer: RAC panic could be caused by many reasons. The scenario in this document is merely one possibility. Any RAC induced panic must be diagnosed by the RAC support team first.
In the example, the RAC software is installed in zpool 'data'. We could observe the following symptoms in the Solaris crash dump: 1. High zpool usage WARNING: ZFS pool "data" @ 0xffffc1c06773a540 usage > 90% 509G/556G
2. Overdue zios WARNING: 213 deadline (103 overdue) 4 read 209 write 10 pending zios for pool "data" vdev /dev/dsk/c3t3d0s0 (run "zfs [-l] zio pending")
WARNING: 138 deadline (30 overdue) 2 read 136 write 10 pending zios for pool "data" vdev /dev/dsk/c3t0d0s0 (run "zfs [-l] zio pending") 3. RAC grid process(es) waiting to write to the zpool ==== user (LWP_SYS) thread: 0xffffc1c0fda4a7e0 PID: 20382 ==== t_wchan: 0xffffc1c06bbd58e6 sobj: condition var (from zfs:txg_wait_open+0x78) idle: 4285160124 hrticks (4.285160124s) unix:_resume_from_idle+0xf5 resume_return() CAT()> getpath 0xffffc1c0f0ef4b00 4. Write in flight "data" @ 0xffffc1c06773a540:
synced ub_timestamp: -4s txg: 10909751 current ub_timestamp: -4s txg: 10909751 sync thread: 0xfffffffc83000c20 SLEEP idle:0.360338689s cv:0xffffc1c0edd4b438 (from zfs:zio_wait+0x5c) sched(zfs:txg_sync_thread) quiesce thread: 0xfffffffc837d6c20 RUNNABLE idle:4.281476232s sched(zfs:txg_quiesce_thread) (in CPU46 pri 60 dispq) open[2] quiesced[1] syncing[0] synced[3] ============ ============ ============ ============ txg 10909754 10909753 10909752 10909751 dp_space_towrite 33841152 35979264 35364864 0 dp_tempreserved 0 0 0 0 dp_sync_write_time: 3.332925963s dp_read_overhead: 0s dp_written_in_sync: 35315712 (33.6M) dp_writes_inflight: 105185280 (100M) dp_throughput: 5005 B/ms dp_write_limit: 33554432 (32M) 5. Fragmentation CAT()> zfs -p data metaslab vdev mirror-0 @ 0xffffc1c06b1f2800 ....
CauseWhen the zpool becomes fragmented and near-full, the write performance is impacted. It causes delay to the RAC software writing logs. SolutionThe system administrator should reduce the zpool usage. Unwanted snapshots and data could be removed. Data could be moved out of the zpool to elsewhere. Attachments This solution has no attachment |
||||||||||||||||
|