![]() | Oracle System Handbook - ISO 7.0 May 2018 Internal/Partner Edition | ||
|
|
![]() |
|||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
Solution Type Technical Instruction Sure Solution 1492368.1 : SuperCluster T4-4 and T5-8 Disk Replacement Guide
Replacement of internal hard drives in SPARC SuperCluster T4-4 In this Document
Applies to:SPARC SuperCluster T4-4 - Version All Versions to All Versions [Release All Releases]Oracle Solaris on SPARC (64-bit) GoalSteps to replace internal hard drive in SuperCluster configurations Please note as other consideration are discovered in various LDom configurations we will be updating this note. So please do not rely on a strict hard copy of it.
Solution1) Confirm zpool has faulted device:As seen from `zpool status` in LDOM where failure is being reported root@orlscdb02:~# zpool status NAME STATE READ WRITE CKSUM Drive showing unavailable in `format` format
2) Detach the failed drive form zpool:Remove drive from zpool in LDOM where failure is being reported zpool detach rpool c0t5000CCA0125099FCd0s0 3) Determine Primary LDOM location:virtinfo -a
Domain role: LDoms guest I/O service root Domain name: ssccn1-app2 Domain UUID: cac06c45-fb37-e0e7-d3cc-8e988be3f16b Control domain: orlscdb01 <-- PRIMARY LDOM Chassis serial#: 1139BDY8C1 4) Check if vdisk device exists on this drive from PRIMARY LDOM:In the example bellow we see that slice 1 of the same drive is also being used al a virtual disk device ldm list-services | grep 5000CCA0125099FC
vdisk2 /dev/dsk/c0t5000CCA0125099FCd0s1 5) Turn on locator beacon from PRIMARY LDOM:
/opt/ipmitool/bin/ipmitool sunoem cli "set /SYS/LOCATE value=fast_blink"
If no vdisk device exists skip to step 9 section
6) Determine the LDOM using the vdisk:root@orlscdb01:~# ldm list -o disk
NAME primary VDS NAME VOLUME OPTIONS MPGROUP DEVICE primary-vds0 vol1 /dev/dsk/c0t5000C5003BE98F4Bd0s1 ------------------------------------------------------------------------------ NAME ssccn1-app1 DISK NAME VOLUME TOUT ID DEVICE SERVER MPGROUP vdisk1 vol1@primary-vds0 0 disk@0 primary vdisk2 vol1@service-vds0 1 disk@1 ssccn1-app2 ------------------------------------------------------------------------------ NAME ssccn1-app2 VDS NAME VOLUME OPTIONS MPGROUP DEVICE service-vds0 vol1 /dev/dsk/c0t5000CCA0125099FCds1 7) Log into Logical domain and detach disk:root@orlscdb01:~# zpool status NAME STATE READ WRITE CKSUM orlsccldm01:~# zpool detach rpool c2t1s0 8) Remove the vdisk and vdisk service device:ldm remove-vdisk vdisk2 ssccn1-app1 ldm remove-vdsdev vol1@servivce-vds0 9) Find device path of drive to use in cfgadm youcfgadm -als "match=partial,select=class(scsi):ap_id(c):type(disk)" -v Ap_Id Receptacle Occupant Condition Information
10) Unconfigure failed drive:cfgadm -c unconfigure c6::w5000cca0125099fd,0 11) Remove failed drive and insert replacement12) Configure in new device:
root@orlscdb02:~# cfgadm -als "match=partial,select=class(scsi):ap_id(c):type(disk)" root@orlscdb02:~# cfgadm -c configure c6::w5000c5003be98f49,0 root@orlscdb02:~# cfgadm -als "match=partial,select=class(scsi):ap_id(c):type(disk)" 13) Update new disk Table of contents from existing zpool member. Ensure you use slice 2 in this command:root@orlscdb02:~# zpool status NAME STATE READ WRITE CKSUM
12a) format -L vtoc -d c0t5000C5003BE98F4Bd0 12b) prtvtoc /dev/rdsk/c0t5000CCA0124C1244d0s2 | fmthard -s - /dev/rdsk/c0t5000C5003BE98F4Bd0s2 14) Attach new disk slice 0 to the rpool:zpool attach rpool c0t5000CCA0124C1244d0s0 c0t5000C5003BE98F4Bd0s0 15) Confirm zpool resilvered and drive status is online:
root@orlscdb02:/var/tmp# zpool status u01-pool NAME STATE READ WRITE CKSUM errors: No known data errors 16) Clear Fault Management:
TIME EVENT-ID MSG-ID SEVERITY Host : orlscdb01 Fault class : fault.fs.zfs.vdev.io Description : The number of I/O errors associated with a ZFS device exceeded Response : The device has been offlined and marked as faulted. An attempt Impact : Fault tolerance of the pool may be compromised. Action : Run 'zpool status -x' and replace the bad device.
fmadm repaired /SYS/MB/HDD3 17) If LDOM is running Solaris 10 and pool is the root pool such as "rpool" or "BIpool" install bootblock on new slice:installboot -F zfs /usr/platform/`uname -i`/lib/fs/zfs/bootblk /dev/rdsk/c0t5000C5003BE98F4Bd0s0 18) Update boot-device with WWN from new drive
#eeprom boot-device=
"/pci@400/pci@1/pci@0/pci@0/LSI,sas@0/disk@w5000ccA0124c1244,0:a /pci@400/pci@1/pci@0/pci@0/LSI,sas@0/disk@w5000c5003be98F4b,0:a disk net"
If a Vdisk was present on the replaced device and step 6 was followed, you must now recreate the vdisk and add it back to the guest LDOM
19) Reconfigure virtual disk:ldm add-vdsdev /dev/dsk/c0t5000C5003BE98F4Bd0s1 vol1@service-vds0 ldm add-vdisk timeout=1 id=0 vdisk2 vol1@service-vds0 ssccn1-app1
20) Connect to Guest LDOM and reattach disk to zpool, mark vdev as repaired and reinstall bootblockzpool attach rpool c2d0s0 c2d1s0 orlscclldm01:~# zpool status NAME STATE READ WRITE CKSUM fmadm repaired zfs://pool=rpool/vdev=2cd55479f321bdae installboot -F zfs /usr/platform/`uname -i`/lib/fs/zfs/bootblk /dev/rdsk/c2d2s0 21) Turn off locator beacon from PRIMARY LDOM:/opt/ipmitool/bin/ipmitool sunoem cli "set /SYS/LOCATE value=off"
22) Example workflow of 3 LDOM config and the loss of HDD5
Attachments This solution has no attachment |
|||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
|