![]() | Oracle System Handbook - ISO 7.0 May 2018 Internal/Partner Edition | ||
|
|
![]() |
||||||||||||||||
Solution Type Technical Instruction Sure Solution 2151140.1 : How to Check the Current Health Status of the Oracle Private Cloud Appliance (PCA) Nodes
In this Document
Applies to:Private Cloud Appliance X5-2 Hardware - Version All Versions and laterOracle Virtual Compute Appliance X3-2 Hardware - Version All Versions and later Oracle Virtual Compute Appliance X4-2 Hardware - Version All Versions and later Information in this document applies to any platform. GoalThe Oracle PCA controller software contains a monitoring service, which is started and stopped with the ovca service on the active management node SolutionThis document shows steps to check the health status on the PCA. #1 How to check the health status on the compute nodes and zfs heads Using ssh and an account with superuser privileges, log into the active management node. # ssh root@xxx.xxx.xxx.xxx Launch the Oracle PCA command line interface. # pca-admin Check the current status of the rack nodes by querying their ILOMs. PCA> diagnose ilom Checking ILOM health............please wait.. IP_Address Status Health_Details Status: Success PCA> exit [root@ovcamn05r1 ~]# For each possible server node slot location, the Status output will show: "OK", "Not Connected", or "Faulty" #2 How to identify the location of the faulted node The output from pca-admin diagnose ilom gives the fault by node ILOM IP address. There is no correlation between the compute node's IP address and it's rack location. The compute node IP addresses are set at provision time, and can change if and when reprovisioned. Use the following command to grep by the node IP (not the ILOM IP) to determine the physical location. [root@ovcamn05r1 ~]# pca-admin list compute-node | grep 192.168.4.5 | cut -f1 -d" " ovcacn07r1
|||| >>>>>>>>>> always ovca ovcacn07r1 || >>>>>> cn = compute node or mn = management node or sn = storage node (zfs storage) ovcacn07r1 || >>>> 07 = the numerical ID of the rack unit (RU) aka the "slot" location. Starting from the bottom with number 1 and counting upwards ovcacn07r1 || >> r1 = rack 1, where r1 = the base rack, r2 = the first expansion rack, r3 = the second expansion rack, etc. In this example, the fault SPX86-8003-73 reported at 192.168.4.105 is located in the base rack, on the compute node that is the 7th RU location up from the bottom. Also note these two IP addresses for the zfs heads are static:
Note: The pca-admin diagnose ilom command currently does not check the ILOM health on either management node.
In order to check the MN ILOMs for faults, Target | Property | Value
---------+------------+------- <<<<<<<<< If blank, no faults are present. -> exit [root@ovcamn05r1 ~]# ssh root@192.168.4.104 Target | Property | Value
---------+------------+------- <<<<<<<<< If blank, no faults are present. -> exit #4 An ILOM snapshot from the faulted node may be required to assist with troubleshooting. Search MOS for the fault code. In this example the fault code is SPX86-8003-73
References<BUG:23125774> - PCA-ADMIN LIST OR SHOW COMPUTE-NODE --ILOM<NOTE:1591197.1> - How to determine OVCA or PCA Management Node, Compute Node, Switch and ILOM service processor IP addresses Attachments This solution has no attachment |
||||||||||||||||
|