Node Eviction
Node Eviction
In addition to the ocssd.bin process which is responsible, among other things, for the
network and disk heartbeats, Oracle Clusterware 11g Release 2 uses two new monitoring
processes cssdagent and cssdmonitor, which run with the highest real-time scheduler
priority and are also able to fence a server.
Since cssdagent and cssdmonitor have scheduling priority of 99 stopping them can reset
a server in case:
Both of them are also associated with an undocumented timeout. In case the execution
of the
processes stop for more than 28 sec.; the node will be evicted.
In this post, I will demonstrate node eviction due to missing network heartbeat i.e. a
node will be evicted from the cluster, if it cant communicate with other nodes in the
cluster. To simulate it, I will stop private network on one of the nodes and then scan alert
logs of the surviving nodes.
Current scenario:
On host02 server stop the execution of all rdbms processes (by sending the STOP signal).
Stop the execution of all rdbms processes (by sending the STOP signal)
[root@host02 ~] # ps -ef | grep ora_ | grep orcl2 | awk {print $2} | while read PID
do
kill -STOP $PID
done
From the client point of view, the Real Application Cluster database is hanging
on both nodes. No queries or DMLs are possible. Try to execute a query. The
query will hang.
Check the database alert log file on host01: LMS process issues a request to
CSSD to reboot the node. The node is evicted and instance is restarted after
node joins the cluster.
[root@host01 trace] # tailf /u01/app/oracle/diag/rdbms/orcl/orcl1/trace/alert_orcl1.log
LMS0 (ospid: 31771) has detected no messaging activity from instance 2
LMS0 (ospid: 31771) issues an IMR to resolve the situation
Please check LMS0 trace file for more detail.
Fri Nov 09 11:15:04 2012
Remote instance kill is issued with system inc 30
Remote instance kill map (size 1): 2
LMON received an instance eviction notification from instance 1
The instance eviction reason is 0x20000000
The instance eviction map is 2
Fri Nov 09 11:15:13 2012
IPC Send timeout detected. Sender: ospid 6308 [[email protected] (PZ97)]
Receiver: inst 2 binc 429420846 ospid 6251
Waiting for instances to leave:
2
Reconfiguration started (old inc 4, new inc 8)
List of instances:
1 (myinst: 1)
.. Recovery of instance 2 starts
Global Resource Directory frozen
.
All grantable enqueues granted
Post SMON to start 1st pass IR
-
Instance recovery: looking for dead threads
Beginning instance recovery of 1 threads
Started redo scan
IPC Send timeout to 2.0 inc 4 for msg type 12 from opid 42
Completed redo scan
read 93 KB redo, 55 data blocks need recovery
Started redo application at
Thread 2: logseq 9, block 42
Recovery of Online Redo Log: Thread 2 Group 3 Seq 9 Reading mem 0
Mem# 0: +DATA/orcl/onlinelog/group_3.266.798828557
Mem# 1: +FRA/orcl/onlinelog/group_3.259.798828561
Completed redo application of 0.05MB
Completed instance recovery at
Thread 2: logseq 9, block 228, scn 1069404
52 data blocks read, 90 data blocks written, 93 redo k-bytes read
Thread 2 advanced to log sequence 10 (thread recovery)
Fri Nov 09 12:18:55 2012
Check the cluster Clusterware alert log of host01 The node is evicted and rebooted to
join the cluster
[grid@host01 host01] $ tailf /u01/app/11.2.0/grid/log/host01/alerthost01.log
[cssd (14493)] CRS-1607: Node host02 is being evicted in cluster incarnation 247848838;
details at (: CSSNM00007:) in
/u01/app/11.2.0/grid/log/host01/cssd/ocssd.log.
2012-11-09 11:15:56.140
[ohasd (12412)] CRS-8011: reboot advisory message from host: host02, component: mo103324,
with time stamp: L-2012-11-09-
11:15:56.580
[ohasd (12412)] CRS-8013: reboot advisory message text: clsnomon_status: need to reboot,
unexpected failure 8 received from
CSS
2012-11-09 11:16:17.365
[cssd (14493)] CRS-1601: CSSD Reconfiguration complete. Active nodes are host01.
2012-11-09 11:16:17.400
[crsd (14820)] CRS-5504: Node down event reported for node host02.
2
Node 2 joins the cluster
[cssd (14493)] CRS-1601: CSSD Reconfiguration complete. Active nodes are host01 host02.
2012-11-09 12:18:52.713
[crsd (14820)] CRS-2772: Server host02 has been assigned to pool Generic.
2012-11-09 12:18:52.713
[crsd (14820)] CRS-2772: Server host02 has been assigned to pool ora.orcl.
After the node rejoins the cluster and the instance is restarted, re-execute the
query it succeeds
SQL> conn sys/oracle@orcl as sysdba
SQL> col host_name for a20
SQL> select instance_name, host_name from gv$instance;
INSTANCE_NAME HOST_NAME
-
orcl1 host01.example.com
orcl2 host02.example.com
In this post, I will demonstrate node eviction due to missing disk heartbeat i.e. a node
will be evicted from the cluster, if it cant access the voting disk. To simulate it, I will stop
iscsi service on one of the nodes and then scan alert logs and ocssd logs of various
nodes.
Current scenario:
scan alert log of host03 Note that I/O error occurs at 03:32:11
Note that ocssd process of host03 is not able to access voting disks
2012-11-17 03:32:11.310
2012-11-17 03:32:11.311
2012-11-17 03:32:11.311
2012-11-17 03:32:11.312
[cssd(5149)]CRS-1649:An I/O error occured for voting file: ORCL:ASMDISK02; details at
(:CSSNM00060:) in /u01/app/11.2.0/grid/log/host03/cssd/ocssd.log.
2012-11-17 03:32:11.310
[client (8048)] CRS-10001: ACFS-9112: The following process IDs have open references on
/u01/app/oracle/acfsmount/11.2.0/sharedhome:
[client (8050)] CRS-10001:6323 6363 6391 6375 6385 6383 6402 6319 6503 6361 6377 6505 6389
6369 6335 6367 6333 6387 6871 6325 6381 6327 6496 6498 6552 6373 7278 6339 6400 6357 6500
6329 6365
2012-11-17 03:33:34.050
At 03:34, voting disk cant be accessed even after waiting for timeout
2012-11-17 03:34:10.718
[cssd (5149)] CRS-1615: No I/O has completed after 50% of the maximum interval. Voting file
ORCL: ASMDISK01 will be considered not functional in 99190 milliseconds
2012-11-17 03:34:10.724
[cssd (5149)] CRS-1615: No I/O has completed after 50% of the maximum interval. Voting file
ORCL: ASMDISK02 will be considered not functional in 99180 milliseconds
2012-11-17 03:34:10.724
[cssd (5149)] CRS-1615: No I/O has completed after 50% of the maximum interval. Voting file
ORCL: ASMDISK03 will be considered not functional in 99180 milliseconds
2012-11-17 03:35:10.666
[cssd (5149)] CRS-1614: No I/O has completed after 75% of the maximum interval. Voting file
ORCL: ASMDISK01 will be considered not functional in 49110 milliseconds
2012-11-17 03:35:10.666
[cssd (5149)] CRS-1614: No I/O has completed after 75% of the maximum interval. Voting file
ORCL: ASMDISK02 will be considered not functional in 49110 milliseconds
2012-11-17 03:35:10.666
[cssd (5149)] CRS-1614: No I/O has completed after 75% of the maximum interval. Voting file
ORCL: ASMDISK03 will be considered not functional in 49110 milliseconds
2012-11-17 03:35:46.654
[cssd (5149)] CRS-1613: No I/O has completed after 90% of the maximum interval. Voting file
ORCL: ASMDISK01 will be considered not functional in 19060 milliseconds
2012-11-17 03:35:46.654
[cssd (5149)] CRS-1613: No I/O has completed after 90% of the maximum interval. Voting file
ORCL: ASMDISK02 will be considered not functional in 19060 milliseconds
2012-11-17 03:35:46.654
[cssd (5149)] CRS-1613: No I/O has completed after 90% of the maximum interval. Voting file
ORCL: ASMDISK03 will be considered not functional in 19060 milliseconds
[cssd (5149)] CRS-1604: CSSD voting file is offline: ORCL: ASMDISK01; details at (:
CSSNM00058:) in /u01/app/11.2.0/grid/log/host03/cssd/ocssd.log.
2012-11-17 03:36:10.596
[cssd (5149)] CRS-1604: CSSD voting file is offline: ORCL: ASMDISK02; details at (:
CSSNM00058:) in /u01/app/11.2.0/grid/log/host03/cssd/ocssd.log.
2012-11-17 03:36:10.596
[cssd (5149)] CRS-1604: CSSD voting file is offline: ORCL: ASMDISK03; details at (:
CSSNM00058:) in /u01/app/11.2.0/grid/log/host03/cssd/ocssd.log.
2012-11-17 03:36:10.596
CSSD of host03 reboots the node as no. of voting disks available (0) is less
than minimum required (2)
[cssd (5149)] CRS-1606: The number of voting files available, 0, is less than the minimum
number of voting files required, 2, resulting in CSSD termination to ensure data integrity;
details at (: CSSNM00018:) in /u01/app/11.2.0/grid/log/host03/cssd/ocssd.log
2012-11-17 03:36:15.645
[ctssd (5236)] CRS-2402: The Cluster Time Synchronization Service aborted on host host03.
Details at (: ctsselect_mmg5_1: in /u01/app/11.2.0/grid/log/host03/ctssd/octssd.log.
I/O fencing for ORCL database is carried out by CSSD at 03:32 (same time as
when host02 got the msg that orcl has failed on host03).
..
..
[ohasd (4942)] CRS-8011: reboot advisory message from host: host03, component: mo031159,
with time stamp: L-2012-11-17-03:36:16.705
[ohasd (4942)] CRS-8013: reboot advisory message text: clsnomon_status: need to reboot,
unexpected failure 8 received from CSS
2012-11-17 03:36:29.610
[cssd (5177)] CRS-1612: Network communication with node host03 (3) missing for 50% of
timeout interval. Removal of this node from cluster in 14.060 seconds
2012-11-17 03:36:37.988
[cssd (5177)] CRS-1611: Network communication with node host03 (3) missing for 75% of
timeout interval. Removal of this node from cluster in 7.050 seconds
2012-11-17 03:36:43.992
[cssd (5177)] CRS-1610: Network communication with node host03 (3) missing for 90% of
timeout interval. Removal of this node from cluster in 2.040 seconds
2012-11-17 03:36:46.441
[cssd (5177)] CRS-1632: Node host03 is being removed from the cluster in cluster
incarnation 232819906
2012-11-17 03:36:46.572
[cssd (5177)] CRS-1601: CSSD Reconfiguration complete. Active nodes are host01 host02.
. At 03:32, CRSD process of host02 receives message that orcl database has failed on host03
as
2012-11-17 03:32:44.303
. CRSD process of host02 receives message that acfs has failed on host03 as
2012-11-17 03:36:16.981
[ohasd (4916)] CRS-8011: reboot advisory message from host: host03, component: ag031159,
with time stamp: L-2012-11-17-03:36:16.705
[ohasd (4916)] CRS-8013: reboot advisory message text: clsnomon_status: need to reboot,
unexpected failure 8 received from CSS
2012-11-17 03:36:16.981
[ohasd (4916)] CRS-8011: reboot advisory message from host: host03, component: mo031159,
with time stamp: L-2012-11-17-03:36:16.705
[ohasd (4916)] CRS-8013: reboot advisory message text: clsnomon_status: need to reboot,
unexpected failure 8 received from CSS
2012-11-17 03:36:28.920
. CSSD process of host02 identifies missing network communication from host03 as host03
has rebooted itself
[cssd (5284)] CRS-1612: Network communication with node host03 (3) missing for 50% of
timeout interval. Removal of this node from cluster in 14.420 seconds
2012-11-17 03:36:37.307
[cssd (5284)] CRS-1611: Network communication with node host03 (3) missing for 75% of
timeout interval. Removal of this node from cluster in 7.410 seconds
2012-11-17 03:36:43.328
[cssd (5284)] CRS-1610: Network communication with node host03 (3) missing for 90% of
timeout interval. Removal of this node from cluster in 2.400 seconds
After network communication cant be established for timeout interval, the node is
removed form cluster
2012-11-17 03:36:46.297
[cssd (5284)] CRS-1601: CSSD Reconfiguration complete. Active nodes are host01 host02.
2012-11-17 03:36:46.470
[crsd (5576)] CRS-5504: Node down event reported for node host03.
2012-11-17 03:36:51.890
[crsd (5576)] CRS-2773: Server host03 has been removed from pool Generic.
2012-11-17 03:36:51.909
[crsd (5576)] CRS-2773: Server host03 has been removed from pool ora.orcl.
[cssd (5284)] CRS-1601: CSSD Reconfiguration complete. Active nodes are host01 host02
host03.
..
2012-11-17 03:36:15.052: [ CSSD] [852125584] clssgmPeerDeactivate: node 3 (host03),
death 0, state 0x1 connstate 0x1e