ClusterWare
Upgrade
Clusterware
upgrade From 11.2.0.3 To 11.2.0.4
Note : First we need to upgrade Clusterware . Once Clusterware get upgraded successfully . We can go ahead
with Database upgrade running on this server
1) Stop OEM agent
,oswatcher on Node1/Node2
2) Take backup of
clusterware Home on one node1/Node2
3) increase
memory_max_target and memory_target from 272M to 1500M for ASM instance using rolling mode. In case of less memory for ASM instance
4)Unset environment
and Install 19 clusterware Home Binary
in a separate Oracle home upgrade
using GUI
5) run rootupgrade.sh on Node1
6) Start cluster ware
and all database instances(if not started automatic)
7)run rootupgrade.sh
on another Node
8) Start cluster ware
and all database instances on Node2 (if not started automatic)
9) Start OEM agent ,oswatcher on Node1/Node2
10) Post verification (health check for clusterware and
databases)
Precheck
1)For Each Node, We need to run Cluster Verification Utility to ensure that
you have completed preinstallation steps.
cluvfy comp software -n all
-verbose
2)check passworless ssh connectivity
grep Node1 /etc/hosts
grep NODE1 /etc/hosts
grep NODE2 /etc/hosts
grep fairfied /etc/hosts
or
date;ssh Node2 date;
date;ssh fairfied-rac date;
date;ssh Node1 date;
date;ssh Node1-rac date;
3) check space in clusterware home on both node
Df -g
4) Ensure that you have the following information which is require
during Clusterware Binary installation
·
An Oracle Grid Base and Oracle Grid Home location for
Oracle Clusterware
Node1[/u02/oracle]$ echo $ORACLE_BASE
/u02/oracle/PROD
Node1[/u02/oracle]$
echo $ORACLE_HOME
/u02/oracle/11.2.0.3/grid
5) download Grid software
6) Check Failover status
select username, sid, serial#,process,failover_type,failover_method FROM
gv$session where upper(failover_method) != 'BASIC' and upper(failover_type)
!='SELECT' and upper(username) not in ('SYS','SYSTEM','SYSMAN','DBSNMP');
7) check space in clusterware home on both node ---------
Node1[/u02/oracle/11.2.0.3/grid]$ df -g .
Filesystem GB blocks Free %Used Iused %Iused Mounted on
/dev/u02_lv 124.00 56.37
55% 924168 7% /u02
Node1[/u02/oracle]$ ls -ltr
total 8
drwxr-xr-x 3 root dba 256 Jul 11 2011 11.2.0
drwxr-xr-x 7 oracrs dba 256 Mar 14 2012 PROD
drwxr-xr-x 4 root dba 256 Sep 15 2013 11.2.0.3
drwxrwx--- 6 oracrs dba 4096 Sep 15 2013 oraInventory
drwxrwxr-x 2 root dba 256 Mar 16 14:01 11.2.0.4
Node1[/u02/oracle]$
Node1[/u02/oracle/11.2.0.3/grid]$
Node2[/u02/oracle/11.2.0.3/grid]$ df -g .
Filesystem GB blocks Free %Used
Iused %Iused Mounted on
/dev/u02_lv 124.00 45.33
64% 2072408 17% /u02
Node2[/u02/oracle]$ ls -ltr
total 8
drwxr-xr-x 3 root dba 256 Jul 11 2011 11.2.0
drwxr-xr-x 7 oracrs dba 256 Mar 14 2012 PROD
drwxr-xr-x 4 root dba 256 Sep 15 2013 11.2.0.3
drwxrwx--- 6 oracrs dba 4096 Sep 15 2013 oraInventory
drwx------ 2 root system 256 Mar 17 09:14 11.2.04
drwxrwxr-x 2 root dba 256 Mar 17 09:15 11.2.0.4
Node2[/u02/oracle]$
Node2[/u02/oracle/11.2.0.3/grid]$
9) check load on both servers and
session count node wise on all database
topas,
select count(1),inst_id from gv$session group by inst_id
10)Check GUI are working on both servers and root access ---------
Node1[/u02/oracle]$ vncserver
New 'X' desktop is Node1:1
Starting applications specified in /home/oracrs/.vnc/xstartup
Log file is /home/oracrs/.vnc/Node1:1.log
Node1[/u02/oracle]$ ps -ef|grep vnc
oracrs 21299840 28836502 0 20:04:10
pts/1 0:00 grep vnc
oracrs 39649952 1
0 20:03:49 pts/1 0:00 Xvnc :1 -desktop X -httpd
/opt/freeware/vnc/classes -auth /home/oracrs/.Xauthority -geometry 1024x768
-depth 8 -rfbwait 120000 -rfbauth /home/oracrs/.vnc/passwd -rfbport 5901
-nolisten local -fp
/usr/lib/X11/fonts/,/usr/lib/X11/fonts/misc/,/usr/lib/X11/fonts/75dpi/,/usr/lib/X11/fonts/100dpi/,/usr/lib/X11/fonts/ibm850/,/usr/lib/X11/fonts/Type1/
Node1[/u02/oracle]$
Node2[/u02/oracle]$ ps -ef|grep vnc
oracrs 46071882 54919540 0 20:06:46
pts/0 0:00 grep vnc
oracrs 13042452 1
0 20:06:34 pts/0 0:00 Xvnc :1 -desktop X -httpd
/opt/freeware/vnc/classes -auth /home/oracrs/.Xauthority -geometry 1024x768
-depth 8 -rfbwait 120000 -rfbauth /home/oracrs/.vnc/passwd -rfbport 5901
-nolisten local -fp
/usr/lib/X11/fonts/,/usr/lib/X11/fonts/misc/,/usr/lib/X11/fonts/75dpi/,/usr/lib/X11/fonts/100dpi/,/usr/lib/X11/fonts/ibm850/,/usr/lib/X11/fonts/Type1/
Node2[/u02/oracle]$
10.160.45.73:5901
password for Node1 =vncforu
11) nsure that you have information you will need during installation,
including the following:
a) An Oracle base location for
Oracle Clusterware.
Node1[/u02/oracle]$
Node1[/u02/oracle]$ echo $ORACLE_BASE
/u02/oracle/PROD
Node1[/u02/oracle]$ echo $ORACLE_HOME
/u02/oracle/11.2.0.3/grid
Node1[/u02/oracle]$ ls -ltr /u02/oracle/PROD
total 0
drwxr-x--- 3 oracrs dba 256 Jul 11 2011 cfgtoollogs
drwxrwxr-x 4 oracrs dba 256 Jul 11 2011 diag
drwxr-xr-x 3 oracrs dba 256 Mar 14 2012 Node1
drwxr-xr-x 2 oracrs dba 256 Mar 14 2012 Clusterware
drwxr-xr-x 2 oracrs dba 256 Mar 14 2012 checkpoints
Node1[/u02/oracle]$
Node2[/u02/oracle]$ echo
$ORACLE_BASE
/u02/oracle/PROD
Node2[/u02/oracle]$ ls -ltr /u02/oracle/PROD
total 0
drwxrwxr-x 4 oracrs dba 256 Jul 11 2011 diag
drwxr-xr-x 2 oracrs dba 256 Jul 11 2011 checkpoints
drwxr-xr-x 3 oracrs dba 256 Mar 14 2012 Node2
drwxr-xr-x 2 oracrs dba 256 Mar 14 2012 Clusterware
drwxr-x--- 5 oracrs dba 256 Mar 14 2012 cfgtoollogs
Node2[/u02/oracle]$
Node2[/u02/oracle]$
b) An Oracle Grid Infrastructure
home location that is different from your existing Oracle Clusterware location
c) A SCAN address
Node1[/u02/oracle]$ srvctl config scan
SCAN name: fusprod-scan, Network: 1/10.160.32.0/255.255.255.192/en0
SCAN VIP name: scan1, IP: /fusprod-scan/10.160.32.41
SCAN VIP name: scan2, IP: /fusprod-scan/10.160.32.42
SCAN VIP name: scan3, IP: /fusprod-scan/10.160.32.40
Node1[/u02/oracle]$
or
Node1[/u02/oracle/11.2.0.3/grid/OPatch]$ srvctl config scan
SCAN name: scan-blaze, Network: 1/10.160.32.128/255.255.255.192/en0
SCAN VIP name: scan1, IP: /scan-blaze/10.160.32.189
SCAN VIP name: scan2, IP: /scan-blaze/10.160.32.188
SCAN VIP name: scan3, IP: /scan-blaze/10.160.32.190
Node1[/u02/oracle/11.2.0.3/grid/OPatch]$
Node1[/u02/oracle/11.2.0.3/grid/OPatch]$ ifconfig -a
en0:
flags=1e084863,480<UP,BROADCAST,NOTRAILERS,RUNNING,SIMPLEX,MULTICAST,GROUPRT,64BIT,CHECKSUM_OFFLOAD(ACTIVE),CHAIN>
inet 10.160.32.178 netmask
0xffffffc0 broadcast 10.160.32.191
inet 192.168.32.178
netmask 0xffffff00 broadcast 192.168.32.255
inet 10.160.32.181 netmask
0xffffffc0 broadcast 10.160.32.191
inet 10.160.32.189 netmask
0xffffffc0 broadcast 10.160.32.191
tcp_sendspace 262144
tcp_recvspace 262144 rfc1323 1
en1:
flags=1e084863,480<UP,BROADCAST,NOTRAILERS,RUNNING,SIMPLEX,MULTICAST,GROUPRT,64BIT,CHECKSUM_OFFLOAD(ACTIVE),CHAIN>
inet 10.160.34.94 netmask
0xffffffc0 broadcast 10.160.34.127
inet 192.168.34.94 netmask 0xffffff00
broadcast 192.168.34.255
tcp_sendspace 262144
tcp_recvspace 262144 rfc1323 1
en2:
flags=1e084863,480<UP,BROADCAST,NOTRAILERS,RUNNING,SIMPLEX,MULTICAST,GROUPRT,64BIT,CHECKSUM_OFFLOAD(ACTIVE),CHAIN>
inet 10.160.36.88 netmask
0xffffffc0 broadcast 10.160.36.127
tcp_sendspace 262144
tcp_recvspace 262144 rfc1323 1
en3:
flags=1e084863,480<UP,BROADCAST,NOTRAILERS,RUNNING,SIMPLEX,MULTICAST,GROUPRT,64BIT,CHECKSUM_OFFLOAD(ACTIVE),CHAIN>
inet 10.160.45.73 netmask
0xffffff00 broadcast 10.160.45.255
inet 10.160.45.85 netmask
0xffffff00 broadcast 10.160.45.255
inet 10.160.45.86 netmask
0xffffff00 broadcast 10.160.45.255
tcp_sendspace 262144
tcp_recvspace 262144 rfc1323 1
en4:
flags=1e084863,480<UP,BROADCAST,NOTRAILERS,RUNNING,SIMPLEX,MULTICAST,GROUPRT,64BIT,CHECKSUM_OFFLOAD(ACTIVE),CHAIN>
inet 192.168.33.2 netmask
0xffffffc0 broadcast 192.168.33.63
inet 169.254.85.13 netmask
0xffff0000 broadcast 169.254.255.255
tcp_sendspace 262144
tcp_recvspace 262144 rfc1323 1
en5:
flags=1e084863,480<UP,BROADCAST,NOTRAILERS,RUNNING,SIMPLEX,MULTICAST,GROUPRT,64BIT,CHECKSUM_OFFLOAD(ACTIVE),CHAIN>
inet 10.160.32.197 netmask
0xffffffc0 broadcast 10.160.32.255
tcp_sendspace 262144
tcp_recvspace 262144 rfc1323 1
lo0:
flags=e08084b,c0<UP,BROADCAST,LOOPBACK,RUNNING,SIMPLEX,MULTICAST,GROUPRT,64BIT,LARGESEND,CHAIN>
inet 127.0.0.1 netmask
0xff000000 broadcast 127.255.255.255
inet6 ::1%1/0
tcp_sendspace 131072 tcp_recvspace 131072
rfc1323 1
Node1[/u02/oracle/11.2.0.3/grid/OPatch]$
and
Node2[/u02/oracle/11.2.0.3/grid/OPatch]$ info
Node2[/u02/oracle/11.2.0.3/grid/OPatch]$ ifconfig -a
en0: flags=1e084863,480<UP,BROADCAST,NOTRAILERS,RUNNING,SIMPLEX,MULTICAST,GROUPRT,64BIT,CHECKSUM_OFFLOAD(ACTIVE),CHAIN>
inet 10.160.32.177 netmask
0xffffffc0 broadcast 10.160.32.191
inet 192.168.32.177
netmask 0xffffff00 broadcast 192.168.32.255
inet 10.160.32.190 netmask
0xffffffc0 broadcast 10.160.32.191
inet 10.160.32.188 netmask
0xffffffc0 broadcast 10.160.32.191
inet 10.160.32.132 netmask
0xffffffc0 broadcast 10.160.32.191
tcp_sendspace 262144
tcp_recvspace 262144 rfc1323 1
en1:
flags=1e084863,480<UP,BROADCAST,NOTRAILERS,RUNNING,SIMPLEX,MULTICAST,GROUPRT,64BIT,CHECKSUM_OFFLOAD(ACTIVE),CHAIN>
inet 10.160.34.93 netmask
0xffffffc0 broadcast 10.160.34.127
inet 192.168.34.93 netmask
0xffffff00 broadcast 192.168.34.255
tcp_sendspace 262144
tcp_recvspace 262144 rfc1323 1
en2:
flags=1e084863,480<UP,BROADCAST,NOTRAILERS,RUNNING,SIMPLEX,MULTICAST,GROUPRT,64BIT,CHECKSUM_OFFLOAD(ACTIVE),CHAIN>
inet 10.160.36.87 netmask
0xffffffc0 broadcast 10.160.36.127
tcp_sendspace 262144
tcp_recvspace 262144 rfc1323 1
en3:
flags=1e084863,480<UP,BROADCAST,NOTRAILERS,RUNNING,SIMPLEX,MULTICAST,GROUPRT,64BIT,CHECKSUM_OFFLOAD(ACTIVE),CHAIN>
inet 10.160.45.72 netmask
0xffffff00 broadcast 10.160.45.255
inet 10.160.45.82 netmask
0xffffff00 broadcast 10.160.45.255
inet 10.160.45.83 netmask
0xffffff00 broadcast 10.160.45.255
inet 10.160.45.84 netmask
0xffffff00 broadcast 10.160.45.255
tcp_sendspace 262144
tcp_recvspace 262144 rfc1323 1
en4:
flags=1e084863,480<UP,BROADCAST,NOTRAILERS,RUNNING,SIMPLEX,MULTICAST,GROUPRT,64BIT,CHECKSUM_OFFLOAD(ACTIVE),CHAIN>
inet 192.168.33.1 netmask
0xffffffc0 broadcast 192.168.33.63
inet 169.254.241.173
netmask 0xffff0000 broadcast 169.254.255.255
tcp_sendspace 262144
tcp_recvspace 262144 rfc1323 1
en5: flags=1e084863,480<UP,BROADCAST,NOTRAILERS,RUNNING,SIMPLEX,MULTICAST,GROUPRT,64BIT,CHECKSUM_OFFLOAD(ACTIVE),CHAIN>
inet 10.160.32.198 netmask
0xffffffc0 broadcast 10.160.32.255
tcp_sendspace 262144
tcp_recvspace 262144 rfc1323 1
lo0: flags=e08084b,c0<UP,BROADCAST,LOOPBACK,RUNNING,SIMPLEX,MULTICAST,GROUPRT,64BIT,LARGESEND,CHAIN>
inet 127.0.0.1 netmask
0xff000000 broadcast 127.255.255.255
inet6 ::1%1/0
tcp_sendspace 131072
tcp_recvspace 131072 rfc1323 1
Node2[/u02/oracle/11.2.0.3/grid/OPatch]$
6) Check opatch version
Node1[/u02/oracle/11.2.0.3/grid]$ OPatch/opatch version
OPatch Version: 11.2.0.3.5
OPatch succeeded.
or
Node1[/u02/oracle/11.2.0.3/grid/OPatch]$ ./opatch version
Invoking OPatch 11.2.0.1.7
OPatch Version: 11.2.0.1.7
OPatch succeeded.
Node1[/u02/oracle/11.2.0.3/grid/OPatch]$
Node1[/u02/oracle/11.2.0.3/grid/OPatch]$ ./opatch lsinventory
Invoking OPatch 11.2.0.1.7
Oracle Interim Patch Installer version 11.2.0.1.7
Copyright (c) 2011, Oracle Corporation.
All rights reserved.
Oracle Home :
/u02/oracle/11.2.0.3/grid
Central Inventory : /u02/oracle/oraInventory
from : /etc/oraInst.loc
OPatch version : 11.2.0.1.7
OUI version : 11.2.0.3.0
Log file location :
/u02/oracle/11.2.0.3/grid/cfgtoollogs/opatch/opatch2016-08-30_09-35-54AM.log
Lsinventory Output file location :
/u02/oracle/11.2.0.3/grid/cfgtoollogs/opatch/lsinv/lsinventory2016-08-30_09-35-54AM.txt
--------------------------------------------------------------------------------
Installed Top-level Products (1):
Oracle Grid Infrastructure
11.2.0.3.0
There are 1 products installed in this Oracle Home.
There are no Interim patches installed in this Oracle Home.
Rac system comprising of multiple nodes
Local node = Node1
Remote node = Node2
--------------------------------------------------------------------------------
OPatch succeeded.
Node1[/u02/oracle/11.2.0.3/grid/OPatch]$
Node1[/u02/oracle/11.2.0.3/grid]$
Node2[/u02/oracle]$ cd $ORACLE_HOME
Node2[/u02/oracle/11.2.0.3/grid]$ OPatch/opatch version
OPatch Version: 11.2.0.3.5
OPatch succeeded.
Node2[/u02/oracle/11.2.0.3/grid]$
or
Node2[/u02/oracle/11.2.0.3/grid/OPatch]$ opatch version
Invoking OPatch 11.2.0.1.7
OPatch Version: 11.2.0.1.7
OPatch succeeded.
Node2[/u02/oracle/11.2.0.3/grid/OPatch]$ opatch lsinventory
Invoking OPatch 11.2.0.1.7
Oracle Interim Patch Installer version 11.2.0.1.7
Copyright (c) 2011, Oracle Corporation.
All rights reserved.
Oracle Home : /u02/oracle/11.2.0.3/grid
Central Inventory : /u02/oracle/oraInventory
from : /etc/oraInst.loc
OPatch version : 11.2.0.1.7
OUI version : 11.2.0.3.0
Log file location :
/u02/oracle/11.2.0.3/grid/cfgtoollogs/opatch/opatch2016-08-30_09-34-40AM.log
Lsinventory Output file location :
/u02/oracle/11.2.0.3/grid/cfgtoollogs/opatch/lsinv/lsinventory2016-08-30_09-34-40AM.txt
--------------------------------------------------------------------------------
Installed Top-level Products (1):
Oracle Grid Infrastructure
11.2.0.3.0
There are 1 products installed in this Oracle Home.
There are no Interim patches installed in this Oracle Home.
Rac system comprising of multiple nodes
Local node = Node2
Remote node = Node1
--------------------------------------------------------------------------------
OPatch succeeded.
Node2[/u02/oracle/11.2.0.3/grid/OPatch]$
Node2[/u02/oracle/11.2.0.3/grid/OPatch]$
Node2[/u02/oracle/11.2.0.3/grid/OPatch]$ df -g .
Filesystem GB blocks Free %Used Iused %Iused Mounted on
/dev/u02 60.00 37.38
38% 217572 3% /u02
Node2[/u02/oracle/11.2.0.3/grid/OPatch]$
clusterware version on Node2
-----------------------------
Node1[/u02/oracle/11.2.0.3/grid]$ crsctl query crs activeversion
Oracle Clusterware active version on the cluster is [11.2.0.3.0]
or
Node2[/u02/oracle/11.2.0.3/grid/OPatch]$ crsctl query crs activeversion
Oracle Clusterware active version on the cluster is [11.2.0.3.0]
Node2[/u02/oracle/11.2.0.3/grid/OPatch]$
Node1[/u02/oracle/11.2.0.3/grid]$ crsctl query crs releaseversion
Oracle High Availability Services release version on the local node is
[11.2.0.3.0]
or
Node2[/u02/oracle/11.2.0.3/grid/OPatch]$
crsctl query crs releaseversion
Oracle High Availability Services release version on the local node is
[11.2.0.3.0]
Node2[/u02/oracle/11.2.0.3/grid/OPatch]$
Node1[/u02/oracle/11.2.0.3/grid]$ crsctl query crs softwareversion
Oracle Clusterware version on node [Node1] is [11.2.0.3.0]
Node1[/u02/oracle/11.2.0.3/grid]$
or
Node2[/u02/oracle/11.2.0.3/grid/OPatch]$ crsctl query crs
softwareversion
Oracle Clusterware version on node [Node2] is [11.2.0.3.0]
Node2[/u02/oracle/11.2.0.3/grid/OPatch]$
clusterware version on Node2
-----------------------------
Node2[/u02/oracle/11.2.0.3/grid]$ crsctl query crs activeversion
Oracle Clusterware active version on the cluster is [11.2.0.3.0]
Node2[/u02/oracle/11.2.0.3/grid]$ crsctl query crs releaseversion
Oracle High Availability Services release version on the local node is
[11.2.0.3.0]
Node2[/u02/oracle/11.2.0.3/grid]$ crsctl query crs softwareversion
Oracle Clusterware version on node [Node2] is [11.2.0.3.0]
Node2[/u02/oracle/11.2.0.3/grid]$
12) Run CVU pre-upgrade check tool to ensure all required prerequisites
are met and check memory max size of ASM ----
14) create folder on both node from root user.
id
[Node1:root:/u02/oracle:] mkdir 11.2.0.4
[Node1:root:/u02/oracle:]
chown root:oinstall
/u02/oracle/11.2.0.4
[Node1:root:/u02/oracle:] ls -ltr
total 8
drwxr-xr-x 3 root oinstall 256 May 07 2013 11.2.0.3
drwxr-xr-x 7 oracrs oinstall 256 May 09 2013 PRD
drwxrwx--- 6 oracrs oinstall 4096 May 10 2013 oraInventory
drwxr-xr-x 5 oracrs oinstall 256 Jul 25 2013 network
drwxr-xr-x 2 root system 256 Aug 30 11:46 11.2.0.4.0
drwxr-xr-x 2 root oinstall 256 Aug 30 11:52 11.2.0.4
[Node1:root:/u02/oracle:]
[Node1:root:/u02/oracle:] chmod
775 /u02/oracle/11.2.0.4/
[Node1:root:/u02/oracle:]
From grid user
----------------
/oswatcher/upgrade/grid/runcluvfy.sh stage -pre crsinst -upgrade -n Node1,Node2
-rolling -src_crshome /u02/oracle/11.2.0.3/grid -dest_crshome
/u02/oracle/11.2.0.4/grid -dest_version 11.2.0.4.0 -fixup -fixupdir
/oswatcher/cluster_upgradlog/fixup -verbose | tee
/oswatcher/upgrade/CH9283/runcluvfy_a_fix14jul.out
or
/oswatcher/upgrade/grid/runcluvfy.sh stage -pre crsinst -upgrade -n Node1,Node2
-rolling -src_crshome /u02/oracle/11.2.0.3/grid -dest_crshome
/u02/oracle/11.2.0.4/grid -dest_version 11.2.0.4.0 -fixup -fixupdir
/oswatcher/upgrade/CH11364/fixup -verbose | tee
/oswatcher/upgrade/CH11364/runcluvfy_a_fix05aug2016.out
cat /oswatcher/upgrade/CH11364/runcluvfy_a_fix05aug2016.out |grep -i failed
cat /oswatcher/upgrade/CH11364/runcluvfy_a_fix05aug2016.out |grep -i war
---failed result on Node1-------
Pre-check for cluster services setup was unsuccessful on all the nodes.
Node1[/u02/oracle/11.2.0.3/grid/OPatch]$ cat
/oswatcher/upgrade/CH11364/runcluvfy_a_fix30aug.out |grep -i failed
Node1 10GB (1.048576E7KB) 16GB (1.6777216E7KB) failed
Node2 10GB (1.048576E7KB) 16GB (1.6777216E7KB) failed
Result: Swap space check failed
Node1 missing Patch IZ97457 failed
Node2 missing Patch IZ97457 failed
Result: Operating system patch check failed for "Patch
IZ97457"
Node1 unknown Patch IZ88711|IZ89514 failed
Node2 unknown Patch IZ88711|IZ89514 failed
Result: Operating system patch check failed for "Patch
IZ88711|IZ89514"
Node1[/u02/oracle/11.2.0.3/grid/OPatch]$ df -g .
Filesystem GB blocks Free %Used Iused %Iused Mounted on
/dev/u02 60.00 36.65
39% 389512 5% /u02
Node1[/u02/oracle/11.2.0.3/grid/OPatch]$ nmon
Node1[/u02/oracle/11.2.0.3/grid/OPatch]$
Node1[/u02/oracle/11.2.0.3/grid/OPatch]$ lsps -s
Total Paging Space Percent Used
10240MB 1%
Node1[/u02/oracle/11.2.0.3/grid/OPatch]$
Node1[/home/oracrs]$
Node1[/home/oracrs]$ env
_=/usr/bin/env
LANG=en_US
LOGIN=oracrs
PATH=/u02/oracle/11.2.0.3/grid/bin:/usr/bin:/etc:/usr/sbin:/usr/ucb:/usr/bin/X11:/sbin:/usr/java6_64/jre/bin:/usr/java6_64/bin:/usr/local/bin
EXTENDED_HISTORY=ON
ORACLE_BASE=/u02/oracle/PRD
LC__FASTMSG=true
HISTFILE=/var/adm/hist/oracrs/.sh_history.eibm994:oracrs
LOGNAME=oracrs
MAIL=/usr/spool/mail/oracrs
ORACLE_SID=+ASM1
LOCPATH=/usr/lib/nls/loc
USER=oracrs
AUTHSTATE=files
SHELL=/usr/bin/ksh
ODMDIR=/etc/objrepos
HISTSIZE=1024
TIMEOUT=600
TMOUT=600
HOME=/home/oracrs
TERM=xterm
MAILMSG=[YOU HAVE NEW MAIL]
ORACLE_HOME=/u02/oracle/11.2.0.3/grid
PWD=/home/oracrs
TZ=GMT0BST,M3.5.0,M10.5.0
AIXTHREAD_SCOPE=S
A__z=! LOGNAME="*TMOUT
NLSPATH=/usr/lib/nls/msg/%L/%N:/usr/lib/nls/msg/%L/%N.cat
Node1[/home/oracrs]$
SQL> show parameter memory
NAME
TYPE VALUE
------------------------------------ -----------
------------------------------
memory_max_target
big integer 272M
memory_target
big integer 272M
SQL>
4)Stop OEM agent ,oswatcher on Node1 and Node2-----------------
Node1[/u02/oracle/11.2.0.3/grid]$ ps -ef|grep pmon
fusprod 3473582
1 0 Jun 15
- 3:43 ora_pmon_FSPROD1
oracrs 12845076 1
0 Jun 15 -
1:26 asm_pmon_+ASM1
fuoprd 15597792 1
0 Jun 15 -
5:17 ora_pmon_fuoprd1
fubprd 17760296 1
0 Jun 15 -
3:20 ora_pmon_fubprd1
oracrs 38862906 28836502 0 20:22:40
pts/1 0:00 grep pmon
Node1[/u02/oracle/11.2.0.3/grid]$ ps -ef|grep tns
oracrs 11141142 1
0 Jun 15 -
2:46 /u02/oracle/11.2.0.3/grid/bin/tnslsnr LISTENER_SCAN1 -inherit
oracrs 11468890 1
0 Jun 15 - 18:53 /u02/oracle/11.2.0.3/grid/bin/tnslsnr
LISTENER -inherit
oracrs 66453796 28836502 0 20:22:49
pts/1 0:00 grep tns
Node1[/u02/oracle/11.2.0.3/grid]$ ps -ef|grep oswatcher
oracrs 49021146 28836502 0 20:23:03
pts/1 0:00 grep oswatcher
Node1[/u02/oracle/11.2.0.3/grid]$
Node2[/u02/oracle/11.2.0.3/grid]$
Node2[/u02/oracle/11.2.0.3/grid]$ ps -ef|grep tns
oracrs 14418170 1
1 Oct 18 - 568:33
/u02/oracle/11.2.0.3/grid/bin/tnslsnr LISTENER -inherit
oracrs 42795434 54919540 0 20:24:52
pts/0 0:00 grep tns
oracrs 21692996 1
0 Jun 15 -
2:49 /u02/oracle/11.2.0.3/grid/bin/tnslsnr LISTENER_SCAN2 -inherit
oracrs 47317996 1
0 Jun 15 -
2:48 /u02/oracle/11.2.0.3/grid/bin/tnslsnr LISTENER_SCAN3 -inherit
Node2[/u02/oracle/11.2.0.3/grid]$ ps -ef|grep pmon
oracrs 12714058 1
0 Oct 18 - 46:59 asm_pmon_+ASM2
fuoprd 14745610 1
0 Oct 18 - 178:42 ora_pmon_fuoprd2
fusprod 15335424 1
0 Oct 18 - 109:26 ora_pmon_FSPROD2
fubprd 19202122 1
0 Oct 18
- 112:28 ora_pmon_fubprd2
oracrs 8913808 54919540 0 20:25:03
pts/0 0:00 grep pmon
Node2[/u02/oracle/11.2.0.3/grid]$ ps -ef|grep oswatcher
oracrs 25427980 4522948
0 Jan 18 - 23:52 /bin/sh ./OSWatcherFM.sh 360
/oswatcher/oswbb/archive
oracrs 7012788 54919540 0 20:25:12
pts/0 0:00 grep oswatcher
Node2[/u02/oracle/11.2.0.3/grid]$
Implementation steps
--------------------
Stop OEM agent ,oswatcher on Node2/Node1
1) stop agent from both node
emctl stop agent
$ORACLE_HOME/bin/emctl stop agent
$ORACLE_HOME/bin/emctl stop dbconsole
2) to stop oswatcher
./stopOSWbb.sh
Node2[/home/oracrs]$ sqlplus
SQL*Plus: Release 11.2.0.3.0 Production on Wed Jun 22 23:14:32 2016
Copyright (c) 1982, 2011, Oracle.
All rights reserved.
Enter user-name: / as sysdba
Connected to:
Oracle Database 11g Enterprise Edition Release 11.2.0.3.0 - 64bit
Production
With the Real Application Clusters and Automatic Storage Management
options
SQL> show parameter db
NAME
TYPE VALUE
------------------------------------ -----------
------------------------------
db_cache_size
big integer 0
db_ultra_safe
string OFF
db_unique_name
string +ASM
SQL> show parameter target
NAME
TYPE VALUE
------------------------------------ -----------
------------------------------
memory_max_target
big integer 304M
memory_target
big integer 304M
pga_aggregate_target
big integer 0
sga_target
big integer 0
SQL> show parameter max
NAME
TYPE VALUE
------------------------------------ -----------
------------------------------
max_dump_file_size
string unlimited
memory_max_target
big integer 304M
sga_max_size
big integer 308M
SQL>
SQL> alter system set memory_max_target=1536m scope=spfile;
System altered.
SQL> alter system set memory_target=1536m scope=spfile;
System altered.
SQL> show parameter memory_max_target
NAME
TYPE VALUE
------------------------------------ -----------
------------------------------
memory_max_target
big integer 272M
SQL>
SQL>
SQL>
SQL>
srvctl stop instance -d testdb -i testdb1 -o immediate
srvctl stop asm -n Node1
or
stop complete
clusterware
cd /u02/oracle/11.2.0.3/grid/bin
./crsctl stop crs
ps -ef|grep pmon
To start clusterware then Database instance from Node1
srvctl start asm -n Node1
or
start complete clusterware
/u02/oracle/11.2.0.3/grid/bin
./crsctl start crs
/crsctl check crs
then database
srvctl start instance -d testdb
-i testdb1
Node1[/home/eibm994]$
Node1[/home/eibm994]$ ps -ef|grep pmon
eibm994 13697098 11731206 0 00:38:15
pts/1 0:00 grep pmon
e2bprd 14614660 1
0 00:25:00 - 0:00 ora_pmon_e2b01prd1
oracrs 9372102
1 0 00:22:50 -
0:00 asm_pmon_+ASM1
Node1[/home/eibm994]$ ps -ef|grep tns
oracrs 8126520
1 0 00:31:34 -
0:00 /u02/oracle/11.2.0.3/grid/bin/tnslsnr LISTENER_SCAN3 -inherit
eibm994 10027022 11731206 0 00:38:20
pts/1 0:00 grep tns
oracrs 13435028 1
0 00:31:35 - 0:00 /u02/oracle/11.2.0.3/grid/bin/tnslsnr
LISTENER_SCAN2 -inherit
oracrs 11337984 1
0 00:23:12 - 0:00 /u02/oracle/11.2.0.3/grid/bin/tnslsnr
LISTENER -inherit
Node1[/home/eibm994]$
and same for Node2
Node2[/home/oracrs]$ id
uid=350(oracrs) gid=207(oinstall)
groups=1(staff),203(dba),210(asmadmin),212(asmdba),213(asmoper)
Node2[/home/oracrs]$
srvctl stop instance -d testdb -i testdb2 -o immediate
srvctl stop asm -n Node2
or
stop complete clusterware
/u02/oracle/11.2.0.3/grid/bin
./crsctl stop crs
srvctl stop asm -n Node2
/crsctl check crs
ps -ef|grep pmon
------start clusterware then database-----------
srvctl start asm -n Node2
or
start complete clusterware
/u02/oracle/11.2.0.3/grid/bin
./crsctl start crs
/crsctl check crs
CRS-4533: Event Manager is online
[Node2:root:/u02/oracle/11.2.0.3/grid/bin:] ps -ef|grep pmon
oracrs 13959340 1
0 00:35:23 - 0:00 asm_pmon_+ASM2
root 29425910 21823716 0 00:36:53
pts/1 0:00 grep pmon
[Node2:root:/u02/oracle/11.2.0.3/grid/bin:] ps -ef|grep tns
oracrs 5374024
1 0 Aug 01
- 0:00 /u02/oracle/11.2.0.3/grid/bin/tnslsnr
listener_adm -inherit
oracrs 11534416 1
0 00:35:49 - 0:00 /u02/oracle/11.2.0.3/grid/bin/tnslsnr
LISTENER_SCAN1 -inherit
oracrs 15073348 1
0 00:35:48 - 0:00 /u02/oracle/11.2.0.3/grid/bin/tnslsnr
LISTENER -inherit
root 29425676 21823716 0 00:36:59
pts/1 0:00 grep tns
[Node2:root:/u02/oracle/11.2.0.3/grid/bin:]
To start Database instance from node2
srvctl start instance -d testdb
-i testdb2
login on asm instance and check memory_target on both node
SQL> show parameter target
NAME
TYPE VALUE
------------------------------------ -----------
------------------------------
memory_max_target
big integer 1536M
memory_target big integer 1536M
pga_aggregate_target
big integer 0
sga_target
big integer 0
SQL> show parameter max
NAME
TYPE VALUE
------------------------------------ -----------
------------------------------
max_dump_file_size
string unlimited
memory_max_target
big integer 1536M
sga_max_size
big integer 1536M
SQL>
SQL>
SQL>
SNode2[/home/oracrs]$
Node2[/home/oracrs]$ ps -ef|grep pmon
oracrs 13959340 1
0 00:35:23 - 0:00 asm_pmon_+ASM2
e2bprd 23134368 1
0 00:38:07 - 0:00 ora_pmon_e2b01prd2
Node2[/home/oracrs]$ ps -ef|grep tns
oracrs 5374024
1 0 Aug 01
- 0:00
/u02/oracle/11.2.0.3/grid/bin/tnslsnr listener_adm -inherit
oracrs 7602332 12648646 0 00:44:33
pts/1 0:00 grep tns
oracrs 11534416 1
0 00:35:49 - 0:00 /u02/oracle/11.2.0.3/grid/bin/tnslsnr
LISTENER_SCAN1 -inherit
oracrs 15073348 1
0 00:35:48 - 0:00 /u02/oracle/11.2.0.3/grid/bin/tnslsnr
LISTENER -inherit
Node2[/home/oracrs]$
Node2[/home/oracrs]$ crst
Node2[/home/oracrs]$ crs_stat -t
Name Type Target State
Host
------------------------------------------------------------
ora....ATA1.dg ora....up.type ONLINE
ONLINE Node2
ora....FRA1.dg ora....up.type ONLINE
ONLINE Node2
ora....EDO1.dg ora....up.type ONLINE
ONLINE Node2
ora....EDO2.dg ora....up.type ONLINE
ONLINE Node2
ora....ER.lsnr ora....er.type ONLINE
ONLINE Node2
ora....N1.lsnr ora....er.type ONLINE
ONLINE Node2
ora....N2.lsnr ora....er.type ONLINE
ONLINE Node1
ora....N3.lsnr ora....er.type ONLINE
ONLINE Node1
ora.VOTING.dg ora....up.type
ONLINE ONLINE Node2
ora.asm ora.asm.type ONLINE
ONLINE Node2
ora.cvu ora.cvu.type ONLINE
ONLINE Node1
ora....1prd.db ora....se.type ONLINE
ONLINE Node1
ora....SM2.asm application
ONLINE ONLINE Node2
ora....LD.lsnr application
ONLINE ONLINE Node2
ora....eld.gsd application
OFFLINE OFFLINE
ora....eld.ons application
ONLINE ONLINE Node2
ora....eld.vip ora....t1.type ONLINE
ONLINE Node2
ora....SM1.asm application
ONLINE ONLINE Node1
ora....OD.lsnr application
ONLINE ONLINE Node1
ora....ood.gsd application OFFLINE OFFLINE
ora....ood.ons application
ONLINE ONLINE Node1
ora....ood.vip ora....t1.type ONLINE
ONLINE Node1
ora.gsd ora.gsd.type OFFLINE
OFFLINE
ora....network ora....rk.type ONLINE
ONLINE Node2
ora.oc4j ora.oc4j.type ONLINE
ONLINE Node1
ora.ons ora.ons.type ONLINE
ONLINE Node2
ora....ry.acfs ora....fs.type ONLINE
ONLINE Node2
ora.scan1.vip ora....ip.type
ONLINE ONLINE Node2
ora.scan2.vip ora....ip.type
ONLINE ONLINE Node1
ora.scan3.vip ora....ip.type
ONLINE ONLINE Node1
Node2[/home/oracrs]$
Node2[/home/oracrs]$ . oraenv
ORACLE_SID = [+ASM2] ?
The Oracle base remains unchanged with value /u02/oracle/PRD
Node2[/home/oracrs]$ cd $ORACLE_HOME
Node2[/u02/oracle/11.2.0.3/grid]$ cd ../..
Node2[/u02/oracle]$ ls -ltr
total 0
drwxr-xr-x 3 root oinstall 256 May 07 2013 11.2.0.3
drwxrwx--- 5 oracrs oinstall 256 May 10 2013 oraInventory
drwxr-xr-x 5 oracrs oinstall 256 May 28 2013 network
drwxr-xr-x 6 oracrs oinstall 256 Jun 25 2013 PRD
drwxrwxr-x 2 root oinstall 256 Aug 30 11:55 11.2.0.4
Node2[/u02/oracle]$ crsctl query css votedisk
## STATE File
Universal Id File Name
Disk group
-- ----- ----------------- --------- ---------
1. ONLINE 42941f8e34314f1ebf0738e18d29e41c
(/dev/rhdisk3) [VOTING]
Located 1 voting disk(s).
Node2[/u02/oracle]$ ocrcheck
Status of Oracle Cluster Registry is as follows :
Version : 3
Total space (kbytes) :
262120
Used space (kbytes) :
3152
Available space (kbytes)
: 258968
ID : 1814067896
Device/File Name :
+VOTING
Device/File integrity check succeeded
Device/File not configured
Device/File not configured
Device/File not configured
Device/File not configured
Cluster registry
integrity check succeeded
Logical corruption check
bypassed due to non-privileged user
Node2[/u02/oracle]$
Node2[/u02/oracle]$ crsctl check cluster -all
**************************************************************
Node2:
CRS-4537: Cluster Ready Services is online
CRS-4529: Cluster Synchronization Services is online
CRS-4533: Event Manager is online
**************************************************************
Node1:
CRS-4537: Cluster Ready Services is online
CRS-4529: Cluster Synchronization Services is online
CRS-4533: Event Manager is online
**************************************************************
Node2[/u02/oracle]$
Node2[/u02/oracle]$ srvctl config database -d
Database unique name:
Database name: e2b01prd
Oracle home: /u01/oracle/testdb/product/11.2.0.3
Oracle user:
Spfile: +E2B01PRD_DATA1/testdb/spfilee2b01prd.ora
Domain: world
Start options: open
Stop options: immediate
Database role: PRIMARY
Management policy: AUTOMATIC
Server pools: e2b01prd
Database instances: testdb2,testdb1
Disk Groups: testdb_DATA1,testdb_FRA1,testdb_REDO1,testdb_REDO2
Mount point paths:
Services:
Type: RAC
Database is administrator managed
Node2[/u02/oracle]$
Verify the qosadmin users :
Known bug related to qosadmin user AUTH failure : So check before
upgrade only :
GI Upgrade from 11.2.0.3.6+ to 11.2.0.4 or 12.1.0.1 Fails with
User(qosadmin) is deactivated. AUTH FAILURE. (Doc ID 1577072.1)
otherwise rootupgrade.sh will fail with error "Failed to perform
J2EE (OC4J) Container Resource upgrade at
/haclu/64bit/11.2.0.4/grid/crs/install/crsconfig_lib.pm line 9323"
[root@node1 ~]# /data01/app/11.2.0/grid_11203/bin/qosctl qosadmin
-listusers
AbstractLoginModule password: ---> Default password is oracle112
oc4jadmin
JtaAdmin
qosadmin
Node2[/u02/oracle]$ qosctl qosadmin -listusers
AbstractLoginModule password: 2
JtaAdmin
oc4jadmin
qosadmin
Node2[/u02/oracle]$
For implementation
Upgradation
-------------------------------
./runInstaller -updateNodelistORACLE_HOME=/u02/oracle/11.2.0.3/grid
"CLUSTER_NODES=Node1,Node2" CRS=true
/oui/bin/runInstaller -updateNodelist ORACLE_HOME= "CLUSTER_NODES={Node1,Node2}"
CRS=true
4) Unzip software in same
directory (don't create different directory for each zip file)
5) Unset Oracle related environment variables especially ORA_CRS_HOME
before starting OUI or before executing any root script for both root and grid
user.
$ unset ORACLE_BASE
$ unset ORACLE_HOME
$ unset ORACLE_HOSTNAME
$ unset ORACLE_SID
$ unset ORACLE_UNQNAM
$ unset TNS_ADMIN
$ unset CRS_HOME
$unset ORA_NLS10
export PATH=$ORACLE_HOME/bin:$ORACLE_HOME/OPatch:$PATH:.
export
PATH=/usr/bin:/etc:/usr/sbin:/usr/ucb:/usr/bin/X11:/sbin:/usr/java14_64/jre/bin:/usr/java5/jre/bin:/usr/java5/bin:/opt/IBM/cmdb/nmap-4.76:/usr/local/bin:/usr/local/bin
export IGNORE_PREADDNODE_CHECKS=Y
6) from root user, create a new directory for the Oracle Grid software
on both node (if already created during clusterverification then not required)
mkdir -p /u02/oracle/11.2.0.4/grid
chown -R oracrs:dba /u02/oracle/11.2.0.4/grid
[Node1:root:/u02/oracle:] ls -ltr
total 8
drwxr-xr-x 3 root oinstall 256 May 07 2013 11.2.0.3
drwxr-xr-x 7 oracrs oinstall 256 May 09 2013 PRD
drwxrwx--- 6 oracrs oinstall
4096 May 10 2013 oraInventory
drwxr-xr-x 5 oracrs oinstall 256 Jul 25 2013 network
drwxr-xr-x 2 root system 256 Aug 30 11:46 11.2.0.4.0
[Node1:root:/u02/oracle:]
or
chown root:oinstall
/u02/oracle/11.2.0.4
ls -ltr
chmod 775 /u02/oracle/11.2.0.4/
7) Install the 11.2.0.4 software using GUI and using the grid user :
$ cd /u02/oracle/11.2.0.4/grid
$ ./runInstaller
Starting Oracle Universal Installer...
The OUI installation log is located at /u02/oracle/oraInventory/logs.
For OUI installations or execution of critical scripts it is recommend
to use VNC to avoid problems in case connection with the server is lost.
run from both node from root user
[Node1:root:/oswatcher/upgrade/grid:] ls -ltr
total 160
drwxr-xr-x 2 oracrs oinstall 256 Oct 09 2013 sshsetup
-rwxr-xr-x 1 oracrs oinstall 4878 Oct 09 2013 runcluvfy.sh
-rwxr-xr-x 1 oracrs oinstall
10306 Oct 09 2013 runInstaller
drwxr-xr-x 2 oracrs oinstall 256 Oct 09 2013 rpm
-rwxr-xr-x 1 oracrs oinstall
14047 Oct 09 2013 rootpre.sh
drwxr-xr-x 3 oracrs oinstall 256 Oct 09 2013 rootpre
drwxr-xr-x 2 oracrs oinstall 256 Oct 09 2013 response
drwxr-xr-x 4 oracrs oinstall 4096 Oct 09 2013 install
drwxr-xr-x 14 oracrs oinstall 4096 Oct 09 2013 stage
-rw-r--r-- 1 oracrs oinstall 500 Oct 09 2013 welcome.html
-rw-r--r-- 1 oracrs oinstall
30016 Oct 09 2013 readme.html
[Node1:root:/oswatcher/upgrade/grid:] ./rootpre.sh
./rootpre.sh output will be logged in /tmp/rootpre.out_16-09-05.01:13:14
Checking if group services should be configured....
Nothing to configure.
[Node1:root:/oswatcher/upgrade/grid:]
[Node2:root:/oswatcher/upgrade/grid:] ls -ltr
total 160
drwxr-xr-x 2 oracrs oinstall 256 Oct 09 2013 sshsetup
-rwxr-xr-x 1 oracrs oinstall 4878 Oct 09 2013 runcluvfy.sh
-rwxr-xr-x 1 oracrs oinstall
10306 Oct 09 2013 runInstaller
drwxr-xr-x 2 oracrs oinstall 256 Oct 09 2013 rpm
-rwxr-xr-x 1 oracrs oinstall
14047 Oct 09 2013 rootpre.sh
drwxr-xr-x 3 oracrs oinstall 256 Oct 09 2013 rootpre
drwxr-xr-x 2 oracrs oinstall
256 Oct 09 2013 response
drwxr-xr-x 4 oracrs oinstall 4096 Oct 09 2013 install
drwxr-xr-x 14 oracrs oinstall 4096 Oct 09 2013 stage
-rw-r--r-- 1 oracrs oinstall 500 Oct 09 2013 welcome.html
-rw-r--r-- 1 oracrs oinstall
30016 Oct 09 2013 readme.html
[Node2:root:/oswatcher/upgrade/grid:] ./rootpre.sh
./rootpre.sh output will be logged in /tmp/rootpre.out_16-09-05.01:14:32
Checking if group services should be configured....
Nothing to configure.
[Node2:root:/oswatcher/upgrade/grid:]
Start the installer, and select the option to upgrade an existing Oracle
Clusterware and Oracle ASM installation.
On the node selection page,
select all nodes.
Select installation options as
prompted.
When prompted, run the
rootupgrade.sh script on each node in the cluster that you want to upgrade.
Run the script on the local
node first. The script shuts down the earlier release installation, replaces it
with the new Oracle Clusterware release, and starts the new Oracle Clusterware
installation.
give the new path as below
Node1[/u02/oracle]$ ls -ltr
total 8
drwxr-xr-x 3 root oinstall 256 May 07 2013 11.2.0.3
drwxr-xr-x 7 oracrs oinstall 256 May 09 2013 PRD
drwxrwx--- 6 oracrs oinstall 4096 May 10 2013 oraInventory
drwxr-xr-x 5 oracrs oinstall 256 Jul 25 2013 network
drwxr-xr-x 2 root system 256 Aug 30 11:46 11.2.0.4.0
drwxrwxr-x 2 root oinstall 256 Aug 30 11:52 11.2.0.4
Node1[/u02/oracle/11.2.0.4]$ pwd
/u02/oracle/11.2.0.4
Node1[/u02/oracle/11.2.0.4]$
software location =/u02/oracle/11.2.0.4/grid
not
/u02/oracle/11.2.0.4/grid/ (not last slash)
From root user
----------------
[root@kens-racnode1 OPatch]# /u01/app/11.2.0.3/grid/rootupgrade.sh
[Node1:root:/u02/oracle/11.2.0.4:]
[Node1:root:/u02/oracle/11.2.0.4:] ls -ltr root*
-rwxr-xr-x 1 oracrs oinstall 480 Sep 05 01:34 rootupgrade.sh
-rwxr-x--- 1 oracrs oinstall 467 Sep 05 01:34 root.sh
[Node1:root:/u02/oracle/11.2.0.4:] id
uid=0(root) gid=0(system) groups=208(tivlogs)
[Node1:root:/u02/oracle/11.2.0.4:] rootupgrade.sh
ksh: rootupgrade.sh: not found.
[Node1:root:/u02/oracle/11.2.0.4:] ./rootupgrade.sh
Performing root user operation for Oracle 11g
The following environment variables are set as:
ORACLE_OWNER= oracrs
ORACLE_HOME= /u02/oracle/11.2.0.4
Enter the full pathname of the local bin directory: [/usr/local/bin]:
The contents of "dbhome" have not changed. No need to
overwrite.
The file "oraenv" already exists in /usr/local/bin. Overwrite it? (y/n) [n]: n
The contents of "coraenv" have not changed. No need to
overwrite.
Entries will be added to the /etc/oratab file as needed by
Database Configuration Assistant when a database is created
Finished running generic part of root script.
Now product-specific root actions will be performed.
Using configuration parameter file:
/u02/oracle/11.2.0.4/crs/install/crsconfig_params
Creating trace directory
User ignored Prerequisites during installation
Installing Trace File Analyzer
Or
srvctl upgrade database -d DATABASENAME -o
‘/u01/app/oracle/12.1.0.2′
If you are upgrading a cluster database from releases 10.2,
11.1, or 11.2, then upgrade the database configuration in Oracle Clusterware
using the following command:
$ srvctl upgrade database -d db-unique-name -o oraclehome
Node1[/home/oracrs]$
Node1[/home/oracrs]$ cat .profile
TH=$PATH:$HOME/bin:/usr/local/bin:.
export PATH
if [ -s $MAIL ] # This
is at Shell startup. In normal
then echo $MAILMSG #
operation, the Shell checks
fi #
periodically.
umask 022
AIXTHREAD_SCOPE=S; export AIXTHREAD_SCOPE
unset ORACLE_HOME ORACLE_SID
ORACLE_BASE
export ORACLE_BASE=/u02/oracle/PRD
export ORACLE_HOME=/u02/oracle/11.2.0.3/grid
export ORACLE_SID=+ASM1
#export ORACLE_UNQNAME=
export PATH=${ORACLE_HOME}/bin:${PATH}
alias oh='cd $ORACLE_HOME'
# alias otrace= 'cd
${ORACLE_BASE}/diag/rdbms/${ORACLE_UNQNAME}/${ORACLE_SID}/trace'
set -o vi
Node1[/home/oracrs]$ ps -ef|grep pmon
oracrs 8060936
1 0 01:57:33 -
0:00 asm_pmon_+ASM1
e2bprd 1245504
1 0 01:58:17 -
0:00 ora_pmon_e2b01prd1
oracrs 9896352
8192044 0 02:00:43 pts/2
0:00 grep pmon
Node1[/home/oracrs]$
Node1[/home/oracrs]$
Node1[/home/oracrs]$
Node1[/home/oracrs]$ ps -ef|grep tns
oracrs 6881398
1 0 01:58:13 -
0:00 /u02/oracle/11.2.0.4/bin/tnslsnr LISTENER -inherit
oracrs 16777426 1
0 01:58:13 - 0:00 /u02/oracle/11.2.0.4/bin/tnslsnr
LISTENER_SCAN1 -inherit
Node1[/home/oracrs]$
SQL> select inst_id,count(1) from gv$session group by inst_id;
INST_ID COUNT(1)
---------- ----------
1 51
2 55
SQL>
Node1[/home/oracrs]$
Node1[/home/oracrs]$ crsctl query crs activeversion
Oracle Clusterware active version on the cluster is [11.2.0.3.0]
Node1[/home/oracrs]$ crsctl query crs releaseversion
Oracle High Availability Services release version on the local node is
[11.2.0.3.0]
Node1[/home/oracrs]$ crsctl query crs softwareversion
Oracle Clusterware version on node [Node1] is [11.2.0.4.0]
Node1[/home/oracrs]$
-rw-r----- 1 oracrs oinstall 8327 Sep 05 01:46
AttachHome2016-09-05_01-14-49AM.log.Node2
-rw-r----- 1 oracrs oinstall 0 Sep 05 02:27
oraInstall2016-09-05_02-27-17AM.err
-rw-r----- 1 oracrs oinstall 0 Sep 05 02:27 oraInstall2016-09-05_02-27-17AM.err.Node2
-rw-r----- 1 oracrs oinstall 33 Sep 05 02:28
oraInstall2016-09-05_02-27-17AM.out.Node2
-rw-r----- 1 oracrs oinstall 8416 Sep 05 02:28
UpdateNodeList2016-09-05_02-27-17AM.log.Node2
-rw-r----- 1 oracrs oinstall 33 Sep 05 02:28
oraInstall2016-09-05_02-27-17AM.out
-rw-r----- 1 oracrs oinstall
12506 Sep 05 02:28 UpdateNodeList2016-09-05_02-27-17AM.log
-rw-r----- 1 oracrs oinstall 0 Sep 05 02:28 oraInstall2016-09-05_02-28-19AM.err
-rw-r----- 1 oracrs oinstall 0 Sep 05 02:28
oraInstall2016-09-05_02-28-19AM.err.Node2
-rw-r----- 1 oracrs oinstall 33 Sep 05 02:29
oraInstall2016-09-05_02-28-19AM.out.Node2
-rw-r----- 1 oracrs oinstall 8446 Sep 05 02:29
UpdateNodeList2016-09-05_02-28-19AM.log.Node2
-rw-r----- 1 oracrs oinstall 33 Sep 05 02:29
oraInstall2016-09-05_02-28-19AM.out
-rw-r----- 1 oracrs oinstall
12654 Sep 05 02:29 UpdateNodeList2016-09-05_02-28-19AM.log
-rw-r----- 1 oracrs oinstall
1818777 Sep 05 02:32 installActions2016-09-05_01-14-49AM.log
Node1[/u02/oracle/oraInventory/logs]$ cat
installActions2016-09-05_01-14-49AM.log
8) Start cluster ware and all database instances(if not started
automatic)
After the script completes
successfully, you can run the script in parallel on all nodes except for one,
which you select as the last node. When the script is run successfully on all
the nodes except the last node, run the script on the last node.
9) and run rootupgrade.sh on Node2
node
10) Change grid home in oratab
Post Implementation
------------------ -
1) Check version
select * from v$version
select group_number,instance_name,db_name,software_version from
gv$asm_client;
2) Verify the cluster health from all nodes
$ crsctl check cluster -all
3) To list all existing nodes in
a cluster
$ olsnodes -n
4) ps -ef |grep css
5) Oracle Clusterware active version on the cluster and ervices release version on both Node.
crsctl query crs
activeversion
crsctl query crs
releaseversion
crsctl query crs
softwareversion
The Oracle base remains unchanged with value /u02/oracle/PRD
Node1[/home/oracrs]$ ocrcheck
Status of Oracle Cluster Registry is as follows :
Version : 3
Total space (kbytes) :
262120
Used space (kbytes) :
3228
Available space (kbytes)
: 258892
ID : 1814067896
Device/File Name :
+VOTING
Device/File integrity check succeeded
Device/File not configured
Device/File not configured
Device/File not configured
Device/File not configured
Cluster registry
integrity check succeeded
Logical corruption check
bypassed due to non-privileged user
6)
crsctl stat res -t
7)
Cluster Health Check with CVU
Starting with 11.2.0.3, the cluster verification utility (cluvfy) is
capable of carrying out the post-Clusterware and Database installation health
checks. With the new –healthcheck argument, cluster and database components
best practices, mandatory requirements, deviation, and proper functionality can
be verified.
The following example collects detailed information about best-practice
recommendations for Clusterware in an HTML file named
cvucheckreport_<timestamp>.htm:
$./cluvfy comp healthcheck –collect cluster –bestpractice -html
When no further arguments are attached with the healthcheck parameter,
the Clusterware and Database checks are carried out. Use the following example
to perform the health checks on the cluster and database because no –html
argument was specified; the output will be stored in a text file:
$./cluvfy comp healthcheck
Node2[/home/oracrs]$ cluvfy comp healthcheck
Verifying OS mandatory requirements
Verifying Multicast check ...passed
Verifying Physical Memory ...passed
Verifying Available Physical Memory ...passed
Verifying Swap Size ...warning
Verifying Free Space: Node1:/u02/oracle/11.2.0.4 ...passed
Verifying Free Space: Node2:/u02/oracle/11.2.0.4 ...passed
Verifying Free Space: Node1:/tmp/ ...passed
Verifying Free Space: Node2:/tmp/ ...passed
Verifying User Existence: oracrs ...passed
Verifying Group Existence: oinstall ...passed
Verifying Group Membership: oinstall(Primary) ...passed
Verifying Run Level ...passed
Verifying Hard Limit: maximum open file descriptors ...passed
Verifying Soft Limit: maximum open file descriptors ...passed
Verifying Hard Limit: maximum user processes ...passed
Verifying Soft Limit: maximum user processes ...passed
Verifying Architecture ...passed
Verifying OS Kernel Version ...passed
Verifying OS Kernel Parameter: ncargs ...passed
Verifying OS Kernel Parameter: maxuproc ...passed
Verifying OS Kernel Parameter: tcp_ephemeral_low ...passed
Verifying OS Kernel Parameter: tcp_ephemeral_high ...passed
Verifying OS Kernel Parameter: udp_ephemeral_low ...passed
Verifying OS Kernel Parameter: udp_ephemeral_high ...passed
Verifying Package: bos.adt.base-... ...passed
Verifying Package: bos.adt.lib-... ...passed
Verifying Package: bos.adt.libm-... ...passed
Verifying Package: bos.perf.libperfstat-6.1.2.1 ...passed
Verifying Package: bos.perf.perfstat-... ...passed
Verifying Package: bos.perf.proctools-... ...passed
Verifying Package: xlC.aix61.rte-10.1.0.0 ...passed
Verifying Package: xlC.rte-10.1.0.0 ...passed
Verifying OS Patch:IZ97457 ...warning
Verifying OS Patch:IZ88711|IZ89514 ...warning
Verifying Users With Same UID ...passed
Verifying Current Group ID ...passed
Verifying Root user consistency ...passed
Verifying Network Time Protocol (NTP) ...passed
Verifying Same core file name pattern ...passed
Verifying User Not In Group: oracrs ...passed
Verifying User Mask ...passed
Verifying Task resolv.conf Integrity ...warning
Verifying Time zone consistency ...passed
Verifying UserID65535 ...passed
Verifying Kernel64Bit ...passed
Verifying Network parameter - ipqmaxlen ...met
Verifying Network parameter - rfc1323 ...met
Verifying Network parameter - sb_max ...met
Verifying Network parameter - tcp_sendspace ...met
Verifying Network parameter - tcp_recvspace ...met
Verifying Network parameter - udp_sendspace ...met
Verifying Network parameter - udp_recvspace ...met
Verifying OS Best Practice
Verifying availability of port 8888 ...met
Verifying Network parameter - ipqmaxlen ...met
Verifying Network parameter - rfc1323 ...met
Verifying Network parameter - sb_max ...met
Verifying Network parameter - tcp_sendspace ...met
Verifying Network parameter - tcp_recvspace ...met
Verifying Network parameter - udp_sendspace ...met
Verifying Network parameter - udp_recvspace ...met
Verifying Clusterware mandatory requirements
Verifying Node Connectivity ...passed
or
./orachk –u –o post
8) change path in OEM
=================
We found Doc ID 1989922.1 and corrected the permissions as suggested. We
did:
preston[/u02/oracle/12.1.0/grid/lib]$ ls -l libclntsh.so
-rwx------ 1 oracrs dba 61812395 May 17 15:44 libclntsh.so
preston[/u02/oracle/12.1.0/grid/lib]$ ls -l libclntshcore.so
-rwx------ 1 oracrs dba 4187247 May 17 15:44 libclntshcore.so
preston[/u02/oracle/12.1.0/grid/lib]$ ls -l libocci.so
-rwx------ 1 oracrs dba 4996222 May 17 15:44 libocci.so
preston[/u02/oracle/12.1.0/grid/lib]$ ls -l libttsh12.so
-rwx------ 1 oracrs dba 83299908 May 17 15:44 libttsh12.so
chmod 755 libclntsh.so libclntshcore.so libocci.so libttsh12.so
This appears to have fixed the issue and we can now start the instances
with srvctl.
I would still like to know if, from the log files, the PSU appears to
have been correctly applied. I noticed some warnings regarding SQL scripts
associated with the patches that could not be run due to the instance being
down. Do I need to run these manually?
downgrad steps
=================
Downgrade Grid Infrastructure from 11.2.0.4 to 11.2.0.3
The environment used for this downgrade had a successful upgrade to
11.2.0.4 from 11.2.0.3
1) Check clusterware version on both nodes
crsctl query crs activeversion
crsctl query crs softwareversion
The old GI home (11.2.0.3 home) was not uninstalled nor removed after
the upgrade to 11.2.0.4
Before proceeding with the downgrade identify the "ocr-node".
OCR node an Oracle term for the node which created the backup of the OCR when
the GI was upgraded. This is usually the node on which the first rootupgrade.sh
was run during the upgrade. Look in the cdata folder inside GI HOME and locate
the backup ocr of the form ocrold_version.
$ pwd
/opt/app/11.2.0.3/grid/cdata
$ ls -l
2)
From Root user
In this two node RAC the rootupgrade.sh was first run on Node1 node and
this become the OCR node. The downgrade should be run last on this node.
On all the nodes other than the OCR-node run the rootcrs.pl with the
downgrade option from the new GI HOME (11.2.0.4 home).
# $GI_HOME/crs/install/rootcrs.pl -downgrade -oldcrshome
/u02/oracle/11.2.0.3/grid/ -version
11.2.0.3.0
or
# $GI_HOME/crs/install/rootcrs.pl -downgrade -force -oldcrshome
/u02/oracle/11.2.0.3/grid/ -version
11.2.0.3.0
If you want to stop a partial or failed 11g release 2 (11.2)
installation and restore the previous release Oracle Clusterware, then use the
-force flag with this command.
From root user
Once all the non-OCR Nodes/local node are downgraded, run the downgrade
on the OCR-Node with the lastnode option.
# $GI_HOME/crs/install/rootcrs.pl -downgrade -force -lastnode
-oldcrshome /u02/oracle/11.2.0.3/grid/
-version 11.2.0.3.0
or
# $GI_HOME/crs/install/rootcrs.pl -downgrade -force -lastnode
-oldcrshome /u02/oracle/11.2.0.3/grid/
-version 11.2.0.3.0
This script downgrades the OCR. If you want to stop a partial or failed
11g Release 2 (11.2) installation and restore the previous release Oracle
Clusterware, then use the -force flag with this command
3)
Before running the crs start there are few things to be done. As part of
the upgrade to 11.2.0.4 trace file analyzer (TFA) is installed. Remove the init
script for TFA
#cd /etc/init
# rm oracle-tfa.conf
rm: remove regular file `oracle-tfa.conf'? y
# rm ../init.d/init.tfa
rm: remove regular file `../init.d/init.tfa'? y
Verify the GRID home in the /etc/oratab is pointing to the old GI home
in both nodes
cat /etc/oratab
+ASM1:/u02/oracle/11.2.0.3/grid/ :N # line added by Agent
Remove gpnp profiles if the rootupgrade.sh was run concurrently on
multiple nodes. Also remove the clusterware checkpoint files.
rm -rf
/u02/oracle/11.2.0.4/grid/gpnp/*
rm -rf $ORACLE_BASE/Clusterware/ckptGridHA_rhel6m1.xml
4)
Start clusterware stack on all the nodes with crsctl start crs. Verify
the active version is the downgraded version.
$ crsctl query crs activeversion
Oracle Clusterware active version on the cluster is [11.2.0.3.0]
$ crsctl query crs softwareversion
Oracle Clusterware version on node [rhel6m2] is [11.2.0.3.0]
Run a ocrcheck to verify the OCR intergrity
# ocrcheck
BTfxlecR
ocrcheck -local
5)
login through Grid owner(oracrs)
The inventory will still have the CRS=true on the 11.2.0.4 home. Run the
runInstaller to set the CRS=true to the 11.2.0.3 home (old home).
a) run the following commands, where /u02/oracle/11.2.0.4/grid/ is the
location of the new (upgraded) Grid home (11.2):
Grid_home/oui/bin/runInstaller -nowait -waitforcompletion -ignoreSysPrereqs
-updateNodeList -silent CRS=false ORACLE_HOME=/u02/oracle/11.2.0.4/grid/
b)run the command ./runInstaller -nowait -waitforcompletion
-ignoreSysPrereqs -updateNodeList -silent CRS=true
ORACLE_HOME=pre11.2_crs_home, where pre11.2_crs_home represents the home
directory of the earlier Oracle Clusterware installation
Grid_home/oui/bin/runInstaller -nowait -waitforcompletion
-ignoreSysPrereqs -updateNodeList -silent CRS=true
ORACLE_HOME=/u02/oracle/11.2.0.3/grid/
As the last step detach the 11.2.0.4 GI Home from the inventory and
remove it
./runInstaller -detachHome
ORACLE_HOME=/u02/oracle/11.2.0.4/grid/ -silent
Note--The ORACLE_HOME is hard coded in the oracrs users .profile script
so you could do with a task to update that so you don't forget.
No comments:
Post a Comment