Summary:
==============================
This presentation will provide detail steps for Oracle DBA and Linux Engineer to add new node to existing 10gR2 (10.2.0.3) database RAC.
The most critical steps that need to be followed are:
- Pre-install checking
- Adding an Oracle clusterware home to new nodes using OUI in interactive node
- Adding an Oracle home to new nodes using OUI in interactive mode
- Reconfigure listener on new node using NETCA
- Adding ASM instance to new nodes manually
1. Verify cluster healthy on existing nodes (ocrcheck, cluvfy)
2. Check OS version, Kernel parameters, /etc/hosts file and ensure they are identical on all nodes
3. Cross check ssh login on all nodes
4. Cross ping hostname and hostname-pn on all nodes
5. Copy .profile from existing node to new node
6. Ensure ocr, css files and shared ASM disk are visible from new node and have right permission
7. Backup crs, ocr and oraInventory on existing nodes
8. Change permission on $ORA_CRS_HOME/bin/vipca.orig and $ORA_CRS_HOME/bin/srvctl.orig on existing node
change permission from
$ ls -lrat *.orig
-rwxr-x--x 1 root users 5380 Feb 27 2007 vipca.orig
-rwxr-x--x 1 root users 5800 Feb 27 2007 srvctl.orig
to
$ls -lrat *.orig
-rwxr-x--x 1 oracle oinstall 5380 2007-02-27 18:47 vipca.orig
-rwxr-x--x 1 oracle oinstall 5800 2007-02-27 18:47 srvctl.orig
9. Comment out two lines below in $ORA_CRS_HOME/bin/srvctl on existing nodes
#LD_ASSUME_KERNEL=2.4.19
#export LD_ASSUME_KERNEL
10. Add "unset LD_ASSUME_KERNEL" in $ORA_CRS_HOME/bin/vipca line 126
Step9 and 10 are 10.2.0.3 specific. The issue is fixed on 10.2.0.4. Please reference Oracle metalink Doc ID: Note:414163.1 -- 10gR2 RAC Install issues on Oracle EL5 or RHEL5 or SLES10 (VIPCA Failures)
11. Ensure $ORACLE_BASE directory exist on new node with correct permission
Part II: Adding an Oracle Clusterware Home to a New Node Using OUI in Interactive Mode
==============================
1. Set the DISPLAY environment variable and run the addNode.sh script from existing node (node1)
. ./ .profile
DISPLAY=ipaddress:0.0; export DISPLAY
cd $ORA_CRS_HOME/oui/bin
./addNode.sh
2.The Oracle Universal Installer (OUI) displays the Node Selection Page
3. Enter the node that you want to add and verify the entries that OUI displays on the Summary Page click Next.
4. Monitor the progress of the copy crs home to new node and verify the total size of the CRS directory
5. Execute configuration scripts by root users
a. Run the orainstRoot.sh script on the new node if OUI prompts you to do so.
ausdfsgriddb10:/u01/app/
Changing permissions of /u01/app/oracle/oraInventory to 770.
Changing groupname of /u01/app/oracle/oraInventory to oinstall.
The execution of the script is complete
b. Run the rootaddNode.sh script from the $ORA_CRS_HOME/install/ directory on the node from which you are running OUI.
ausdfsgriddb01:/u01/app/
clscfg: EXISTING configuration version 3 detected.
clscfg: version 3 is 10G Release 2.
Attempting to add 1 new nodes to the configuration
Using ports: CSS=49895 CRS=49896 EVMC=49898 and EVMR=49897.
node
node 10: ausdfsgriddb10 ausdfsgriddb10-pn ausdfsgriddb10
Creating OCR keys for user 'root', privgrp 'root'..
Operation successful.
/u01/app/oracle/product/10.2.
c. Run the root.sh script on the new node from $ORA_CRS_HOME to start Oracle Clusterware on the new node.
ausdfsgriddb10:/u01/app/
WARNING: directory '/u01/app/oracle/product/10.2.
WARNING: directory '/u01/app/oracle/product' is not owned by root
WARNING: directory '/u01/app/oracle' is not owned by root
Checking to see if Oracle CRS stack is already configured
/etc/oracle does not exist. Creating it now.
OCR LOCATIONS = /u02/oradata/ocr1,/u02/
OCR backup directory '/u01/app/oracle/product/10.2.
Setting the permissions on OCR backup directory
Setting up NS directories
Oracle Cluster Registry configuration upgraded successfully
WARNING: directory '/u01/app/oracle/product/10.2.
WARNING: directory '/u01/app/oracle/product' is not owned by root
WARNING: directory '/u01/app/oracle' is not owned by root
clscfg: EXISTING configuration version 3 detected.
clscfg: version 3 is 10G Release 2.
Successfully accumulated necessary OCR keys.
Using ports: CSS=49895 CRS=49896 EVMC=49898 and EVMR=49897.
node
node 1: ausdfsgriddb01 ausdfsgriddb01-pn ausdfsgriddb01
node 2: ausdfsgriddb02 ausdfsgriddb02-pn ausdfsgriddb02
node 3: ausdfsgriddb03 ausdfsgriddb03-pn ausdfsgriddb03
node 4: ausdfsgriddb04 ausdfsgriddb04-pn ausdfsgriddb04
node 5: ausdfsgriddb05 ausdfsgriddb05-pn ausdfsgriddb05
node 6: ausdfsgriddb06 ausdfsgriddb06-pn ausdfsgriddb06
node 7: ausdfsgriddb07 ausdfsgriddb07-pn ausdfsgriddb07
node 8: ausdfsgriddb08 ausdfsgriddb08-pn ausdfsgriddb08
clscfg: Arguments check out successfully.
NO KEYS WERE WRITTEN. Supply -force parameter to override.
-force is destructive and will destroy any previous cluster
configuration.
Oracle Cluster Registry for cluster has already been initialized
Startup will be queued to init within 30 seconds.
Adding daemons to inittab
Expecting the CRS daemons to be up within 600 seconds.
CSS is active on these nodes.
ausdfsgriddb01
ausdfsgriddb02
ausdfsgriddb03
ausdfsgriddb04
ausdfsgriddb05
ausdfsgriddb06
ausdfsgriddb07
ausdfsgriddb08
ausdfsgriddb09
ausdfsgriddb10
CSS is active on all nodes.
Waiting for the Oracle CRSD and EVMD to start
Oracle CRS stack installed and running under init(1M)
Running vipca(silent) for configuring nodeapps
Creating VIP application resource on (0) nodes.
Creating GSD application resource on (0) nodes.
Creating ONS application resource on (0) nodes.
Starting VIP application resource on (8) nodes.........
Starting GSD application resource on (8) nodes.........
Starting ONS application resource on (8) nodes.........
Done.
6. Verify crs is started on new node and nodeapps are started except for listener, and then exit OUI.
7. Obtain the remote port identifier, which you need to know for the next step, by running the following command on the existing node from the $ORA_CRS_HOME/opmn/conf directory:
$ cat $ORA_CRS_HOME/opmn/conf/ons.
localport=6113
remoteport=6201
loglevel=3
useocr=on
9.From the $ORA_CRS_HOME/bin directory on an existing node, run the Oracle Notification Service (RACGONS) utility as in the following example where remote_port is the port number from the previous step and node2 is the name of the node that you are adding:
./racgons add_config New_Node:Remote_Port
Ex:
$ ./racgons add_config ausdfsgriddb10:6201
10. Move the Oracle created S96init.crs script to S11 and chkconfig init.crs off and then back on to ensure CRS will start properly during a reboot. Finally, reboot the node and confirm. (SE)
mv /etc/init.d/rc3.d/S96init.crs /etc/init.d/rc3.d/S11init.crs
chkconfig -e init.crs (Now within vi change to on and “:wq”)
chkconfig -e init.crs (Now within vi change to off and “:wq”)
chkconfig -e init.crs (Now within vi change to on and “:wq”)
Part III: Adding an Oracle Home to a New Node Using OUI in Interactive Mode
==============================
Ensure that you have successfully installed Oracle with the Oracle RAC software on at least one node in your cluster environment. To use these procedures as shown, your $ORACLE_HOME environment variable must identify your successfully installed Oracle home.
- Go to $ORACLE_HOME/oui/bin and run the addNode.sh script on node1.
. ./ .profile
DISPLAY=ipaddress:0.0; export DISPLAY
cd $ORACLE_HOME/oui/bin
./addNode.sh
2. When OUI displays the Node Selection Page, select the node to be added and click Next.
3. Verify the entries that OUI displays on the Summary Page and click Next.
4.Run the root.sh script on the new node from Oracle_home when OUI prompts you to do so.
ausdfsgriddb10:/u01/app/
Running Oracle10 root.sh script...
The following environment variables are set as:
ORACLE_OWNER= oracle
ORACLE_HOME= /u01/app/oracle/product/10.2.
Enter the full pathname of the local bin directory: [/usr/local/bin]:
Copying dbhome to /usr/local/bin ...
Copying oraenv to /usr/local/bin ...
Copying coraenv to /usr/local/bin ...
Creating /etc/oratab file...
Entries will be added to the /etc/oratab file as needed by
Database Configuration Assistant when a database is created
Finished running generic part of root.sh script.
Now product-specific root actions will be performed.
Part IV: reconfigure listener on new node
On the new node, run the Oracle Net Configuration Assistant (NETCA) to add a Listener.
export DISPLAY=ipaddress:0.0
./netca &
Waiting for listener configuration to complete. Verify listener is up using srvctl
Part V: add ASM instance on new node manually
==============================
a. add ASM entry in /etc/oratab on new node
+ASM10:/u01/app/
b. modify init+ASM10.ora in /u01/app/oracle/product/10.2.
cd /u01/app/oracle/product/10.2.
mv init+ASM1.ora init+ASM10.ora
mv orapw+ASM1 orapw+ASM10
c. create admin directories for ASM instance
mkdir /u01/app/oracle/admin/+ASM/
d. copy init.ora to new node
cd /u01/app/oracle/admin/+ASM/
scp ausdfsgriddb01:/u01/app/
e. add new ASM instance in /u01/app/oracle/admin/+ASM/
+ASM10.instance_
f. add ASM to cluster
srvctl add asm -n ausdfsgriddb10 -i +ASM10 –o /u01/app/oracle/product/10.2.
g. start asm instance
srvctl start asm -n ausdfsgriddb10
No comments:
Post a Comment