Following steps needs to be done to extend or add node to existing RAC with ASM.
- Configure hardware on Node 3 ( rac3)
- Install Oracle Cluster software –services on rac3
- Install and create ASM instance for storage on rac3
- Install Oracle Software (RAC) database
- Configure Listener for RAC on rac3
- Add Database instance using DBCA
– Configure hardware on Node 3 ( rac3)
– Make sure private, public network card has been configure correct IP address.
– Make sure proper entries /etc/hosts file for all three IP address
– User id has been created and base directory has correct permission.
– All the OS patches has been applied
– All the nodes are accessible using ssh without password prompt.
– Adding the public and private node names for the new node to the /etc/hosts file on the existing nodes, rac1 and rac2
– Verifying the new node can be accessed (using the ping command) from the existing nodes
– Running the following command on either docrac1 or docrac2 to verify the new node has been properly configured:
Execute following from node “rac1” or “rac2”
cluvfy stage -pre crsinst -n rac3
– Install Oracle Cluster software –services on rac3
Add node or extend cluster ware – Oracle Clusterware home to the node
This will add clusterware software on “rac3” node and enable all the clusterready services
– Source the Oracle Home for Oracle Cluster ware home
– Go to CRS_home/oui/bin and run the addNode.sh script.
– OUI starts and first displays the Welcome window.
– The Specify Cluster Nodes to Add to Installation window appears.
– Select the node or nodes that you want to add. After selecting rac3, click Next.
– Verify the entries that OUI displays on the Summary Page and click Next.
– Run the rootaddNode.sh script from the CRS_home/install/ directory on rac1 when prompted to do so.
– Basically, this script adds the node applications of the new node to the OCR configuration.
– Run the orainstRoot.sh script on the node rac3
– Run the CRS_home/root.sh script on the node rac3 to start Oracle Clusterware on the new node.
– Add the new node’s Oracle Notification Services (ONS) configuration information to the shared Oracle Cluster Registry (OCR).
– Obtain the ONS port identifier used by the new node, which you need to know for the next step, by running the
– following command from the CRS_home/opmn/conf directory on the rac1 node:
– After you locate the ONS port number for the new node, you must make sure that the ONS on rac1 can communicate with the ONS on the new node, rac3.
– To verify the installation of Oracle Clusterware on the new node, you can run the following command as the root user on the newly configured node, rac3:
CRS_home/bin/cluvfy stage -post crsinst -n docrac3 –verbose
3 Install and create ASM instance for storage on rac3
Extending the Oracle Automatic Storage Management Home Directory
Next step is to configure the shared storage for the new database instances that will be
created on new node. You should be able to access sharedas well to avoid SPOF failure needs separate ASM instance.
On rac1 , source ASM oracle home
– $ASM_HOME/oui/bin directory on rac1 and
– run the addNode.sh script.
– OUI displays the Node Selection window, select the node to be added (rac3), then click Next.
– Run the root.sh script on the new node, rac3, from the ASM home directory.
4 Oracle RAC Software Home Directory
Following process create oracle database home on rac3
Source oracle home database
– Ensure that you have successfully installed the Oracle RAC software on at least one node in your cluster environment.
– To use these procedures as shown, your $ORACLE_HOME environment variable must identify your successfully installed Oracle RAC home directory.
– Go to the $ORACLE_HOME/oui/bin directory on rac1 and run the addNode.sh script.
– When OUI displays the Specify Cluster Nodes to Add to Installation window, select the node to be added (rac3), then click Next.
– Verify the entries that OUI displays in the Cluster Node Addition Summary window, then click Next.
– Run the root.sh script on the new node, rac3, from the $ORACLE_HOME directory on that node when OUI prompts you to do so.
– Now rac3 Oracle RAC is ready
- Configure Listener
Create proper new Listener on the new node using Oracle Net Configuration Assistant:
Start the Oracle Net Configuration Assistant by entering netca at the system prompt from the $ORACLE_HOME/bin directory.
- NETCA Select Listener configuration, and click Next.
- NETCA displays the Listener Configuration, Listener window.
- Select Add to create a new Listener, then click Next.
- NETCA displays the Listener Configuration, Listener Name window.
- Accept the default value of LISTENER for the Listener name by clicking Next.
- NETCA displays the Listener Configuration, Select Protocols window.
- Choose TCP and move it to the Selected Protocols area, then click Next.
- NETCA displays the Listener Configuration, TCP/IP Protocol window.
- Choose Use the standard port number of 1521, then click Next.
- NETCA displays the Real Application Clusters window.
- Select Cluster configuration for the type of configuration to perform, then click Next.
- NETCA displays the Real Application Clusters, Active Nodes window.
- Create Cluster Instance on the New Node
– On rac3 , make sure you have all the Oracle Home for database is properly sourced
– Start DBCA from $ORACLE_HOME/bin
Select Oracle Real Application Clusters database,
Select Instance Management, and then click Next.
DBCA displays the Instance Management window.
Select Add an Instance, then click Next.
DBCA displays the List of Cluster Databases window, which shows the databases and their current status
After you terminate your DBCA session, you should run the following command to verify the administrative privileges
on the new node and obtain detailed information about these privileges:
CRS_home/bin/cluvfy comp admprv -o db_config -d oracle_home -n docrac3 -verbose