2016-04-23 Install GI 12.1.0.2

Overview

Following the successful creation of two (2) VM images under XEN with OEL7 and using the revised network design, I now go ahead with the installation of Grid Infrastructure 12.1.0.2 on that cluster.

References

Preparation for Installation

Pre-installation RPM Installed

Following the procedure in 3.2 Installing the Oracle Preinstallation RPM From Unbreakable Linux Network, I ran the following command on both REDFERN1 and REDFERN2 (assuming what is valid for OEL6 is also valid for OEL7):

yum install oracle-rdbms-server-12cR1-preinstall

There was a large amount of output which I did not capture, but the installation appeared to be successful.

Changed ORACLE Password

Changed the password for the oracle user on both REDFERN1 and REDFERN2.

Permanently Change Host Name

Unfortunately, the use of hostname does not permanently change the hostname. I had to do the following (as root) on REDFERN1, and a similar thing on REDFERN2:

cat >>/etc/sysctl.conf <<DONE kernel.hostname = redfern1.yaocm.id.au DONE reboot

Create Oracle GI Software Home

Create the Oracle Grid Infrastructure Software Home as follows on both REDFERN1 and REDFERN2:

sudo mkdir /opt/app/grid_infra sudo chown oracle:oinstall /opt/app/grid_infra

Installed OEM Agent

From CRONULLA, the OEM agent was pushed out to both REDFERN1 and REDFERN2. I ignored the warning about RHEL 7 not being supported.

Change Ownership for Shared Disks

Ran the following command to change the ownership of the shared disks so that they can be discovered by ASM:

sudo chown oracle:dba /dev/xvd[d-h]

The ownership of these shared disks was verified as follows:

ls -l /dev/xvd[d-h]

And the ouput is:

brw-rw----. 1 oracle dba 202, 48 Jan 1 20:48 /dev/xvdd brw-rw----. 1 oracle dba 202, 64 Jan 1 20:48 /dev/xvde brw-rw----. 1 oracle dba 202, 80 Jan 1 20:48 /dev/xvdf brw-rw----. 1 oracle dba 202, 96 Jan 1 20:48 /dev/xvdg brw-rw----. 1 oracle dba 202, 112 Jan 1 20:48 /dev/xvdh

Locate GI Software

Instead of unloading the GRID software onto the local host, I now use the NFS directory as set up through Use NFS for Oracle Software.

Start Installation

From PENRITH, I ran the following commands in an XTerm session:

xhost + 192.168.1.140 ssh -Y oracle@192.168.1.140

Once on REDFERN1, I ran the following commands:

cd /opt/share/Software/grid/linuxamd64_12102/grid ./runInstaller

Installation Attempt #1

Step 1: Installation Option

Left the default option as Install and Configure Oracle Grid Infrastructure:

Clicked Next.

Step 2: Cluster Type

Left the default option as Configure a Standard Cluster:

Clicked Next.

Step 3: Installation Type

Changed option to Advanced Installation:

Clicked Next.

Step 4: Select Product Language

Left the default language English:

Clicked Next.

Step 5: Grid Plug and Play Information

Made the following changes:

Scan Name:

Configure GNS:

redfern-crs.yaocm.id.au

No

Clicked Next.

Step 6: Cluster Node Information

Got the following screen:

Clicked Add....

Added the following details for REDFERN2 as follows:

Clicked OK to get the following screen:

Clicked SSH connectivity..., and entered the password:

Clicked Setup. After a few minutes, the following message appears:

Clicked OK. Then clicked Next.

Step 7: Network Interface Usage

Got the following screen—no changes were made:

Clicked Next.

Step 8: Storage Option Information

No changes were made on the following screen:

Clicked Next.

Got a warning message which was expanded to get the following details:

Clicked OK.

Clicked Back to go back to Step 7 and made the following change to use eth1 as private only:

Clicked Next and on the next screen as well.

Step 9: Create ASM Disk Group

On the following screen,

    • Changed Redundancy to External.

Clicked on Change Discovery Path... to /dev/xvd* as follows:

Clicked OK.

Created VOTE disk group on /dev/xvdh:

Clicked Next.

Got a warning message which was expanded to get the following details:

Clicked OK.

Clicked Cancel to stop the installation.

Correct Size of VOTE Disk

Shutdown REDFERN Cluster

I shutdown the REDFERN cluster.

Increased Size of VOTE Disk Group

On VICTORIA, I increased the size of the shared VOTE_01 disk as follows:

cd /OVS/running_pool/REDFERN/shared dd if=/dev/zero of=VOTE_01 bs=1G count=6

Start Up REDFERN Cluster

The REDFERN cluster was then started.

Change Disk Ownership

Once the cluster was running, I had to change the ownership of the shared disks again:

sudo chown oracle:dba /dev/xvd[d-h]

Installation Attempt #2

The installation process was started again.

Step 09: Create ASM Disk Group with Larger Disk

Now the screen is:

Clicked Next.

Step 10: ASM Password

Filled in the passwords on the following screen:

Clicked Next.

Step 11: Failure Isolation

Only one (1) option was available on the following screen:

Clicked Next.

Step 12: Management Options

Filled in the OEM details as follows:

Clicked Next.

Step 13: Operating System Groups

Filled in the OS group details as follows:

Clicked Next.

Got the following message after expanding the details:

Clicked Yes.

Step 14: Installation Location

Set up the software location as follows:

Clicked Next.

Step 15: Root script execution

Since this is a home system, I can use root's password as follows:

Clicked Next.

Step 16: Prerequisite Checks

The following progress screen appears:

After a while, the following results appeared:

There is only one (1) failed check. The details for the resolv.conf integrity are:

This can be ignored as the domain (yaocm.id.au) is only defined on my home network.

Clicked Close, then Fix & Check Again. The following screen appears:

Clicked OK to let OUI run the fix-up script using the credentials provided in Step 15. The following screen shows that the fix-up was successful:

However, the following problems remain:

Chose to ignore all of these problems, and clicked Next. The following warning appears:

Clicked Yes.

Step 17: Summary

The following summary screen appears:

Clicked Install.

Step 18: Install Product

The following progress screen appears:

After a while, the following message appears:

Clicked Yes.

However, this failed with the following error message:

According to the log at ~/oraInventory/logs/installActions2016-01-02_12-50-14PM.log:

2016/01/02 21:01:27 CLSRSC-12: The ASM resource ora.asm did not start 2016/01/02 21:01:27 CLSRSC-258: Failed to configure and start ASM Died at /opt/app/grid_infra/12.1.0/grid/crs/install/crsinstall.pm line 2017. The command '/opt/app/grid_infra/12.1.0/grid/perl/bin/perl -I/opt/app/grid_infra/12.1.0/grid/perl/lib -I/opt/app/grid_infra/12.1.0/grid/crs/install /opt/app/grid_infra/12.1.0/grid/crs /install/rootcrs.pl -auto -lang=en_AU.UTF-8' execution failed

And in /opt/app/grid_infra/12.1.0/grid/cfgtoollogs/crsconfig/rootcrs_redfern1_2016-01-02_08-53-53PM.log:

2016-01-02 21:01:21: Executing cmd: /opt/app/grid_infra/12.1.0/grid/bin/crsctl status resource ora.asm -init 2016-01-02 21:01:22: Checking the status of ora.asm 2016-01-02 21:01:27: Executing cmd: /opt/app/grid_infra/12.1.0/grid/bin/clsecho -p has -f clsrsc -m 12 2016-01-02 21:01:27: Command output: > CLSRSC-12: The ASM resource ora.asm did not start >End Command output