11g RAC完全安装图解(VM Workstation8.0+Centos5.5)01-Grid安装

Created By Cryking 转载请注明出处,谢谢

环境:
VMware Workstation8.0 +Centos 5.5(32位)+Oracle 11.2.0.1.0
两个节点crydb01(节点1)和crydb02(节点2)

一、设置主机名及 网络环境配置
在节点1上:
设置主机名

# hostname crydb01
# vi /etc/sysconfig/network
NETWORKING=yes
NETWORKING_IPV6=yes
HOSTNAME=crydb01
GATEWAY=192.168.123.1
# vi /etc/hosts
127.0.0.1               localhost.localdomain localhost
#Public IP
192.168.123.109 crydb01
192.168.123.108 crydb02
#Private IP
10.0.0.100      crydb01-pri
10.0.0.101      crydb02-pri
#VIP
192.168.123.209 crydb01-vip
192.168.123.208 crydb02-vip
#scan-ip
192.168.123.200 crydb-scan
::1             localhost6.localdomain6 localhost6

# vi /etc/sysconfig/network-scripts/ifcfg-eth0
DEVICE=eth0
BOOTPROTO=static
ONBOOT=yes
IPADDR=192.168.123.109
NETMASK=255.255.255.0
GATEWAY=192.168.123.1
HWADDR=00:0c:29:6c:59:9c

节点1的虚拟机添加网卡:
11g RAC完全安装图解(VM Workstation8.0+Centos5.5)01-Grid安装_第1张图片
11g RAC完全安装图解(VM Workstation8.0+Centos5.5)01-Grid安装_第2张图片
11g RAC完全安装图解(VM Workstation8.0+Centos5.5)01-Grid安装_第3张图片

添加完网卡后重启系统并设置eth1的ip:

# reboot
# vi /etc/sysconfig/network-scripts/ifcfg-eth1
DEVICE=eth1
BOOTPROTO=static
ONBOOT=yes
IPADDR=10.0.0.100
NETMASK=255.255.255.0
HWADDR=00:0c:29:6c:59:a6
# service network restart

在节点2上:

# hostname crydb02
# vi /etc/sysconfig/network
NETWORKING=yes
NETWORKING_IPV6=yes
HOSTNAME=crydb02
GATEWAY=192.168.123.1

# vi /etc/hosts
127.0.0.1               localhost.localdomain localhost
#Public IP
192.168.123.109 crydb01
192.168.123.108 crydb02
#Private IP
10.0.0.100      crydb01-pri
10.0.0.101      crydb02-pri
#VIP
192.168.123.209 crydb01-vip
192.168.123.208 crydb02-vip
#scan-ip
192.168.123.200 scan
::1             localhost6.localdomain6 localhost6

# vi /etc/sysconfig/network-scripts/ifcfg-eth0
DEVICE=eth0
BOOTPROTO=static
ONBOOT=yes
IPADDR=192.168.123.108
NETMASK=255.255.255.0
GATEWAY=192.168.123.1
HWADDR=00:0c:29:6c:59:9c

重复上面的添加网卡过程,然后重启系统,然后:

# vi /etc/sysconfig/network-scripts/ifcfg-eth1
DEVICE=eth1
BOOTPROTO=static
ONBOOT=yes
IPADDR=10.0.0.101
NETMASK=255.255.255.0
HWADDR=00:0c:29:85:3b:3f

二、 用户及用户组设置
分别在两个节点上以root用户进行如下操作:

# groupadd -g 500 oinstall
# groupadd -g 501 dba
# groupadd -g 1020 asmadmin
# groupadd -g 1022 asmoper
# groupadd -g 1021 asmdba
# useradd -u 1100 -g oinstall -G dba,asmadmin,asmdba,asmoper -d /home/grid -s /bin/bash grid
# useradd -u 1101 -g oinstall -G dba,asmdba -d /home/oracle -s /bin/bash oracle
# passwd grid
# passwd oracle
# mkdir -p /u01/app/11.2.0/grid
# mkdir -p /u01/app/grid
# mkdir -p /u01/app/oracle
# chown grid:oinstall /u01/app/11.2.0/grid
# chown grid:oinstall /u01/app/grid
# chown oracle:oinstall /u01/app/oracle
# chmod -R 775 /u01/

三、配置用户SSH等效性
先配置Oracle用户SSH等效性.
在节点1上:

[oracle@crydb01 ~]$ su - oracle
[oracle@crydb01 ~]$ mkdir ./ssh
[oracle@crydb01 ~]$ chmod 700 ./ssh
[oracle@crydb01 ~]$ /usr/bin/ssh-keygen -t rsa
Generating public/private rsa key pair.
Enter file in which to save the key (/home/oracle/.ssh/id_rsa): 
Created directory '/home/oracle/.ssh'.
Enter passphrase (empty for no passphrase): 
Enter same passphrase again: 
Your identification has been saved in /home/oracle/.ssh/id_rsa.
Your public key has been saved in /home/oracle/.ssh/id_rsa.pub.
The key fingerprint is:
d1:85:66:a2:78:65:8c:db:81:f6:90:52:21:4b:b2:6b oracle@crydb01
[oracle@crydb01 ~]$ /usr/bin/ssh-keygen -t dsa
Generating public/private dsa key pair.
Enter file in which to save the key (/home/oracle/.ssh/id_dsa): 
Enter passphrase (empty for no passphrase): 
Enter same passphrase again: 
Your identification has been saved in /home/oracle/.ssh/id_dsa.
Your public key has been saved in /home/oracle/.ssh/id_dsa.pub.
The key fingerprint is:
0f:10:50:57:c5:2f:06:b2:61:52:5a:61:5b:18:9c:1d oracle@crydb01
[oracle@crydb01 ssh]$ cat ~/.ssh/id_rsa.pub >>~/.ssh/authorized_keys
[oracle@crydb01 ssh]$ cat ~/.ssh/id_dsa.pub >>~/.ssh/authorized_keys
[oracle@crydb01 ~]$ chmod 600 ~/.ssh/authorized_keys

在节点2上进行同样操作:

[root@crydb02 ~]# su - oracle
[oracle@crydb02 ~]$ mkdir ./ssh
[oracle@crydb02 ~]$ chmod 700 ./ssh
[oracle@crydb02 ~]$ /usr/bin/ssh-keygen -t rsa
Generating public/private rsa key pair.
Enter file in which to save the key (/home/oracle/.ssh/id_rsa): 
/home/oracle/.ssh/id_rsa already exists.
Overwrite (y/n)? y
Enter passphrase (empty for no passphrase): 
Enter same passphrase again: 
Your identification has been saved in /home/oracle/.ssh/id_rsa.
Your public key has been saved in /home/oracle/.ssh/id_rsa.pub.
The key fingerprint is:
ff:6b:3e:b5:39:80:2b:37:88:d6:fe:cf:c6:b5:3b:ec oracle@crydb02
[oracle@crydb02 ~]$ /usr/bin/ssh-keygen -t dsa
Generating public/private dsa key pair.
Enter file in which to save the key (/home/oracle/.ssh/id_dsa): 
/home/oracle/.ssh/id_dsa already exists.
Overwrite (y/n)? y
Enter passphrase (empty for no passphrase): 
Enter same passphrase again: 
Your identification has been saved in /home/oracle/.ssh/id_dsa.
Your public key has been saved in /home/oracle/.ssh/id_dsa.pub.
The key fingerprint is:
aa:b8:46:02:60:aa:83:72:df:43:d1:1a:b6:56:a9:14 oracle@crydb02

再回到节点1上:

[oracle@crydb01 ~]$ cat ~/.ssh/id_rsa.pub >>~/.ssh/authorized_keys
[oracle@crydb01 ~]$ cat ~/.ssh/id_dsa.pub >>~/.ssh/authorized_keys
[oracle@crydb01 ssh]$ ssh crydb02 cat ~/.ssh/id_rsa.pub >>~/.ssh/authorized_keys
oracle@crydb02's password: [oracle@crydb01 ssh]$ ssh crydb02 cat ~/.ssh/id_dsa.pub >>~/.ssh/authorized_keys oracle@crydb02's password:
[oracle@crydb01 ssh]$ scp ~/.ssh/authorized_keys crydb02:~/.ssh/authorized_keys
oracle@crydb02's password:

验证在节点2上:

[oracle@crydb02 ~]$ ssh crydb01 date
[oracle@crydb02 ~]$ ssh crydb02 date

在节点1上:

[oracle@crydb01 ~]$ ssh crydb01 date
[oracle@crydb01 ~]$ ssh crydb02 date

如果都不用输入密码则配置成功.
注意:如果确认已按上面的执行,但配置总是不成功的时候,请确保两个节点的/home/oracle的权限一致,或直接使用root在所有节点执行chmod -R 700 /home/oracle.

在grid用户下也重复进行上面的配置.
这样SSH用户等效性就配置完成了.

四、 创建共享磁盘
如图:

11g RAC完全安装图解(VM Workstation8.0+Centos5.5)01-Grid安装_第4张图片
这里我只给了10G磁盘,勾选上立即分配所有空间,分配后点击增加的磁盘,在右边选择高级设置(Advanced),如下:
11g RAC完全安装图解(VM Workstation8.0+Centos5.5)01-Grid安装_第5张图片

关闭节点1的系统
[root@crydb01 ~]# halt

编辑crydb01的虚拟机文件(*.vmx):
新增的磁盘内容如下:

scsi1.present = “TRUE”
scsi1:0.present = “TRUE”
scsi1:0.fileName = “Red Hat Linux02_shared.vmdk”
scsi1:0.mode = “independent-persistent”
scsi1:0.redo = “”
在最后加上:

diskLib.dataCacheMaxSize=0
diskLib.dataCacheMaxReadAheadSize=0
diskLib.dataCacheMinReadAheadSize=0
diskLib.dataCachePageSize=4096
diskLib.maxUnsyncedWrites="0"

disk.locking="FALSE"
scsi1.sharedBus="virtual"
scsi1.virtualDev = "lsilogic"

启动节点1,发现没问题后再在节点2的虚拟机上进行如下操作:
11g RAC完全安装图解(VM Workstation8.0+Centos5.5)01-Grid安装_第6张图片
这里注意选使用现有的虚拟磁盘
11g RAC完全安装图解(VM Workstation8.0+Centos5.5)01-Grid安装_第7张图片
11g RAC完全安装图解(VM Workstation8.0+Centos5.5)01-Grid安装_第8张图片
然后关闭节点2的虚拟机后,进行和节点1一样的磁盘高级设置:
11g RAC完全安装图解(VM Workstation8.0+Centos5.5)01-Grid安装_第9张图片
11g RAC完全安装图解(VM Workstation8.0+Centos5.5)01-Grid安装_第10张图片
这里虚拟驱动代码要和节点1的保持一致.
然后编辑节点2的虚拟机文件(*.vmx),添加和节点1一样的内容.

期间发生一个问题,就是新增的SCSI盘系统总是无法认到,后来在vmx文件中添加这一行后就正常了:
scsi1.virtualDev = “lsilogic”

回到节点1,以root登录:

[root@crydb01 ~]# fdisk -l

Disk /dev/hda: 26.8 GB, 26843545600 bytes
255 heads, 63 sectors/track, 3263 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes

   Device Boot      Start         End      Blocks   Id  System
/dev/hda1   *           1          13      104391   83  Linux
/dev/hda2              14        1044     8281507+  83  Linux
/dev/hda3            1045        3002    15727635   83  Linux
/dev/hda4            3003        3263     2096482+   5  Extended
/dev/hda5            3003        3263     2096451   82  Linux swap / Solaris

Disk /dev/sda: 10.7 GB, 10737418240 bytes
255 heads, 63 sectors/track, 1305 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes

Disk /dev/sda doesn't contain a valid partition table

可以看到新加的磁盘/dev/sda,对它进行分区,如下:

[root@crydb01 ~]# fdisk /dev/sda
Device contains neither a valid DOS partition table, nor Sun, SGI or OSF disklabel
Building a new DOS disklabel. Changes will remain in memory only,
until you decide to write them. After that, of course, the previous
content won't be recoverable.


The number of cylinders for this disk is set to 1305. There is nothing wrong with that, but this is larger than 1024, and could in certain setups cause problems with: 1) software that runs at boot time (e.g., old versions of LILO) 2) booting and partitioning software from other OSs (e.g., DOS FDISK, OS/2 FDISK) Warning: invalid flag 0x0000 of partition table 4 will be corrected by w(rite) Command (m for help): n Command action e extended p primary partition (1-4) p Partition number (1-4): 1 First cylinder (1-1305, default 1): Using default value 1 Last cylinder or +size or +sizeM or +sizeK (1-1305, default 1305): Using default value 1305 Command (m for help): p Disk /dev/sda: 10.7 GB, 10737418240 bytes 255 heads, 63 sectors/track, 1305 cylinders Units = cylinders of 16065 * 512 = 8225280 bytes Device Boot Start End Blocks Id System /dev/sda1 1 1305 10482381 83 Linux Command (m for help): w The partition table has been altered! Calling ioctl() to re-read partition table. Syncing disks. [root@crydb01 rpm]# partprobe

再去到节点2上查看:

[root@crydb02 ~]# fdisk -l

Disk /dev/hda: 26.8 GB, 26843545600 bytes
255 heads, 63 sectors/track, 3263 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes

   Device Boot      Start         End      Blocks   Id  System
/dev/hda1   *           1          13      104391   83  Linux
/dev/hda2              14        1044     8281507+  83  Linux
/dev/hda3            1045        3002    15727635   83  Linux
/dev/hda4            3003        3263     2096482+   5  Extended
/dev/hda5            3003        3263     2096451   82  Linux swap / Solaris

Disk /dev/sda: 10.7 GB, 10737418240 bytes
255 heads, 63 sectors/track, 1305 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes

   Device Boot      Start         End      Blocks   Id  System
/dev/sda1               1        1305    10482381   83  Linux

说明已经成功建立共享磁盘.

五、 集群配置检查
如下检查响应的rpm包是否安装:
[root@crydb01 dev]# rpm -qa | grep -i binutils-2.*
binutils-2.17.50.0.6-14.el5

如果没有就需要安装.
安装11g RAC需要的依赖包列表(也可在Concept上去找):

    binutils-2.17.50.0.6 
    compat-libstdc++-33-3.2.3 
    compat-libstdc++-33-3.2.3 (32 bit) 
    elfutils-libelf-0.125 
    elfutils-libelf-devel-0.125 
    elfutils-libelf-devel-static-0.125 
    gcc-4.1.2 
    gcc-c++-4.1.2 
    glibc-2.5-24 
    glibc-2.5-24 (32 bit) 
    glibc-common-2.5 
    glibc-devel-2.5 
    glibc-devel-2.5 (32 bit) 
    glibc-headers-2.5 
    ksh-20060214 
    libaio-0.3.106 
    libaio-0.3.106 (32 bit) 
    libaio-devel-0.3.106 
    libaio-devel-0.3.106 (32 bit) 
    libgcc-4.1.2 
    libgcc-4.1.2 (32 bit) 
    libstdc++-4.1.2 
    libstdc++-4.1.2 (32 bit) 
    libstdc++-devel 4.1.2 
    make-3.81 
    pdksh-5.2.14 
    sysstat-7.0.2 
    unixODBC-2.2.11 
    unixODBC-2.2.11 (32 bit) 
    unixODBC-devel-2.2.11 
    unixODBC-devel-2.2.11 (32 bit) 

在节点1上执行检查:

[root@crydb01 oracle]# su - grid
[grid@crydb01 grid]$ ./grid/runcluvfy.sh stage -pre crsinst -n crydb01,crydb02 -fixup -verbose
Performing pre-checks for cluster services setup 

Checking node reachability...

Check: Node reachability from node "crydb01"
  Destination Node                      Reachable?              
  ------------------------------------  ------------------------
  crydb01                               yes                     
  crydb02                               yes                     
Result: Node reachability check passed from node "crydb01"
...

仔细检查各个未通过的项目并解决.
注:如两个节点时间不同步,则可能出现
PRVF-5415 : Check to see if NTP daemon is running failed
在两个节点上启动ntpd服务: service ntpd start
如果还出现:
PRVF-5436 : The NTP daemon running on one or more nodes lacks the slewing option
则分别在两个节点上编辑文件如下(添加-x):

# vi /etc/sysconfig/ntpd
# Drop root to id 'ntp:ntp' by default. OPTIONS="-x -u ntp:ntp -p /var/run/ntpd.pid" # Set to 'yes' to sync hw clock after successful ntpdate SYNC_HWCLOCK=no # Additional options for ntpdate NTPDATE_OPTIONS=""

这里我有个Result: Swap space check failed,不用理它.
验证OS和硬件:

$ ./runcluvfy.sh stage -post hwos -n crydb01,crydb02 -verbose
...
Post-check for hardware and operating system setup was successful

六、用户环境配置
在两个节点上关闭防火墙:
/etc/rc.d/init.d/iptables stop

分别在两个节点上:

# su - grid
$ vi .bash_profile
# .bash_profile

# Get the aliases and functions
if [ -f ~/.bashrc ]; then
        . ~/.bashrc
fi
ORACLE_SID=+ASM1;export ORACLE_SID
GRID_BASE=/u01/app/grid;export GRID_BASE
ORACLE_BASE=$GRID_BASE;export ORACLE_BASE

GRID_HOME=/u01/app/11.2.0/grid; export GRID_HOME
ORACLE_HOME=$GRID_HOME;export ORACLE_HOME

NLS_DATE_FORMAT="YYYY-MM-DD HH24:MI:SS";export NLS_DATE_FORMAT

TNS_ADMIN=$GRID_HOME/network/admin;export TNS_ADMIN

CLASSPATH=$GRID_HOME/JRE
CLASSPATH=${CLASSPATH}:$GRID_HOME/jdbc/lib/ojdbc6.jar
CLASSPATH=${CLASSPATH}:$GRID_HOME/jlib
CLASSPATH=${CLASSPATH}:$GRID_HOME/rdbms/jlib
CLASSPATH=${CLASSPATH}:$GRID_HOME/network/jlib
export CLASSPATH

export TEMP=/tmp
export TMPDIR=/tmp
# User specific environment and startup programs

PATH=$GRID_HOME/bin:$PATH:$HOME/bin

export PATH

umask 022

:wq

$ su - oracle
$ vi .bash_profile
# Get the aliases and functions
if [ -f ~/.bashrc ]; then
        . ~/.bashrc
fi

# User specific environment and startup programs
ORACLE_UNQNAME=crydb;export ORACLE_UNQNAME
ORACLE_BASE=/u01/app/oracle;export ORACLE_BASE
ORACLE_SID=crydb01;export ORACLE_SID
ORACLE_HOME=$ORACLE_BASE/product/11.2.0/dbhome_1;export ORACLE_HOME
ORACLE_HOME_LISTNER=$ORACLE_HOME;export ORACLE_LISTNER
NLS_DATE_FORMAT="YYYY-MM-DD HH24:MI:SS"; export NLS_DATE_FORMAT
TNS_ADMIN=$ORACLE_HOME/network/admin;export TNS_ADMIN
ORACLE_TRACE=$ORACLE_BASE/diag/rdbms/$ORACLE_SID/$ORACLE_SID/trace;export ORACLE_TRACE
PATH=$ORACLE_HOME/bin:$PATH:$HOME/bin

export PATH

alias sqlplus='rlwrap sqlplus'

export TEMP=/tmp
export TMPDIR=/tmp
umask 022

:wq

设置资源限制,分别在两个节点下添加:

[root@crydb01 /]# vi /etc/security/limits.conf
# Oracle
oracle              soft    nproc   2047
oracle              hard    nproc   16384
oracle              soft    nofile  1024
oracle              hard    nofile  65536
oracle              soft    stack   10240
# Grid
grid              soft    nproc   2047
grid              hard    nproc   16384
grid              soft    nofile  1024
grid              hard    nofile  65536
grid              soft    stack   10240

在/etc/profile文件最后加上:

# Oracle
if [ $USER = "oracle" ]|| [ $USER = "grid" ]; then
   if [ $SHELL = "/bin/ksh" ]; then
      ulimit -p 16384
      ulimit -n 65536
   else
      ulimit -u 16384 -n 65536
   fi
   umask 022
fi

七、为ASM安装、配置ASMLib
提示:配置ASM不一定需要ASMLib,也可使用标准Linux I/O调用来管理裸设备.
用root在两个节点上执行:

[root@crydb01 ~]# rpm -Uvh oracleasm-2.6.18-194.el5-2.0.5-1.el5.i686.rpm \
> oracleasmlib-2.0.4-1.el5.i386.rpm \
> oracleasm-support-2.1.7-1.el5.i386.rpm

验证是否成功安装:

[root@crydb01 ~]# rpm -qa --queryformat "%{NAME}-%{VERSION}-%{RELEASE} (%{ARCH})\n"| grep oracleasm | sort
oracleasm-2.6.18-194.el5-2.0.5-1.el5 (i686)
oracleasmlib-2.0.4-1.el5 (i386)
oracleasm-support-2.1.7-1.el5 (i386)

配置ASMLib:

[root@crydb01 ~]# /usr/sbin/oracleasm configure -i
Configuring the Oracle ASM library driver.

This will configure the on-boot properties of the Oracle ASM library
driver.  The following questions will determine whether the driver is
loaded on boot and what permissions it will have.  The current values
will be shown in brackets ('[]').  Hitting <ENTER> without typing an
answer will keep that current value.  Ctrl-C will abort.

Default user to own the driver interface []: grid
Default group to own the driver interface []: asmadmin
Start Oracle ASM library driver on boot (y/n) [n]: y
Scan for Oracle ASM disks on boot (y/n) [y]: y
Writing Oracle ASM library driver configuration: done
[root@crydb01 ~]# /usr/sbin/oracleasm init
Creating /dev/oracleasm mount point: /dev/oracleasm
Loading module "oracleasm": oracleasm
Mounting ASMlib driver filesystem: /dev/oracleasm

创建ASM磁盘(只在节点1执行):

[root@crydb01 ~]# /usr/sbin/oracleasm createdisk DISK1 /dev/sda1
Writing disk header: done
Instantiating disk: done

在节点2上执行扫描磁盘:

[root@crydb02 ~]# /usr/sbin/oracleasm scandisks
Reloading disk partitions: done
Cleaning any stale ASM disks...
Scanning system for ASM disks...
Instantiating disk "DISK1"

查看下当前的ASM磁盘(分别在两个节点上执行):

[root@crydb01 ~]# /usr/sbin/oracleasm listdisks
DISK1

八、正式安装Grid
下面开始安装grid和ASM(只在节点1上执行):
在此之前需要保证X11远程显示正常(DISPLAY设置正确),我这里是直接使用的Xshell,不用配置

[grid@crydb01 ~]$ cd /home/grid/grid
[grid@crydb01 grid]$ ./runInstaller
Starting Oracle Universal Installer...

Checking Temp space: must be greater than 80 MB.   Actual 3283 MB    Passed
Checking swap space: must be greater than 150 MB.   Actual 2047 MB    Passed
Checking monitor: must be configured to display at least 256 colors.    Actual 16777216    Passed
Preparing to launch Oracle Universal Installer from /tmp/OraInstall2015-02-12_11-10-08PM. Please wait ...


11g RAC完全安装图解(VM Workstation8.0+Centos5.5)01-Grid安装_第11张图片

11g RAC完全安装图解(VM Workstation8.0+Centos5.5)01-Grid安装_第12张图片
注意上图中的SCAN Name要和hosts中的一致
11g RAC完全安装图解(VM Workstation8.0+Centos5.5)01-Grid安装_第13张图片
下面一步是添加节点并配置SSH的连通性:
11g RAC完全安装图解(VM Workstation8.0+Centos5.5)01-Grid安装_第14张图片

上图中点击”Setup”.
11g RAC完全安装图解(VM Workstation8.0+Centos5.5)01-Grid安装_第15张图片
SSH连通性建立成功:
11g RAC完全安装图解(VM Workstation8.0+Centos5.5)01-Grid安装_第16张图片
11g RAC完全安装图解(VM Workstation8.0+Centos5.5)01-Grid安装_第17张图片
11g RAC完全安装图解(VM Workstation8.0+Centos5.5)01-Grid安装_第18张图片
因为只创建了一块共享磁盘,所以选外部(External)
11g RAC完全安装图解(VM Workstation8.0+Centos5.5)01-Grid安装_第19张图片
11g RAC完全安装图解(VM Workstation8.0+Centos5.5)01-Grid安装_第20张图片
11g RAC完全安装图解(VM Workstation8.0+Centos5.5)01-Grid安装_第21张图片
11g RAC完全安装图解(VM Workstation8.0+Centos5.5)01-Grid安装_第22张图片

下面是先决条件检查:
11g RAC完全安装图解(VM Workstation8.0+Centos5.5)01-Grid安装_第23张图片

这里我的ntpd有问题,重启的时候报了:
ntpd: Synchronizing with time server: [FAILED]
后来发现是DNS配置(/etc/resolv.conf)有问题,改好后就正常了.

还有个错误(是dba组不是grid的主组导致),点击”Fix & Check Again”,找到修复脚本路径,以root登录执行下就可以了–文中上面的脚本已修改,应不会出现此错误了.
如我这里是:

[root@crydb01 ntp]# cd /tmp/CVU_11.2.0.1.0_grid/
[root@crydb01 CVU_11.2.0.1.0_grid]# ./runfixup.sh 
Response file being used is :./fixup.response
Enable file being used is :./fixup.enable
Log file location: ./orarun.log
uid=1100(grid) gid=500(oinstall) groups=500(oinstall),501(dba),1020(asmadmin),1022(asmoper),1021(asmdba)

11g RAC完全安装图解(VM Workstation8.0+Centos5.5)01-Grid安装_第24张图片
这里可以保存响应文件,方便后面做静默安装使用.

开始漫长的安装了…

11g RAC完全安装图解(VM Workstation8.0+Centos5.5)01-Grid安装_第25张图片

分别在节点1、2上以root执行图中提示的脚本:

[root@crydb01 CVU_11.2.0.1.0_grid]# cd /u01/app/11.2.0/grid/
[root@crydb01 grid]# ./root.sh
Running Oracle 11g root.sh script...

The following environment variables are set as: ORACLE_OWNER= grid ORACLE_HOME= /u01/app/11.2.0/grid Enter the full pathname of the local bin directory: [/usr/local/bin]: The file "dbhome" already exists in /usr/local/bin. Overwrite it? (y/n) [n]: y Copying dbhome to /usr/local/bin ... The file "oraenv" already exists in /usr/local/bin. Overwrite it? (y/n) [n]: y Copying oraenv to /usr/local/bin ... The file "coraenv" already exists in /usr/local/bin. Overwrite it? (y/n) [n]: y Copying coraenv to /usr/local/bin ... Entries will be added to the /etc/oratab file as needed by Database Configuration Assistant when a database is created Finished running generic part of root.sh script. Now product-specific root actions will be performed. 2015-02-13 00:38:49: Parsing the host name 2015-02-13 00:38:49: Checking for super user privileges 2015-02-13 00:38:49: User has super user privileges Using configuration parameter file: /u01/app/11.2.0/grid/crs/install/crsconfig_params Creating trace directory LOCAL ADD MODE Creating OCR keys for user 'root', privgrp 'root'.. Operation successful. root wallet root wallet cert root cert export peer wallet profile reader wallet pa wallet peer wallet keys pa wallet keys peer cert request pa cert request peer cert pa cert peer root cert TP profile reader root cert TP pa root cert TP peer pa cert TP pa peer cert TP profile reader pa cert TP profile reader peer cert TP peer user cert pa user cert Adding daemon to inittab CRS-4123: Oracle High Availability Services has been started. ohasd is starting ADVM/ACFS is not supported on centos-release-5-5.el5.centos CRS-2672: Attempting to start 'ora.gipcd' on 'crydb01' CRS-2672: Attempting to start 'ora.mdnsd' on 'crydb01' CRS-2676: Start of 'ora.gipcd' on 'crydb01' succeeded CRS-2676: Start of 'ora.mdnsd' on 'crydb01' succeeded CRS-2672: Attempting to start 'ora.gpnpd' on 'crydb01' CRS-2676: Start of 'ora.gpnpd' on 'crydb01' succeeded CRS-2672: Attempting to start 'ora.cssdmonitor' on 'crydb01' CRS-2676: Start of 'ora.cssdmonitor' on 'crydb01' succeeded CRS-2672: Attempting to start 'ora.cssd' on 'crydb01' CRS-2672: Attempting to start 'ora.diskmon' on 'crydb01' CRS-2676: Start of 'ora.diskmon' on 'crydb01' succeeded CRS-2676: Start of 'ora.cssd' on 'crydb01' succeeded CRS-2672: Attempting to start 'ora.ctssd' on 'crydb01' CRS-2676: Start of 'ora.ctssd' on 'crydb01' succeeded ASM created and started successfully. DiskGroup DATA created successfully. clscfg: -install mode specified Successfully accumulated necessary OCR keys. Creating OCR keys for user 'root', privgrp 'root'.. Operation successful. CRS-2672: Attempting to start 'ora.crsd' on 'crydb01' CRS-2676: Start of 'ora.crsd' on 'crydb01' succeeded CRS-4256: Updating the profile Successful addition of voting disk 23c34f816b8a4ffabf6414ab4123ef3d. Successfully replaced voting disk group with +DATA. CRS-4256: Updating the profile CRS-4266: Voting file(s) successfully replaced ## STATE File Universal Id File Name Disk group -- ----- ----------------- --------- --------- 1. ONLINE 23c34f816b8a4ffabf6414ab4123ef3d (ORCL:DISK1) [DATA] Located 1 voting disk(s). CRS-2673: Attempting to stop 'ora.crsd' on 'crydb01' CRS-2677: Stop of 'ora.crsd' on 'crydb01' succeeded CRS-2673: Attempting to stop 'ora.asm' on 'crydb01' CRS-2677: Stop of 'ora.asm' on 'crydb01' succeeded CRS-2673: Attempting to stop 'ora.ctssd' on 'crydb01' CRS-2677: Stop of 'ora.ctssd' on 'crydb01' succeeded CRS-2673: Attempting to stop 'ora.cssdmonitor' on 'crydb01' CRS-2677: Stop of 'ora.cssdmonitor' on 'crydb01' succeeded CRS-2673: Attempting to stop 'ora.cssd' on 'crydb01' CRS-2677: Stop of 'ora.cssd' on 'crydb01' succeeded CRS-2673: Attempting to stop 'ora.gpnpd' on 'crydb01' CRS-2677: Stop of 'ora.gpnpd' on 'crydb01' succeeded CRS-2673: Attempting to stop 'ora.gipcd' on 'crydb01' CRS-2677: Stop of 'ora.gipcd' on 'crydb01' succeeded CRS-2673: Attempting to stop 'ora.mdnsd' on 'crydb01' CRS-2677: Stop of 'ora.mdnsd' on 'crydb01' succeeded CRS-2672: Attempting to start 'ora.mdnsd' on 'crydb01' CRS-2676: Start of 'ora.mdnsd' on 'crydb01' succeeded CRS-2672: Attempting to start 'ora.gipcd' on 'crydb01' CRS-2676: Start of 'ora.gipcd' on 'crydb01' succeeded CRS-2672: Attempting to start 'ora.gpnpd' on 'crydb01' CRS-2676: Start of 'ora.gpnpd' on 'crydb01' succeeded CRS-2672: Attempting to start 'ora.cssdmonitor' on 'crydb01' CRS-2676: Start of 'ora.cssdmonitor' on 'crydb01' succeeded CRS-2672: Attempting to start 'ora.cssd' on 'crydb01' CRS-2672: Attempting to start 'ora.diskmon' on 'crydb01' CRS-2676: Start of 'ora.diskmon' on 'crydb01' succeeded CRS-2676: Start of 'ora.cssd' on 'crydb01' succeeded CRS-2672: Attempting to start 'ora.ctssd' on 'crydb01' CRS-2676: Start of 'ora.ctssd' on 'crydb01' succeeded CRS-2672: Attempting to start 'ora.asm' on 'crydb01' CRS-2676: Start of 'ora.asm' on 'crydb01' succeeded CRS-2672: Attempting to start 'ora.crsd' on 'crydb01' CRS-2676: Start of 'ora.crsd' on 'crydb01' succeeded CRS-2672: Attempting to start 'ora.evmd' on 'crydb01' CRS-2676: Start of 'ora.evmd' on 'crydb01' succeeded CRS-2672: Attempting to start 'ora.asm' on 'crydb01' CRS-2676: Start of 'ora.asm' on 'crydb01' succeeded CRS-2672: Attempting to start 'ora.DATA.dg' on 'crydb01' CRS-2676: Start of 'ora.DATA.dg' on 'crydb01' succeeded crydb01 2015/02/13 00:46:02 /u01/app/11.2.0/grid/cdata/crydb01/backup_20150213_004602.olr Configure Oracle Grid Infrastructure for a Cluster ... succeeded Updating inventory properties for clusterware Starting Oracle Universal Installer... Checking swap space: must be greater than 500 MB. Actual 2047 MB Passed The inventory pointer is located at /etc/oraInst.loc The inventory is located at /u01/app/oraInventory 'UpdateNodeList' was successful.

在最后出现了这个错误:

查看日志:

...
INFO: ERROR: 
INFO: PRVF-4664 : Found inconsistent name resolution entries for SCAN name "crydb-scan"
INFO: ERROR: 
INFO: PRVF-4657 : Name resolution setup check for "crydb-scan" (IP address: 192.168.123.200) failed
INFO: ERROR: 
INFO: PRVF-4664 : Found inconsistent name resolution entries for SCAN name "crydb-scan"
INFO: Verification of SCAN VIP and Listener setup failed
...

看来是scan-ip解析的问题.原因可能是使用了hosts解析scan-ip而不是使用oracle推荐的DNS或GNS;或者配置了DNS但nslookup crydb-scan不通.
由于我这里在两个节点上ping crydb-scan是通的:

[root@crydb01 grid]# ping crydb-scan
PING crydb-scan (192.168.123.200) 56(84) bytes of data.
64 bytes from crydb-scan (192.168.123.200): icmp_seq=1 ttl=64 time=0.025 ms
64 bytes from crydb-scan (192.168.123.200): icmp_seq=2 ttl=64 time=0.018 ms
...

故可忽略此错误.
至此11g Grid安装完成:

九、安装Grid后的检查和脚本备份
安装完成后,检查crs(Cluster Ready Services)状态:

[grid@crydb01 ~]$ crsctl check crs
CRS-4638: Oracle High Availability Services is online
CRS-4537: Cluster Ready Services is online
CRS-4529: Cluster Synchronization Services is online
CRS-4533: Event Manager is online

查看集群节点:

[grid@crydb01 ~]$ olsnodes -n
crydb01 1
crydb02 2

查看ASM运行状态:

[grid@crydb01 ~]$ srvctl status asm -a
ASM is running on crydb01,crydb02
ASM is enabled.

检查OCR(Oracle Cluster Registry):

[grid@crydb01 ~]$ ocrcheck
Status of Oracle Cluster Registry is as follows :
     Version                  :          3
     Total space (kbytes)     :     262120
     Used space (kbytes)      :       2252
     Available space (kbytes) :     259868
     ID                       :  677623650
     Device/File Name         :      +DATA
                                    Device/File integrity check succeeded

                                    Device/File not configured

                                    Device/File not configured

                                    Device/File not configured

                                    Device/File not configured

     Cluster registry integrity check succeeded

     Logical corruption check bypassed due to non-privileged user

检查表决磁盘:

[grid@crydb01 ~]$ crsctl query css votedisk
##  STATE    File Universal Id                File Name Disk group
--  -----    -----------------                --------- ---------
 1. ONLINE   23c34f816b8a4ffabf6414ab4123ef3d (ORCL:DISK1) [DATA]
Located 1 voting disk(s).

一切正常.
备份root.sh脚本:

[root@crydb01 ~]# cp /u01/app/11.2.0/grid/root.sh ~/root.sh.crydb01.150213
[root@crydb02 ~]# cp /u01/app/11.2.0/grid/root.sh ~/root.sh.crydb02.150213

至此,Grid安装全部完成,下一篇将在此基础上安装Oracle数据库.

你可能感兴趣的:(RAC,oracle11g,图解教程)