redhat5.5_11gR2_RAC_安装,这篇主要记录RAC安装的执行步骤,最烦琐的就是前期配置,到后面图形界面runInstaller,asmca,dbca就很容易了。
--hostname检查--
[root@node1 Server]# hostname
node1
--hostname修改,假定两个节点,node1,node2—
[root@node1 Server]#vi /etc/sysconfig/network
[root@node1 Server]#hostname xxx
--修改hosts文件,注意对应的hostname和命名,注意127那两列,不要出现主机名,每台机器需要两个网卡,一个绑定public-ip,一个绑定private-ip,virtual-ip和scan-ip会在安装完grid后自动开通--
[root@node1 Server]# cat /etc/hosts
# Do not remove the following line, or various programs # that require network functionality will fail. 127.0.0.1 localhost.localdomain localhost ::1 localhost6.localdomain6 localhost6 #public ip 192.168.100.101 node1 192.168.100.102 node2 #virtual ip 192.168.100.201 node1-vip 192.168.100.202 node2-vip #private ip 10.10.10.1 node1-priv 10.10.10.2 node2-priv #scan ip 192.168.100.250 racscan
--创建grid(管理grid)和oracle(管理database)用户,以及oinstall、dba、oper、asmadmin、asmoper、asmdba组—
--如果只创建一个oracle用户我试验也可以,但是会造成一定的混乱,因为grid和database两个的软件的目录中,存在相同的命令--
[root@node1 Server]#groupadd -g 1000 oinstall
[root@node1 Server]#groupadd -g 1001 dba
[root@node1 Server]#groupadd -g 1002 oper
[root@node1 Server]#groupadd -g 1003 asmadmin
[root@node1 Server]#groupadd -g 1004 asmoper
[root@node1 Server]#groupadd -g 1001 asmdba
[root@node1 Server]#useradd –u 1000 –g oinstall –G dba,asmdba,oper oracle
[root@node1 Server]#passwd oracle
[root@node1 Server]#useradd –u 1001 –g oinstall –G dba,asmadmin,asmoper,asmdba,oper grid
[root@node1 Server]#passwd grid
[oracle@node6 ~]$ id
uid=1000(oracle) gid=1000(oinstall) groups=1000(oinstall),1001(dba),1002(oper),1005(asmdba)
[grid@node6 ~]$ id
uid=1001(grid) gid=1000(oinstall) groups=1000(oinstall),1001(dba),1002(oper),1003(asmadmin),1004(asmoper),1005(asmdba)
--创建相关目录--
[root@node1 Server]# mkdir -p /u01/ora11g/oracle
[root@node1 Server]# mkdir -p /u01/ora11g/grid
[root@node1 Server]# mkdir -p /opt/ora11g
[root@node1 Server]# chown -R oracle:oinstall /u01/ora11g
[root@node1 Server]# chown -R grid:oinstall /u01/ora11g/grid
[root@node1 Server]# chown -R grid:oinstall /opt/ora11g
[root@node1 Server]# chmod 777 /u01
--关闭系统ntp服务(同步时间),11gR2自带同步时间的服务,移除ntp服务的配置文件,不然GRID还是会找ntp--
[root@node1 /]# service ntpd stop
[root@node1 /]# chkconfig ntpd off
[root@node1 /]# chkconfig sendmail off
[root@node1 /]# mv /etc/ntp.conf /etc/ntp.conf.bk
--安装相关软件,注意32位和64位,现在基本上都是64位的吧--
[root@node1 Server]# pwd /media/RHEL_5.5 i386 DVD/Server [root@node1 Server]# rpm -ivh compat-gcc-34-3.4.6-4.i386.rpm gcc-4.1.2-48.el5.i386.rpm gcc-c++-4.1.2-48.el5.i386.rpm glibc-devel-2.5-49.i386.rpm libstdc++-devel-4.1.2-48.el5.i386.rpm libXp-1.0.0-8.1.el5.i386.rpm openmotif22-2.2.3-18.i386.rpm elfutils-libelf-devel-0.137-3.el5.i386.rpm elfutils-libelf-devel-static-0.137-3.el5.i386.rpm sysstat-7.0.2-3.el5.i386.rpm libaio-devel-0.3.106-5.i386.rpm libgomp-4.4.0-6.el5.i386.rpm glibc-headers-2.5-49.i386.rpm kernel-headers-2.6.18-194.el5.i386.rpm [root@localhost Server]# rpm -ivh compat-gcc-34-3.4.6-4.x86_64.rpm gcc-4.1.2-48.el5.x86_64.rpm gcc-c++-4.1.2-48.el5.x86_64.rpm glibc-devel-2.5-49.x86_64.rpm libstdc++-devel-4.1.2-48.el5.x86_64.rpm libXp-1.0.0-8.1.el5.x86_64.rpm openmotif22-2.2.3-18.x86_64.rpm elfutils-libelf-devel-0.137-3.el5.x86_64.rpm elfutils-libelf-devel-static-0.137-3.el5.x86_64.rpm sysstat-7.0.2-3.el5.x86_64.rpm libaio-devel-0.3.106-5.x86_64.rpm libgomp-4.4.0-6.el5.x86_64.rpm glibc-headers-2.5-49.x86_64.rpm kernel-headers-2.6.18-194.el5.x86_64.rpm
--设置系统参数,添加到最下方,可以根据自身系统进行修改--
[root@node1 Server]# vi /etc/sysctl.conf kernel.shmall = 2097152 kernel.shmmax = 2147483648 kernel.shmmni = 4096 kernel.sem = 250 32000 100 128 fs.file-max = 6815744 fs.aio-max-nr = 1048576 net.ipv4.ip_local_port_range = 9000 65000 net.core.rmem_default = 1048576 net.core.rmem_max = 4194304 net.core.wmem_default = 262144 net.core.wmem_max = 1048576 [root@node1 Server]# sysctl -p [root@node1 Server]# vi /etc/security/limits.conf oracle soft nproc 2047 oracle hard nproc 16384 oracle soft nofile 1024 oracle hard nofile 65536 grid soft nproc 2047 grid hard nproc 16384 grid soft nofile 1024 grid hard nofile 65536 [root@node1 /]# vi /etc/pam.d/login session required pam.limits.so [root@node1 Server]# vi /etc/selinux/config SELINUX=disabled (重启生效)
--设置tmpfs的大小,/dev/shm限制了数据库可以配置的内存--
[root@node1 Server]# vi /etc/fstab
tmpfs /dev/shm /tmpfs defaults,size=30G 0 0
[root@node1 Server]# mount –o remount /dev/shm
--两个用户的环境变量,grid用户BASE与HOME不要设置为上下级关系,不然安装时会报错ins-32026。oracle用户添加一个ORACLE_UNQNAME,此环境变量可以打开em,emctl start dbconsole。根据需要添加NLS_LANG和NLS_DATE_FORMAT。注意修改各节点的SID—
export NLS_LANG="SIMPLIFIED CHINESE_CHINA”.ZHS16GBK & UTF8 export NLS_DATE_FORMAT='YYYY-MM-DD HH24:MI:SS' [grid@node1 ~]$ vi .bash_profile export TMP=/tmp export TMPDIR=$TMP export ORACLE_SID=+ASM1 export ORACLE_BASE=/opt/ora11g export ORACLE_HOME=/u01/ora11g/grid/product/11.2.0 export PATH=$ORACLE_HOME/bin:/usr/sbin:$PATH if [ $USER = "oracle" ] || [ $USER = "grid" ];then if [ $SHELL = "/bin/ksh" ];then ulimit -p 16384 ulimit -n 65536 else ulimit -u 16384 -n 65536 fi umask 022 fi [oracle@node1 ~]$ vi .bash_profile export TMP=/tmp export TMPDIR=$TMP export ORACLE_SID=racdb1 export ORACLE_TREM=xterm export ORACLE_BASE=/u01/ora11g export ORACLE_HOME=$ORACLE_BASE/oracle/product/11.2.0 export PATH=$ORACLE_HOME/bin:/usr/sbin:$PATH export LD_LIBRARY_PATH=$ORACLE_HOME/lib:/lib:/usr/lib export CLASSPATH=$ORACLE_HOME/JRE:$ORACLE_HOME/jlib:$ORACLE_HOME/rdbms/jlib export ORACLE_UNQNAME=racdb if [ $USER = "oracle" ] || [ $USER = "grid" ];then if [ $SHELL = "/bin/ksh" ];then ulimit -p 16384 ulimit -n 65536 else ulimit -u 16384 -n 65536 fi umask 022 fi
--查看共享磁盘,此时还未格式化,两节点必须相同,11gRAC一般要有3个共享磁盘,一个grid用,一个data,一个recovery--
[root@node1 ~]# fdisk -l Disk /dev/sda: 21.4 GB, 21474836480 bytes 255 heads, 63 sectors/track, 2610 cylinders Units = cylinders of 16065 * 512 = 8225280 bytes Device Boot Start End Blocks Id System /dev/sda1 * 1 13 104391 83 Linux /dev/sda2 14 274 2096482+ 82 Linux swap / Solaris /dev/sda3 275 2610 18763920 83 Linux Disk /dev/sdb: 1073 MB, 1073741824 bytes 255 heads, 63 sectors/track, 130 cylinders Units = cylinders of 16065 * 512 = 8225280 bytes Disk /dev/sdb doesn't contain a valid partition table Disk /dev/sdc: 1073 MB, 1073741824 bytes 255 heads, 63 sectors/track, 130 cylinders Units = cylinders of 16065 * 512 = 8225280 bytes Disk /dev/sdc doesn't contain a valid partition table Disk /dev/sdd: 1073 MB, 1073741824 bytes 255 heads, 63 sectors/track, 130 cylinders Units = cylinders of 16065 * 512 = 8225280 bytes Disk /dev/sdd doesn't contain a valid partition table Disk /dev/sde: 10.7 GB, 10737418240 bytes 255 heads, 63 sectors/track, 1305 cylinders Units = cylinders of 16065 * 512 = 8225280 bytes Disk /dev/sde doesn't contain a valid partition table Disk /dev/sdf: 10.7 GB, 10737418240 bytes 255 heads, 63 sectors/track, 1305 cylinders Units = cylinders of 16065 * 512 = 8225280 bytes Disk /dev/sdf doesn't contain a valid partition table
----------------------------以下方式是利用asmlib进行asm创建---------------------------------------------------------------------
--安装asm相关软件,注意安装32位或64位。也可以不用安装,利用系统自带的udev进行asm创建(udev我在redhat5.5中没有试验成功,在OEL6.4中成功)--
[root@node2 redhat]# uname -a Linux node2 2.6.18-194.el5 #1 SMP Tue Mar 16 21:52:43 EDT 2010 i686 i686 i386 GNU/Linux [root@node2 redhat]# ls oracleasm-2.6.18-194.el5-2.0.5-1.el5.i686.rpm oracleasmlib-2.0.4-1.el5.i386.rpm oracleasm-support-2.1.8-1.el5.i386.rpm [root@node1 x86_64]# ls oracleasm-2.6.18-194.el5-2.0.5-1.el5.x86_64.rpm oracleasmlib-2.0.4-1.el5.x86_64.rpm oracleasm-support-2.1.8-1.el5.x86_64.rpm [root@node2 redhat]# rpm –ivh *
--格式化过程,sdb~sdf,按顺序:n、p、1、enter、enter、w--
[root@node1 ~]# fdisk /dev/sdb Device contains neither a valid DOS partition table, nor Sun, SGI or OSF disklabel Building a new DOS disklabel. Changes will remain in memory only, until you decide to write them. After that, of course, the previous content won't be recoverable. Warning: invalid flag 0x0000 of partition table 4 will be corrected by w(rite) Command (m for help): n Command action e extended p primary partition (1-4) p Partition number (1-4): 1 First cylinder (1-130, default 1): Using default value 1 Last cylinder or +size or +sizeM or +sizeK (1-130, default 130): Using default value 130 Command (m for help): w The partition table has been altered! Calling ioctl() to re-read partition table. Syncing disks.
--再次查看磁盘,已经格式化完毕,两节点必须一样--
[root@node1 ~]# fdisk -l Disk /dev/sda: 21.4 GB, 21474836480 bytes 255 heads, 63 sectors/track, 2610 cylinders Units = cylinders of 16065 * 512 = 8225280 bytes Device Boot Start End Blocks Id System /dev/sda1 * 1 13 104391 83 Linux /dev/sda2 14 274 2096482+ 82 Linux swap / Solaris /dev/sda3 275 2610 18763920 83 Linux Disk /dev/sdb: 1073 MB, 1073741824 bytes 255 heads, 63 sectors/track, 130 cylinders Units = cylinders of 16065 * 512 = 8225280 bytes Device Boot Start End Blocks Id System /dev/sdb1 1 130 1044193+ 83 Linux Disk /dev/sdc: 1073 MB, 1073741824 bytes 255 heads, 63 sectors/track, 130 cylinders Units = cylinders of 16065 * 512 = 8225280 bytes Device Boot Start End Blocks Id System /dev/sdc1 1 130 1044193+ 83 Linux Disk /dev/sdd: 1073 MB, 1073741824 bytes 255 heads, 63 sectors/track, 130 cylinders Units = cylinders of 16065 * 512 = 8225280 bytes Device Boot Start End Blocks Id System /dev/sdd1 1 130 1044193+ 83 Linux Disk /dev/sde: 10.7 GB, 10737418240 bytes 255 heads, 63 sectors/track, 1305 cylinders Units = cylinders of 16065 * 512 = 8225280 bytes Device Boot Start End Blocks Id System /dev/sde1 1 1305 10482381 83 Linux Disk /dev/sdf: 10.7 GB, 10737418240 bytes 255 heads, 63 sectors/track, 1305 cylinders Units = cylinders of 16065 * 512 = 8225280 bytes Device Boot Start End Blocks Id System /dev/sdf1 1 1305 10482381 83 Linux
--查看节点node1的磁盘文件--
[root@node1 ~]# ll /dev/sd*
brw-r----- 1 root disk 8, 0 2013-03-28 /dev/sda
brw-r----- 1 root disk 8, 1 03-28 09:16 /dev/sda1
brw-r----- 1 root disk 8, 2 2013-03-28 /dev/sda2
brw-r----- 1 root disk 8, 3 03-28 09:16 /dev/sda3
brw-r----- 1 root disk 8, 16 03-28 09:24 /dev/sdb
brw-r----- 1 root disk 8, 17 03-28 09:24 /dev/sdb1
brw-r----- 1 root disk 8, 32 03-28 09:24 /dev/sdc
brw-r----- 1 root disk 8, 33 03-28 09:25 /dev/sdc1
brw-r----- 1 root disk 8, 48 03-28 09:26 /dev/sdd
brw-r----- 1 root disk 8, 49 03-28 09:26 /dev/sdd1
brw-r----- 1 root disk 8, 64 03-28 11:27 /dev/sde
brw-r----- 1 root disk 8, 65 03-28 11:27 /dev/sde1
brw-r----- 1 root disk 8, 80 03-28 11:29 /dev/sdf
brw-r----- 1 root disk 8, 81 03-28 11:29 /dev/sdf1
--发现节点node2看不到全部格式化好的磁盘文件,重启--
[root@node2 ~]# ll /dev/sd*
brw-r----- 1 root disk 8, 0 2013-03-28 /dev/sda
brw-r----- 1 root disk 8, 1 03-28 09:16 /dev/sda1
brw-r----- 1 root disk 8, 2 2013-03-28 /dev/sda2
brw-r----- 1 root disk 8, 3 03-28 09:16 /dev/sda3
brw-r----- 1 root disk 8, 16 2013-03-28 /dev/sdb
brw-r----- 1 root disk 8, 32 2013-03-28 /dev/sdc
brw-r----- 1 root disk 8, 48 2013-03-28 /dev/sdd
brw-r----- 1 root disk 8, 64 2013-03-28 /dev/sde
brw-r----- 1 root disk 8, 80 2013-03-28 /dev/sdf
[root@node2 ~]#init 6
[root@node2 ~]# ll /dev/sd*
brw-r----- 1 root disk 8, 0 2013-03-28 /dev/sda
brw-r----- 1 root disk 8, 1 03-28 11:40 /dev/sda1
brw-r----- 1 root disk 8, 2 2013-03-28 /dev/sda2
brw-r----- 1 root disk 8, 3 03-28 11:40 /dev/sda3
brw-r----- 1 root disk 8, 16 2013-03-28 /dev/sdb
brw-r----- 1 root disk 8, 17 2013-03-28 /dev/sdb1
brw-r----- 1 root disk 8, 32 2013-03-28 /dev/sdc
brw-r----- 1 root disk 8, 33 2013-03-28 /dev/sdc1
brw-r----- 1 root disk 8, 48 2013-03-28 /dev/sdd
brw-r----- 1 root disk 8, 49 2013-03-28 /dev/sdd1
brw-r----- 1 root disk 8, 64 2013-03-28 /dev/sde
brw-r----- 1 root disk 8, 65 2013-03-28 /dev/sde1
brw-r----- 1 root disk 8, 80 2013-03-28 /dev/sdf
brw-r----- 1 root disk 8, 81 2013-03-28 /dev/sdf1
--节点node1,初始化oracleasm,并创建asm--
[root@node1 ~]# /etc/init.d/oracleasm configure Configuring the Oracle ASM library driver. This will configure the on-boot properties of the Oracle ASM library driver. The following questions will determine whether the driver is loaded on boot and what permissions it will have. The current values will be shown in brackets ('[]'). Hitting <ENTER> without typing an answer will keep that current value. Ctrl-C will abort. Default user to own the driver interface []: grid Default group to own the driver interface []: oinstall Start Oracle ASM library driver on boot (y/n) [n]: y Scan for Oracle ASM disks on boot (y/n) [y]: y Writing Oracle ASM library driver configuration: done Initializing the Oracle ASMLib driver: [ OK ] Scanning the system for Oracle ASMLib disks: [ OK ] [root@node1 ~]# /etc/init.d/oracleasm createdisk CRS1 /dev/sdb1 Marking disk "CRS1" as an ASM disk: [ OK ] [root@node1 ~]# /etc/init.d/oracleasm createdisk CRS2 /dev/sdc1 Marking disk "CRS2" as an ASM disk: [ OK ] [root@node1 ~]# /etc/init.d/oracleasm createdisk CRS3 /dev/sdd1 Marking disk "CRS3" as an ASM disk: [ OK ] [root@node1 ~]# /etc/init.d/oracleasm createdisk DATA1 /dev/sde1 Marking disk "DATA1" as an ASM disk: [ OK ] [root@node1 ~]# /etc/init.d/oracleasm createdisk DATA2 /dev/sdf1 Marking disk "DATA2" as an ASM disk: [ OK ] [root@node1 ~]# /etc/init.d/oracleasm scandisks Scanning the system for Oracle ASMLib disks: [ OK ] [root@node1 ~]# /etc/init.d/oracleasm listdisks CRS1 CRS2 CRS3 DATA1 DATA2
--节点node2,初始化oracleasm,并扫描asm--
[root@node2 ~]# /etc/init.d/oracleasm configure Configuring the Oracle ASM library driver. This will configure the on-boot properties of the Oracle ASM library driver. The following questions will determine whether the driver is loaded on boot and what permissions it will have. The current values will be shown in brackets ('[]'). Hitting <ENTER> without typing an answer will keep that current value. Ctrl-C will abort. Default user to own the driver interface []: grid Default group to own the driver interface []: oinstall Start Oracle ASM library driver on boot (y/n) [n]: y Scan for Oracle ASM disks on boot (y/n) [y]: y Writing Oracle ASM library driver configuration: done Initializing the Oracle ASMLib driver: [ OK ] Scanning the system for Oracle ASMLib disks: [ OK ] [root@node2 ~]# /etc/init.d/oracleasm scandisks Scanning the system for Oracle ASMLib disks: [ OK ]
[root@node2 ~]# /etc/init.d/oracleasm listdisks CRS1 CRS2 CRS3 DATA1 DATA2
--------------以下方式是利用udev进行asm创建,redhat5.5没有找到利用udev进行设置的方法,用OEL6.4做个例子------------------
首先确认是否安装了udev软件包 [root@node6 ~]# rpm -qa|grep udev udev-095-14.21.el5 通过scsi_id获取设备的块设备的唯一标识名(在OEL6.4上面试验成功) scsi_id -g -u /dev/sd* 在Oracle Linux 5下,可以使用如下命令(redhat5没成功): # /sbin/scsi_id -g -u -s /block/sdb/sdb1 SATA_VBOX_HARDDISK_VBd306dbe0-df3367e3_ 创建并配置UDEVRules 文件 [root@rac1 rules.d]# touch /etc/udev/rules.d/99-oracle-asmdevices.rules 添加如下内容: KERNEL=="sd?1",BUS=="scsi", PROGRAM=="/sbin/scsi_id -g -u /dev/$name",RESULT=="1ATA_VBOX_HARDDISK_VB83552343-28d5a489",NAME="asm-disk1", OWNER="oracle", GROUP="dba",MODE="0660" KERNEL=="sd?1",BUS=="scsi", PROGRAM=="/sbin/scsi_id -g -u /dev/$name", RESULT=="1ATA_VBOX_HARDDISK_VBb96d5ecb-4eae1d96",NAME="asm-disk2", OWNER="oracle", GROUP="dba",MODE="0660" KERNEL=="sd?1",BUS=="scsi", PROGRAM=="/sbin/scsi_id -g -u /dev/$name",RESULT=="1ATA_VBOX_HARDDISK_VBfd7bba6c-b91fba70",NAME="asm-disk3", OWNER="oracle", GROUP="dba",MODE="0660" KERNEL=="sd?1",BUS=="scsi", PROGRAM=="/sbin/scsi_id -g -u /dev/$name",RESULT=="1ATA_VBOX_HARDDISK_VB3239ed0d-db15bbec",NAME="asm-disk4", OWNER="oracle", GROUP="dba",MODE="0660"
--设置oracle与grid用户不需要密码ssh互联,并且要分别执行ssh xxx1/2 date,以防runcluvfy过不去--
--在10g RAC中必须手动配置成功,在11g RAC中不必,因为grid和db的图形安装中可以进行自动配置,简单粗暴。但是为了跑检查脚本,还是手动配置上吧--
[oracle@node1 ~]$ mkdir .ssh [oracle@node1 ~]$ ssh-keygen -t rsa Generating public/private rsa key pair. Enter file in which to save the key (/home/oracle/.ssh/id_rsa): Enter passphrase (empty for no passphrase): Enter same passphrase again: Your identification has been saved in /home/oracle/.ssh/id_rsa. Your public key has been saved in /home/oracle/.ssh/id_rsa.pub. The key fingerprint is: e3:71:9f:fa:da:14:a9:cb:c1:01:17:a2:6d:9d:9d:5f oracle@node1 [oracle@node1 ~]$ ssh-keygen -t dsa Generating public/private dsa key pair. Enter file in which to save the key (/home/oracle/.ssh/id_dsa): Enter passphrase (empty for no passphrase): Enter same passphrase again: Your identification has been saved in /home/oracle/.ssh/id_dsa. Your public key has been saved in /home/oracle/.ssh/id_dsa.pub. The key fingerprint is: 0b:bf:0e:7f:1c:c7:e6:59:9c:62:d4:42:ec:f7:3f:b0 oracle@node1 [oracle@node1 ~]$ cd .ssh [oracle@node1 .ssh]$ ls id_dsa id_dsa.pub id_rsa id_rsa.pub [oracle@node2 ~]$ mkdir .ssh [oracle@node2 ~]$ ssh-keygen -t rsa Generating public/private rsa key pair. Enter file in which to save the key (/home/oracle/.ssh/id_rsa): Enter passphrase (empty for no passphrase): Enter same passphrase again: Your identification has been saved in /home/oracle/.ssh/id_rsa. Your public key has been saved in /home/oracle/.ssh/id_rsa.pub. The key fingerprint is: 82:6c:3f:6f:20:ec:6f:e8:cc:1a:18:73:0b:cf:f5:04 oracle@node2 [oracle@node2 ~]$ ssh-keygen -t dsa Generating public/private dsa key pair. Enter file in which to save the key (/home/oracle/.ssh/id_dsa): Enter passphrase (empty for no passphrase): Enter same passphrase again: Your identification has been saved in /home/oracle/.ssh/id_dsa. Your public key has been saved in /home/oracle/.ssh/id_dsa.pub. The key fingerprint is: 8f:e9:90:9e:d3:77:4f:87:a4:c6:5d:ff:ce:50:42:4f oracle@node2 [oracle@node1 ~]$ cat ~/.ssh/id_rsa.pub >> ~/.ssh/authorized_keys [oracle@node1 ~]$ cat ~/.ssh/id_dsa.pub >> ~/.ssh/authorized_keys [oracle@node1 ~]$ ssh node2 cat ~/.ssh/id_rsa.pub >> ~/.ssh/authorized_keys [oracle@node1 ~]$ ssh node2 cat ~/.ssh/id_dsa.pub >> ~/.ssh/authorized_keys [oracle@node1 ~]$ scp ~/.ssh/authorized_keys node2:~/.ssh/authorized_keys [oracle@node1 ~]$ ssh node2 date [oracle@node1 ~]$ ssh node1 date [oracle@node2 ~]$ ssh node1 date [oracle@node2 ~]$ ssh node2 date
[grid@node1 ~]$ mkdir .ssh [grid@node1 ~]$ ssh-keygen -t rsa Generating public/private rsa key pair. Enter file in which to save the key (/home/grid/.ssh/id_rsa): Enter passphrase (empty for no passphrase): Enter same passphrase again: Your identification has been saved in /home/grid/.ssh/id_rsa. Your public key has been saved in /home/grid/.ssh/id_rsa.pub. The key fingerprint is: ce:e0:24:b0:4c:f2:25:7d:b9:e5:48:19:10:23:33:97 grid@node1 [grid@node1 ~]$ ssh-keygen -t dsa
Generating public/private dsa key pair. Enter file in which to save the key (/home/grid/.ssh/id_dsa): Enter passphrase (empty for no passphrase): Enter same passphrase again. Your identification has been saved in /home/grid/.ssh/id_dsa. Your public key has been saved in /home/grid/.ssh/id_dsa.pub. The key fingerprint is: 11:78:88:1a:47:ff:c3:9f:a1:b1:d2:d9:d9:27:48:59 grid@node1 [grid@node2 ~]$ mkdir .ssh [grid@node2 ~]$ ssh-keygen -t rsa Generating public/private rsa key pair. Enter file in which to save the key (/home/grid/.ssh/id_rsa): Enter passphrase (empty for no passphrase): Enter same passphrase again: Your identification has been saved in /home/grid/.ssh/id_rsa. Your public key has been saved in /home/grid/.ssh/id_rsa.pub. The key fingerprint is: 53:d7:d5:e8:63:5a:c6:15:c4:81:11:80:9b:5f:f9:42 grid@node2 [grid@node2 ~]$ ssh-keygen -t dsa Generating public/private dsa key pair. Enter file in which to save the key (/home/grid/.ssh/id_dsa): Enter passphrase (empty for no passphrase): Enter same passphrase again: Your identification has been saved in /home/grid/.ssh/id_dsa. Your public key has been saved in /home/grid/.ssh/id_dsa.pub. The key fingerprint is: 9f:d7:3e:cf:e4:ff:29:df:c5:8c:85:46:6b:e2:07:1b grid@node2
[grid@node1 ~]$ cat ~/.ssh/id_rsa.pub >> ~/.ssh/authorized_keys [grid@node1 ~]$ cat ~/.ssh/id_dsa.pub >> ~/.ssh/authorized_keys [grid@node1 ~]$ ssh node2 cat ~/.ssh/id_rsa.pub >> ~/.ssh/authorized_keys [grid@node1 ~]$ ssh node2 cat ~/.ssh/id_dsa.pub >> ~/.ssh/authorized_keys [grid@node1 ~]$ scp ~/.ssh/authorized_keys node2:~/.ssh/authorized_keys [grid@node1 ~]$ ssh node2 date [grid@node1 ~]$ ssh node1 date [grid@node2 ~]$ ssh node1 date [grid@node2 ~]$ ssh node2 date
--在两个节点安装cvuqdisk包并验证--
在两个 Oracle RAC 节点上安装操作系统程序包 cvuqdisk。
如果没有 cvuqdisk,集群验证实用程序就无法发现共享磁盘,当运行(手动运行或在 Oracle Grid Infrastructure 安装结束时自动运行)集群验证实用程序时,会收到这样的错误消息:“Package cvuqdisk not installed”。使用适用于硬件体系结构(例如,x86_64 或 i386)的 cvuqdisk RPM。
cvuqdisk RPM 包含在 Oracle Grid Infrastructure 安装介质上的 rpm 目录中。
设置环境变量 CVUQDISK_GRP,使其指向作为 cvuqdisk 的所有者所在的组(本文为 oinstall):
[root@node1 rpm]# pwd
/oracle-tools/oracle11gR2_linux/grid/rpm
[root@node1 rpm]# ll
总计 12
-rwxr-xr-x 1 root root 8233 2011-09-22 cvuqdisk-1.0.9-1.rpm
[root@node1 rpm]# rpm -ivh *
Preparing... (100%########################################### [100%]
Using default group oinstall to install package
1:cvuqdisk ( 99%########################################### [100####3
[grid@node1 ~]$ export CVUQDISK_GRP=oinstall
--安装前检查,最后会因为dns过不去,可以忽略。最后就是图形界面安装,先用grid用户安装grid,再用oracle用户安装db,之后用grid用户设置asm,最后oracle用户dbca建立数据库--
[grid@node1 grid]$ ./runcluvfy.sh stage -pre crsinst -n node1,node2 -fixup -verbose >> /opt/ora11g/check