[terry笔记]redhat5.5_11gR2_RAC_安装

 redhat5.5_11gR2_RAC_安装,这篇主要记录RAC安装的执行步骤,最烦琐的就是前期配置,到后面图形界面runInstaller,asmca,dbca就很容易了。

 

--hostname检查--

[root@node1 Server]# hostname

node1

 

--hostname修改,假定两个节点,node1,node2—

[root@node1 Server]#vi /etc/sysconfig/network

[root@node1 Server]#hostname xxx

 

--修改hosts文件,注意对应的hostname和命名,注意127那两列,不要出现主机名,每台机器需要两个网卡,一个绑定public-ip,一个绑定private-ip,virtual-ip和scan-ip会在安装完grid后自动开通--

[root@node1 Server]# cat /etc/hosts

# Do not remove the following line, or various programs

# that require network functionality will fail.

127.0.0.1        localhost.localdomain localhost

::1             localhost6.localdomain6 localhost6



#public ip

192.168.100.101 node1

192.168.100.102 node2



#virtual ip

192.168.100.201 node1-vip

192.168.100.202 node2-vip



#private ip

10.10.10.1 node1-priv

10.10.10.2 node2-priv



#scan ip

192.168.100.250 racscan

 

--创建grid(管理grid)和oracle(管理database)用户,以及oinstall、dba、oper、asmadmin、asmoper、asmdba组—

--如果只创建一个oracle用户我试验也可以,但是会造成一定的混乱,因为grid和database两个的软件的目录中,存在相同的命令--

[root@node1 Server]#groupadd -g 1000 oinstall

[root@node1 Server]#groupadd -g 1001 dba

[root@node1 Server]#groupadd -g 1002 oper

[root@node1 Server]#groupadd -g 1003 asmadmin

[root@node1 Server]#groupadd -g 1004 asmoper

[root@node1 Server]#groupadd -g 1001 asmdba

[root@node1 Server]#useradd –u 1000 –g oinstall –G dba,asmdba,oper oracle

[root@node1 Server]#passwd oracle

[root@node1 Server]#useradd –u 1001 –g oinstall –G dba,asmadmin,asmoper,asmdba,oper grid

[root@node1 Server]#passwd grid

[oracle@node6 ~]$ id
uid=1000(oracle) gid=1000(oinstall) groups=1000(oinstall),1001(dba),1002(oper),1005(asmdba)

[grid@node6 ~]$ id
uid=1001(grid) gid=1000(oinstall) groups=1000(oinstall),1001(dba),1002(oper),1003(asmadmin),1004(asmoper),1005(asmdba)

 

--创建相关目录--

[root@node1 Server]# mkdir -p /u01/ora11g/oracle

[root@node1 Server]# mkdir -p /u01/ora11g/grid

[root@node1 Server]# mkdir -p /opt/ora11g

[root@node1 Server]# chown -R oracle:oinstall /u01/ora11g

[root@node1 Server]# chown -R grid:oinstall /u01/ora11g/grid

[root@node1 Server]# chown -R grid:oinstall /opt/ora11g

[root@node1 Server]# chmod 777 /u01

 

--关闭系统ntp服务(同步时间),11gR2自带同步时间的服务,移除ntp服务的配置文件,不然GRID还是会找ntp--

[root@node1 /]# service ntpd stop

[root@node1 /]# chkconfig ntpd off

[root@node1 /]# chkconfig sendmail off

[root@node1 /]# mv /etc/ntp.conf /etc/ntp.conf.bk

 

--安装相关软件,注意32位和64位,现在基本上都是64位的吧--

[root@node1 Server]# pwd

/media/RHEL_5.5 i386 DVD/Server



[root@node1 Server]# rpm -ivh compat-gcc-34-3.4.6-4.i386.rpm gcc-4.1.2-48.el5.i386.rpm gcc-c++-4.1.2-48.el5.i386.rpm glibc-devel-2.5-49.i386.rpm libstdc++-devel-4.1.2-48.el5.i386.rpm libXp-1.0.0-8.1.el5.i386.rpm openmotif22-2.2.3-18.i386.rpm elfutils-libelf-devel-0.137-3.el5.i386.rpm elfutils-libelf-devel-static-0.137-3.el5.i386.rpm sysstat-7.0.2-3.el5.i386.rpm libaio-devel-0.3.106-5.i386.rpm libgomp-4.4.0-6.el5.i386.rpm glibc-headers-2.5-49.i386.rpm kernel-headers-2.6.18-194.el5.i386.rpm



[root@localhost Server]# rpm -ivh compat-gcc-34-3.4.6-4.x86_64.rpm gcc-4.1.2-48.el5.x86_64.rpm gcc-c++-4.1.2-48.el5.x86_64.rpm glibc-devel-2.5-49.x86_64.rpm libstdc++-devel-4.1.2-48.el5.x86_64.rpm libXp-1.0.0-8.1.el5.x86_64.rpm openmotif22-2.2.3-18.x86_64.rpm elfutils-libelf-devel-0.137-3.el5.x86_64.rpm elfutils-libelf-devel-static-0.137-3.el5.x86_64.rpm sysstat-7.0.2-3.el5.x86_64.rpm libaio-devel-0.3.106-5.x86_64.rpm libgomp-4.4.0-6.el5.x86_64.rpm glibc-headers-2.5-49.x86_64.rpm kernel-headers-2.6.18-194.el5.x86_64.rpm

 

--设置系统参数,添加到最下方,可以根据自身系统进行修改--

[root@node1 Server]# vi /etc/sysctl.conf

kernel.shmall = 2097152

kernel.shmmax = 2147483648

kernel.shmmni = 4096

kernel.sem = 250 32000 100 128

fs.file-max = 6815744

fs.aio-max-nr = 1048576

net.ipv4.ip_local_port_range = 9000 65000

net.core.rmem_default = 1048576

net.core.rmem_max = 4194304

net.core.wmem_default = 262144

net.core.wmem_max = 1048576



[root@node1 Server]# sysctl -p



[root@node1 Server]# vi /etc/security/limits.conf

oracle soft nproc 2047

oracle hard nproc 16384

oracle soft nofile 1024

oracle hard nofile 65536

grid soft nproc 2047

grid hard nproc 16384

grid soft nofile 1024

grid hard nofile 65536



[root@node1 /]# vi /etc/pam.d/login

session    required     pam.limits.so



[root@node1 Server]# vi /etc/selinux/config

SELINUX=disabled

(重启生效)

 

--设置tmpfs的大小,/dev/shm限制了数据库可以配置的内存--

[root@node1 Server]# vi /etc/fstab

tmpfs  /dev/shm  /tmpfs  defaults,size=30G  0 0

[root@node1 Server]# mount –o remount /dev/shm

 

--两个用户的环境变量,grid用户BASE与HOME不要设置为上下级关系,不然安装时会报错ins-32026。oracle用户添加一个ORACLE_UNQNAME,此环境变量可以打开em,emctl start dbconsole。根据需要添加NLS_LANG和NLS_DATE_FORMAT。注意修改各节点的SID—

export NLS_LANG="SIMPLIFIED CHINESE_CHINA”.ZHS16GBK & UTF8

export NLS_DATE_FORMAT='YYYY-MM-DD HH24:MI:SS'



[grid@node1 ~]$ vi .bash_profile

export TMP=/tmp

export TMPDIR=$TMP

export ORACLE_SID=+ASM1

export ORACLE_BASE=/opt/ora11g

export ORACLE_HOME=/u01/ora11g/grid/product/11.2.0

export PATH=$ORACLE_HOME/bin:/usr/sbin:$PATH

if [ $USER = "oracle" ] || [ $USER = "grid" ];then

   if [ $SHELL = "/bin/ksh" ];then

      ulimit -p 16384

      ulimit -n 65536

  else

    ulimit -u 16384 -n 65536

   fi

   umask 022

fi





[oracle@node1 ~]$ vi .bash_profile

export TMP=/tmp

export TMPDIR=$TMP

export ORACLE_SID=racdb1

export ORACLE_TREM=xterm

export ORACLE_BASE=/u01/ora11g

export ORACLE_HOME=$ORACLE_BASE/oracle/product/11.2.0

export PATH=$ORACLE_HOME/bin:/usr/sbin:$PATH

export LD_LIBRARY_PATH=$ORACLE_HOME/lib:/lib:/usr/lib

export CLASSPATH=$ORACLE_HOME/JRE:$ORACLE_HOME/jlib:$ORACLE_HOME/rdbms/jlib

export ORACLE_UNQNAME=racdb

if [ $USER = "oracle" ] || [ $USER = "grid" ];then

   if [ $SHELL = "/bin/ksh" ];then

      ulimit -p 16384

      ulimit -n 65536

  else

    ulimit -u 16384 -n 65536

   fi

   umask 022

fi

 

--查看共享磁盘,此时还未格式化,两节点必须相同,11gRAC一般要有3个共享磁盘,一个grid用,一个data,一个recovery--

[root@node1 ~]# fdisk -l



Disk /dev/sda: 21.4 GB, 21474836480 bytes

255 heads, 63 sectors/track, 2610 cylinders

Units = cylinders of 16065 * 512 = 8225280 bytes



   Device Boot      Start         End      Blocks   Id  System

/dev/sda1   *           1          13      104391   83  Linux

/dev/sda2              14         274     2096482+  82  Linux swap / Solaris

/dev/sda3             275        2610    18763920   83  Linux





Disk /dev/sdb: 1073 MB, 1073741824 bytes

255 heads, 63 sectors/track, 130 cylinders

Units = cylinders of 16065 * 512 = 8225280 bytes



Disk /dev/sdb doesn't contain a valid partition table





Disk /dev/sdc: 1073 MB, 1073741824 bytes

255 heads, 63 sectors/track, 130 cylinders

Units = cylinders of 16065 * 512 = 8225280 bytes



Disk /dev/sdc doesn't contain a valid partition table



 

Disk /dev/sdd: 1073 MB, 1073741824 bytes

255 heads, 63 sectors/track, 130 cylinders

Units = cylinders of 16065 * 512 = 8225280 bytes



Disk /dev/sdd doesn't contain a valid partition table





Disk /dev/sde: 10.7 GB, 10737418240 bytes

255 heads, 63 sectors/track, 1305 cylinders

Units = cylinders of 16065 * 512 = 8225280 bytes



Disk /dev/sde doesn't contain a valid partition table





Disk /dev/sdf: 10.7 GB, 10737418240 bytes

255 heads, 63 sectors/track, 1305 cylinders

Units = cylinders of 16065 * 512 = 8225280 bytes



Disk /dev/sdf doesn't contain a valid partition table

 

 

----------------------------以下方式是利用asmlib进行asm创建---------------------------------------------------------------------

--安装asm相关软件,注意安装32位或64位。也可以不用安装,利用系统自带的udev进行asm创建(udev我在redhat5.5中没有试验成功,在OEL6.4中成功)--

[root@node2 redhat]# uname -a

Linux node2 2.6.18-194.el5 #1 SMP Tue Mar 16 21:52:43 EDT 2010 i686 i686 i386 GNU/Linux



[root@node2 redhat]# ls

oracleasm-2.6.18-194.el5-2.0.5-1.el5.i686.rpm

oracleasmlib-2.0.4-1.el5.i386.rpm

oracleasm-support-2.1.8-1.el5.i386.rpm



[root@node1 x86_64]# ls

oracleasm-2.6.18-194.el5-2.0.5-1.el5.x86_64.rpm

oracleasmlib-2.0.4-1.el5.x86_64.rpm

oracleasm-support-2.1.8-1.el5.x86_64.rpm



[root@node2 redhat]# rpm –ivh *

 

 

--格式化过程,sdb~sdf,按顺序:n、p、1、enter、enter、w--

[root@node1 ~]# fdisk /dev/sdb



Device contains neither a valid DOS partition table, nor Sun, SGI or OSF disklabel

Building a new DOS disklabel. Changes will remain in memory only,

until you decide to write them. After that, of course, the previous

content won't be recoverable.



Warning: invalid flag 0x0000 of partition table 4 will be corrected by w(rite)



Command (m for help): n



Command action

   e   extended

   p   primary partition (1-4)

p



Partition number (1-4): 1

First cylinder (1-130, default 1):

Using default value 1

Last cylinder or +size or +sizeM or +sizeK (1-130, default 130):

Using default value 130





Command (m for help): w

The partition table has been altered!



Calling ioctl() to re-read partition table.

Syncing disks.

 

--再次查看磁盘,已经格式化完毕,两节点必须一样--

[root@node1 ~]# fdisk -l



Disk /dev/sda: 21.4 GB, 21474836480 bytes

255 heads, 63 sectors/track, 2610 cylinders

Units = cylinders of 16065 * 512 = 8225280 bytes



   Device Boot      Start         End      Blocks   Id  System

/dev/sda1   *           1          13      104391   83  Linux

/dev/sda2              14         274     2096482+  82  Linux swap / Solaris

/dev/sda3             275        2610    18763920   83  Linux



 

Disk /dev/sdb: 1073 MB, 1073741824 bytes

255 heads, 63 sectors/track, 130 cylinders

Units = cylinders of 16065 * 512 = 8225280 bytes



   Device Boot      Start         End      Blocks   Id  System

/dev/sdb1               1         130     1044193+  83  Linux



 

Disk /dev/sdc: 1073 MB, 1073741824 bytes

255 heads, 63 sectors/track, 130 cylinders

Units = cylinders of 16065 * 512 = 8225280 bytes



   Device Boot      Start         End      Blocks   Id  System

/dev/sdc1               1         130     1044193+  83  Linux



 

Disk /dev/sdd: 1073 MB, 1073741824 bytes

255 heads, 63 sectors/track, 130 cylinders

Units = cylinders of 16065 * 512 = 8225280 bytes



   Device Boot      Start         End      Blocks   Id  System

/dev/sdd1               1         130     1044193+  83  Linux



 

Disk /dev/sde: 10.7 GB, 10737418240 bytes

255 heads, 63 sectors/track, 1305 cylinders

Units = cylinders of 16065 * 512 = 8225280 bytes



   Device Boot      Start         End      Blocks   Id  System

/dev/sde1               1        1305    10482381   83  Linux





Disk /dev/sdf: 10.7 GB, 10737418240 bytes

255 heads, 63 sectors/track, 1305 cylinders

Units = cylinders of 16065 * 512 = 8225280 bytes

 

   Device Boot      Start         End      Blocks   Id  System

/dev/sdf1               1        1305    10482381   83  Linux

 

 

--查看节点node1的磁盘文件--

[root@node1 ~]# ll /dev/sd*

brw-r----- 1 root disk 8,  0 2013-03-28 /dev/sda

brw-r----- 1 root disk 8,  1 03-28 09:16 /dev/sda1

brw-r----- 1 root disk 8,  2 2013-03-28 /dev/sda2

brw-r----- 1 root disk 8,  3 03-28 09:16 /dev/sda3

brw-r----- 1 root disk 8, 16 03-28 09:24 /dev/sdb

brw-r----- 1 root disk 8, 17 03-28 09:24 /dev/sdb1

brw-r----- 1 root disk 8, 32 03-28 09:24 /dev/sdc

brw-r----- 1 root disk 8, 33 03-28 09:25 /dev/sdc1

brw-r----- 1 root disk 8, 48 03-28 09:26 /dev/sdd

brw-r----- 1 root disk 8, 49 03-28 09:26 /dev/sdd1

brw-r----- 1 root disk 8, 64 03-28 11:27 /dev/sde

brw-r----- 1 root disk 8, 65 03-28 11:27 /dev/sde1

brw-r----- 1 root disk 8, 80 03-28 11:29 /dev/sdf

brw-r----- 1 root disk 8, 81 03-28 11:29 /dev/sdf1

 

--发现节点node2看不到全部格式化好的磁盘文件,重启--

[root@node2 ~]# ll /dev/sd*

brw-r----- 1 root disk 8,  0 2013-03-28 /dev/sda

brw-r----- 1 root disk 8,  1 03-28 09:16 /dev/sda1

brw-r----- 1 root disk 8,  2 2013-03-28 /dev/sda2

brw-r----- 1 root disk 8,  3 03-28 09:16 /dev/sda3

brw-r----- 1 root disk 8, 16 2013-03-28 /dev/sdb

brw-r----- 1 root disk 8, 32 2013-03-28 /dev/sdc

brw-r----- 1 root disk 8, 48 2013-03-28 /dev/sdd

brw-r----- 1 root disk 8, 64 2013-03-28 /dev/sde

brw-r----- 1 root disk 8, 80 2013-03-28 /dev/sdf

[root@node2 ~]#init 6

 

[root@node2 ~]# ll /dev/sd*

brw-r----- 1 root disk 8,  0 2013-03-28 /dev/sda

brw-r----- 1 root disk 8,  1 03-28 11:40 /dev/sda1

brw-r----- 1 root disk 8,  2 2013-03-28 /dev/sda2

brw-r----- 1 root disk 8,  3 03-28 11:40 /dev/sda3

brw-r----- 1 root disk 8, 16 2013-03-28 /dev/sdb

brw-r----- 1 root disk 8, 17 2013-03-28 /dev/sdb1

brw-r----- 1 root disk 8, 32 2013-03-28 /dev/sdc

brw-r----- 1 root disk 8, 33 2013-03-28 /dev/sdc1

brw-r----- 1 root disk 8, 48 2013-03-28 /dev/sdd

brw-r----- 1 root disk 8, 49 2013-03-28 /dev/sdd1

brw-r----- 1 root disk 8, 64 2013-03-28 /dev/sde

brw-r----- 1 root disk 8, 65 2013-03-28 /dev/sde1

brw-r----- 1 root disk 8, 80 2013-03-28 /dev/sdf

brw-r----- 1 root disk 8, 81 2013-03-28 /dev/sdf1

 

--节点node1,初始化oracleasm,并创建asm--

[root@node1 ~]# /etc/init.d/oracleasm configure



Configuring the Oracle ASM library driver.



This will configure the on-boot properties of the Oracle ASM library

driver.  The following questions will determine whether the driver is

loaded on boot and what permissions it will have.  The current values

will be shown in brackets ('[]').  Hitting <ENTER> without typing an

answer will keep that current value.  Ctrl-C will abort.





Default user to own the driver interface []: grid

Default group to own the driver interface []: oinstall

Start Oracle ASM library driver on boot (y/n) [n]: y

Scan for Oracle ASM disks on boot (y/n) [y]: y

Writing Oracle ASM library driver configuration: done

Initializing the Oracle ASMLib driver:                     [  OK  ]

Scanning the system for Oracle ASMLib disks:               [  OK  ]





[root@node1 ~]# /etc/init.d/oracleasm createdisk CRS1 /dev/sdb1

Marking disk "CRS1" as an ASM disk:                        [  OK  ]

[root@node1 ~]# /etc/init.d/oracleasm createdisk CRS2 /dev/sdc1

Marking disk "CRS2" as an ASM disk:                        [  OK  ]

[root@node1 ~]# /etc/init.d/oracleasm createdisk CRS3 /dev/sdd1

Marking disk "CRS3" as an ASM disk:                        [  OK  ]

[root@node1 ~]# /etc/init.d/oracleasm createdisk DATA1 /dev/sde1

Marking disk "DATA1" as an ASM disk:                       [  OK  ]

[root@node1 ~]# /etc/init.d/oracleasm createdisk DATA2 /dev/sdf1

Marking disk "DATA2" as an ASM disk:                       [  OK  ]

[root@node1 ~]# /etc/init.d/oracleasm scandisks

Scanning the system for Oracle ASMLib disks:               [  OK  ]

[root@node1 ~]# /etc/init.d/oracleasm listdisks



CRS1

CRS2

CRS3

DATA1

DATA2

 

--节点node2,初始化oracleasm,并扫描asm--

[root@node2 ~]# /etc/init.d/oracleasm configure

Configuring the Oracle ASM library driver.



This will configure the on-boot properties of the Oracle ASM library

driver.  The following questions will determine whether the driver is

loaded on boot and what permissions it will have.  The current values

will be shown in brackets ('[]').  Hitting <ENTER> without typing an

answer will keep that current value.  Ctrl-C will abort.



Default user to own the driver interface []: grid

Default group to own the driver interface []: oinstall

Start Oracle ASM library driver on boot (y/n) [n]: y

Scan for Oracle ASM disks on boot (y/n) [y]: y

Writing Oracle ASM library driver configuration: done

Initializing the Oracle ASMLib driver:                     [  OK  ]

Scanning the system for Oracle ASMLib disks:               [  OK  ]



[root@node2 ~]# /etc/init.d/oracleasm scandisks

Scanning the system for Oracle ASMLib disks:               [  OK  ]
[root
@node2 ~]# /etc/init.d/oracleasm listdisks CRS1 CRS2 CRS3 DATA1 DATA2

 

--------------以下方式是利用udev进行asm创建,redhat5.5没有找到利用udev进行设置的方法,用OEL6.4做个例子------------------

首先确认是否安装了udev软件包



[root@node6 ~]# rpm -qa|grep udev

udev-095-14.21.el5



通过scsi_id获取设备的块设备的唯一标识名(在OEL6.4上面试验成功)

scsi_id -g -u /dev/sd*



在Oracle Linux 5下,可以使用如下命令(redhat5没成功):

# /sbin/scsi_id -g -u -s /block/sdb/sdb1

SATA_VBOX_HARDDISK_VBd306dbe0-df3367e3_



创建并配置UDEVRules 文件

[root@rac1 rules.d]# touch /etc/udev/rules.d/99-oracle-asmdevices.rules



添加如下内容:

KERNEL=="sd?1",BUS=="scsi", PROGRAM=="/sbin/scsi_id -g -u /dev/$name",RESULT=="1ATA_VBOX_HARDDISK_VB83552343-28d5a489",NAME="asm-disk1", OWNER="oracle", GROUP="dba",MODE="0660"

KERNEL=="sd?1",BUS=="scsi", PROGRAM=="/sbin/scsi_id -g -u /dev/$name", RESULT=="1ATA_VBOX_HARDDISK_VBb96d5ecb-4eae1d96",NAME="asm-disk2", OWNER="oracle", GROUP="dba",MODE="0660"

KERNEL=="sd?1",BUS=="scsi", PROGRAM=="/sbin/scsi_id -g -u /dev/$name",RESULT=="1ATA_VBOX_HARDDISK_VBfd7bba6c-b91fba70",NAME="asm-disk3", OWNER="oracle", GROUP="dba",MODE="0660"

KERNEL=="sd?1",BUS=="scsi", PROGRAM=="/sbin/scsi_id -g -u /dev/$name",RESULT=="1ATA_VBOX_HARDDISK_VB3239ed0d-db15bbec",NAME="asm-disk4", OWNER="oracle", GROUP="dba",MODE="0660"

 

--设置oracle与grid用户不需要密码ssh互联,并且要分别执行ssh xxx1/2 date,以防runcluvfy过不去--

--在10g RAC中必须手动配置成功,在11g RAC中不必,因为grid和db的图形安装中可以进行自动配置,简单粗暴。但是为了跑检查脚本,还是手动配置上吧--

[oracle@node1 ~]$ mkdir .ssh

[oracle@node1 ~]$ ssh-keygen -t rsa



Generating public/private rsa key pair.

Enter file in which to save the key (/home/oracle/.ssh/id_rsa):

Enter passphrase (empty for no passphrase):

Enter same passphrase again:

Your identification has been saved in /home/oracle/.ssh/id_rsa.

Your public key has been saved in /home/oracle/.ssh/id_rsa.pub.

The key fingerprint is:

e3:71:9f:fa:da:14:a9:cb:c1:01:17:a2:6d:9d:9d:5f oracle@node1



[oracle@node1 ~]$ ssh-keygen -t dsa



Generating public/private dsa key pair.

Enter file in which to save the key (/home/oracle/.ssh/id_dsa):

Enter passphrase (empty for no passphrase):

Enter same passphrase again:

Your identification has been saved in /home/oracle/.ssh/id_dsa.

Your public key has been saved in /home/oracle/.ssh/id_dsa.pub.

The key fingerprint is:

0b:bf:0e:7f:1c:c7:e6:59:9c:62:d4:42:ec:f7:3f:b0 oracle@node1



[oracle@node1 ~]$ cd .ssh

[oracle@node1 .ssh]$ ls

id_dsa  id_dsa.pub  id_rsa  id_rsa.pub





[oracle@node2 ~]$ mkdir .ssh

[oracle@node2 ~]$ ssh-keygen -t rsa



Generating public/private rsa key pair.

Enter file in which to save the key (/home/oracle/.ssh/id_rsa):

Enter passphrase (empty for no passphrase):

Enter same passphrase again:

Your identification has been saved in /home/oracle/.ssh/id_rsa.

Your public key has been saved in /home/oracle/.ssh/id_rsa.pub.

The key fingerprint is:

82:6c:3f:6f:20:ec:6f:e8:cc:1a:18:73:0b:cf:f5:04 oracle@node2



[oracle@node2 ~]$ ssh-keygen -t dsa



Generating public/private dsa key pair.

Enter file in which to save the key (/home/oracle/.ssh/id_dsa):

Enter passphrase (empty for no passphrase):

Enter same passphrase again:

Your identification has been saved in /home/oracle/.ssh/id_dsa.

Your public key has been saved in /home/oracle/.ssh/id_dsa.pub.

The key fingerprint is:

8f:e9:90:9e:d3:77:4f:87:a4:c6:5d:ff:ce:50:42:4f oracle@node2

 



[oracle@node1 ~]$ cat ~/.ssh/id_rsa.pub >> ~/.ssh/authorized_keys

[oracle@node1 ~]$ cat ~/.ssh/id_dsa.pub >> ~/.ssh/authorized_keys

[oracle@node1 ~]$ ssh node2 cat ~/.ssh/id_rsa.pub >> ~/.ssh/authorized_keys

[oracle@node1 ~]$ ssh node2 cat ~/.ssh/id_dsa.pub >> ~/.ssh/authorized_keys

[oracle@node1 ~]$ scp ~/.ssh/authorized_keys node2:~/.ssh/authorized_keys



[oracle@node1 ~]$ ssh node2 date

[oracle@node1 ~]$ ssh node1 date

[oracle@node2 ~]$ ssh node1 date

[oracle@node2 ~]$ ssh node2 date

 

[grid@node1 ~]$ mkdir .ssh

[grid@node1 ~]$ ssh-keygen -t rsa



Generating public/private rsa key pair.

Enter file in which to save the key (/home/grid/.ssh/id_rsa):

Enter passphrase (empty for no passphrase):

Enter same passphrase again:

Your identification has been saved in /home/grid/.ssh/id_rsa.

Your public key has been saved in /home/grid/.ssh/id_rsa.pub.

The key fingerprint is:

ce:e0:24:b0:4c:f2:25:7d:b9:e5:48:19:10:23:33:97 grid@node1



[grid@node1 ~]$ ssh-keygen -t dsa
Generating public
/private dsa key pair. Enter file in which to save the key (/home/grid/.ssh/id_dsa): Enter passphrase (empty for no passphrase): Enter same passphrase again. Your identification has been saved in /home/grid/.ssh/id_dsa. Your public key has been saved in /home/grid/.ssh/id_dsa.pub. The key fingerprint is: 11:78:88:1a:47:ff:c3:9f:a1:b1:d2:d9:d9:27:48:59 grid@node1 [grid@node2 ~]$ mkdir .ssh [grid@node2 ~]$ ssh-keygen -t rsa Generating public/private rsa key pair. Enter file in which to save the key (/home/grid/.ssh/id_rsa): Enter passphrase (empty for no passphrase): Enter same passphrase again: Your identification has been saved in /home/grid/.ssh/id_rsa. Your public key has been saved in /home/grid/.ssh/id_rsa.pub. The key fingerprint is: 53:d7:d5:e8:63:5a:c6:15:c4:81:11:80:9b:5f:f9:42 grid@node2 [grid@node2 ~]$ ssh-keygen -t dsa Generating public/private dsa key pair. Enter file in which to save the key (/home/grid/.ssh/id_dsa): Enter passphrase (empty for no passphrase): Enter same passphrase again: Your identification has been saved in /home/grid/.ssh/id_dsa. Your public key has been saved in /home/grid/.ssh/id_dsa.pub. The key fingerprint is: 9f:d7:3e:cf:e4:ff:29:df:c5:8c:85:46:6b:e2:07:1b grid@node2
[grid@node1 ~]$ cat ~/.ssh/id_rsa.pub >> ~/.ssh/authorized_keys [grid@node1 ~]$ cat ~/.ssh/id_dsa.pub >> ~/.ssh/authorized_keys [grid@node1 ~]$ ssh node2 cat ~/.ssh/id_rsa.pub >> ~/.ssh/authorized_keys [grid@node1 ~]$ ssh node2 cat ~/.ssh/id_dsa.pub >> ~/.ssh/authorized_keys [grid@node1 ~]$ scp ~/.ssh/authorized_keys node2:~/.ssh/authorized_keys [grid@node1 ~]$ ssh node2 date [grid@node1 ~]$ ssh node1 date [grid@node2 ~]$ ssh node1 date [grid@node2 ~]$ ssh node2 date

 

--在两个节点安装cvuqdisk包并验证--

在两个 Oracle RAC 节点上安装操作系统程序包 cvuqdisk。

如果没有 cvuqdisk,集群验证实用程序就无法发现共享磁盘,当运行(手动运行或在 Oracle Grid Infrastructure 安装结束时自动运行)集群验证实用程序时,会收到这样的错误消息:“Package cvuqdisk not installed”。使用适用于硬件体系结构(例如,x86_64 或 i386)的 cvuqdisk RPM。

cvuqdisk RPM 包含在 Oracle Grid Infrastructure 安装介质上的 rpm 目录中。

设置环境变量 CVUQDISK_GRP,使其指向作为 cvuqdisk 的所有者所在的组(本文为 oinstall):

[root@node1 rpm]# pwd

/oracle-tools/oracle11gR2_linux/grid/rpm

[root@node1 rpm]# ll

总计 12

-rwxr-xr-x 1 root root 8233 2011-09-22 cvuqdisk-1.0.9-1.rpm

[root@node1 rpm]# rpm -ivh *

Preparing...                                                            (100%########################################### [100%]

Using default group oinstall to install package

   1:cvuqdisk                                                           ( 99%########################################### [100####3

[grid@node1 ~]$ export CVUQDISK_GRP=oinstall

 

--安装前检查,最后会因为dns过不去,可以忽略。最后就是图形界面安装,先用grid用户安装grid,再用oracle用户安装db,之后用grid用户设置asm,最后oracle用户dbca建立数据库--

[grid@node1 grid]$ ./runcluvfy.sh stage -pre crsinst -n node1,node2 -fixup -verbose >> /opt/ora11g/check

你可能感兴趣的:(redhat)