ocfs2 rhel5.9 oracle10g rac总成问题汇总

问题一:

在配置ocfs2的过程中出现如下错误:
 
 
# ocfs2console

Configure Nodes --> Add --> 输入NODE名和IP --> OK --> Apply
出现如下错误:
o2cb_ctl: Unable to access cluster service while creating node
       Could not add node node1

解决办法:

将/etc/ocfs2/下的不正确的cluster.conf文件删掉,重新用ocfs2console 配置

问题二:

在安装CRS时最后执行root.sh时出现错误:
/usr/appsoft/oracle/product/10.2.0/crs_1/jdk/jre//bin/java: error while loading shared libraries: libpthread.so.0: cannot open shared object file: No such file or directory

解决办法:
编辑vipca与srvctl这两个文件,修改如下:
For vipca
if [ "$arch" = "i686" -o "$arch" = "ia64" -o "$arch" = "x86_64" ]
then
  LD_ASSUME_KERNEL=2.4.19
  export LD_ASSUME_KERNEL
fi

unset LD_ASSUME_KERNEL    //添加一条

For srvctl :
LD_ASSUME_KERNEL=2.4.19
export LD_ASSUME_KERNEL

unset LD_ASSUME_KERNEL   //添加一条


问题三:

运行vipca时出现如下错误:
# ./vipca
Error 0(Native: listNetInterfaces:[3])
[Error 0(Native: listNetInterfaces:[3])]
Don't mind,if  if you still have the OUI window open, click OK and it will create the "oifcfg" information, then cluvfy will fail due to vipca not completed successfully.On this time you can run vipca again,and it will normal,then return to the installer and cluvfy will succeed.

解决办法:

[root@his01a bin]# ./oifcfg iflist  //查看网卡信息

//添加网卡信息
[root@his01a bin]# ./oifcfg setif -global bond0/192.168.1.0:public 
[root@his01a bin]# ./oifcfg setif -global bond1/10.1.1.0:cluster_interconnect

//再次相看:
[root@his01a bin]# ./oifcfg getif
bond0  192.168.1.0  global  public
bond1  10.1.1.0  global  cluster_interconnect

再次运行正常:
[root@his01a bin]# ./vipca

问题四:(可以避免的错误)

在安装时意外中断, 重新安装clusterware 软件, 安装过程中没有问题,在最后执行root.sh 脚本的时候出现异常, 脚本一闪而过,没有执行成功:

[root@his01a oracle]# /oracle/product/10.2.0/crs_1/root.sh
WARNING: directory '/oracle/product/10.2.0' is not owned by root
WARNING: directory '/oracle/product' is not owned by root
WARNING: directory '/oracle' is not owned by root
Checking to see if Oracle CRS stack is already configured
Oracle CRS stack is already configured and will be running under init(1M)

解决方法:
 
删除每台机器上的/etc/oracle/scls_scr/rac1/oracle/cssfatal 文件即可.

问题五:

在测试主机与存储的过程中,断开一根连接主机与存储的光纤,主机在60s后自动重启,测试过程如下:

his01b:

第一次:于14:35:03拔掉 A控3口---连接his01b的pci1,60s后his01b重启
第二次:于14:45:02拔掉 B控3口---连接his01b的pci6  60s后his01b重启

his01a:
第一次:于15:26:01拔掉 A控4口---连接his01a的pci1,60s后his01a重启
第二次:于15:38:  拔掉 B控4口---连接his01a的pci6,60s后his01a重启

解决办法:
 经反复测试,问题由在ocfs2的心跳时间上,其心跳时间默认配置为60s,而我们在断开一根光纤后,主机切换连接存储链路的时间恰好也是60s,心跳在60s内无法正常工作,相应主机便会重启,这样我们只要修改其心跳时间就可以了:

[root@his01a ocfs2]# /etc/init.d/o2cb configure
Configuring the O2CB driver.

This will configure the on-boot properties of the O2CB driver.
The following questions will determine whether the driver is loaded on
boot.  The current values will be shown in brackets ('[]').  Hitting
<ENTER> without typing an answer will keep that current value.  Ctrl-C
will abort.

Load O2CB driver on boot (y/n) [y]:
Cluster stack backing O2CB [o2cb]:
Cluster to start on boot (Enter "none" to clear) [ocfs2]:
Specify heartbeat dead threshold (>=7) [31]:61          ****这个地方是ocfs心跳时间设置为120s,默认31即为60s;
Specify network idle timeout in ms (>=5000) [30000]:
Specify network keepalive delay in ms (>=1000) [2000]:
Specify network reconnect delay in ms (>=2000) [2000]:
Writing O2CB configuration: OK
Cluster ocfs2 already online

修改后,经测试正常。

 

你可能感兴趣的:(oracle10g)