标签:node 11gR2 RAC root etc sh DATA1 ORA
oracle 11gR2 RAC root.sh 错误 ORA-15072 ORA-15018
Oracle 11gR2 RAC 安装Clusterware 结束, 在第二个节点执行root.sh脚本的是报如下错误:
DiskGroup DATA1 creation failed with the following message:
ORA-15018: diskgroup cannot be created
ORA-15072: command requires at least 1 regular failure groups, discovered only 0
在Oracle 的官网搜了一下, 相关解释如下
Applies to:
Oracle Server - Enterprise Edition - Version: 11.2.0.1 and later [Release: 11.2 and later ]
Information in this document applies to any platform.
Symptoms
While installing Oracle Grid Infrastructure with ASM, root.sh ran successfully in first node, but fails on the second node.
Error example
1. root.sh failed on second node with following errors
-------------------------------------------------------
DiskGroup DATA1 creation failed with the following message:
ORA-15018: diskgroup cannot be created
ORA-15072: command requires at least 1 regular failure groups, discovered only 0
Configuration of ASM failed, see logs for details
2. rootcrs_nodename.log
-----------------------
2010-02-03 13:40:43: Configuring ASM via ASMCA
2010-02-03 13:40:43: Executing as oracle: /u01/app/1120/grid/bin/asmca -silent -diskGroupName DATA1 -diskList ORCL:DATA1 -redundancy EXTERNAL -configureLocalASM
2010-02-03 13:40:43: Running as user oracle: /u01/app/1120/grid/bin/asmca -silent -diskGroupName DATA1 -diskList ORCL:DATA1 -redundancy EXTERNAL -configureLocalASM
2010-02-03 13:40:43: Invoking "/u01/app/1120/grid/bin/asmca -silent -diskGroupName DATA1 -diskList ORCL:DATA1 -redundancy EXTERNAL -configureLocalASM" as user "oracle"
2010-02-03 13:40:51: Configuration of ASM failed, see logs for details
3. On the 2nd node
/etc/oratab files shows +ASM1, rather than +ASM2
4. The following commands on the 2nd node show the ASM disk information correctly
/etc/init.d/oracleasm listdisks
/etc/init.d/oracleasm scandisks
ls -ltr /dev/oracleasm/disks
Cause
After configuring multipath disks on Linux x86-64, proper parameters have not been configured in /etc/sysconfig/oracleasm
Solution
On all nodes,
1. Modify the /etc/sysconfig/oracleasm with:
ORACLEASM_SCANORDER="dm"
ORACLEASM_SCANEXCLUDE="sd"
2. restart the asmlib by :
# /etc/init.d/oracleasm restart
3. Run root.sh on the 2nd node
在这里要注意的事, 在运行root.sh 脚本的时候,会报如下错误:
Using configuration parameter file: /oracle/grid/crs/install/crsconfig_params
CRS is already configured on this node for crshome=0
Cannot configure two CRS instances on the same cluster.
Please deconfigure before proceeding with the configuration of new home.
因为之前已经运行过了。 我们需要把之前注册的信息 删除之后,在运行root.sh 脚本。 执行如下命令,删除节点注册信息:
# $ORACLE_HOME/crs/install/rootcrs.pl -deconfig -force -verbose
待删除完成后在执行,就不会报错了。
标签:node,11gR2,RAC,root,etc,sh,DATA1,ORA 来源: https://blog.51cto.com/lhrbest/2703321
本站声明: 1. iCode9 技术分享网(下文简称本站)提供的所有内容,仅供技术学习、探讨和分享; 2. 关于本站的所有留言、评论、转载及引用,纯属内容发起人的个人观点,与本站观点和立场无关; 3. 关于本站的所有言论和文字,纯属内容发起人的个人观点,与本站观点和立场无关; 4. 本站文章均是网友提供,不完全保证技术分享内容的完整性、准确性、时效性、风险性和版权归属;如您发现该文章侵犯了您的权益,可联系我们第一时间进行删除; 5. 本站为非盈利性的个人网站,所有内容不会用来进行牟利,也不会利用任何形式的广告来间接获益,纯粹是为了广大技术爱好者提供技术内容和技术思想的分享性交流网站。