免费注册 查看新帖 |

Chinaunix

  平台 论坛 博客 文库
12下一页
最近访问板块 发新帖
查看: 3748 | 回复: 10
打印 上一主题 下一主题

双机起不来了,跪求sun cluster3.0和agent安装包 [复制链接]

论坛徽章:
0
跳转到指定楼层
1 [收藏(0)] [报告]
发表于 2012-05-09 09:57 |只看该作者 |倒序浏览
愁死我了,两台sun880和两台3510阵列,solaris8

双机启动不了,boot -x没有问题。
启动信息提示如下:
SUNW,pci-gem0: Using Gigabit SERDES Interface
SUNW,pci-gem0: Auto-Negotiated 1000 Mbps Full-Duplex Link Up
Could not read symbolic link for: /dev/rdsk/c2t44d1s2 path not loaded
        No such file or directory
Could not read symbolic link for: /dev/rdsk/c2t44d0s2 path not loaded
        No such file or directory
Booting as part of a cluster
NOTICE: CMM: Node hh_db1 (nodeid = 1) with votecount = 1 added.
NOTICE: CMM: Node hh_db2 (nodeid = 2) with votecount = 1 added.
NOTICE: CMM: Quorum device 1 (/dev/did/rdsk/d17s2) added; votecount = 1, bitmask of nodes with configured paths = 0x3.
WARNING: CMM: Open failed with error '(No such device or address)' and errno = 6 for quorum device 1 with gdevname '/dev/did/rdsk/d17s2'.
NOTICE: clcomm: Adapter eri0 constructed
NOTICE: clcomm: Path hh_db2:eri0 - hh_db1:eri0 being constructed
NOTICE: clcomm: Adapter qfe0 constructed
NOTICE: clcomm: Path hh_db2:qfe0 - hh_db1:qfe0 being constructed
NOTICE: CMM: Node hh_db2: attempting to join cluster.
NOTICE: clcomm: Path hh_db2:qfe0 - hh_db1:qfe0 being initiated
NOTICE: clcomm: Path hh_db2:qfe0 - hh_db1:qfe0 online
NOTICE: CMM: Node hh_db1 (nodeid: 1, incarnation #: 1336489999) has become reachable.
NOTICE: CMM: Cluster has reached quorum.
NOTICE: CMM: Node hh_db1 (nodeid = 1) is up; new incarnation number = 1336489999.
NOTICE: CMM: Node hh_db2 (nodeid = 2) is up; new incarnation number = 1336490017.
NOTICE: CMM: Cluster members: hh_db1 hh_db2.
NOTICE: CMM: node reconfiguration #1 completed.
NOTICE: CMM: Node hh_db2: joined cluster.
NOTICE: CCR: Waiting for repository synchronization to finish.
NOTICE: clcomm: Path hh_db2:eri0 - hh_db1:eri0 being initiated
NOTICE: clcomm: Path hh_db2:eri0 - hh_db1:eri0 online
Could not read symbolic link for: /dev/rdsk/c2t44d1s2 path not loaded
        No such file or directory
Could not read symbolic link for: /dev/rdsk/c2t44d0s2 path not loaded
        No such file or directory
VxVM general startup...
The system is coming up.  Please wait.
Starting cpudiagd ... done.
starting rpc services: rpcbind done.
Setting netmask of lo0:1 to 255.255.255.255
Setting netmask of ce1 to 255.255.255.0
Setting netmask of ge0 to 255.255.255.0
Setting netmask of eri0 to 255.255.255.128
Setting netmask of eri0:1 to 255.255.255.252
Setting netmask of qfe0 to 255.255.255.128
Setting default IPv4 interface for multicast: add net 224.0/4: gateway hh_db2
syslog service starting.
obtaining access to all attached disks
Configuring the /dev/global directory (global devices)
Print services started.
volume management starting.

panic[cpu0]/thread=30005a2e3a0: CMM: Cluster lost operational quorum; aborting.

000002a1026d7450 cl_runtime:__0FZsc_syslog_msg_log_no_argsPviTCPCcTB+60 (30005f74000, 3, 0, 784c8a7c, 2a1026d7650, 3)
  %l0-3: 00000300055900e0 0000000000000e90 0000000000000001 0000030003348538
  %l4-7: 0000000000000002 0000000000000000 0000000000000000 0000030003348810
000002a1026d7500 cl_runtime:__0f5CosNsc_syslog_msgDlogiTBPCce+1c (30004fcc4d8, 3, 0, 784c8a7c, 0, 3000334853
  %l0-3: 000002a1026d77c0 0000000000000001 0000030003348538 0000000000000002
  %l4-7: 0000000000000000 0000000000000000 0000000000000000 0000000000000000
000002a1026d75b0 cl_comm:__0fOautomaton_implMqcheck_statev+6ec (30004f2c008, 1, 2a1026d76bc, 784c9790, 30004f2c080, 0)
  %l0-3: 00000000783a6ff0 0000030004f2c1c0 0000030004f2c008 0000000000000000
  %l4-7: 0000000000000041 0000030001d75ea8 0000000000000002 0000030001d75ed0


怀疑是did和quorum device有问题了,但是双机起不来,怎么更改did和quorum device啊?

是不是只能重新装cluster了?没有3.0的安装包啊。

论坛徽章:
20
申猴
日期:2013-09-12 19:39:05狮子座
日期:2014-07-20 21:19:51寅虎
日期:2014-08-16 18:37:47水瓶座
日期:2014-10-15 18:58:25天蝎座
日期:2015-01-22 18:19:15NBA常规赛纪念章
日期:2015-05-04 22:32:032015亚冠之胡齐斯坦钢铁
日期:2015-06-03 11:28:502015亚冠之吉达阿赫利
日期:2015-09-19 12:41:47午马
日期:2013-09-18 14:36:40戌狗
日期:2013-09-18 14:44:39处女座
日期:2013-09-24 17:46:41CU十二周年纪念徽章
日期:2013-10-24 15:41:34
2 [报告]
发表于 2012-05-09 10:07 |只看该作者
boot  -x进系统看3510是否正常,看网络是否正常

论坛徽章:
0
3 [报告]
发表于 2012-05-09 10:13 |只看该作者
boot  -x进系统后3510正常,网络也正常,手工挂载了veritas的相关卷,暂时单机恢复了业务

注:两台主机boot -x状态都正常

论坛徽章:
2
2015年迎新春徽章
日期:2015-03-04 09:58:11IT运维版块每日发帖之星
日期:2015-10-04 06:20:00
4 [报告]
发表于 2012-05-09 10:36 |只看该作者
检查链路吧,双机心跳和仲裁都是好的。可能是资源组挂载卷有问题,导致宕机。

论坛徽章:
0
5 [报告]
发表于 2012-05-09 11:08 |只看该作者
rf00147 发表于 2012-05-09 10:36
检查链路吧,双机心跳和仲裁都是好的。可能是资源组挂载卷有问题,导致宕机。



启动提示Quorum device 报错

还有就是veritas的dg也不能import进来,我怀疑是did出问题了,导致的

现在双机一启动就马上重启,没有办法进cluster状态,怎么修改啊

论坛徽章:
2
2015年迎新春徽章
日期:2015-03-04 09:58:11IT运维版块每日发帖之星
日期:2015-10-04 06:20:00
6 [报告]
发表于 2012-05-09 11:57 |只看该作者
仲裁报错,也不会宕机,心跳通的就不会宕,还有就是看你的日志他是加入cluster后宕的回复 5# abbend


   

论坛徽章:
0
7 [报告]
发表于 2012-05-09 19:39 |只看该作者
实在没有办法了,最后用cluster3.1重建了,故障解决

论坛徽章:
2
双鱼座
日期:2014-02-23 12:10:03操作系统版块每日发帖之星
日期:2015-12-17 06:20:00
8 [报告]
发表于 2012-05-10 00:21 |只看该作者
二台主机都报这个错吗:panic[cpu0]/thread=30005a2e3a0: CMM: Cluster lost operational quorum; aborting.

论坛徽章:
1
狮子座
日期:2013-09-02 12:10:41
9 [报告]
发表于 2012-05-10 10:02 |只看该作者
本帖最后由 byuq 于 2012-05-10 10:12 编辑

/dev/rdsk/c2t44d1s2
/dev/rdsk/c2t44d0s2
其中一块硬盘肯定是/dev/did/rdsk/d17s2,所以导致票盘失效。


Scenario:
     * The quorum device has failed.
     * All nodes are out of cluster mode.

Solution:

Use the following procedures to remove the existing quorum device and
create a new quorum device.

Example for two node cluster, named node-0 and node-1:
------------------------------------------------------

   1.  Reboot each node with boot -x command.
   2.  Edit the /etc/cluster/ccr/infrastructure file on all nodes.
   3.  Change the installmode from disabled to enabled:

        cluster.properties.installmode  enabled

   4.  Delete all quorum devices by removing lines that start with
       cluster.quorum_devices and save the file.

        cluster.quorum_devices.1.name   d4
        cluster.quorum_devices.1.state  enabled
        cluster.quorum_devices.1.properties.votecount   1
        cluster.quorum_devices.1.properties.gdevname    /dev/did/rdsk/d4s2
        cluster.quorum_devices.1.properties.path_1      enabled
        cluster.quorum_devices.1.properties.path_2      enabled

   5.  Regenerate the checksum of the infrastructure file by running the
       following command on node-0:

        /usr/cluster/lib/sc/ccradm -i /etc/cluster/ccr/infrastructure -o
       
   6.  Regenerate the checksum of the infrastructure file by running the
       following command on node-1:

        /usr/cluster/lib/sc/ccradm -i /etc/cluster/ccr/infrastructure

   7.  Boot both nodes back into cluster mode.
   8.  Rerun scsetup to reset installmode and recreate the quorum device.

Example for three node cluster (OPS/RAC environment:
--------------------------------------------------

   1.  Delete all references to the quorum device by removing lines that
       start with cluster.quorum_devices on all nodes and save the file

        cluster.quorum_devices.1.name   d4
        cluster.quorum_devices.1.state  enabled
        cluster.quorum_devices.1.properties.votecount   1
        cluster.quorum_devices.1.properties.gdevname    /dev/did/rdsk/d4s2
        cluster.quorum_devices.1.properties.path_1      enabled
        cluster.quorum_devices.1.properties.path_2      enabled

   2.  Regenerate the checksum of the infrastructure file by running the
       following command on node-0:

        /usr/cluster/lib/sc/ccradm -i /etc/cluster/ccr/infrastructure -o
       
   3.  Regenerate the checksum of the infrastructure file by running the
       following command on node-1:

        /usr/cluster/lib/sc/ccradm -i /etc/cluster/ccr/infrastructure

   4.  Regenerate the checksum of the infrastructure file by running the
       following command on node-2:

        /usr/cluster/lib/sc/ccradm -i /etc/cluster/ccr/infrastructure

   5.  Boot first node into cluster mode. Node will hang at:

        NOTICE: CMM: Cluster doesn't have operational quorum yet; waiting for
quorum.

   6.  Boot second node into cluster mode.
   7.  The first and second nodes should meet quorum requirements and boot
       into cluster mode.
   8.  Boot third node into cluster mode. This node should join cluster
       without issue.

   NOTE:  With configurations larger than three nodes, continue booting
   nodes until quorum count is reached.

   9.  Once all nodes are in cluster mode, run scsetup to create new quorum
       device.
   10. Ensure new quorum device is allotted correct number of votes based
       on number of nodes able to access the quorum device.

论坛徽章:
0
10 [报告]
发表于 2012-05-10 15:47 |只看该作者
东方蜘蛛 发表于 2012-05-10 00:21
二台主机都报这个错吗:panic[cpu0]/thread=30005a2e3a0: CMM: Cluster lost operational quorum; abort ...


另一台当时主要是dg无法import,但是我尝试手动可以import的。报错如下:
May  9 15:16:44  Cluster.Framework: stderr: /usr/cluster/lib/sc/run_reserve[26]: 886 Segmentation Fault(coredump)
May  9 15:16:44  Cluster.Framework: stderr: vxvm:vxdg: ERROR
May  9 15:16:44  Cluster.Framework: stderr: : Disk group ipasdg: No such disk group is imported
Fatal error: could not deport VxVM diskgroup ipasdg. Halting node.

root@ # May  9 15:16:44 /usr/lib/snmp/snmpdx: received signal 15
May  9 15:16:44 rpcbind: rpcbind terminating on signal.
syncing file systems... done
WARNING: CMM: Node being shut down.

感谢各位
您需要登录后才可以回帖 登录 | 注册

本版积分规则 发表回复

  

北京盛拓优讯信息技术有限公司. 版权所有 京ICP备16024965号-6 北京市公安局海淀分局网监中心备案编号:11010802020122 niuxiaotong@pcpop.com 17352615567
未成年举报专区
中国互联网协会会员  联系我们:huangweiwei@itpub.net
感谢所有关心和支持过ChinaUnix的朋友们 转载本站内容请注明原作者名及出处

清除 Cookies - ChinaUnix - Archiver - WAP - TOP