SUN CLUSTER问题
在配置WEBLOGIC应用到sun cluster总是出现错误信息,但是网上没有看见太多的解释,希望各位大拿帮忙。bash-3.00# ./scstat -pv
------------------------------------------------------------------
-- Cluster Nodes --
Node name Status
--------- ------
Cluster node: SUN10A Online
Cluster node: SUN10B Online
------------------------------------------------------------------
-- Cluster Transport Paths --
Endpoint Endpoint Status
-------- -------- ------
Transport path: SUN10A:e1000g2 SUN10B:e1000g3 Path online
Transport path: SUN10A:e1000g3 SUN10B:e1000g2 Path online
------------------------------------------------------------------
-- Quorum Summary from latest node reconfiguration --
Quorum votes possible: 3
Quorum votes needed: 2
Quorum votes present: 2
-- Quorum Votes by Node (current status) --
Node Name Present Possible Status
--------- ------- -------- ------
Node votes: SUN10A 1 1 Online
Node votes: SUN10B 1 1 Online
-- Quorum Votes by Device (current status) --
Device Name Present Possible Status
----------- ------- -------- ------
Device votes: /dev/did/rdsk/d1s21 1 Online
------------------------------------------------------------------
-- Device Group Servers --
Device Group Primary Secondary
------------ ------- ---------
-- Device Group Spares --
Device Group Spare Nodes
------------ -----------
-- Device Group Inactives --
Device Group Inactive Nodes
------------ --------------
-- Device Group Transitions --
Device Group In Transition Nodes
------------ -------------------
-- Device Group Status --
Device Group Status
------------ ------
-- Multi-owner Device Groups --
Device Group Online Status
------------ -------------
------------------------------------------------------------------
-- Resource Groups and Resources --
Group Name Resources
---------- ---------
Resources: scgroup srvip weblogicpool
-- Resource Groups --
Group Name Node Name State Suspended
---------- --------- ----- ---------
Group: scgroup SUN10A Online Yes
Group: scgroup SUN10B Offline Yes
-- Resources --
Resource NameNode Name State Status Message
---------------------- ----- --------------
Resource: srvip SUN10A Online Online - LogicalHostname online.
Resource: srvip SUN10B Offline Offline
Resource: weblogicpool SUN10A Online Online
Resource: weblogicpool SUN10B Offline Offline
------------------------------------------------------------------
-- IPMP Groups --
Node Name Group Status Adapter Status
--------- ----- ------ ------- ------
IPMP Group: SUN10A sc_ipmp0 Online e1000g0 Online
IPMP Group: SUN10B sc_ipmp0 Online e1000g0 Online
-- IPMP Groups in Zones --
Zone Name Group Status Adapter Status
--------- ----- ------ ------- ------
------------------------------------------------------------------
bash-3.00# ./scdidadm -L
1 SUN10A:/dev/rdsk/c1t0d0 /dev/did/rdsk/d1
1 SUN10B:/dev/rdsk/c1t0d0 /dev/did/rdsk/d1
2 SUN10A:/dev/rdsk/c0t0d0 /dev/did/rdsk/d2
3 SUN10B:/dev/rdsk/c0t1d0 /dev/did/rdsk/d3
4 SUN10B:/dev/rdsk/c0t0d0 /dev/did/rdsk/d4
bash-3.00# ./scconf -pv
Cluster name: cluster
Cluster ID: 0x53CD3FBE
Cluster install mode: disabled
Cluster private net: 172.16.0.0
Cluster private netmask: 255.255.252.0
Cluster maximum nodes: 62
Cluster maximum private networks: 10
Cluster new node authentication: unix
Cluster authorized-node list: <. - Exclude all nodes>
Cluster transport heart beat timeout: 10000
Cluster transport heart beat quantum: 1000
Round Robin Load Balancing UDP session timeout: 480
Cluster nodes: SUN10A SUN10B
Cluster node name: SUN10A
(SUN10A) Node ID: 1
(SUN10A) Node enabled: yes
(SUN10A) Node private hostname: clusternode1-priv
(SUN10A) Node quorum vote count: 1
(SUN10A) Node reservation key: 0x53CD3FBE00000001
(SUN10A) Node zones: <NULL>
(SUN10A) CPU shares for global zone: 1
(SUN10A) Minimum CPU requested for global zone:1
(SUN10A) Node transport adapters: e1000g3 e1000g2
(SUN10A) Node transport adapter: e1000g3
(SUN10A:e1000g3) Adapter enabled: yes
(SUN10A:e1000g3) Adapter transport type: dlpi
(SUN10A:e1000g3) Adapter property: device_name=e1000g
(SUN10A:e1000g3) Adapter property: device_instance=3
(SUN10A:e1000g3) Adapter property: lazy_free=1
(SUN10A:e1000g3) Adapter property: dlpi_heartbeat_timeout=10000
(SUN10A:e1000g3) Adapter property: dlpi_heartbeat_quantum=1000
(SUN10A:e1000g3) Adapter property: nw_bandwidth=80
(SUN10A:e1000g3) Adapter property: bandwidth=70
(SUN10A:e1000g3) Adapter property: ip_address=172.16.0.65
(SUN10A:e1000g3) Adapter property: netmask=255.255.255.192
(SUN10A:e1000g3) Adapter port names: 0
(SUN10A:e1000g3) Adapter port: 0
(SUN10A:e1000g3@0) Port enabled: yes
(SUN10A) Node transport adapter: e1000g2
(SUN10A:e1000g2) Adapter enabled: yes
(SUN10A:e1000g2) Adapter transport type: dlpi
(SUN10A:e1000g2) Adapter property: device_name=e1000g
(SUN10A:e1000g2) Adapter property: device_instance=2
(SUN10A:e1000g2) Adapter property: lazy_free=1
(SUN10A:e1000g2) Adapter property: dlpi_heartbeat_timeout=10000
(SUN10A:e1000g2) Adapter property: dlpi_heartbeat_quantum=1000
(SUN10A:e1000g2) Adapter property: nw_bandwidth=80
(SUN10A:e1000g2) Adapter property: bandwidth=70
(SUN10A:e1000g2) Adapter property: ip_address=172.16.0.129
(SUN10A:e1000g2) Adapter property: netmask=255.255.255.192
(SUN10A:e1000g2) Adapter port names: 0
(SUN10A:e1000g2) Adapter port: 0
(SUN10A:e1000g2@0) Port enabled: yes
Cluster node name: SUN10B
(SUN10B) Node ID: 2
(SUN10B) Node enabled: yes
(SUN10B) Node private hostname: clusternode2-priv
(SUN10B) Node quorum vote count: 1
(SUN10B) Node reservation key: 0x53CD3FBE00000002
(SUN10B) Node zones: <NULL>
(SUN10B) CPU shares for global zone: 1
(SUN10B) Minimum CPU requested for global zone:1
(SUN10B) Node transport adapters: e1000g2 e1000g3
(SUN10B) Node transport adapter: e1000g2
(SUN10B:e1000g2) Adapter enabled: yes
(SUN10B:e1000g2) Adapter transport type: dlpi
(SUN10B:e1000g2) Adapter property: device_name=e1000g
(SUN10B:e1000g2) Adapter property: device_instance=2
(SUN10B:e1000g2) Adapter property: lazy_free=1
(SUN10B:e1000g2) Adapter property: dlpi_heartbeat_timeout=10000
(SUN10B:e1000g2) Adapter property: dlpi_heartbeat_quantum=1000
(SUN10B:e1000g2) Adapter property: nw_bandwidth=80
(SUN10B:e1000g2) Adapter property: bandwidth=70
(SUN10B:e1000g2) Adapter property: ip_address=172.16.0.66
(SUN10B:e1000g2) Adapter property: netmask=255.255.255.192
(SUN10B:e1000g2) Adapter port names: 0
(SUN10B:e1000g2) Adapter port: 0
(SUN10B:e1000g2@0) Port enabled: yes
(SUN10B) Node transport adapter: e1000g3
(SUN10B:e1000g3) Adapter enabled: yes
(SUN10B:e1000g3) Adapter transport type: dlpi
(SUN10B:e1000g3) Adapter property: device_name=e1000g
(SUN10B:e1000g3) Adapter property: device_instance=3
(SUN10B:e1000g3) Adapter property: lazy_free=1
(SUN10B:e1000g3) Adapter property: dlpi_heartbeat_timeout=10000
(SUN10B:e1000g3) Adapter property: dlpi_heartbeat_quantum=1000
(SUN10B:e1000g3) Adapter property: nw_bandwidth=80
(SUN10B:e1000g3) Adapter property: bandwidth=70
(SUN10B:e1000g3) Adapter property: ip_address=172.16.0.130
(SUN10B:e1000g3) Adapter property: netmask=255.255.255.192
(SUN10B:e1000g3) Adapter port names: 0
(SUN10B:e1000g3) Adapter port: 0
(SUN10B:e1000g3@0) Port enabled: yes
Cluster transport switches: switch1 switch2
Cluster transport switch: switch1
(switch1) Switch enabled: yes
(switch1) Switch type: switch
(switch1) Switch port names: 1 2
(switch1) Switch port: 1
(switch1@1) Port enabled: yes
(switch1) Switch port: 2
(switch1@2) Port enabled: yes
Cluster transport switch: switch2
(switch2) Switch enabled: yes
(switch2) Switch type: switch
(switch2) Switch port names: 1 2
(switch2) Switch port: 1
(switch2@1) Port enabled: yes
(switch2) Switch port: 2
(switch2@2) Port enabled: yes
Cluster transport cables
Endpoint Endpoint State
-------- -------- -----
Transport cable: SUN10A:e1000g3@0 switch1@1 Enabled
Transport cable: SUN10A:e1000g2@0 switch2@1 Enabled
Transport cable: SUN10B:e1000g2@0 switch1@2 Enabled
Transport cable: SUN10B:e1000g3@0 switch2@2 Enabled
Quorum devices: d1
Quorum device name: d1
(d1) Quorum device votes: 1
(d1) Quorum device enabled: yes
(d1) Quorum device name: /dev/did/rdsk/d1s2
(d1) Quorum device type: shared_disk
(d1) Quorum device hosts (enabled): SUN10A SUN10B
(d1) Quorum device hosts (disabled):
(d1) Quorum device access mode: scsi2
bash-3.00#
bash-3.00# ./scrgadm -a -j weblogic -g scgroup -t SUNW.wls:4.1 \
> -x Confdir_list=/weblogic/wlserver_10.3 \
> -x Server_url=http://srvip:7001 \
> -x Start_script=/weblogic/user_projects/domains/base_domain/bin/startWeblogic.sh
SUN10A - Cannot access file: /weblogic/user_projects/domains/base_domain/bin/startWeblogic.sh (Bad file number)
(C189917) VALIDATE on resource weblogic, resource group scgroup, exited with non-zero exit status.
(C720144) Validation of resource weblogic in resource group scgroup on node SUN10A failed.
Jul 22 15:50:46 SUN10A Cluster.RGM.global.rgmd: method <hafoip_start> completed successfully for resource <srvip>, resource group <scgroup>, node <SUN10A>, time used: 0% of timeout <500 seconds>
Jul 22 15:50:46 SUN10A Cluster.RGM.global.rgmd: launching method <hafoip_monitor_start> for resource <srvip>, resource group <scgroup>, node <SUN10A>, timeout <300> seconds
Jul 22 15:50:46 SUN10A Cluster.RGM.global.rgmd: method <hafoip_monitor_start> completed successfully for resource <srvip>, resource group <scgroup>, node <SUN10A>, time used: 0% of timeout <300 seconds>
Jul 22 15:52:42 SUN10A Cluster.RGM.global.rgmd: launching method <wls_validate> for resource <weblogic>, resource group <scgroup>, node <SUN10A>, timeout <180> seconds
Jul 22 15:52:43 SUN10A SC[,SUNW.wls:4.1,scgroup,weblogic,wls_validate]: Cannot access file: /weblogic/user_projects/domains/base_domain/bin/startWeblogic.sh (No such file or directory)
Jul 22 15:52:43 SUN10A Cluster.RGM.global.rgmd: VALIDATE failed on resource <weblogic>, resource group <scgroup>, time used: 0% of timeout <180, seconds>
bash-3.00# zfs list
NAME USEDAVAILREFERMOUNTPOINT
rpool 9.28G6.35G32.5K/rpool
rpool/ROOT 6.79G6.35G 21Klegacy
rpool/ROOT/s10x_u9wos_14a6.79G6.35G6.79G/
rpool/dump 768M6.35G 768M-
rpool/export 1.21G6.35G 23K/export
rpool/export/home 1.21G6.35G1.21G/export/home
rpool/swap 544M6.73G 154M-
weblogicpool 648M9.15G 21K/weblogicpool
weblogicpool/weblogic 648M9.15G 648M/weblogic
bash-3.00# ls /
Desktop core export lib platform system vol
Documents dev global mnt proc tmp weblogic
bin devices home net rpool usr weblogicpool
boot etc kernel opt sbin var zfs.log
Jul 22 15:52:43 SUN10A SC[,SUNW.wls:4.1,scgroup,weblogic,wls_validate]: Cannot access file: /weblogic/user_projects/domains/base_domain/bin/startWeblogic.sh (No such file or directory) 回复 2# q1208c
字母敲错。。。。
bash-3.00# ./scrgadm -a -j weblogic -g scgroup -t SUNW.wls:4.1 \
> -x Confdir_list=/weblogic/wlserver_10.3 \
> -x Server_url=http://srvip:7001 \
> -x Start_script=/weblogic/user_projects/domains/base_domain/bin/startWebLogic.sh
SUN10B - stat of file /weblogic/wlserver_10.3/server/bin/setWLSEnv.sh failed: <Bad file number>.
(C189917) VALIDATE on resource weblogic, resource group scgroup, exited with non-zero exit status.
(C720144) Validation of resource weblogic in resource group scgroup on node SUN10B failed.
是不是因为我现在的共享存储放在SUN10A上,所以造成SUN10B不能访问,而出现错误?该怎么解决呢,大师 回复 3# seallove
我只配过两回 cluster.
共享存储 应该两台都可以访问的吧?
不然你怎么切换呀? 没有存储资源? ./scrgadm -a -j weblogic -g scgroup -t SUNW.wls:4.1 \
> -x Confdir_list=/weblogic/wlserver_10.3 \
> -x Server_url=http://srvip:7001 \
> -x Start_script=/weblogic/user_projects/domains/base_domain/bin/startWebLogic.sh
> -x Server_name=SUN10A
加了这个,现在好像好了。但是还是会报错 回复 4# q1208c
估计是我配置的有问题。配置的ZFS只能单边访问和挂载。 回复 7# seallove
如果是nfs的话, 好象要把 locker 放在 共享的存储上.
zfs 我没玩过 ...
回复 3# seallove
zfs不是全局文件系统,so 你只能在两台机器的本地都配置weblogic,用HA软件来控制两边的启停~
页:
[1]