- 论坛徽章:
- 1
|
提出我的node1,node2上 /etc/hosts
node1
[root@ecp-app1 ~]# more /etc/hosts
# Do not remove the following line, or various programs
# that require network functionality will fail.
127.0.0.1 localhost.localdomain localhost
::1 localhost6.localdomain6 localhost6
#134.175.13.76 ecp-app-cluster-node1.ecp
#134.175.13.80 ecp-app-cluster-node2.ecp
#134.175.13.77 ecp_app_cluster_service
134.175.13.76 ecp-app1
134.175.13.80 ecp-app2
192.168.10.1 ecp-app1.cluster.com ecp-app1
192.168.10.2 ecp-app2.cluster.com ecp-app2
#134.175.13.76 ecp-app1.cluster ecp-app1
#134.175.13.80 ecp-app2.cluster ecp-app2
#134.175.13.77 ecp_app_cluster_service
node2
[root@ecp-app2 etc]# more hosts
# Do not remove the following line, or various programs
# that require network functionality will fail.
127.0.0.1 localhost.localdomain localhost
::1 localhost6.localdomain6 localhost6
#134.175.13.76 ecp-app-cluster-node1.ecp
#134.175.13.80 ecp-app-cluster-node2.ecp
#134.175.13.77 ecp_app_cluster_service
134.175.13.76 ecp-app1
134.175.13.80 ecp-app2
192.168.10.1 ecp-app1.cluster.com ecp-app1
192.168.10.2 ecp-app2.cluster.com ecp-app2
#134.175.13.76 ecp-app1.cluster ecp-app1
#134.175.13.80 ecp-app2.cluster ecp-app2
#134.175.13.77 ecp_app_cluster_service
[root@ecp-app2 etc]# more /etc/cluster/cluster.conf
<?xml version="1.0"?>
<cluster alias="ecp-app-ha" config_version="12" name="ecp-app-ha">
<fence_daemon clean_start="0" post_fail_delay="15" post_join_delay="25"/>
<clusternodes>
<clusternode name="ecp-app2.cluster.com" nodeid="1" votes="1">
<fence>
<method name="1">
<device name="ecp_app_fence2"/>
</method>
</fence>
</clusternode>
<clusternode name="ecp-app1.cluster.com" nodeid="2" votes="1">
<fence>
<method name="1">
<device name="ecp_app_fence1"/>
</method>
</fence>
</clusternode>
</clusternodes>
<cman expected_votes="1" two_node="1"/>
<fencedevices>
<fencedevice agent="fence_ipmilan" ipaddr="134.175.13.78" login="root" name="ecp_app_fence1" passwd="science"/>
<fencedevice agent="fence_ipmilan" ipaddr="134.175.13.82" login="root" name="ecp_app_fence2" passwd="science"/>
</fencedevices>
<rm>
<failoverdomains>
<failoverdomain name="ecp_failover_domain" nofailback="0" ordered="0" restricted="1">
<failoverdomainnode name="ecp-app2.cluster.com" priority="1"/>
<failoverdomainnode name="ecp-app1.cluster.com" priority="1"/>
</failoverdomain>
</failoverdomains>
<resources>
<ip address="134.175.13.77" monitor_link="1"/>
<clusterfs device="/dev/mapper/appvg-lv_ecp_98g" force_unmount="1" fsid="49425" fstype="gfs" mountpoint="/ec
p" name="ecp_app_storage" self_fence="1"/>
</resources>
<service autostart="1" domain="ecp_failover_domain" exclusive="0" name="ecp_app_service_ip" recovery="relocate">
<ip ref="134.175.13.77"/>
</service>
<service autostart="1" domain="ecp_failover_domain" exclusive="0" name="ecp_app_service_storage" recovery="relocate"
>
<clusterfs ref="ecp_app_storage"/>
</service>
</rm>
</cluster>
在这样的配置,我将gfs和ip分别配置到不同之资源中,这样ip地址就可以启动,但是那个gfs的文件系统始终无法正常挂载。
more /var/log/messages
Jul 27 10:49:51 ecp-app2 openais[12618]: [CMAN ] CMAN 2.0.84 (built Apr 15 2008 16:19:14) started
Jul 27 10:49:51 ecp-app2 openais[12618]: [SYNC ] Not using a virtual synchrony filter.
Jul 27 10:49:51 ecp-app2 openais[12618]: [TOTEM] Creating commit token because I am the rep.
Jul 27 10:49:51 ecp-app2 openais[12618]: [TOTEM] Saving state aru 0 high seq received 0
Jul 27 10:49:51 ecp-app2 openais[12618]: [TOTEM] Storing new sequence id for ring 54
Jul 27 10:49:51 ecp-app2 openais[12618]: [TOTEM] entering COMMIT state.
Jul 27 10:49:51 ecp-app2 openais[12618]: [TOTEM] entering RECOVERY state.
Jul 27 10:49:51 ecp-app2 openais[12618]: [TOTEM] position [0] member 192.168.10.2:
Jul 27 10:49:51 ecp-app2 openais[12618]: [TOTEM] previous ring seq 80 rep 192.168.10.2
Jul 27 10:49:51 ecp-app2 openais[12618]: [TOTEM] aru 0 high delivered 0 received flag 1
Jul 27 10:49:51 ecp-app2 openais[12618]: [TOTEM] Did not need to originate any messages in recovery.
Jul 27 10:49:51 ecp-app2 openais[12618]: [TOTEM] Sending initial ORF token
Jul 27 10:49:51 ecp-app2 openais[12618]: [CLM ] CLM CONFIGURATION CHANGE
Jul 27 10:49:51 ecp-app2 openais[12618]: [CLM ] New Configuration:
Jul 27 10:49:51 ecp-app2 openais[12618]: [CLM ] Members Left:
Jul 27 10:49:51 ecp-app2 openais[12618]: [CLM ] Members Joined:
Jul 27 10:49:51 ecp-app2 openais[12618]: [CLM ] CLM CONFIGURATION CHANGE
Jul 27 10:49:51 ecp-app2 openais[12618]: [CLM ] New Configuration:
Jul 27 10:49:51 ecp-app2 openais[12618]: [CLM ] r(0) ip(192.168.10.2)
Jul 27 10:49:51 ecp-app2 openais[12618]: [CLM ] Members Left:
Jul 27 10:49:51 ecp-app2 openais[12618]: [CLM ] Members Joined:
Jul 27 10:49:51 ecp-app2 openais[12618]: [CLM ] r(0) ip(192.168.10.2)
Jul 27 10:49:51 ecp-app2 openais[12618]: [SYNC ] This node is within the primary component and will provide service.
Jul 27 10:49:51 ecp-app2 openais[12618]: [TOTEM] entering OPERATIONAL state.
Jul 27 10:49:51 ecp-app2 openais[12618]: [CMAN ] quorum regained, resuming activity
Jul 27 10:49:51 ecp-app2 openais[12618]: [CLM ] got nodejoin message 192.168.10.2
Jul 27 10:49:52 ecp-app2 openais[12618]: [TOTEM] entering GATHER state from 11.
Jul 27 10:49:52 ecp-app2 openais[12618]: [TOTEM] Saving state aru 9 high seq received 9
Jul 27 10:49:52 ecp-app2 openais[12618]: [TOTEM] Storing new sequence id for ring 5c
Jul 27 10:49:52 ecp-app2 openais[12618]: [TOTEM] entering COMMIT state.
Jul 27 10:49:52 ecp-app2 openais[12618]: [TOTEM] entering RECOVERY state.
Jul 27 10:49:52 ecp-app2 openais[12618]: [TOTEM] position [0] member 192.168.10.1:
Jul 27 10:49:52 ecp-app2 openais[12618]: [TOTEM] previous ring seq 88 rep 192.168.10.1
Jul 27 10:49:52 ecp-app2 openais[12618]: [TOTEM] aru 9 high delivered 9 received flag 1
Jul 27 10:49:52 ecp-app2 openais[12618]: [TOTEM] position [1] member 192.168.10.2:
Jul 27 10:49:52 ecp-app2 openais[12618]: [TOTEM] previous ring seq 84 rep 192.168.10.2
Jul 27 10:49:52 ecp-app2 openais[12618]: [TOTEM] aru 9 high delivered 9 received flag 1
Jul 27 10:49:52 ecp-app2 openais[12618]: [TOTEM] Did not need to originate any messages in recovery.
Jul 27 10:49:52 ecp-app2 openais[12618]: [CLM ] CLM CONFIGURATION CHANGE
Jul 27 10:49:52 ecp-app2 openais[12618]: [CLM ] New Configuration:
Jul 27 10:49:52 ecp-app2 openais[12618]: [CLM ] r(0) ip(192.168.10.2)
Jul 27 10:49:52 ecp-app2 openais[12618]: [CLM ] Members Left:
Jul 27 10:49:52 ecp-app2 openais[12618]: [CLM ] Members Joined:
Jul 27 10:49:52 ecp-app2 openais[12618]: [CLM ] CLM CONFIGURATION CHANGE
Jul 27 10:49:52 ecp-app2 openais[12618]: [CLM ] New Configuration:
Jul 27 10:49:52 ecp-app2 openais[12618]: [CLM ] r(0) ip(192.168.10.1)
Jul 27 10:49:52 ecp-app2 openais[12618]: [CLM ] r(0) ip(192.168.10.2)
Jul 27 10:49:52 ecp-app2 openais[12618]: [CLM ] Members Left:
Jul 27 10:49:52 ecp-app2 openais[12618]: [CLM ] Members Joined:
Jul 27 10:49:52 ecp-app2 openais[12618]: [CLM ] r(0) ip(192.168.10.1)
Jul 27 10:49:52 ecp-app2 openais[12618]: [SYNC ] This node is within the primary component and will provide service.
Jul 27 10:49:52 ecp-app2 openais[12618]: [TOTEM] entering OPERATIONAL state.
Jul 27 10:49:52 ecp-app2 openais[12618]: [CLM ] got nodejoin message 192.168.10.1
Jul 27 10:49:52 ecp-app2 openais[12618]: [CLM ] got nodejoin message 192.168.10.2
Jul 27 10:49:52 ecp-app2 ccsd[12612]: Initial status:: Quorate
Jul 27 10:49:56 ecp-app2 kernel: dlm: Using TCP for communications
Jul 27 10:49:56 ecp-app2 kernel: dlm: connecting to 2
Jul 27 10:49:57 ecp-app2 clvmd: Cluster LVM daemon started - connected to CMAN
Jul 27 10:49:59 ecp-app2 clurgmgrd[12755]: <notice> Resource Group Manager Starting
Jul 27 10:50:58 ecp-app2 ccsd[12612]: Update of cluster.conf complete (version 1 -> 2).
Jul 27 10:51:02 ecp-app2 ccsd[12612]: Update of cluster.conf complete (version 2 -> 3).
Jul 27 10:51:09 ecp-app2 clurgmgrd[12755]: <notice> Reconfiguring
Jul 27 11:05:42 ecp-app2 ccsd[12612]: Update of cluster.conf complete (version 3 -> 4).
Jul 27 11:05:49 ecp-app2 clurgmgrd[12755]: <notice> Reconfiguring
Jul 27 11:06:47 ecp-app2 ccsd[12612]: Update of cluster.conf complete (version 4 -> 5).
Jul 27 11:06:59 ecp-app2 clurgmgrd[12755]: <notice> Reconfiguring
Jul 27 11:08:15 ecp-app2 ccsd[12612]: Update of cluster.conf complete (version 5 -> 6).
Jul 27 11:08:30 ecp-app2 clurgmgrd[12755]: <notice> Reconfiguring
Jul 27 11:09:21 ecp-app2 ccsd[12612]: Update of cluster.conf complete (version 6 -> 7).
Jul 27 11:09:30 ecp-app2 clurgmgrd[12755]: <notice> Reconfiguring
Jul 27 11:10:35 ecp-app2 ccsd[12612]: Update of cluster.conf complete (version 7 -> .
Jul 27 11:10:40 ecp-app2 clurgmgrd[12755]: <notice> Reconfiguring
Jul 27 11:12:29 ecp-app2 ccsd[12612]: Update of cluster.conf complete (version 8 -> 9).
Jul 27 11:12:40 ecp-app2 clurgmgrd[12755]: <notice> Reconfiguring
Jul 27 11:12:41 ecp-app2 clurgmgrd[12755]: <notice> Initializing service:ecp_app_service
Jul 27 11:12:49 ecp-app2 clurgmgrd[12755]: <notice> Recovering failed service service:ecp_app_service
Jul 27 11:12:49 ecp-app2 gfs_controld[12654]: mount: fs requires cluster="ecp_app_cluster" current="ecp-app-ha"
Jul 27 11:12:49 ecp-app2 clurgmgrd: [12755]: <err> 'mount -t gfs /dev/mapper/appvg-lv_ecp_98g /ecp' failed, error=1
Jul 27 11:12:52 ecp-app2 gfs_controld[12654]: mount: fs requires cluster="ecp_app_cluster" current="ecp-app-ha"
Jul 27 11:12:52 ecp-app2 clurgmgrd: [12755]: <err> 'mount -t gfs /dev/mapper/appvg-lv_ecp_98g /ecp' failed, error=1
Jul 27 11:12:55 ecp-app2 gfs_controld[12654]: mount: fs requires cluster="ecp_app_cluster" current="ecp-app-ha"
Jul 27 11:12:55 ecp-app2 clurgmgrd: [12755]: <err> 'mount -t gfs /dev/mapper/appvg-lv_ecp_98g /ecp' failed, error=1
Jul 27 11:12:58 ecp-app2 clurgmgrd[12755]: <notice> start on clusterfs "ecp_app_storage" returned 2 (invalid argument(s))
Jul 27 11:12:58 ecp-app2 clurgmgrd[12755]: <warning> #68: Failed to start service:ecp_app_service; return value: 1
Jul 27 11:12:58 ecp-app2 clurgmgrd[12755]: <notice> Stopping service service:ecp_app_service
Jul 27 11:13:00 ecp-app2 clurgmgrd[12755]: <notice> Service service:ecp_app_service is recovering
Jul 27 11:14:38 ecp-app2 ccsd[12612]: Update of cluster.conf complete (version 9 -> 10).
Jul 27 11:14:46 ecp-app2 clurgmgrd[12755]: <notice> Reconfiguring
Jul 27 11:15:33 ecp-app2 ccsd[12612]: Update of cluster.conf complete (version 10 -> 11).
Jul 27 11:15:33 ecp-app2 ccsd[12612]: Unable to parse updated config file.
Jul 27 11:15:46 ecp-app2 clurgmgrd[12755]: <notice> Reconfiguring
Jul 27 11:15:46 ecp-app2 clurgmgrd[12755]: <notice> Initializing service:ecp_app_service_ip
Jul 27 11:17:12 ecp-app2 ccsd[12612]: Update of cluster.conf complete (version 11 -> 12).
Jul 27 11:17:20 ecp-app2 clurgmgrd[12755]: <notice> Reconfiguring
Jul 27 11:17:20 ecp-app2 clurgmgrd[12755]: <notice> Initializing service:ecp_app_service_storage
Jul 27 11:17:34 ecp-app2 clurgmgrd[12755]: <notice> Recovering failed service service:ecp_app_service_storage
Jul 27 11:17:34 ecp-app2 gfs_controld[12654]: mount: fs requires cluster="ecp_app_cluster" current="ecp-app-ha"
Jul 27 11:17:34 ecp-app2 clurgmgrd: [12755]: <err> 'mount -t gfs /dev/mapper/appvg-lv_ecp_98g /ecp' failed, error=1
Jul 27 11:17:37 ecp-app2 gfs_controld[12654]: mount: fs requires cluster="ecp_app_cluster" current="ecp-app-ha"
Jul 27 11:17:37 ecp-app2 clurgmgrd: [12755]: <err> 'mount -t gfs /dev/mapper/appvg-lv_ecp_98g /ecp' failed, error=1
Jul 27 11:17:40 ecp-app2 gfs_controld[12654]: mount: fs requires cluster="ecp_app_cluster" current="ecp-app-ha"
Jul 27 11:17:40 ecp-app2 clurgmgrd: [12755]: <err> 'mount -t gfs /dev/mapper/appvg-lv_ecp_98g /ecp' failed, error=1
Jul 27 11:17:43 ecp-app2 clurgmgrd[12755]: <notice> start on clusterfs "ecp_app_storage" returned 2 (invalid argument(s))
Jul 27 11:17:43 ecp-app2 clurgmgrd[12755]: <warning> #68: Failed to start service:ecp_app_service_storage; return value: 1
Jul 27 11:17:43 ecp-app2 clurgmgrd[12755]: <notice> Stopping service service:ecp_app_service_storage
Jul 27 11:17:46 ecp-app2 clurgmgrd[12755]: <notice> Service service:ecp_app_service_storage is recovering
Jul 27 11:22:13 ecp-app2 luci[7278]: Unable to retrieve batch 273858179 status from ecp-app2.cluster.com:11111: module scheduled for execution
Jul 27 11:22:52 ecp-app2 clurgmgrd[12755]: <notice> Starting disabled service service:ecp_app_service_ip
Jul 27 11:22:54 ecp-app2 in.rdiscd[19896]: setsockopt (IP_ADD_MEMBERSHIP): Address already in use
Jul 27 11:22:54 ecp-app2 in.rdiscd[19896]: Failed joining addresses
Jul 27 11:22:54 ecp-app2 clurgmgrd[12755]: <notice> Service service:ecp_app_service_ip started
Jul 27 11:22:55 ecp-app2 clurgmgrd[12755]: <notice> Stopping service service:ecp_app_service_ip
Jul 27 11:22:57 ecp-app2 luci[7278]: Unable to retrieve batch 1293856098 status from ecp-app2.cluster.com:11111: module scheduled for execution
Jul 27 11:23:05 ecp-app2 clurgmgrd[12755]: <notice> Service service:ecp_app_service_ip is stopped
Jul 27 11:23:06 ecp-app2 luci[7278]: Unable to retrieve batch 1293856098 status from ecp-app2.cluster.com:11111: module scheduled for execution
Jul 27 11:23:06 ecp-app2 clurgmgrd[12755]: <notice> Service service:ecp_app_service_ip is now running on member 2
Jul 27 11:24:50 ecp-app2 clurgmgrd[12755]: <notice> Starting stopped service service:ecp_app_service_ip
Jul 27 11:24:51 ecp-app2 clurgmgrd[12755]: <notice> Service service:ecp_app_service_ip started
Jul 27 11:25:53 ecp-app2 clurgmgrd[12755]: <notice> Starting stopped service service:ecp_app_service_storage
Jul 27 11:25:53 ecp-app2 gfs_controld[12654]: mount: fs requires cluster="ecp_app_cluster" current="ecp-app-ha"
Jul 27 11:25:53 ecp-app2 clurgmgrd: [12755]: <err> 'mount -t gfs /dev/mapper/appvg-lv_ecp_98g /ecp' failed, error=1
Jul 27 11:25:55 ecp-app2 luci[7278]: Unable to retrieve batch 1917832722 status from ecp-app2.cluster.com:11111: module scheduled for execution
Jul 27 11:25:56 ecp-app2 gfs_controld[12654]: mount: fs requires cluster="ecp_app_cluster" current="ecp-app-ha"
Jul 27 11:25:56 ecp-app2 clurgmgrd: [12755]: <err> 'mount -t gfs /dev/mapper/appvg-lv_ecp_98g /ecp' failed, error=1
Jul 27 11:25:59 ecp-app2 gfs_controld[12654]: mount: fs requires cluster="ecp_app_cluster" current="ecp-app-ha"
Jul 27 11:25:59 ecp-app2 clurgmgrd: [12755]: <err> 'mount -t gfs /dev/mapper/appvg-lv_ecp_98g /ecp' failed, error=1
Jul 27 11:26:02 ecp-app2 clurgmgrd[12755]: <notice> start on clusterfs "ecp_app_storage" returned 2 (invalid argument(s))
Jul 27 11:26:02 ecp-app2 clurgmgrd[12755]: <warning> #68: Failed to start service:ecp_app_service_storage; return value: 1
Jul 27 11:26:02 ecp-app2 clurgmgrd[12755]: <notice> Stopping service service:ecp_app_service_storage
Jul 27 11:26:04 ecp-app2 luci[7278]: Unable to retrieve batch 1917832722 status from ecp-app2.cluster.com:11111: module scheduled for execution
Jul 27 11:26:04 ecp-app2 clurgmgrd[12755]: <notice> Service service:ecp_app_service_storage is recovering
Jul 27 11:26:04 ecp-app2 clurgmgrd[12755]: <warning> #71: Relocating failed service service:ecp_app_service_storage
Jul 27 11:26:13 ecp-app2 luci[7278]: Unable to retrieve batch 1917832722 status from ecp-app2.cluster.com:11111: module scheduled for execution
Jul 27 11:26:16 ecp-app2 clurgmgrd[12755]: <notice> Service service:ecp_app_service_storage is stopped
Jul 27 11:26:22 ecp-app2 luci[7278]: Unable to retrieve batch 1917832722 status from ecp-app2.cluster.com:11111: clusvcadm start failed to start ecp_app_service_storage: |
|