- 论坛徽章:
- 0
|
历史数据库以前磁盘阵列做的是raid0+1,然后oracle 10g RAC、HACMP这些都已经部署了,但还没正式用,但最近磁盘阵列又改成raid5+1了,这样OCRdisk,votedisk及oracle数据库的数据文件就都被丢失了。
现在想重新部署oracle rac,我的想法是CRS,HACMP、数据库软件都是装在本地的,我就不重装了,我只需要重装CRS及安装oracle数据库,ocrdisk,votedisk及数据库文件仍按重做磁盘阵列之前进行配置,不知这种想法是否正确?
假设正确的话,我又应当做些什么事?我的理解是先把新的磁盘阵列先认出来,然后加到卷组,然后按之前的重建祼设备,然后再安装CRS及oracle数据库,也不知道这样理解对不对,如对的话还请教下具体的详细步骤
$ lspv
hdisk0 00c9eb642f01d1db rootvg active
hdisk1 none None
hdisk2 00c9eb6408b1b480 datavg
说明:hdisk2为原来的磁盘阵列盘,datavg为原来的vg,但重做Raid后肯定都不能用了,如何能恢复原状(重新识别磁盘然后加到卷组?)
$ lsvg -l datavg
0516-010 : Volume group must be varied on; use varyonvg command.
果然用不了了,连上正式库,参考一下正式库(历史库是基本按正式库来进行配置的),正式库上执行结果如下
# lsvg -l datavg
datavg:
LV NAME TYPE LPs PPs PVs LV STATE MOUNT POINT
OCRFile raw 8 8 1 open/syncd N/A
OCRFileMirror raw 8 8 1 open/syncd N/A
VOTFile raw 8 8 1 open/syncd N/A
VOTFile2 raw 8 8 1 open/syncd N/A
VOTFile3 raw 8 8 1 open/syncd N/A
acc_sys_2g raw 8 8 1 open/syncd N/A
acc_sau_2g raw 8 8 1 open/syncd N/A
acc_usr_1g raw 4 4 1 open/syncd N/A
acc_tmp_10g raw 40 40 1 open/syncd N/A
acc_udo1_10g raw 40 40 1 open/syncd N/A
acc_udo2_10g raw 40 40 1 open/syncd N/A
acc_rdo1_1a_512 raw 2 2 1 open/syncd N/A
acc_rdo1_2a_512 raw 2 2 1 open/syncd N/A
acc_rdo1_3a_512 raw 2 2 1 open/syncd N/A
acc_rdo2_1a_512 raw 2 2 1 open/syncd N/A
acc_rdo2_2a_512 raw 2 2 1 open/syncd N/A
acc_rdo2_3a_512 raw 2 2 1 open/syncd N/A
acc_ctl1_512M raw 2 2 1 open/syncd N/A
acc_ctl2_512M raw 2 2 1 open/syncd N/A
acc_spfile_256M raw 1 1 1 closed/syncd N/A
acc_pwd_256M raw 1 1 1 closed/syncd N/A
acc_rdo1_1b_512 raw 2 2 1 open/syncd N/A
acc_rdo1_2b_512 raw 2 2 1 open/syncd N/A
acc_rdo1_3b_512 raw 2 2 1 open/syncd N/A
acc_rdo2_1b_512 raw 2 2 1 open/syncd N/A
acc_rdo2_2b_512 raw 2 2 1 open/syncd N/A
acc_rdo2_3b_512 raw 2 2 1 open/syncd N/A
tivoliorts_512 raw 2 2 1 open/syncd N/A
tivolitemp_512 raw 2 2 1 open/syncd N/A
lvdata001 raw 8 8 1 open/syncd N/A
lvdata002 raw 8 8 1 open/syncd N/A
lvdata003 raw 8 8 1 open/syncd N/A
lvdata004 raw 8 8 1 open/syncd N/A
lvdata005 raw 8 8 1 open/syncd N/A
lvqfdata001 raw 128 128 1 open/syncd N/A
lvqfdata002 raw 128 128 1 open/syncd N/A
lvqfdata003 raw 128 128 1 open/syncd N/A
lvqfdata004 raw 128 128 1 open/syncd N/A
lvqfpar001 raw 40 40 1 open/syncd N/A
lvqfdex001 raw 128 128 1 open/syncd N/A
ftelogdata001 raw 16 16 1 open/syncd N/A
lvaccdata001 raw 128 128 1 open/syncd N/A
lvxddata001 raw 16 16 1 open/syncd N/A
历史库上再执行
$ lsdev -Cc disk
hdisk0 Available 07-08-00 SAS Disk Drive
hdisk1 Available 07-08-00 SAS Disk Drive
hdisk2 Available 04-00-02 MPIO Other DS4K Array Disk
hdisk3 Defined 04-00-02 MPIO Other DS4K Array Disk
为什么这里看到的是hdisk0到3,而lspv里只看到hdisk0到2?
是否后面这两个盘就是磁盘阵列?
# mpio_get_config -Av
Frame id 0:
Storage Subsystem worldwide name: 60ab80047b7aa00004b21771b
Controller count: 2
Partition count: 1
Partition 0:
Storage Subsystem Name = ''
hdisk LUN # Ownership User Label
hdisk2 0 A (preferred) 1
# lspv hdisk2
0516-010 : Volume group must be varied on; use varyonvg command.
PHYSICAL VOLUME: hdisk2 VOLUME GROUP: datavg
PV IDENTIFIER: 00c9eb6408b1b480 VG IDENTIFIER 00c9eb6400004c000000012708f08d28
PV STATE: ???????
STALE PARTITIONS: ??????? ALLOCATABLE: ???????
PP SIZE: ??????? LOGICAL VOLUMES: ???????
TOTAL PPs: ??????? VG DESCRIPTORS: ???????
FREE PPs: ??????? HOT SPARE: ???????
USED PPs: ??????? MAX REQUEST: 256 kilobytes
FREE DISTRIBUTION: ???????
USED DISTRIBUTION: ???????
MIRROR POOL: ???????
# lspv hdisk3
0516-304 : Unable to find device id hdisk3 in the Device
Configuration Database.
$ lsattr -E -l hdisk0
PCM PCM/friend/scsiscsd Path Control Module False
algorithm fail_over Algorithm True
dist_err_pcnt 0 Distributed Error Percentage True
dist_tw_width 50 Distributed Error Sample Time True
hcheck_interval 0 Health Check Interval True
hcheck_mode nonactive Health Check Mode True
max_transfer 0x100000 Maximum TRANSFER Size True
pvid 00c9eb642f01d1db0000000000000000 Physical volume identifier False
queue_depth 16 Queue DEPTH False
reserve_policy no_reserve Reserve Policy True
size_in_mb 300000 Size in Megabytes False
unique_id 2A1135000C50012BA64BF0BST3300656SS03IBMsas Unique device identifier False
ww_id 5000c50012ba64bf World Wide Identifier False
$ lsattr -E -l hdisk1
PCM PCM/friend/scsiscsd Path Control Module False
algorithm fail_over Algorithm True
dist_err_pcnt 0 Distributed Error Percentage True
dist_tw_width 50 Distributed Error Sample Time True
hcheck_interval 0 Health Check Interval True
hcheck_mode nonactive Health Check Mode True
max_transfer 0x100000 Maximum TRANSFER Size True
pvid none Physical volume identifier False
queue_depth 16 Queue DEPTH False
reserve_policy no_reserve Reserve Policy True
size_in_mb 300000 Size in Megabytes False
unique_id 2A1135000C50012E698070BST3300656SS03IBMsas Unique device identifier False
ww_id 5000c50012e69807 World Wide Identifier False
$ lsattr -E -l hdisk2
PCM PCM/friend/otherapdisk Path Control Module False
PR_key_value none Persistant Reserve Key Value True
algorithm fail_over Algorithm True
autorecovery no Path/Ownership Autorecovery True
clr_q no Device CLEARS its Queue on error True
cntl_delay_time 0 Controller Delay Time True
cntl_hcheck_int 0 Controller Health Check Interval True
dist_err_pcnt 0 Distributed Error Percentage True
dist_tw_width 50 Distributed Error Sample Time True
hcheck_cmd inquiry Health Check Command True
hcheck_interval 60 Health Check Interval True
hcheck_mode nonactive Health Check Mode True
location Location Label True
lun_id 0x0 Logical Unit Number ID False
lun_reset_spt yes LUN Reset Supported True
max_retry_delay 60 Maximum Quiesce Time True
max_transfer 0x40000 Maximum TRANSFER Size True
node_name 0x200400a0b847ba48 FC Node Name False
pvid 00c9eb6408b1b4800000000000000000 Physical volume identifier False
q_err yes Use QERR bit True
q_type simple Queuing TYPE True
queue_depth 10 Queue DEPTH True
reassign_to 120 REASSIGN time out value True
reserve_policy single_path Reserve Policy True
rw_timeout 30 READ/WRITE time out value True
scsi_id 0x10200 SCSI ID False
start_timeout 60 START unit time out value True
unique_id 3E213600A0B800047BA4800000F844BF307970F1815 FAStT03IBMfcp Unique device identifier False
ww_name 0x202500a0b847ba48 FC World Wide Name False
$ lsattr -E -l hdisk3
PCM PCM/friend/otherapdisk Path Control Module False
PR_key_value none Persistant Reserve Key Value True
algorithm fail_over Algorithm True
autorecovery no Path/Ownership Autorecovery True
clr_q no Device CLEARS its Queue on error True
cntl_delay_time 0 Controller Delay Time True
cntl_hcheck_int 0 Controller Health Check Interval True
dist_err_pcnt 0 Distributed Error Percentage True
dist_tw_width 50 Distributed Error Sample Time True
hcheck_cmd inquiry Health Check Command True
hcheck_interval 60 Health Check Interval True
hcheck_mode nonactive Health Check Mode True
location Location Label True
lun_id 0x1000000000000 Logical Unit Number ID False
lun_reset_spt yes LUN Reset Supported True
max_retry_delay 60 Maximum Quiesce Time True
max_transfer 0x40000 Maximum TRANSFER Size True
node_name 0x200400a0b847ba48 FC Node Name False
pvid none Physical volume identifier False
q_err yes Use QERR bit True
q_type simple Queuing TYPE True
queue_depth 10 Queue DEPTH True
reassign_to 120 REASSIGN time out value True
reserve_policy single_path Reserve Policy True
rw_timeout 30 READ/WRITE time out value True
scsi_id 0x10200 SCSI ID False
start_timeout 60 START unit time out value True
unique_id 3E213600A0B800047BA48000004314B3ACBDF0F1815 FAStT03IBMfcp Unique device identifier False
ww_name 0x202500a0b847ba48 FC World Wide Name False
# ls -lrt /dev/r*
发现原来建的祼设备都还在,这里看到估计也只是个快捷方式一样的东西,那我是不是要先把这些设备删除,如何删除呢?然后再重建裸设备,是这样弄么?
我也不知道我有没有表达清楚我的意思,我刚学这个也没多长时间,所以可能提问题也不专业,我想达到的目的就是重新识别到新的磁盘阵列,然后全加到datavg卷,然后将需要的ocrdisk,votedisk及数据文件对应的祼设备文件重新按原来的配置建起来,最后只安装CRS及数据库就达到恢复到重做磁盘阵列前的状态,用来存放历史数据。希望大家能帮我支支招,或者有没有人有过类似的经历,说得越详细越好,我的水平真的是很菜。 |
|