- 论坛徽章:
- 0
|
历史数据库以前磁盘阵列做的是raid0+1,然后oracle 10g RAC、HACMP这些都已经部署了,但还没正式用,但最近磁盘阵列又改成raid5+1了,这样OCRdisk,votedisk及oracle数据库的数据文件就都被丢失了。\r\n\r\n\r\n\r\n 现在想重新部署oracle rac,我的想法是CRS,HACMP、数据库软件都是装在本地的,我就不重装了,我只需要重装CRS及安装oracle数据库,ocrdisk,votedisk及数据库文件仍按重做磁盘阵列之前进行配置,不知这种想法是否正确?\r\n\r\n\r\n\r\n假设正确的话,我又应当做些什么事?我的理解是先把新的磁盘阵列先认出来,然后加到卷组,然后按之前的重建祼设备,然后再安装CRS及oracle数据库,也不知道这样理解对不对,如对的话还请教下具体的详细步骤\r\n\r\n$ lspv\r\nhdisk0 00c9eb642f01d1db rootvg active\r\nhdisk1 none None \r\nhdisk2 00c9eb6408b1b480 datavg \r\n\r\n\r\n\r\n说明:hdisk2为原来的磁盘阵列盘,datavg为原来的vg,但重做Raid后肯定都不能用了,如何能恢复原状(重新识别磁盘然后加到卷组?)\r\n\r\n\r\n\r\n$ lsvg -l datavg\r\n0516-010 : Volume group must be varied on; use varyonvg command.\r\n\r\n\r\n\r\n果然用不了了,连上正式库,参考一下正式库(历史库是基本按正式库来进行配置的),正式库上执行结果如下\r\n\r\n# lsvg -l datavg\r\ndatavg:\r\nLV NAME TYPE LPs PPs PVs LV STATE MOUNT POINT\r\nOCRFile raw 8 8 1 open/syncd N/A\r\nOCRFileMirror raw 8 8 1 open/syncd N/A\r\nVOTFile raw 8 8 1 open/syncd N/A\r\nVOTFile2 raw 8 8 1 open/syncd N/A\r\nVOTFile3 raw 8 8 1 open/syncd N/A\r\nacc_sys_2g raw 8 8 1 open/syncd N/A\r\nacc_sau_2g raw 8 8 1 open/syncd N/A\r\nacc_usr_1g raw 4 4 1 open/syncd N/A\r\nacc_tmp_10g raw 40 40 1 open/syncd N/A\r\nacc_udo1_10g raw 40 40 1 open/syncd N/A\r\nacc_udo2_10g raw 40 40 1 open/syncd N/A\r\nacc_rdo1_1a_512 raw 2 2 1 open/syncd N/A\r\nacc_rdo1_2a_512 raw 2 2 1 open/syncd N/A\r\nacc_rdo1_3a_512 raw 2 2 1 open/syncd N/A\r\nacc_rdo2_1a_512 raw 2 2 1 open/syncd N/A\r\nacc_rdo2_2a_512 raw 2 2 1 open/syncd N/A\r\nacc_rdo2_3a_512 raw 2 2 1 open/syncd N/A\r\nacc_ctl1_512M raw 2 2 1 open/syncd N/A\r\nacc_ctl2_512M raw 2 2 1 open/syncd N/A\r\nacc_spfile_256M raw 1 1 1 closed/syncd N/A\r\nacc_pwd_256M raw 1 1 1 closed/syncd N/A\r\nacc_rdo1_1b_512 raw 2 2 1 open/syncd N/A\r\nacc_rdo1_2b_512 raw 2 2 1 open/syncd N/A\r\nacc_rdo1_3b_512 raw 2 2 1 open/syncd N/A\r\nacc_rdo2_1b_512 raw 2 2 1 open/syncd N/A\r\nacc_rdo2_2b_512 raw 2 2 1 open/syncd N/A\r\nacc_rdo2_3b_512 raw 2 2 1 open/syncd N/A\r\ntivoliorts_512 raw 2 2 1 open/syncd N/A\r\ntivolitemp_512 raw 2 2 1 open/syncd N/A\r\nlvdata001 raw 8 8 1 open/syncd N/A\r\nlvdata002 raw 8 8 1 open/syncd N/A\r\nlvdata003 raw 8 8 1 open/syncd N/A\r\nlvdata004 raw 8 8 1 open/syncd N/A\r\nlvdata005 raw 8 8 1 open/syncd N/A\r\nlvqfdata001 raw 128 128 1 open/syncd N/A\r\nlvqfdata002 raw 128 128 1 open/syncd N/A\r\nlvqfdata003 raw 128 128 1 open/syncd N/A\r\nlvqfdata004 raw 128 128 1 open/syncd N/A\r\nlvqfpar001 raw 40 40 1 open/syncd N/A\r\nlvqfdex001 raw 128 128 1 open/syncd N/A\r\nftelogdata001 raw 16 16 1 open/syncd N/A\r\nlvaccdata001 raw 128 128 1 open/syncd N/A\r\nlvxddata001 raw 16 16 1 open/syncd N/A\r\n\r\n\r\n\r\n历史库上再执行\r\n\r\n$ lsdev -Cc disk\r\nhdisk0 Available 07-08-00 SAS Disk Drive\r\nhdisk1 Available 07-08-00 SAS Disk Drive\r\nhdisk2 Available 04-00-02 MPIO Other DS4K Array Disk\r\nhdisk3 Defined 04-00-02 MPIO Other DS4K Array Disk\r\n\r\n\r\n\r\n为什么这里看到的是hdisk0到3,而lspv里只看到hdisk0到2?\r\n\r\n是否后面这两个盘就是磁盘阵列?\r\n\r\n\r\n\r\n# mpio_get_config -Av\r\nFrame id 0:\r\n Storage Subsystem worldwide name: 60ab80047b7aa00004b21771b\r\n Controller count: 2\r\n Partition count: 1\r\n Partition 0:\r\n Storage Subsystem Name = \'\'\r\n hdisk LUN # Ownership User Label\r\n hdisk2 0 A (preferred) 1\r\n\r\n\r\n\r\n# lspv hdisk2\r\n0516-010 : Volume group must be varied on; use varyonvg command.\r\nPHYSICAL VOLUME: hdisk2 VOLUME GROUP: datavg\r\nPV IDENTIFIER: 00c9eb6408b1b480 VG IDENTIFIER 00c9eb6400004c000000012708f08d28\r\nPV STATE: ??????? \r\nSTALE PARTITIONS: ??????? ALLOCATABLE: ???????\r\nPP SIZE: ??????? LOGICAL VOLUMES: ???????\r\nTOTAL PPs: ??????? VG DESCRIPTORS: ???????\r\nFREE PPs: ??????? HOT SPARE: ???????\r\nUSED PPs: ??????? MAX REQUEST: 256 kilobytes\r\nFREE DISTRIBUTION: ??????? \r\nUSED DISTRIBUTION: ??????? \r\nMIRROR POOL: ??????? \r\n \r\n# lspv hdisk3\r\n0516-304 : Unable to find device id hdisk3 in the Device\r\n Configuration Database.\r\n\r\n\r\n\r\n$ lsattr -E -l hdisk0\r\nPCM PCM/friend/scsiscsd Path Control Module False\r\nalgorithm fail_over Algorithm True\r\ndist_err_pcnt 0 Distributed Error Percentage True\r\ndist_tw_width 50 Distributed Error Sample Time True\r\nhcheck_interval 0 Health Check Interval True\r\nhcheck_mode nonactive Health Check Mode True\r\nmax_transfer 0x100000 Maximum TRANSFER Size True\r\npvid 00c9eb642f01d1db0000000000000000 Physical volume identifier False\r\nqueue_depth 16 Queue DEPTH False\r\nreserve_policy no_reserve Reserve Policy True\r\nsize_in_mb 300000 Size in Megabytes False\r\nunique_id 2A1135000C50012BA64BF0BST3300656SS03IBMsas Unique device identifier False\r\nww_id 5000c50012ba64bf World Wide Identifier False\r\n\r\n\r\n$ lsattr -E -l hdisk1\r\nPCM PCM/friend/scsiscsd Path Control Module False\r\nalgorithm fail_over Algorithm True\r\ndist_err_pcnt 0 Distributed Error Percentage True\r\ndist_tw_width 50 Distributed Error Sample Time True\r\nhcheck_interval 0 Health Check Interval True\r\nhcheck_mode nonactive Health Check Mode True\r\nmax_transfer 0x100000 Maximum TRANSFER Size True\r\npvid none Physical volume identifier False\r\nqueue_depth 16 Queue DEPTH False\r\nreserve_policy no_reserve Reserve Policy True\r\nsize_in_mb 300000 Size in Megabytes False\r\nunique_id 2A1135000C50012E698070BST3300656SS03IBMsas Unique device identifier False\r\nww_id 5000c50012e69807 World Wide Identifier False\r\n\r\n\r\n$ lsattr -E -l hdisk2\r\nPCM PCM/friend/otherapdisk Path Control Module False\r\nPR_key_value none Persistant Reserve Key Value True\r\nalgorithm fail_over Algorithm True\r\nautorecovery no Path/Ownership Autorecovery True\r\nclr_q no Device CLEARS its Queue on error True\r\ncntl_delay_time 0 Controller Delay Time True\r\ncntl_hcheck_int 0 Controller Health Check Interval True\r\ndist_err_pcnt 0 Distributed Error Percentage True\r\ndist_tw_width 50 Distributed Error Sample Time True\r\nhcheck_cmd inquiry Health Check Command True\r\nhcheck_interval 60 Health Check Interval True\r\nhcheck_mode nonactive Health Check Mode True\r\nlocation Location Label True\r\nlun_id 0x0 Logical Unit Number ID False\r\nlun_reset_spt yes LUN Reset Supported True\r\nmax_retry_delay 60 Maximum Quiesce Time True\r\nmax_transfer 0x40000 Maximum TRANSFER Size True\r\nnode_name 0x200400a0b847ba48 FC Node Name False\r\npvid 00c9eb6408b1b4800000000000000000 Physical volume identifier False\r\nq_err yes Use QERR bit True\r\nq_type simple Queuing TYPE True\r\nqueue_depth 10 Queue DEPTH True\r\nreassign_to 120 REASSIGN time out value True\r\nreserve_policy single_path Reserve Policy True\r\nrw_timeout 30 READ/WRITE time out value True\r\nscsi_id 0x10200 SCSI ID False\r\nstart_timeout 60 START unit time out value True\r\nunique_id 3E213600A0B800047BA4800000F844BF307970F1815 FAStT03IBMfcp Unique device identifier False\r\nww_name 0x202500a0b847ba48 FC World Wide Name False\r\n\r\n\r\n$ lsattr -E -l hdisk3\r\nPCM PCM/friend/otherapdisk Path Control Module False\r\nPR_key_value none Persistant Reserve Key Value True\r\nalgorithm fail_over Algorithm True\r\nautorecovery no Path/Ownership Autorecovery True\r\nclr_q no Device CLEARS its Queue on error True\r\ncntl_delay_time 0 Controller Delay Time True\r\ncntl_hcheck_int 0 Controller Health Check Interval True\r\ndist_err_pcnt 0 Distributed Error Percentage True\r\ndist_tw_width 50 Distributed Error Sample Time True\r\nhcheck_cmd inquiry Health Check Command True\r\nhcheck_interval 60 Health Check Interval True\r\nhcheck_mode nonactive Health Check Mode True\r\nlocation Location Label True\r\nlun_id 0x1000000000000 Logical Unit Number ID False\r\nlun_reset_spt yes LUN Reset Supported True\r\nmax_retry_delay 60 Maximum Quiesce Time True\r\nmax_transfer 0x40000 Maximum TRANSFER Size True\r\nnode_name 0x200400a0b847ba48 FC Node Name False\r\npvid none Physical volume identifier False\r\nq_err yes Use QERR bit True\r\nq_type simple Queuing TYPE True\r\nqueue_depth 10 Queue DEPTH True\r\nreassign_to 120 REASSIGN time out value True\r\nreserve_policy single_path Reserve Policy True\r\nrw_timeout 30 READ/WRITE time out value True\r\nscsi_id 0x10200 SCSI ID False\r\nstart_timeout 60 START unit time out value True\r\nunique_id 3E213600A0B800047BA48000004314B3ACBDF0F1815 FAStT03IBMfcp Unique device identifier False\r\nww_name 0x202500a0b847ba48 FC World Wide Name False\r\n\r\n\r\n\r\n\r\n\r\n# ls -lrt /dev/r*\r\n\r\n发现原来建的祼设备都还在,这里看到估计也只是个快捷方式一样的东西,那我是不是要先把这些设备删除,如何删除呢?然后再重建裸设备,是这样弄么?\r\n\r\n\r\n\r\n我也不知道我有没有表达清楚我的意思,我刚学这个也没多长时间,所以可能提问题也不专业,我想达到的目的就是重新识别到新的磁盘阵列,然后全加到datavg卷,然后将需要的ocrdisk,votedisk及数据文件对应的祼设备文件重新按原来的配置建起来,最后只安装CRS及数据库就达到恢复到重做磁盘阵列前的状态,用来存放历史数据。希望大家能帮我支支招,或者有没有人有过类似的经历,说得越详细越好,我的水平真的是很菜。 |
|