--1 修改參數(shù) vi /etc/inittab #h1:35:respawn:/etc/init.d/init.ohasd run /dev/null 21 /dev/null #h1:35:respawn:/etc/init.d/init.cssd run /dev/null 21 /dev/null --2 清理文件 mv init.ohasd init.ohasd20141219 mv ohasd ohasd20141219 mv init.cssd
--1 修改參數(shù)
vi /etc/inittab
#h1:35:respawn:/etc/init.d/init.ohasd run >/dev/null 2>&1
#h1:35:respawn:/etc/init.d/init.cssd run >/dev/null 2>&1
--2 清理文件
mv init.ohasd init.ohasd20141219
mv ohasd ohasd20141219
mv init.cssd init.cssd20141219
--3 啟動cssd
[root@oratest init.d]# /opt/oracle/product/10.2.0/bin/localconfig add
Successfully accumulated necessary OCR keys.
Creating OCR keys for user 'root', privgrp 'root'..
Operation successful.
Configuration for local CSS has been initialized
Adding to inittab
Startup will be queued to init within 30 seconds.
Checking the status of new Oracle init process...
Expecting the CRS daemons to be up within 600 seconds.
CSS is active on these nodes.
oratest
CSS is active on all nodes.
Oracle CSS service is installed and running under init(1M)
[root@oratest init.d]# ps -ef|grep css
root 6171 1 0 00:58 ? 00:00:00 /bin/sh /etc/init.d/init.cssd run
oracle 6412 6171 4 00:58 ? 00:00:01 /opt/oracle/product/10.2.0/bin/ocssd.bin
root 6439 4631 0 00:59 pts/2 00:00:00 grep css
[root@oratest init.d]# clear
--5 刪除
[root@oratest init.d]# /opt/oracle/product/10.2.0/bin/localconfig delete
Stopping CSSD.
Shutting down CSS daemon.
--6 關(guān)鍵文件被移除
[root@oratest init.d]# ll |grep css
-rwxr-xr-x 1 root root 55164 Nov 17 23:32 init.cssd20141219
[root@oratest init.d]# ll |grep has
-rwxr-x--- 1 root root 8880 Oct 9 14:06 init.ohasd20141219
-rwxr-x--- 1 root root 6749 Oct 9 14:06 ohasd20141219
[root@oratest init.d]#
--7 inittab文件更新
# Run xdm in runlevel 5
x:3:respawn:/etc/X11/prefdm -nodaemon
#h1:35:respawn:/etc/init.d/init.ohasd run >/dev/null 2>&1
--8 重新添加css
[root@oratest init.d]# /opt/oracle/product/10.2.0/bin/localconfig add
Successfully accumulated necessary OCR keys.
Creating OCR keys for user 'root', privgrp 'root'..
Operation successful.
Configuration for local CSS has been initialized
Adding to inittab
Startup will be queued to init within 30 seconds.
Checking the status of new Oracle init process...
Expecting the CRS daemons to be up within 600 seconds.
CSS is active on these nodes.
oratest
CSS is active on all nodes.
Oracle CSS service is installed and running under init(1M)
[root@oratest init.d]# ps -ef|grep css
root 6961 1 0 01:04 ? 00:00:00 /bin/sh /etc/init.d/init.cssd run
oracle 7201 6961 5 01:04 ? 00:00:01 /opt/oracle/product/10.2.0/bin/ocssd.bin
root 7221 4631 0 01:04 pts/2 00:00:00 grep css
[root@oratest init.d]#
Shutdown request successfully issued.
Shutdown has begun. The daemons should exit soon.
[root@oratest init.d]# ps -ef|grep css
root 6665 4631 0 01:02 pts/2 00:00:00 grep css
[root@oratest init.d]#
--9 創(chuàng)建asm磁盤vol9
[root@oratest init.d]# oracleasm init
Creating /dev/oracleasm mount point: /dev/oracleasm
Loading module "oracleasm": oracleasm
Mounting ASMlib driver filesystem: /dev/oracleasm
[root@oratest init.d]# oracleasm scandisks
Reloading disk partitions: done
Cleaning any stale ASM disks...
Scanning system for ASM disks...
Instantiating disk "VOL1"
Instantiating disk "VOL2"
[root@oratest init.d]# oracleasm listdisks
VOL1
VOL2
[root@oratest init.d]# oracleasm createdisk vol9 /dev/sdg1
Writing disk header: done
Instantiating disk: done
[root@oratest init.d]# oracleasm listdisks
VOL1
VOL2
VOL9
[root@oratest init.d]#
--10 確認(rèn)asm磁盤信息
[root@oratest init.d]# oracleasm configure
ORACLEASM_ENABLED=false
ORACLEASM_UID=
ORACLEASM_GID=
ORACLEASM_SCANBOOT=true
ORACLEASM_SCANORDER=""
ORACLEASM_SCANEXCLUDE=""
[root@oratest init.d]#
--11 重新配置
[root@oratest ~]# oracleasm configure -i
Configuring the Oracle ASM library driver.
This will configure the on-boot properties of the Oracle ASM library
driver. The following questions will determine whether the driver is
loaded on boot and what permissions it will have. The current values
will be shown in brackets ('[]'). Hitting
without typing an
answer will keep that current value. Ctrl-C will abort.
Default user to own the driver interface []: oracle
Default group to own the driver interface []: oinstall
Start Oracle ASM library driver on boot (y/n) [n]: y
Scan for Oracle ASM disks on boot (y/n) [y]: y
[root@oratest disks]# ll
total 0
[root@oratest disks]# oracleasm scandisks
Reloading disk partitions: done
Cleaning any stale ASM disks...
Scanning system for ASM disks...
Instantiating disk "VOL1"
Instantiating disk "VOL2"
Instantiating disk "VOL9"
[root@oratest disks]# ll
total 0
brw-rw---- 1 oracle oinstall 8, 17 Nov 18 01:44 VOL1
brw-rw---- 1 oracle oinstall 8, 33 Nov 18 01:44 VOL2
brw-rw---- 1 oracle oinstall 8, 97 Nov 18 01:44 VOL9
[root@oratest disks]# cd ..
[root@oratest oracleasm]# l
-bash: l: command not found
[root@oratest oracleasm]# ll
total 0
drwxr-xr-x 1 root root 0 Nov 18 01:39 disks
drwxrwx--- 1 root root 0 Nov 18 01:39 iid
[root@oratest oracleasm]#
--創(chuàng)建dg
SQL> create diskgroup dgtest
2 external redundancy disk '/dev/oracleasm/disks/VOL9'
3 /
Diskgroup created.
SQL> select name from v$asm_diskgroup;
NAME
------------------------------
DG1
DGTEST
-- DB 無法連接asm
[root@oratest ~]# ps -ef|grep css
root 4451 1 0 02:25 ? 00:00:00 /bin/sh /etc/init.d/init.cssd run
oracle 5016 4451 0 02:25 ? 00:00:06 /opt/oracle/product/10.2.0/bin/ocssd.bin
root 7777 7735 0 05:20 pts/5 00:00:00 grep css
[root@oratest ~]# ps -ef|grep asmb
root 7779 7735 0 05:20 pts/5 00:00:00 grep asmb
[root@oratest ~]# ps -ef|grep asm
oracle 5109 1 0 02:26 ? 00:00:02 asm_pmon_+ASM
oracle 5111 1 0 02:26 ? 00:00:00 asm_psp0_+ASM
oracle 5113 1 0 02:26 ? 00:00:00 asm_mman_+ASM
oracle 5115 1 0 02:26 ? 00:00:00 asm_dbw0_+ASM
oracle 5117 1 0 02:26 ? 00:00:00 asm_lgwr_+ASM
oracle 5119 1 0 02:26 ? 00:00:00 asm_ckpt_+ASM
oracle 5121 1 0 02:26 ? 00:00:00 asm_smon_+ASM
oracle 5123 1 0 02:26 ? 00:00:00 asm_rbal_+ASM
oracle 5125 1 0 02:26 ? 00:00:02 asm_gmon_+ASM
root 7781 7735 0 05:20 pts/5 00:00:00 grep asm
[root@oratest ~]#
--重新安裝后可以
--10g 磁盤頭備份
1051 kfed read /dev/oracleasm/disks/VOL9 > vol9_10g_kefd.bk
1052 kfed read /dev/oracleasm/disks/VOL1 > vol1_10g_kefd.bk
1053 kfed read /dev/oracleasm/disks/VOL2 > vol2_10g_kefd.bk
1054 dd if=/dev/oracleasm/disks/VOL9 of=vol9_10g_dd.bk bs=4096 count=1
1055 dd if=/dev/oracleasm/disks/VOL2 of=vol2_10g_dd.bk bs=4096 count=1
1056 dd if=/dev/oracleasm/disks/VOL1 of=vol1_10g_dd.bk bs=4096 count=1
--備份注冊信息
[root@oratest etc]# history|grep cp
cp -r -p oracle oracle10g
cp oraInst.loc oraInst.loc10g
--停止css
[root@oratest /]# /opt/oracle/product/10g/bin/localconfig delete
Stopping CSSD.
Shutting down CSS daemon.
Shutdown request successfully issued.
Shutdown has begun. The daemons should exit soon.
[root@oratest /]#
[root@oratest /]#
[root@oratest ~]# /opt/oracle/product/11g/grid/crs/install/roothas.pl -deconfig -force
Using configuration parameter file: /opt/oracle/product/11g/grid/crs/install/crsconfig_params
Creating trace directory
Failure in execution (rc=-1, 256, No such file or directory) for command /opt/oracle/product/11g/grid/bin/crsctl stop resource ora.cssd -f
Failure in execution (rc=-1, 256, No such file or directory) for command /opt/oracle/product/11g/grid/bin/crsctl delete resource ora.cssd -f
Failure in execution (rc=-1, 256, No such file or directory) for command /opt/oracle/product/11g/grid/bin/crsctl stop has -f
You must kill ohasd processes or reboot the system to properly
cleanup the processes started by Oracle clusterware
Either /etc/oracle/olr.loc does not exist or is not readable
Make sure the file exists and it has read and execute access
Failure in execution (rc=-1, 256, No such file or directory) for command /etc/init.d/ohasd deinstall
Successfully deconfigured Oracle Restart stack
--安裝CSS
[root@oratest ~]# /opt/oracle/product/11g/grid/crs/install/roothas.pl
Using configuration parameter file: /opt/oracle/product/11g/grid/crs/install/crsconfig_params
User ignored Prerequisites during installation
LOCAL ADD MODE
Creating OCR keys for user 'oracle', privgrp 'oinstall'..
Operation successful.
LOCAL ONLY MODE
Successfully accumulated necessary OCR keys.
Creating OCR keys for user 'root', privgrp 'root'..
Operation successful.
CRS-4664: Node oratest successfully pinned.
Adding Clusterware entries to inittab
oratest 2014/11/18 08:42:57 /opt/oracle/product/11g/grid/cdata/oratest/backup_20141118_084257.olr
Successfully configured Oracle Grid Infrastructure for a Standalone Server
[root@oratest ~]#
[oracle@oratest ~]$ crsctl check has
CRS-4638: Oracle High Availability Services is online
[oracle@oratest ~]$
[oracle@oratest ~]$ crs_stat -t
Name Type Target State Host
------------------------------------------------------------
ora.cssd ora.cssd.type OFFLINE OFFLINE
ora.diskmon ora....on.type OFFLINE OFFLINE
ora.evmd ora.evm.type ONLINE ONLINE oratest
ora.ons ora.ons.type OFFLINE OFFLINE
netca
asmca
--10g database startup
[oracle@oratest kefd]$ sqlplus / as sysdba
SQL*Plus: Release 10.2.0.5.0 - Production on Tue Nov 18 09:00:36 2014
Copyright (c) 1982, 2010, Oracle. All Rights Reserved.
Connected to an idle instance.
SQL> startup
ORACLE instance started.
Total System Global Area 1052770304 bytes
Fixed Size 1277676 bytes
Variable Size 260047124 bytes
Database Buffers 788529152 bytes
Redo Buffers 2916352 bytes
Database mounted.
Database opened.
SQL> create table hj as select * from dba_objects;
Table created.
SQL>
SQL> select name from v$datafile;
NAME
--------------------------------------------------------------------------------
+DG10/d10g/datafile/system.256.863939381
+DG10/d10g/datafile/undotbs1.258.863939383
+DG10/d10g/datafile/sysaux.257.863939381
+DG10/d10g/datafile/users.259.863939383
SQL>
kfed read /dev/oracleasm/disks/VOL9 > vol9_11g_kefd.bk
--asm磁盤升級后信息比較
c7
< kfbh.check: 3158608421 ; 0x00c: 0xbc448a25
---
> kfbh.check: 2117325354 ; 0x00c: 0x7e33d22a
29,30c29,30
< kfdhdb.mntstmp.hi: 33009223 ; 0x0b0: HOUR=0x7 DAYS=0x12 MNTH=0xb YEAR=0x7de
< kfdhdb.mntstmp.lo: 567114752 ; 0x0b4: USEC=0x0 MSEC=0x35f SECS=0x1c MINS=0x8
---
> kfdhdb.mntstmp.hi: 33009224 ; 0x0b0: HOUR=0x8 DAYS=0x12 MNTH=0xb YEAR=0x7de
> kfdhdb.mntstmp.lo: 3820626944 ; 0x0b4: USEC=0x0 MSEC=0x289 SECS=0x3b MINS=0x38
[oracle@oratest kefd]$ kfed read /dev/oracleasm/disks/VOL2 > vol2_11g_kefd.bk
[oracle@oratest kefd]$ ll
total 48
-rw-r--r-- 1 oracle oinstall 4096 Nov 18 07:50 vol1_10g_dd.bk
-rw-r--r-- 1 oracle oinstall 655 Nov 18 07:49 vol1_10g_kefd.bk
-rw-r--r-- 1 oracle oinstall 4096 Nov 18 07:50 vol2_10g_dd.bk
-rw-r--r-- 1 oracle oinstall 6608 Nov 18 07:49 vol2_10g_kefd.bk
-rw-r--r-- 1 oracle oinstall 6605 Nov 18 09:08 vol2_11g_kefd.bk
-rw-r--r-- 1 oracle oinstall 4096 Nov 18 07:50 vol9_10g_dd.bk
-rw-r--r-- 1 oracle oinstall 6607 Nov 18 07:49 vol9_10g_kefd.bk
-rw-r--r-- 1 oracle oinstall 6608 Nov 18 09:02 vol9_11g_kefd.bk
[oracle@oratest kefd]$ diff vol2_11g_kefd.bk vol9_11g_kefd.bk
7c7
< kfbh.check: 3081658923 ; 0x00c: 0xb7ae622b
---
> kfbh.check: 2117325354 ; 0x00c: 0x7e33d22a
12,13c12,13
< kfdhdb.driver.provstr: ORCLDISKVOL2 ; 0x000: length=12
< kfdhdb.driver.reserved[0]: 843861846 ; 0x008: 0x324c4f56
---
> kfdhdb.driver.provstr: ORCLDISKVOL9 ; 0x000: length=12
> kfdhdb.driver.reserved[0]: 961302358 ; 0x008: 0x394c4f56
19c19
< kfdhdb.compat: 186646528 ; 0x020: 0x0b200000
---
> kfdhdb.compat: 168820736 ; 0x020: 0x0a100000
23,25c23,25
< kfdhdb.dskname: DG11_0000 ; 0x028: length=9
< kfdhdb.grpname: DG11 ; 0x048: length=4
< kfdhdb.fgname: DG11_0000 ; 0x068: length=9
---
> kfdhdb.dskname: DG10_0000 ; 0x028: length=9
> kfdhdb.grpname: DG10 ; 0x048: length=4
> kfdhdb.fgname: DG10_0000 ; 0x068: length=9
27,30c27,30
< kfdhdb.crestmp.hi: 33009225 ; 0x0a8: HOUR=0x9 DAYS=0x12 MNTH=0xb YEAR=0x7de
< kfdhdb.crestmp.lo: 542999552 ; 0x0ac: USEC=0x0 MSEC=0x361 SECS=0x5 MINS=0x8
< kfdhdb.mntstmp.hi: 33009225 ; 0x0b0: HOUR=0x9 DAYS=0x12 MNTH=0xb YEAR=0x7de
< kfdhdb.mntstmp.lo: 555575296 ; 0x0b4: USEC=0x0 MSEC=0x35a SECS=0x11 MINS=0x8
---
> kfdhdb.crestmp.hi: 33009223 ; 0x0a8: HOUR=0x7 DAYS=0x12 MNTH=0xb YEAR=0x7de
> kfdhdb.crestmp.lo: 558790656 ; 0x0ac: USEC=0x0 MSEC=0x39e SECS=0x14 MINS=0x8
> kfdhdb.mntstmp.hi: 33009224 ; 0x0b0: HOUR=0x8 DAYS=0x12 MNTH=0xb YEAR=0x7de
> kfdhdb.mntstmp.lo: 3820626944 ; 0x0b4: USEC=0x0 MSEC=0x289 SECS=0x3b MINS=0x38
45,46c45,46
< kfdhdb.grpstmp.hi: 33009225 ; 0x0e4: HOUR=0x9 DAYS=0x12 MNTH=0xb YEAR=0x7de
< kfdhdb.grpstmp.lo: 542457856 ; 0x0e8: USEC=0x0 MSEC=0x150 SECS=0x5 MINS=0x8
---
> kfdhdb.grpstmp.hi: 33009223 ; 0x0e4: HOUR=0x7 DAYS=0x12 MNTH=0xb YEAR=0x7de
> kfdhdb.grpstmp.lo: 558645248 ; 0x0e8: USEC=0x0 MSEC=0x310 SECS=0x14 MINS=0x8
--10g 數(shù)據(jù)庫因為版本不一樣,只能讀取一個文件
SQL> select path from v$asm_disk;
PATH
--------------------------------------------------------------------------------
/dev/oracleasm/disks/VOL9
SQL>
[oracle@oratest kefd]$ ll
total 56
-rw-r--r-- 1 oracle oinstall 4096 Nov 18 07:50 vol1_10g_dd.bk
-rw-r--r-- 1 oracle oinstall 655 Nov 18 07:49 vol1_10g_kefd.bk
-rw-r--r-- 1 oracle oinstall 6607 Nov 18 09:19 vol1_11g_kefd.bk
-rw-r--r-- 1 oracle oinstall 4096 Nov 18 07:50 vol2_10g_dd.bk
-rw-r--r-- 1 oracle oinstall 6608 Nov 18 07:49 vol2_10g_kefd.bk
-rw-r--r-- 1 oracle oinstall 6605 Nov 18 09:08 vol2_11g_kefd.bk
-rw-r--r-- 1 oracle oinstall 4096 Nov 18 07:50 vol9_10g_dd.bk
-rw-r--r-- 1 oracle oinstall 6607 Nov 18 07:49 vol9_10g_kefd.bk
-rw-r--r-- 1 oracle oinstall 6608 Nov 18 09:02 vol9_11g_kefd.bk
[oracle@oratest kefd]$ diff vol1_11g_kefd.bk vol2_11g_kefd.bk
7c7
< kfbh.check: 3738323499 ; 0x00c: 0xded24a2b
---
> kfbh.check: 3081658923 ; 0x00c: 0xb7ae622b
12,13c12,13
< kfdhdb.driver.provstr: ORCLDISKVOL1 ; 0x000: length=12
< kfdhdb.driver.reserved[0]: 827084630 ; 0x008: 0x314c4f56
---
> kfdhdb.driver.provstr: ORCLDISKVOL2 ; 0x000: length=12
> kfdhdb.driver.reserved[0]: 843861846 ; 0x008: 0x324c4f56
19c19
< kfdhdb.compat: 185597952 ; 0x020: 0x0b100000
---
> kfdhdb.compat: 186646528 ; 0x020: 0x0b200000
23,25c23,25
< kfdhdb.dskname: DG12_0000 ; 0x028: length=9
< kfdhdb.grpname: DG12 ; 0x048: length=4
< kfdhdb.fgname: DG12_0000 ; 0x068: length=9
---
> kfdhdb.dskname: DG11_0000 ; 0x028: length=9
> kfdhdb.grpname: DG11 ; 0x048: length=4
> kfdhdb.fgname: DG11_0000 ; 0x068: length=9
28c28
< kfdhdb.crestmp.lo: 1220320256 ; 0x0ac: USEC=0x0 MSEC=0x327 SECS=0xb MINS=0x12
---
> kfdhdb.crestmp.lo: 542999552 ; 0x0ac: USEC=0x0 MSEC=0x361 SECS=0x5 MINS=0x8
30c30
< kfdhdb.mntstmp.lo: 1230435328 ; 0x0b4: USEC=0x0 MSEC=0x1bd SECS=0x15 MINS=0x12
---
> kfdhdb.mntstmp.lo: 555575296 ; 0x0b4: USEC=0x0 MSEC=0x35a SECS=0x11 MINS=0x8
44c44
< kfdhdb.dbcompat: 185597952 ; 0x0e0: 0x0b100000
---
> kfdhdb.dbcompat: 168820736 ; 0x0e0: 0x0a100000
46c46
< kfdhdb.grpstmp.lo: 1219750912 ; 0x0e8: USEC=0x0 MSEC=0xfb SECS=0xb MINS=0x12
---
> kfdhdb.grpstmp.lo: 542457856 ; 0x0e8: USEC=0x0 MSEC=0x150 SECS=0x5 MINS=0x8
[oracle@oratest kefd]$ kfed read /dev/oracleasm/disks/VOL1 > vol1_11g_kefd.bk1
[oracle@oratest kefd]$ diff vol1_11g_kefd.bk vol1_11g_kefd.bk1
7c7
< kfbh.check: 3738323499 ; 0x00c: 0xded24a2b
---
> kfbh.check: 3739372075 ; 0x00c: 0xdee24a2b
19c19
< kfdhdb.compat: 185597952 ; 0x020: 0x0b100000
---
> kfdhdb.compat: 186646528 ; 0x020: 0x0b200000
[oracle@oratest kefd]$
[oracle@oratest ~]$ crsctl stop has
CRS-2791: Starting shutdown of Oracle High Availability Services-managed resources on 'oratest'
CRS-2673: Attempting to stop 'ora.LISTENER.lsnr' on 'oratest'
CRS-2673: Attempting to stop 'ora.DG10.dg' on 'oratest'
CRS-2673: Attempting to stop 'ora.DG11.dg' on 'oratest'
CRS-2673: Attempting to stop 'ora.DG12.dg' on 'oratest'
CRS-2677: Stop of 'ora.DG10.dg' on 'oratest' succeeded
CRS-2677: Stop of 'ora.DG11.dg' on 'oratest' succeeded
CRS-2677: Stop of 'ora.DG12.dg' on 'oratest' succeeded
CRS-2673: Attempting to stop 'ora.asm' on 'oratest'
CRS-2677: Stop of 'ora.LISTENER.lsnr' on 'oratest' succeeded
CRS-2677: Stop of 'ora.asm' on 'oratest' succeeded
CRS-2673: Attempting to stop 'ora.cssd' on 'oratest'
CRS-2677: Stop of 'ora.cssd' on 'oratest' succeeded
CRS-2673: Attempting to stop 'ora.evmd' on 'oratest'
CRS-2677: Stop of 'ora.evmd' on 'oratest' succeeded
CRS-2793: Shutdown of Oracle High Availability Services-managed resources on 'oratest' has completed
CRS-4133: Oracle High Availability Services has been stopped.
--卸載11g asm 啟動10g asm
[root@oratest ~]# /opt/oracle/product/11g/grid/crs/install/roothas.pl -deconfig -force
Using configuration parameter file: /opt/oracle/product/11g/grid/crs/install/crsconfig_params
CRS-4639: Could not contact Oracle High Availability Services
CRS-4000: Command Stop failed, or completed with errors.
CRS-4639: Could not contact Oracle High Availability Services
CRS-4000: Command Delete failed, or completed with errors.
CRS-4544: Unable to connect to OHAS
CRS-4000: Command Stop failed, or completed with errors.
Successfully deconfigured Oracle Restart stack
[root@oratest ~]# /opt/oracle/product/10.2.0/bin/localconfig add
-bash: /opt/oracle/product/10.2.0/bin/localconfig: No such file or directory
[root@oratest ~]# /opt/oracle/product/10g/bin/localconfig add
Successfully accumulated necessary OCR keys.
Creating OCR keys for user 'root', privgrp 'root'..
Operation successful.
Configuration for local CSS has been initialized
Adding to inittab
Startup will be queued to init within 30 seconds.
Checking the status of new Oracle init process...
Expecting the CRS daemons to be up within 600 seconds.
CSS is active on these nodes.
oratest
CSS is active on all nodes.
Oracle CSS service is installed and running under init(1M)
[root@oratest ~]#
SQL> l
1* select name,state from v$asm_diskgroup
SQL> /
NAME STATE
------------------------------ -----------
DG10 MOUNTED
DG11 DISMOUNTED
DG12 DISMOUNTED
SQL> SELECT PATH FROM V$AMS_DISK;
SELECT PATH FROM V$AMS_DISK
*
ERROR at line 1:
ORA-01219: database not open: queries allowed on fixed tables/views only
SQL> SELECT PATH FROM V$ASM_DISK;
PATH
--------------------------------------------------------------------------------
/dev/oracleasm/disks/VOL2
/dev/oracleasm/disks/VOL1
/dev/oracleasm/disks/VOL9
SQL>
--數(shù)據(jù)庫可以直接掛載
[oracle@oratest kefd]$ sqlplus / as sysdba
SQL*Plus: Release 10.2.0.5.0 - Production on Tue Nov 18 09:41:39 2014
Copyright (c) 1982, 2010, Oracle. All Rights Reserved.
Connected to an idle instance.
SQL> startup
ORACLE instance started.
Total System Global Area 1052770304 bytes
Fixed Size 1277676 bytes
Variable Size 260047124 bytes
Database Buffers 788529152 bytes
Redo Buffers 2916352 bytes
Database mounted.
Database opened.
SQL> create table hj10g as select * from dba_objects;
Table created.
SQL>
--磁盤重新掛在后信息的差異:ASM實例的修改
kfed read /dev/oracleasm/disks/VOL9 > vol9_10g_kefd.bk1
[oracle@oratest kefd]$ diff vol9_10g_kefd.bk vol9_10g_kefd.bk1
7c7
< kfbh.check: 3158608421 ; 0x00c: 0xbc448a25
---
> kfbh.check: 127173163 ; 0x00c: 0x0794822b
29,30c29,30
< kfdhdb.mntstmp.hi: 33009223 ; 0x0b0: HOUR=0x7 DAYS=0x12 MNTH=0xb YEAR=0x7de
< kfdhdb.mntstmp.lo: 567114752 ; 0x0b4: USEC=0x0 MSEC=0x35f SECS=0x1c MINS=0x8
---
> kfdhdb.mntstmp.hi: 33009225 ; 0x0b0: HOUR=0x9 DAYS=0x12 MNTH=0xb YEAR=0x7de
> kfdhdb.mntstmp.lo: 2585621504 ; 0x0b4: USEC=0x0 MSEC=0x35d SECS=0x21 MINS=0x26
[oracle@oratest kefd]$
ASM磁盤頭信息保存在每個磁盤的前4K里面,這個信息的備份對于ASM的恢復(fù)非常重要,有下面的幾種方法
1.直接做dd來備份磁盤的前4K,磁盤頭信息丟失時,dd回來
備份:dd if=/dev/raw/raw1 of=/gyj/asmheader.dd bs=4096 count=1
恢復(fù):dd if=/gyj/asmheader.dd of=/dev/raw/raw1
2.利用oracle的kfed工具來備份,將磁盤頭信息轉(zhuǎn)換成文本明文形式記錄下來,恢復(fù)時使用kfed merge進(jìn)去
備份:kfed read /dev/raw/raw1 aunum=0 blknum=0 text=raw1.txt
恢復(fù):kfed write /dev/raw/raw1 aunum=0 blknum=0 text=raw1.txt
3.11G asmcmd里面增加了md_backup,md_restore備份指令,但這個只是備份DG的meta結(jié)構(gòu)信息,他的恢復(fù)必須是DG已經(jīng)mount,然
后導(dǎo)入結(jié)構(gòu)信息,類似于exp/imp,這種方式對于DG損壞是不起作用的
備份:asmcmd md_backup -b /gyj/asmheader.20130130.txt
恢復(fù):asmcmd md_restore -t full -g dg2 -i -b /gyj/asmheader.20130130.txt
--自動備份的位置
[oracle@oratest kefd]$ kfed read /dev/oracleasm/disks/VOL9 blkn=0 |grep KFBTYP
kfbh.type: 1 ; 0x002: KFBTYP_DISKHEAD
[oracle@oratest kefd]$ kfed read /dev/oracleasm/disks/VOL9 blkn=510 |grep KFBTYP
kfbh.type: 1 ; 0x002: KFBTYP_DISKHEAD
[oracle@oratest kefd]$
[oracle@oratest kefd]$ kfed repair /dev/oracleasm/disks/VOL9
[oracle@oratest kefd]$
結(jié)論:
從Oracle 10.2.0.5開始,ASM磁盤已經(jīng)開始自動將頭塊進(jìn)行備份,備份塊的位置在第2個AU的倒數(shù)第2個塊上(對于默認(rèn)1M的AU來說,是第510個塊), 如果頭塊損壞,可以用kfed repair命令來修復(fù)。因此對于選用ASM存儲作為生產(chǎn)環(huán)境的用戶來說,盡快升級到10.2.0.5是明智的選擇。
[root@oratest 11g]# /opt/oracle/grid/product/11.2.0/grid/rootupgrade.sh
Performing root user operation for Oracle 11g
The following environment variables are set as:
ORACLE_OWNER= oracle
ORACLE_HOME= /opt/oracle/grid/product/11.2.0/grid
Enter the full pathname of the local bin directory: [/usr/local/bin]:
The file "dbhome" already exists in /usr/local/bin. Overwrite it? (y/n)
[n]:
The file "oraenv" already exists in /usr/local/bin. Overwrite it? (y/n)
[n]:
The file "coraenv" already exists in /usr/local/bin. Overwrite it? (y/n)
[n]:
Entries will be added to the /etc/oratab file as needed by
Database Configuration Assistant when a database is created
Finished running generic part of root script.
Now product-specific root actions will be performed.
Using configuration parameter file: /opt/oracle/grid/product/11.2.0/grid/crs/install/crsconfig_params
Creating trace directory
LOCAL ADD MODE
Creating OCR keys for user 'oracle', privgrp 'oinstall'..
Operation successful.
CSS appears healthy
Stopping CSSD.
Shutting down CSS daemon.
Shutdown request successfully issued.
Shutdown has begun. The daemons should exit soon.
LOCAL ONLY MODE
Successfully accumulated necessary OCR keys.
Creating OCR keys for user 'root', privgrp 'root'..
Operation successful.
CRS-4664: Node oratest successfully pinned.
Replacing Clusterware entries in inittab
oratest 2014/11/18 10:35:07 /opt/oracle/grid/product/11.2.0/grid/cdata/oratest/backup_20141118_103507.olr
Successfully configured Oracle Grid Infrastructure for a Standalone Server
[root@oratest 11g]#
聲明:本網(wǎng)頁內(nèi)容旨在傳播知識,若有侵權(quán)等問題請及時與本網(wǎng)聯(lián)系,我們將在第一時間刪除處理。TEL:177 7030 7066 E-MAIL:11247931@qq.com