记一次删除ocr与dbfile的恢复记录
自己造成的一个案例:
场景:ocr磁盘组被我dd掉了,dbfile磁盘组也被我dd掉了。Rac起不来。之前ocr的DATA磁盘组被替换到了ABC磁盘。所幸的是有备份。
重新加载OCR磁盘
[root@rhel1 ~]# /u01/app/11.2.0/grid/crs/install/roothas.pl -deconfig -force -verbose
defined(@array) is deprecated at /u01/app/11.2.0/grid/crs/install/crsconfig_lib.pm line 1702.
(Maybe you should just omit the defined()?)
defined(@array) is deprecated at /u01/app/11.2.0/grid/crs/install/crsconfig_lib.pm line 1702.
(Maybe you should just omit the defined()?)
defined(@array) is deprecated at /u01/app/11.2.0/grid/crs/install/crsconfig_lib.pm line 1780.
(Maybe you should just omit the defined()?)
2019-05-22 10:31:57: Checking for super user privileges
2019-05-22 10:31:57: User has super user privileges
2019-05-22 10:31:57: Parsing the host name
Using configuration parameter file: /u01/app/11.2.0/grid/crs/install/crsconfig_params
CRS-4047: No Oracle Clusterware components configured.
CRS-4000: Command Stop failed, or completed with errors.
CRS-4047: No Oracle Clusterware components configured.
CRS-4000: Command Delete failed, or completed with errors.
CRS-4047: No Oracle Clusterware components configured.
CRS-4000: Command Stop failed, or completed with errors.
You must kill ohasd processes or reboot the system to properly
cleanup the processes started by Oracle clusterware
ADVM/ACFS is not supported on redhat-release-server-7.1-1.el7.x86_64
ACFS-9201: Not Supported
Either /etc/oracle/olr.loc does not exist or is not readable
Make sure the file exists and it has read and execute access
/bin/dd: failed to open ‘’: No such file or directory
Successfully deconfigured Oracle Restart stack
[root@rhel1 ~]# less /etc/oracle/olr.loc
/etc/oracle/olr.loc: No such file or directory
[root@rhel1 ~]# cd /u01/app/11.2.0/grid/
[root@rhel1 grid]# ./root.sh
Running Oracle 11g root.sh script...
The following environment variables are set as:
ORACLE_OWNER= grid
ORACLE_HOME= /u01/app/11.2.0/grid
Enter the full pathname of the local bin directory: [/usr/local/bin]:
The file "dbhome" already exists in /usr/local/bin. Overwrite it? (y/n)
[n]:
The file "oraenv" already exists in /usr/local/bin. Overwrite it? (y/n)
[n]:
The file "coraenv" already exists in /usr/local/bin. Overwrite it? (y/n)
[n]:
Entries will be added to the /etc/oratab file as needed by
Database Configuration Assistant when a database is created
Finished running generic part of root.sh script.
Now product-specific root actions will be performed.
2019-05-22 10:33:37: Parsing the host name
2019-05-22 10:33:37: Checking for super user privileges
2019-05-22 10:33:37: User has super user privileges
Using configuration parameter file: /u01/app/11.2.0/grid/crs/install/crsconfig_params --->root.sh执行的是这个脚本
LOCAL ADD MODE
Creating OCR keys for user 'root', privgrp 'root'..
Operation successful.
Adding daemon to inittab ---->(11.2.0.1需要执行dd if=/var/tmp/.oracle/npohasd of=/dev/null bs=1024 count=1)
CRS-4123: Oracle High Availability Services has been started.
ohasd is starting
ADVM/ACFS is not supported on redhat-release-server-7.1-1.el7.x86_64
CRS-2672: Attempting to start 'ora.gipcd' on 'rhel1'
CRS-2672: Attempting to start 'ora.mdnsd' on 'rhel1'
CRS-2676: Start of 'ora.gipcd' on 'rhel1' succeeded
CRS-2676: Start of 'ora.mdnsd' on 'rhel1' succeeded
CRS-2672: Attempting to start 'ora.gpnpd' on 'rhel1'
CRS-2676: Start of 'ora.gpnpd' on 'rhel1' succeeded
CRS-2672: Attempting to start 'ora.cssdmonitor' on 'rhel1'
CRS-2676: Start of 'ora.cssdmonitor' on 'rhel1' succeeded
CRS-2672: Attempting to start 'ora.cssd' on 'rhel1'
CRS-2672: Attempting to start 'ora.diskmon' on 'rhel1'
CRS-2676: Start of 'ora.diskmon' on 'rhel1' succeeded
CRS-2676: Start of 'ora.cssd' on 'rhel1' succeeded
CRS-2672: Attempting to start 'ora.ctssd' on 'rhel1'
CRS-2676: Start of 'ora.ctssd' on 'rhel1' succeeded
Disk Group DATA already exists. Cannot be created again --->这个磁盘组之前替换后并没有被我dd,所以asm加载失败。
Configuration of ASM failed, see logs for details
Did not succssfully configure and start ASM
CRS-2500: Cannot stop resource 'ora.crsd' as it is not running
CRS-4000: Command Stop failed, or completed with errors.
Command return code of 1 (256) from command: /u01/app/11.2.0/grid/bin/crsctl stop resource ora.crsd -init
Stop of resource "ora.crsd -init" failed
Failed to stop CRSD
CRS-2500: Cannot stop resource 'ora.asm' as it is not running
CRS-4000: Command Stop failed, or completed with errors.
Command return code of 1 (256) from command: /u01/app/11.2.0/grid/bin/crsctl stop resource ora.asm -init
Stop of resource "ora.asm -init" failed
Failed to stop ASM
CRS-2673: Attempting to stop 'ora.ctssd' on 'rhel1'
CRS-2677: Stop of 'ora.ctssd' on 'rhel1' succeeded
CRS-2673: Attempting to stop 'ora.cssdmonitor' on 'rhel1'
CRS-2677: Stop of 'ora.cssdmonitor' on 'rhel1' succeeded
CRS-2673: Attempting to stop 'ora.cssd' on 'rhel1'
CRS-2677: Stop of 'ora.cssd' on 'rhel1' succeeded
CRS-2673: Attempting to stop 'ora.gpnpd' on 'rhel1'
CRS-2677: Stop of 'ora.gpnpd' on 'rhel1' succeeded
CRS-2679: Attempting to clean 'ora.gpnpd' on 'rhel1'
CRS-2681: Clean of 'ora.gpnpd' on 'rhel1' succeeded
CRS-2673: Attempting to stop 'ora.gipcd' on 'rhel1'
CRS-2677: Stop of 'ora.gipcd' on 'rhel1' succeeded
CRS-2673: Attempting to stop 'ora.mdnsd' on 'rhel1'
CRS-2677: Stop of 'ora.mdnsd' on 'rhel1' succeeded
Initial cluster configuration failed. See /u01/app/11.2.0/grid/cfgtoollogs/crsconfig/rootcrs_rhel1.log for details
[root@rhel1 grid]# /u01/app/11.2.0/grid/crs/install/roothas.pl -deconfig -force –verbose
defined(@array) is deprecated at /u01/app/11.2.0/grid/crs/install/crsconfig_lib.pm line 1702.
(Maybe you should just omit the defined()?)
defined(@array) is deprecated at /u01/app/11.2.0/grid/crs/install/crsconfig_lib.pm line 1702.
(Maybe you should just omit the defined()?)
defined(@array) is deprecated at /u01/app/11.2.0/grid/crs/install/crsconfig_lib.pm line 1780.
(Maybe you should just omit the defined()?)
2019-05-22 10:38:55: Checking for super user privileges
2019-05-22 10:38:55: User has super user privileges
2019-05-22 10:38:55: Parsing the host name
Using configuration parameter file: /u01/app/11.2.0/grid/crs/install/crsconfig_params
CRS-4535: Cannot communicate with Cluster Ready Services
CRS-4000: Command Stop failed, or completed with errors.
CRS-4535: Cannot communicate with Cluster Ready Services
CRS-4000: Command Delete failed, or completed with errors.
CRS-4133: Oracle High Availability Services has been stopped.
ADVM/ACFS is not supported on redhat-release-server-7.1-1.el7.x86_64
ACFS-9201: Not Supported
Successfully deconfigured Oracle Restart stack
[root@rhel1 ~]# vi /u01/app/11.2.0/grid/crs/install/crsconfig_params ----->更改root.sh执行的脚本,在此我修改了asm的组
(
ASM_DISK_GROUP=ABC
ASM_DISCOVERY_STRING=
ASM_DISKS=ORCL:AAA,ORCL:BBB,ORCL:CCC
)
[root@rhel1 grid]# ./root.sh
Running Oracle 11g root.sh script...
The following environment variables are set as:
ORACLE_OWNER= grid
ORACLE_HOME= /u01/app/11.2.0/grid
Enter the full pathname of the local bin directory: [/usr/local/bin]:
The file "dbhome" already exists in /usr/local/bin. Overwrite it? (y/n)
[n]:
The file "oraenv" already exists in /usr/local/bin. Overwrite it? (y/n)
[n]:
The file "coraenv" already exists in /usr/local/bin. Overwrite it? (y/n)
[n]:
Entries will be added to the /etc/oratab file as needed by
Database Configuration Assistant when a database is created
Finished running generic part of root.sh script.
Now product-specific root actions will be performed.
2019-05-22 10:40:41: Parsing the host name
2019-05-22 10:40:41: Checking for super user privileges
2019-05-22 10:40:41: User has super user privileges
Using configuration parameter file: /u01/app/11.2.0/grid/crs/install/crsconfig_params
LOCAL ADD MODE
Creating OCR keys for user 'root', privgrp 'root'..
Operation successful.
Adding daemon to inittab
CRS-4123: Oracle High Availability Services has been started.
ohasd is starting
ADVM/ACFS is not supported on redhat-release-server-7.1-1.el7.x86_64
CRS-2672: Attempting to start 'ora.gipcd' on 'rhel1'
CRS-2672: Attempting to start 'ora.mdnsd' on 'rhel1'
CRS-2676: Start of 'ora.gipcd' on 'rhel1' succeeded
CRS-2676: Start of 'ora.mdnsd' on 'rhel1' succeeded
CRS-2672: Attempting to start 'ora.gpnpd' on 'rhel1'
CRS-2676: Start of 'ora.gpnpd' on 'rhel1' succeeded
CRS-2672: Attempting to start 'ora.cssdmonitor' on 'rhel1'
CRS-2676: Start of 'ora.cssdmonitor' on 'rhel1' succeeded
CRS-2672: Attempting to start 'ora.cssd' on 'rhel1'
CRS-2672: Attempting to start 'ora.diskmon' on 'rhel1'
CRS-2676: Start of 'ora.diskmon' on 'rhel1' succeeded
CRS-2676: Start of 'ora.cssd' on 'rhel1' succeeded
CRS-2672: Attempting to start 'ora.ctssd' on 'rhel1'
CRS-2676: Start of 'ora.ctssd' on 'rhel1' succeeded
ASM created and started successfully.
DiskGroup ABC created successfully.
clscfg: -install mode specified
Successfully accumulated necessary OCR keys.
Creating OCR keys for user 'root', privgrp 'root'..
Operation successful.
CRS-2672: Attempting to start 'ora.crsd' on 'rhel1'
CRS-2676: Start of 'ora.crsd' on 'rhel1' succeeded
Successful addition of voting disk 5967fa6e88834f56bf236d0128957259.
Successful addition of voting disk 49c2405c6c6e4f96bf7a417022905afb.
Successful addition of voting disk 9f7ea2d3d0fc4f94bf9e7fd81a2228d2.
Successfully replaced voting disk group with +ABC.
CRS-4266: Voting file(s) successfully replaced
## STATE File Universal Id File Name Disk group
-- ----- ----------------- --------- ---------
1. ONLINE 5967fa6e88834f56bf236d0128957259 (ORCL:AAA) [ABC]
2. ONLINE 49c2405c6c6e4f96bf7a417022905afb (ORCL:BBB) [ABC]
3. ONLINE 9f7ea2d3d0fc4f94bf9e7fd81a2228d2 (ORCL:CCC) [ABC]
Located 3 voting disk(s).
CRS-2673: Attempting to stop 'ora.crsd' on 'rhel1'
CRS-2677: Stop of 'ora.crsd' on 'rhel1' succeeded
CRS-2679: Attempting to clean 'ora.crsd' on 'rhel1'
CRS-2681: Clean of 'ora.crsd' on 'rhel1' succeeded
CRS-2673: Attempting to stop 'ora.asm' on 'rhel1'
CRS-2677: Stop of 'ora.asm' on 'rhel1' succeeded
CRS-2673: Attempting to stop 'ora.ctssd' on 'rhel1'
CRS-2677: Stop of 'ora.ctssd' on 'rhel1' succeeded
CRS-2673: Attempting to stop 'ora.cssdmonitor' on 'rhel1'
CRS-2677: Stop of 'ora.cssdmonitor' on 'rhel1' succeeded
CRS-2673: Attempting to stop 'ora.cssd' on 'rhel1'
CRS-2677: Stop of 'ora.cssd' on 'rhel1' succeeded
CRS-2673: Attempting to stop 'ora.gpnpd' on 'rhel1'
CRS-2677: Stop of 'ora.gpnpd' on 'rhel1' succeeded
CRS-2679: Attempting to clean 'ora.gpnpd' on 'rhel1'
CRS-2681: Clean of 'ora.gpnpd' on 'rhel1' succeeded
CRS-2673: Attempting to stop 'ora.gipcd' on 'rhel1'
CRS-2677: Stop of 'ora.gipcd' on 'rhel1' succeeded
CRS-2673: Attempting to stop 'ora.mdnsd' on 'rhel1'
CRS-2677: Stop of 'ora.mdnsd' on 'rhel1' succeeded
CRS-2672: Attempting to start 'ora.mdnsd' on 'rhel1'
CRS-2676: Start of 'ora.mdnsd' on 'rhel1' succeeded
CRS-2672: Attempting to start 'ora.gipcd' on 'rhel1'
CRS-2676: Start of 'ora.gipcd' on 'rhel1' succeeded
CRS-2672: Attempting to start 'ora.gpnpd' on 'rhel1'
CRS-2676: Start of 'ora.gpnpd' on 'rhel1' succeeded
CRS-2672: Attempting to start 'ora.cssdmonitor' on 'rhel1'
CRS-2676: Start of 'ora.cssdmonitor' on 'rhel1' succeeded
CRS-2672: Attempting to start 'ora.cssd' on 'rhel1'
CRS-2672: Attempting to start 'ora.diskmon' on 'rhel1'
CRS-2676: Start of 'ora.diskmon' on 'rhel1' succeeded
CRS-2676: Start of 'ora.cssd' on 'rhel1' succeeded
CRS-2672: Attempting to start 'ora.ctssd' on 'rhel1'
CRS-2676: Start of 'ora.ctssd' on 'rhel1' succeeded
CRS-2672: Attempting to start 'ora.asm' on 'rhel1'
CRS-2676: Start of 'ora.asm' on 'rhel1' succeeded
CRS-2672: Attempting to start 'ora.crsd' on 'rhel1'
CRS-2676: Start of 'ora.crsd' on 'rhel1' succeeded
CRS-2672: Attempting to start 'ora.evmd' on 'rhel1'
CRS-2676: Start of 'ora.evmd' on 'rhel1' succeeded
CRS-2672: Attempting to start 'ora.asm' on 'rhel1'
CRS-2676: Start of 'ora.asm' on 'rhel1' succeeded
CRS-2672: Attempting to start 'ora.ABC.dg' on 'rhel1'
CRS-2676: Start of 'ora.ABC.dg' on 'rhel1' succeeded
rhel1 2019/05/22 10:46:57 /u01/app/11.2.0/grid/cdata/rhel1/backup_20190522_104657.olr
Preparing packages...
cvuqdisk-1.0.7-1.x86_64
Configure Oracle Grid Infrastructure for a Cluster ... succeeded
Updating inventory properties for clusterware
Starting Oracle Universal Installer...
Checking swap space: must be greater than 500 MB. Actual 3054 MB Passed
The inventory pointer is located at /etc/oraInst.loc
The inventory is located at /u01/app/oraInventory
'UpdateNodeList' was successful.
[root@rhel1 grid]#
节点二执行相同步骤。
可是查看asm磁盘组的时候,之前的dbfile与FRA磁盘的信息却不在了。
[grid@rhel2 ~]$ asmcmd
ASMCMD> ls
ABC/
ASMCMD> exit
两个节点分别执行前面的操作
还原OCR磁盘
[root@rhel2 ~]# /u01/app/11.2.0/grid/bin/crsctl stop crs
[root@rhel1 grid]# ll /u1/ocrdisk/ ---->查看自动备份ocr的信息,其中/u1/ocrdisk/备份路径是我之前定义的,
backup00.ocr day.ocr
backup01.ocr ocr
backup02.ocr week_.ocr
backup_20190507_111618.ocr week.ocr
backup_20190521_161536.ocr
[root@rhel1 grid]# ./bin/ocrcheck
Status of Oracle Cluster Registry is as follows :
Version : 3
Total space (kbytes) : 262120
Used space (kbytes) : 2032
Available space (kbytes) : 260088
ID : 849232275
Device/File Name : +ABC
Device/File integrity check succeeded
Device/File not configured
Device/File not configured
Device/File not configured
Device/File not configured
Cluster registry integrity check succeeded
Logical corruption check succeeded
[root@rhel1 grid]# ./bin/ocrconfig -showbackup
PROT-24: Auto backups for the Oracle Cluster Registry are not available
PROT-25: Manual backups for the Oracle Cluster Registry are not available
[root@rhel1 grid]# ./bin/crsctl start crs -excl
CRS-4123: Oracle High Availability Services has been started.
CRS-2672: Attempting to start 'ora.gipcd' on 'rhel1'
CRS-2672: Attempting to start 'ora.mdnsd' on 'rhel1'
CRS-2676: Start of 'ora.mdnsd' on 'rhel1' succeeded
CRS-2676: Start of 'ora.gipcd' on 'rhel1' succeeded
CRS-2672: Attempting to start 'ora.gpnpd' on 'rhel1'
CRS-2676: Start of 'ora.gpnpd' on 'rhel1' succeeded
CRS-2672: Attempting to start 'ora.cssdmonitor' on 'rhel1'
CRS-2676: Start of 'ora.cssdmonitor' on 'rhel1' succeeded
CRS-2672: Attempting to start 'ora.cssd' on 'rhel1'
CRS-2679: Attempting to clean 'ora.diskmon' on 'rhel1'
CRS-2681: Clean of 'ora.diskmon' on 'rhel1' succeeded
CRS-2672: Attempting to start 'ora.diskmon' on 'rhel1'
CRS-2676: Start of 'ora.diskmon' on 'rhel1' succeeded
CRS-2676: Start of 'ora.cssd' on 'rhel1' succeeded
CRS-2672: Attempting to start 'ora.ctssd' on 'rhel1'
CRS-2676: Start of 'ora.ctssd' on 'rhel1' succeeded
CRS-2672: Attempting to start 'ora.asm' on 'rhel1'
CRS-2676: Start of 'ora.asm' on 'rhel1' succeeded
CRS-2672: Attempting to start 'ora.crsd' on 'rhel1'
CRS-2676: Start of 'ora.crsd' on 'rhel1' succeeded
[root@rhel1 grid]# ll /u1/ocrdisk/ ---->这是之前的ocr备份,通过该备份还原
backup00.ocr day.ocr
backup01.ocr ocr
backup02.ocr week_.ocr
backup_20190507_111618.ocr week.ocr
backup_20190521_161536.ocr
[root@rhel1 grid]# ll /u1/ocrdisk/backup00.ocr
-rw------- 1 root root 7012352 May 20 16:37 /u1/ocrdisk/backup00.ocr
[root@rhel1 grid]# ./bin/ocrconfig -restore /u1/ocrdisk/backup00.ocr
PROT-19: Cannot proceed while the Cluster Ready Service is running
[root@rhel1 grid]# ./bin/crsctl stop resource ora.crsd –init
CRS-2673: Attempting to stop 'ora.crsd' on 'rhel1'
CRS-2677: Stop of 'ora.crsd' on 'rhel1' succeeded
CRS-2679: Attempting to clean 'ora.crsd' on 'rhel1'
CRS-2681: Clean of 'ora.crsd' on 'rhel1' succeeded
[root@rhel1 grid]# ./bin/ocrconfig -restore /u1/ocrdisk/backup00.ocr
[root@rhel1 grid]# ./bin/crsctl stop crs
CRS-2791: Starting shutdown of Oracle High Availability Services-managed resources on 'rhel1'
CRS-2673: Attempting to stop 'ora.cssdmonitor' on 'rhel1'
CRS-2673: Attempting to stop 'ora.ctssd' on 'rhel1'
CRS-2673: Attempting to stop 'ora.asm' on 'rhel1'
CRS-2673: Attempting to stop 'ora.mdnsd' on 'rhel1'
CRS-2677: Stop of 'ora.cssdmonitor' on 'rhel1' succeeded
CRS-2677: Stop of 'ora.mdnsd' on 'rhel1' succeeded
CRS-2677: Stop of 'ora.ctssd' on 'rhel1' succeeded
CRS-2677: Stop of 'ora.asm' on 'rhel1' succeeded
CRS-2673: Attempting to stop 'ora.cssd' on 'rhel1'
CRS-2677: Stop of 'ora.cssd' on 'rhel1' succeeded
CRS-2673: Attempting to stop 'ora.gpnpd' on 'rhel1'
CRS-2673: Attempting to stop 'ora.diskmon' on 'rhel1'
CRS-2677: Stop of 'ora.gpnpd' on 'rhel1' succeeded
CRS-2679: Attempting to clean 'ora.gpnpd' on 'rhel1'
CRS-2681: Clean of 'ora.gpnpd' on 'rhel1' succeeded
CRS-2673: Attempting to stop 'ora.gipcd' on 'rhel1'
CRS-2677: Stop of 'ora.gipcd' on 'rhel1' succeeded
CRS-2677: Stop of 'ora.diskmon' on 'rhel1' succeeded
CRS-2793: Shutdown of Oracle High Availability Services-managed resources on 'rhel1' has completed
CRS-4133: Oracle High Availability Services has been stopped.
[root@rhel1 grid]# ./bin/crsctl start crs
CRS-4123: Oracle High Availability Services has been started.
[root@rhel1 grid]# su - grid
Last login: Wed May 22 10:46:59 CST 2019 on pts/0
[grid@rhel1 ~]$ asmcmd --->如下存在FRA磁盘组。
ASMCMD> ls
ABC/
DATA/
FRA/
ASMCMD>
ocr磁盘组还原成功,之后就是还原数据库。
还原数据库
[grid@rhel1 ~]$ srvctl start database -d ORCL -o nomount
PRCR-1079 : Failed to start resource ora.orcl.db
ORA-15032: not all alterations performed
ORA-15017: diskgroup "DBFILE" cannot be mounted
ORA-15063: ASM discovered an insufficient number of disks for diskgroup "DBFILE"
CRS-2674: Start of 'ora.DBFILE.dg' on 'rhel1' failed
ORA-15032: not all alterations performed
ORA-15017: diskgroup "DBFILE" cannot be mounted
ORA-15063: ASM discovered an insufficient number of disks for diskgroup "DBFILE"
CRS-2674: Start of 'ora.DBFILE.dg' on 'rhel2' failed
CRS-2632: There are no more servers to try to place resource 'ora.orcl.db' on that would satisfy its placement policy
[grid@rhel1 ~]$ sqlplus "/as sysasm"
SQL*Plus: Release 11.2.0.1.0 Production on Wed May 22 11:25:00 2019
Copyright (c) 1982, 2009, Oracle. All rights reserved.
Connected to:
Oracle Database 11g Enterprise Edition Release 11.2.0.1.0 - 64bit Production
With the Real Application Clusters and Automatic Storage Management options
SQL> create diskgroup DBFILE external redundancy disk 'ORCL:DBFILE' name DBFILE;
Diskgroup created.
[oracle@rhel1 ~]$ cat /u01/app/oracle/product/11.2.0/db_1/dbs/initORCL1.ora --->查看pfile参数文件,调用的是dbfile磁盘组下面的spfile文件。
SPFILE='+DBFILE/ORCL/spfileORCL.ora'
[oracle@rhel1 ~]$ cat /u01/app/oracle/admin/ORCL/pfile/init.ora.3182019191943 --->该目录下会有最开始的参数文件,通过这个打开至nomount状态
##############################################################################
# Copyright (c) 1991, 2001, 2002 by Oracle Corporation
##############################################################################
###########################################
# Cache and I/O
###########################################
db_block_size=8192
###########################################
# Cluster Database
###########################################
remote_listener=rhel-cluster-scan.grid.example.com:1521
###########################################
# Cursors and Library Cache
###########################################
open_cursors=300
###########################################
# Database Identification
###########################################
db_domain=""
db_name=ORCL
###########################################
# File Configuration
###########################################
db_create_file_dest=+DBFILE
db_recovery_file_dest=+FRA
db_recovery_file_dest_size=5218762752
###########################################
# Miscellaneous
###########################################
compatible=11.2.0.0.0
diagnostic_dest=/u01/app/oracle
memory_target=764411904
###########################################
# Processes and Sessions
###########################################
processes=150
###########################################
# Security and Auditing
###########################################
audit_file_dest=/u01/app/oracle/admin/ORCL/adump
audit_trail=db
remote_login_passwordfile=exclusive
###########################################
# Shared Server
###########################################
dispatchers="(PROTOCOL=TCP) (SERVICE=ORCLXDB)"
control_files=("+DBFILE/orcl/controlfile/current.256.1005931173", "+FRA/orcl/controlfile/current.256.1005931177")
cluster_database=true
ORCL1.instance_number=1
ORCL2.instance_number=2
ORCL2.thread=2
ORCL1.undo_tablespace=UNDOTBS1
ORCL2.undo_tablespace=UNDOTBS2
ORCL1.thread=1
[oracle@rhel1 ~]$ sqlplus "/as sysdba"
SQL > startup nomount pfile='/u01/app/oracle/admin/ORCL/pfile/init.ora.3182019191943'; --->可以将创建成spfile,再复制到asm磁盘文件中。
ORACLE instance started.
Total System Global Area 764121088 bytes
Fixed Size 2217264 bytes
Variable Size 511707856 bytes
Database Buffers 247463936 bytes
Redo Buffers 2732032 bytes
[oracle@rhel1 ~]$ rman target /
RMAN> restore controlfile from '+FRA/orcl/controlfile/current.256.1005931177'; --->因为fra盘并未被我删除。所以控制文件,归档日志文件还存在里面
Starting restore at 22-MAY-19
using target database control file instead of recovery catalog
allocated channel: ORA_DISK_1
channel ORA_DISK_1: SID=30 instance=ORCL1 device type=DISK
channel ORA_DISK_1: copied control file copy
output file name=+DBFILE/orcl/controlfile/current.256.1008936907
output file name=+FRA/orcl/controlfile/current.256.1005931177
Finished restore at 22-MAY-19
RMAN> alter database mount;
database mounted
released channel: ORA_DISK_1
RMAN> list backup;
List of Backup Sets
===================
BS Key Type LV Size Device Type Elapsed Time Completion Time
------- ---- -- ---------- ----------- ------------ ---------------
3 Full 947.01M DISK 00:02:47 14-MAY-19
BP Key: 3 Status: AVAILABLE Compressed: NO Tag: TAG20190514T170446
Piece Name: /home/oracle/03u1hntf_1_1.bak
List of Datafiles in backup set 3
File LV Type Ckp SCN Ckp Time Name
---- -- ---- ---------- --------- ----
1 Full 1240032 14-MAY-19 +DBFILE/orcl/datafile/system.259.1005931193
2 Full 1240032 14-MAY-19 +DBFILE/orcl/datafile/sysaux.260.1005931273
3 Full 1240032 14-MAY-19 +DBFILE/orcl/datafile/undotbs1.261.1005931315
4 Full 1240032 14-MAY-19 +DBFILE/orcl/datafile/undotbs2.263.1005931371
5 Full 1240032 14-MAY-19 +DBFILE/orcl/datafile/users.264.1005931391
6 Full 1240032 14-MAY-19 +DBFILE/orcl/datafile/tbs1.269.1008256827
BS Key Type LV Size Device Type Elapsed Time Completion Time
------- ---- -- ---------- ----------- ------------ ---------------
4 Full 17.70M DISK 00:00:15 14-MAY-19
BP Key: 4 Status: AVAILABLE Compressed: NO Tag: TAG20190514T170446
Piece Name: /home/oracle/04u1ho2o_1_1.bak --->刚好在路径下面有备份。
SPFILE Included: Modification time: 14-MAY-19
SPFILE db_unique_name: ORCL
Control File Included: Ckp SCN: 1240638 Ckp time: 14-MAY-19
RMAN> restore database;
Starting restore at 22-MAY-19
allocated channel: ORA_DISK_1
channel ORA_DISK_1: SID=30 instance=ORCL1 device type=DISK
channel ORA_DISK_1: starting datafile backup set restore
channel ORA_DISK_1: specifying datafile(s) to restore from backup set
channel ORA_DISK_1: restoring datafile 00001 to +DBFILE/orcl/datafile/system.259.1005931193
channel ORA_DISK_1: restoring datafile 00002 to +DBFILE/orcl/datafile/sysaux.260.1005931273
channel ORA_DISK_1: restoring datafile 00003 to +DBFILE/orcl/datafile/undotbs1.261.1005931315
channel ORA_DISK_1: restoring datafile 00004 to +DBFILE/orcl/datafile/undotbs2.263.1005931371
channel ORA_DISK_1: restoring datafile 00005 to +DBFILE/orcl/datafile/users.264.1005931391
channel ORA_DISK_1: restoring datafile 00006 to +DBFILE/orcl/datafile/tbs1.269.1008256827
channel ORA_DISK_1: reading from backup piece /home/oracle/03u1hntf_1_1.bak
channel ORA_DISK_1: piece handle=/home/oracle/03u1hntf_1_1.bak tag=TAG20190514T170446
channel ORA_DISK_1: restored backup piece 1
channel ORA_DISK_1: restore complete, elapsed time: 00:02:30
Finished restore at 22-MAY-19
RMAN> recover database;
Starting recover at 22-MAY-19
using channel ORA_DISK_1
starting media recovery
archived log for thread 1 with sequence 66 is already on disk as file +FRA/orcl/archivelog/2019_05_14/thread_1_seq_66.263.1008264651
archived log for thread 1 with sequence 67 is already on disk as file +FRA/orcl/archivelog/2019_05_20/thread_1_seq_67.262.1008754305
archived log for thread 1 with sequence 68 is already on disk as file +FRA/orcl/archivelog/2019_05_20/thread_1_seq_68.269.1008756045
archived log for thread 1 with sequence 69 is already on disk as file +FRA/orcl/archivelog/2019_05_20/thread_1_seq_69.270.1008757911
archived log for thread 1 with sequence 70 is already on disk as file +FRA/orcl/archivelog/2019_05_20/thread_1_seq_70.272.1008757951
archived log for thread 1 with sequence 71 is already on disk as file +FRA/orcl/archivelog/2019_05_20/thread_1_seq_71.273.1008762899
archived log for thread 1 with sequence 72 is already on disk as file +FRA/orcl/archivelog/2019_05_20/thread_1_seq_72.276.1008764651
archived log for thread 1 with sequence 73 is already on disk as file +FRA/orcl/archivelog/2019_05_20/thread_1_seq_73.277.1008774719
archived log for thread 1 with sequence 74 is already on disk as file +FRA/orcl/archivelog/2019_05_21/thread_1_seq_74.278.1008858631
archived log for thread 1 with sequence 75 is already on disk as file +FRA/orcl/archivelog/2019_05_21/thread_1_seq_75.282.1008863235
archived log for thread 1 with sequence 76 is already on disk as file +FRA/orcl/archivelog/2019_05_21/thread_1_seq_76.283.1008864805
archived log for thread 2 with sequence 13 is already on disk as file +FRA/orcl/archivelog/2019_05_20/thread_2_seq_13.267.1008754309
archived log for thread 2 with sequence 14 is already on disk as file +FRA/orcl/archivelog/2019_05_20/thread_2_seq_14.268.1008754309
archived log for thread 2 with sequence 15 is already on disk as file +FRA/orcl/archivelog/2019_05_20/thread_2_seq_15.271.1008757919
archived log for thread 2 with sequence 16 is already on disk as file +FRA/orcl/archivelog/2019_05_20/thread_2_seq_16.274.1008762921
archived log for thread 2 with sequence 17 is already on disk as file +FRA/orcl/archivelog/2019_05_20/thread_2_seq_17.275.1008762923
archived log for thread 2 with sequence 18 is already on disk as file +FRA/orcl/archivelog/2019_05_21/thread_2_seq_18.279.1008863229
archived log for thread 2 with sequence 19 is already on disk as file +FRA/orcl/archivelog/2019_05_21/thread_2_seq_19.280.1008863233
archived log for thread 2 with sequence 20 is already on disk as file +FRA/orcl/archivelog/2019_05_21/thread_2_seq_20.281.1008863233
archived log file name=+FRA/orcl/archivelog/2019_05_14/thread_1_seq_66.263.1008264651 thread=1 sequence=66
archived log file name=+FRA/orcl/archivelog/2019_05_20/thread_2_seq_13.267.1008754309 thread=2 sequence=13
archived log file name=+FRA/orcl/archivelog/2019_05_20/thread_1_seq_67.262.1008754305 thread=1 sequence=67
archived log file name=+FRA/orcl/archivelog/2019_05_20/thread_2_seq_14.268.1008754309 thread=2 sequence=14
archived log file name=+FRA/orcl/archivelog/2019_05_20/thread_1_seq_68.269.1008756045 thread=1 sequence=68
archived log file name=+FRA/orcl/archivelog/2019_05_20/thread_2_seq_15.271.1008757919 thread=2 sequence=15
archived log file name=+FRA/orcl/archivelog/2019_05_20/thread_1_seq_69.270.1008757911 thread=1 sequence=69
archived log file name=+FRA/orcl/archivelog/2019_05_20/thread_1_seq_70.272.1008757951 thread=1 sequence=70
archived log file name=+FRA/orcl/archivelog/2019_05_20/thread_2_seq_16.274.1008762921 thread=2 sequence=16
archived log file name=+FRA/orcl/archivelog/2019_05_20/thread_1_seq_71.273.1008762899 thread=1 sequence=71
archived log file name=+FRA/orcl/archivelog/2019_05_20/thread_2_seq_17.275.1008762923 thread=2 sequence=17
archived log file name=+FRA/orcl/archivelog/2019_05_20/thread_1_seq_72.276.1008764651 thread=1 sequence=72
archived log file name=+FRA/orcl/archivelog/2019_05_20/thread_1_seq_73.277.1008774719 thread=1 sequence=73
archived log file name=+FRA/orcl/archivelog/2019_05_21/thread_2_seq_18.279.1008863229 thread=2 sequence=18
archived log file name=+FRA/orcl/archivelog/2019_05_21/thread_1_seq_74.278.1008858631 thread=1 sequence=74
archived log file name=+FRA/orcl/archivelog/2019_05_21/thread_2_seq_19.280.1008863233 thread=2 sequence=19
media recovery complete, elapsed time: 00:00:42
Finished recover at 22-MAY-19
RMAN> alter database open;
RMAN-00571: ===========================================================
RMAN-00569: =============== ERROR MESSAGE STACK FOLLOWS ===============
RMAN-00571: ===========================================================
RMAN-03002: failure of alter db command at 05/22/2019 12:48:55
ORA-19751: could not create the change tracking file
ORA-19750: change tracking file: '+DBFILE/orcl/changetracking/ctf.268.1008159369'
ORA-17502: ksfdcre:4 Failed to create file +DBFILE/orcl/changetracking/ctf.268.1008159369
ORA-15046: ASM file name '+DBFILE/orcl/changetracking/ctf.268.1008159369' is not in single-file creation form
tracking file是优化增量备份的
连接oracle
SQL> ALTER DATABASE DISABLE BLOCK CHANGE TRACKING;----->可以先关闭
Database altered.
SQL> alter database open resetlogs;
Database altered.
SQL> archive log list;
Database log mode Archive Mode
Automatic archival Enabled
Archive destination USE_DB_RECOVERY_FILE_DEST
Oldest online log sequence 2
Next log sequence to archive 3
Current log sequence 3
SQL> ALTER DATABASE ENABLE BLOCK CHANGE TRACKING;
Database altered.
SQL> SELECT STATUS, FILENAME FROM V$BLOCK_CHANGE_TRACKING;---->因为是OMF管理的,可以不用using file来创建
STATUS
--------------------
FILENAME
--------------------------------------------------------------------------------
ENABLED
+DBFILE/orcl/changetracking/ctf.277.1008949385
[grid@rhel1 ~]$ crsctl stat res -t
--------------------------------------------------------------------------------
NAME TARGET STATE SERVER STATE_DETAILS
--------------------------------------------------------------------------------
Local Resources
--------------------------------------------------------------------------------
ora.ABC.dg
ONLINE ONLINE rhel1
ONLINE ONLINE rhel2
ora.DATA.dg
ONLINE ONLINE rhel1
ONLINE ONLINE rhel2
ora.DBFILE.dg
ONLINE ONLINE rhel1
ONLINE ONLINE rhel2
ora.FRA.dg
ONLINE ONLINE rhel1
ONLINE ONLINE rhel2
ora.LISTENER.lsnr
ONLINE ONLINE rhel1
ONLINE ONLINE rhel2
ora.asm
ONLINE ONLINE rhel1 Started
ONLINE ONLINE rhel2 Started
ora.eons
ONLINE ONLINE rhel1
ONLINE ONLINE rhel2
ora.gsd
OFFLINE OFFLINE rhel1
OFFLINE OFFLINE rhel2
ora.net1.network
ONLINE ONLINE rhel1
ONLINE ONLINE rhel2
ora.ons
ONLINE ONLINE rhel1
ONLINE ONLINE rhel2
--------------------------------------------------------------------------------
Cluster Resources
--------------------------------------------------------------------------------
ora.LISTENER_SCAN1.lsnr
1 ONLINE ONLINE rhel2
ora.LISTENER_SCAN2.lsnr
1 ONLINE ONLINE rhel1
ora.LISTENER_SCAN3.lsnr
1 ONLINE ONLINE rhel1
ora.oc4j
1 OFFLINE OFFLINE
ora.orcl.db
1 ONLINE ONLINE rhel1 Open
2 ONLINE ONLINE rhel2 Open
ora.rhel1.vip
1 ONLINE ONLINE rhel1
ora.rhel2.vip
1 ONLINE ONLINE rhel2
ora.scan1.vip
1 ONLINE ONLINE rhel2
ora.scan2.vip
1 ONLINE ONLINE rhel1
ora.scan3.vip
1 ONLINE ONLINE rhel1