run {
SET nocfau;
ALLOCATE CHANNEL CH0 DEVICE TYPE DISK FORMAT '/nfs/datafile/%U_%T';
ALLOCATE CHANNEL CH1 DEVICE TYPE DISK FORMAT '/nfs/datafile/%U_%T';
ALLOCATE CHANNEL CH2 DEVICE TYPE DISK FORMAT '/nfs/datafile/%U_%T';
ALLOCATE CHANNEL CH3 DEVICE TYPE DISK FORMAT '/nfs/datafile/%U_%T';
BACKUP AS COPY INCREMENTAL LEVEL 0 TAG 'INCR_MERGE' DATABASE;
BACKUP ARCHIVELOG ALL FORMAT '/nfs/archivelogs/%U' NOT BACKED UP 1 TIMES;
RELEASE CHANNEL CH0;
RELEASE CHANNEL CH1;
RELEASE CHANNEL CH2;
RELEASE CHANNEL CH3;
}
The above is the rman script to run backup for Oracle RAC. I have NFS share that is mounted with following options: rw,user,hard,noac,vers=4
The NFS Server has the following export: NFSv4,rw,no_root_squash,no_all_squash,secure
The issue is that nothing gets written to the NFS share. I can create files, copy large files, but when RMAN is run nothing is copied to the NFS share.
The RMAN process runs till it's terminated.
When I do nfsiostat -a it shows 5 mandatory GETATTR requests.
Does it mean the Oracle datafiles are not being written in NFS share because the file attributes along with file uid/gid is not being successfully retrieved?
Update: tried running the backup on Oracle 19c RAC setup
Get the following RMAN error
RMAN-03009: failure of backup command on CH0 channel at 08/24/2020 16:59:11
ORA-19504: failed to create file "/home/oracle/nfs/ractest/1178998226/full/datafile/data_D-ORAMS19C_I-1178998226_TS-SYSAUX_FNO-3_rhv8m5mt_20200824"
ORA-17503: ksfdopn:11 Failed to open file /home/oracle/nfs/ractest/1178998226/full/datafile/data_D-ORAMS19C_I-1178998226_TS-SYSAUX_FNO-3_rhv8m5mt_20200824
ORA-17500: ODM err: No such file or directory
I ran the rman backup on a local disk they succeed and the datafiles have the group asmadmin. Do backups fail on NFS share becuase the datafile attributes/group information is not being copied to the NFS share?
The oracle user is not part of the asmadmin group.
Your NFS options are wrong for RMAN. For Oracle RAC, they should be as follows:
rw,bg,hard,nointr,rsize=32768,wsize=32768,tcp,vers=3,timeo=600
These options are required, per Oracle documentation. Also, the use of the "noac" option is specifically a problem for RMAN and should not be used.
It seems to be an issue with Oracle 12.1, Following the oracle support document and applying the patch(20720667) provided has resolved the issue. Here is the link
Related
I am trying to export a dump file and log file on a remote machine using oracle expdp.
However i am getting the following error :
Connected to: Oracle Database 11g
Enterprise Edition Release 11.2.0.4.0 - 64bit Production
With the Partitioning, OLAP, Data Mining and Real Application Testing
options
ORA-39002: invalid operation
ORA-39070: Unable to open the log file.
ORA-29283: invalid file operation
ORA-06512: at "SYS.UTL_FILE", line 536
ORA-29283: invalid file operation
Command run on remote machine host-name 'Local' using oracle client are :
SQL> create directory expdp_dir as '/vault2/expdp_dir';
SQL> grant read,write on directory expdp_dir to dbuser;
expdp dbuser/dbpwd#SID SCHEMAS=dbuser DIRECTORY=expdp_dir DUMPFILE=testDB24NOV17.dmp logfile=testDB24NOV17.log EXCLUDE=STATISTICS
Note vault 2 is mounted on a remote machine with hostname 'Local'. The database is on a machine with hostname TestDB.
The OS is RHEL6.
Any thoughts /ideas on making this operation successful would be appreciated.
Please check this one:
as per Oracle Doc.ID Doc ID 1305166.1
The errors can have multiple causes. Known causes are listed below.
One of the usual reasons for this problem to occur is when the listener process has not been started under the same account as the database instance service. The listener forks the new server process and when this runs under a different security context as the database, then access to directories and files are likely impacted.
Please verify the following information:
1) the output of:
ps -ef | grep SMON
2) the output of:
ps -ef | grep tnslsnr
3) the output of:
ps -ef|grep LIST
4) the output of:
ls -ld
Note:
When using ASM, the listener may have been started from the ASM Home instead of the RDBMS Home. Depending on your security settings, this may give to this issue.
One more:
4. Directory path/folder exists but create directory is executed by a different user in the database and the import is run by a different user.
Solution:
1. Make sure the listener and instance services are started from the same account.
Make sure that the directory are shared between nodes so that the directory can be accessed on any instance, or, create a folder similar to the other nodes locally, if there is already a folder created locally on the all the node with the same file directory path structure check if the permission are correct.
Make sure the folder exist has specified in during creation in the "CREATE DIRECTORY" syntax command.
Grant the required permission to the importing user to use the directory.
grant read, write on directory to ;
If above 4 possible causes and solutions are not applicable at your end, please check if the user has proper permission to export to run utl_file package.
Hope it helps.
I am trying to build a standby database in a Data Guard configuration using RMAN Duplicate Target Database For Standby Dorecover.
All the rman backupsets are there in Netbackup and I can see the inventory of backup pieces using the RMAN views.
However, the Duplicate Target Database command is failing with:
channel p4: starting datafile backup set restore
channel p4: restoring control file
channel p4: reading from backup piece c-2404308342-20160306-02
channel p4: ORA-19870: error while restoring backup piece c-2404308342-20160306-02
ORA-19507: failed to retrieve sequential file, handle="c-2404308342-20160306-02", parms=""
ORA-27029: skgfrtrv: sbtrestore returned error
ORA-19511: Error received from media manager layer, error text:
Backup file <c-2404308342-20160306-02> not found in NetBackup catalog
Basically, the backup pieces are there, but RMAN/Netbackup is saying that they are not there.
Do you know why?
Thank you.
Please try validating the restore with a validate restore command. It should take some time but will notify you of any backup corruptions.
e.g. RMAN > restore database validate;
Thanks
Sabiha
I am facing Database timed-out issue while creating a view and/or rebasing a stream on vobs
ERROR -> cleartool: Error: Database timed out -- aborting transaction: "\TP".
ERROR -> cleartool: Error: Unable to generate config spec for stream "stream name".
I am using CCRC to connect.
CCRC possibly means CC8.0.X+, but you would still need to access logs (that was cleartool getlog in previous CC versions, since you mentioned before using CC 7.1), in order to know more about that error message.
A "Database timed out -- aborting transaction" usually is followed on the logs by the actual root cause (like, for instance, db_VISTA database error -926 - problem in shared memory lock manager)
In the OP's case, it was (from this technote), a corrupt transaction file.
Error: db_VISTA error -925
Database timed out error is resolved after following this http://www-01.ibm.com/support/docview.wss?uid=swg21133944
Lock the VOB (if you receive errors from the lock command, proceed to step 2.)
Unmount the VOB
Kill any lingering VOB processes on the server using the albd_list utility.
Note: Review technote 1148639 for more details on albd_list and its usage.
Should some ClearCase processes still exist that are accessing the transaction files and preventing you from deleting them, you must completely - stop ClearCase on the server to end those processes.
Move the vista.tjf, vista.tcf, vista.taf files out of the VOB db directory
Refer to technote 1149322 About ClearCase database vista.* files for further information about these files.
Remount the VOB
The files will be recreated and the VOB will be functional
Note: If the files are not recreated, stop and restart ClearCase services on the host.
I have created a standby database on Windows machine with version 10.2.0.1
For the first time for the database to sync I have copied the archive from the primary database. But on the switch from the primary database. The archive log doesn't ship to the standby database.
SQL> recover standby database;
ORA-00279: change 1248007 generated at 07/29/2012 22:46:37 needed for thread 1
ORA-00289: suggestion : E:\ARCHIVE\ARC00043_0778845461.001
ORA-00280: change 1248007 for thread 1 is in sequence #43
Specify log: {<RET>=suggested | filename | AUTO | CANCEL}
ORA-00308: cannot open archived log 'E:\ARCHIVE\ARC00043_0778845461.001'
ORA-27041: unable to open file
OSD-04002: unable to open file
O/S-Error: (OS 2) The system cannot find the file specified.
ORA-01547: warning: RECOVER succeeded but OPEN RESETLOGS would get error below
ORA-01152: file 1 was not restored from a sufficiently old backup
ORA-01110: data file 1: 'E:\ORACLE\PRODUCT\10.2.0\ORADATA\STANDBY\SYSTEM01.DBF'
Earlier I took the cold backup to move the files.
I am trying to restore oracle RAC RAMN backup to another RAC where only software is installed. I am able to restore control files some how on "+DATA/..."
I have some doubt about RAC setup.
Que 1) What are +DATA, +FRA.
Que 2) I am so curious about exploring +DATA, +FRA ,how can i see content of it in linux?
Que 3)How to change oracle "+DATA/CDBRAC/ARCHIVELOG/2016_02_20" archive destination to local storage e.g. "/u02/archive/2016_02_20" during recovery?
because i am not able to startup mount because of below failure
Oracle instance started
RMAN-00571: ===========================================================
RMAN-00569: =============== ERROR MESSAGE STACK FOLLOWS ===============
RMAN-00571: ===========================================================
RMAN-03002: failure of startup command at 02/25/2016 15:25:45
ORA-00205: error in identifying control file, check alert log for more info
I am pretty sure this is because there are no +DATA/DBRAC/ARCHIVELOG/2016_02_20 path on +DATA.
similar problem i was able to resolve in oracle single machine recovery to other location by creating required directory.
+DATA and +FRA are ASM Disk Groups, you can find out more about them by querying the catalog view v$asm_diskgroup.
You can see the contents of ASM Disk Groups with the asmcmd utility, e.g.
$ amscmd
ASMCMD> cd DATA
ASMCMD> ls
YOURDB1/
YOURDB2/
ASMCMD> cd YOURDB1/
ASMCMD> ls
CONTROLFILE/
DATAFILE/
ONLINELOG/
PARAMETERFILE/
TEMPFILE/
spfileYOURDB1.ora
The error you got is related to the control file, you can check the value of the control_files parameter with the SQL*Plus command show parameter control_files.