oracle export failed due to unable to close dump file - oracle

The export is failing due to below error
ORA-31693: Table data object "x"."x_t" failed to load/unload and is being skipped due to error:
ORA-31643: unable to close dump file ""
Anyone has knowledge regarding this issue, please help out

Related

reg: error in dataguard setup core dump eception

Getting below error in dataguard is setup, any suggestion how to resolve it.
getting below error , We used with duplicate and without duplicate command but error is same. ORA-07445: exception encountered core dump [gslusslStrlen()+60] [SIGSEGV] [ADDR:0x0] [PC:0x62882AC] [SI_KERNEL(general_protection)] []

how to resolve thousand of errors in oracle import using impdp where I don't know the what parameters were used during expdp?

I am trying to import an oracle 11g dump file using impdp utility but while doing so, inter alia, I am facing two major errors:
First, It is showing the following error:
Processing object type DATABASE_EXPORT/TABLESPACE
ORA-39083: Object type TABLESPACE:"HIS_USER" failed to create with error:
ORA-01119: error in creating database file '/oracle/app/oracle/oradata/dwhrajdr1/his_user13.dbf'
ORA-27040: file create error, unable to create file
OSD-04002: unable to open file
O/S-Error: (OS 3) The system cannot find the path specified.
so to solve this, I have created the tablesapce with same name but now it is showing that 'HIS_USER' tablespace already exists.
Second, I am getting thousands of errors, where it is showing user or role does not exist:
Failing sql is:
GRANT EXECUTE ANY ASSEMBLY TO "DSS"
ORA-39083: Object type SYSTEM_GRANT failed to create with error:
ORA-01917: user or role 'DSS' does not exist
Please suggest how to solve these errors!
How can I import the dumpfile without making hundreds of users/roles or tablespaces?
you can generate sql statement using impdp the following way.
http://www.dba-oracle.com/t_convert_expdp_dmp_file_sql.htm
then adjust parameter accordingly.
scott

Missing productTypeIdentifier error using new Apple Build System (xcbuild)

I have an XCode project that builds with the legacy build system, but when using the new build system under xcode11 immediately fails with the err:
.. forcing incremental PIF cache clear after loading error: Required key 'productTypeIdentifier' is missing in StandardTarget dictionary
note: Using new build system
error: Build service could not start build operation: unknown error while handling message: MsgHandlingError(message: "no PIF transfer has been initiated")
I'm looking for any info that might help clarify this.
What is a 'StandardTarget dictionary'? Can I add the missing key directly? Or is this a consequence some other setting?
Are there any 'PIF transfer' experts out there?

Getting System error code of 126 in Informatica

The setup routine for the BMC ODBC for Oracle(default) Odbc driver couldn't be loaded due to system error 126. The specified module could not be found.
I'm getting this error in imformatica 9.5.1
Can anyone suggest any solution..

java.sql.SQLException: ORA-01157: cannot identify/lock data file

I was getting the below given error, When I run the application:
Caused by: org.hibernate.exception.GenericJDBCException: could not execute native bulk manipulation query
.
.
Caused by: java.sql.SQLException: ORA-01157: cannot identify/lock data file - see DBWR trace file
ORA-01110: data file : '/fld1/fld2/mytemp_tablespace.dbf'
I tried to find out this files and came to know that there is no folders. I have ,
then created the respective folders and a new empty mytemptemp_tablespace.dbf file. But still the same error is getting over there.
Any idea why this error is happening?If it is an SQL exception it could have happened at the right beginning itself.
What I have done is, I have created a new schema and exported the db from the old to this new one.
Also how can I see or get the DBWR trace file.
This could be the result of a restored database and during the restore rman was not able to create the tempfiles because of a missing directory.
Solution is quite simple, once the directories are created, just add one or more tempfiles:
alter tablespace mytemp_tablespace add tempfile '/fld1/fld2/mytemp_tablespace01.dbf';
when the temp tablespace has it's storage, your actions can succeed.

Resources