Oracle data pump import job list - oracle

Once monthly we are doing full export of production database and import those data into develop environment. Version of Oracle database is 11.2.0.3 on RHEL 6.4 (64bit).
My question is can i get list of those completed import/export data pump jobs, by querying data dictionary view or by using some other way?
Thank you

Related

oracle data pump export dump file to UAT environment

I am trying to learn data pump export and import of tables in oracle. I am asked to export from the device to UAT/sit environment. I have a doubt. I have been practicing this on my laptop with 2 pluggable databases. As it's on the same device, the directory object has no problem. what if I had to export it to the UAT environment? how would I use the impdp to import?
i am still a beginner in oracle.

Oracle 12C: ORA-00406: COMPATIBLE parameter needs to be 12.2.0.0.0

We have a 2 current Oracle 10G(10.2.0.1) production databases, and we are planning to migrate to a new database server with Oracle 12C. Since data is for each database is only around 5GB, the best way to create a new instance and use data pump to transfer data. To achieve this, I created a database link from the Oracle 12C database to the 10G, and use that to expdp the data from the 12C database. However when I import the exported data I have around tables that have an error like this:
ORA-39083: Object type TABLE:"USER"."WH_SEARCH_ACT" failed to create with error:
ORA-00406: COMPATIBLE parameter needs to be 12.2.0.0.0 or greater
ORA-00722: Feature "Partition Read Only"
Is there any solution to this other than adding the COMPATIBLE parameter in the production database? This is production so I can't really just update/modify the current production database. Any other solutions here, because i don't like just create the tables before importing the data pump file.
Use the version parameter in the data pump utility. For example:
expdp hr/hr TABLES=hr.employees VERSION=10.2
DIRECTORY=data_pump_dir DUMPFILE=emp10g.dmp LOGFILE=emp.log
Just make sure the version of the export utility is 10g and import utility is 12G. Hope this helps.
In the production database . you can use exp program
exp username/password buffer=64000 file=/path/to/path.dmp full=y
copy exp dump file to oracle 12c and use
imp username/password buffer=64000 file=/path/to/path.dmp full=y

How to import a csv into a remote Oracle database?

I have a dump from a database in a csv format ('|' character as the delimiter), and I want to import that into a remote Oracle database. I am using AWS and the csv is on an EC2 instance running amazon linux, and the remote Oracle database is an RDS instance. This is the first time I'm touching an Oracle database.
I expected this to be fairly simple, but trying to find info I kind of got lost. Some people say to use SQL*Loader, but I can't manage to even install that thing. Other's say that SQL*Loader is not supposed to even be installed on something that isn't the actual database server. So far I've only managed to install sqlplus and connect to the database, but no importing so far.
Basically I'm looking for an equivalent of \COPY in psql, but for Oracle. And how on earth would I use it in this context.
If you don't want to use SQL*Loader and only need to import CSV as a one time task, I would recommend using SQL Developer. If you want an automated process, SQL*Loader would probably be your best bet.
Here are some links (note that an Excel import and a CSV import are nearly identical):
How to move csv file into oracle database using sql developer? [closed]
How to Import from Excel to a New Oracle Table with SQL Developer
How to Import from Excel to Oracle with SQL Developer

Import the Cognos 10 Trial GOSALES schema (with data) into Oracle Database

I have installed the Cognos 10.2 Trial version. The installation comes a sample schema called "GoSales".
I would like to import the GoSales schema into my Oracle database. This will enable me to view the GoSales Schema in a database query tool (i.e. TOAD)
My ultimate goal is to:
- Understand the GoSales schema.
- Make sure that the reports I develop using Cognos Studio are correct by checking the data in the database.
So, how can I import the "Go Sales" schema into my Oracle database?
There is a compressed file with the Cognos BI samples. I guess you can download this file from IBM. Its name is bi_smps_10.2.1_mp_ml.tar.gz.
You have to uncompress this file and execute the script ...\win\GOSaleConfig.bat (for a Windows machine). This script will ask you the connection parameters of your Oracle database and create the schemas with sample data.

How can I replicate an Oracle 11g database(data+structure) on my local machine for development?

I am working on a test server with an Oracle 11g installed. I was wondering if there is anyway I can replicate the database(environment + data) on my local Linux machine. I am using a CentOS 5.3 on Windows XP with SUN Virtual Box. On Windows I am using sqldeveloper client to connect to the 11g database.
There are a number of ways to move the data over:
Restore an RMAN backup on your test server
Export and import the data using exp/expdp/imp/impdp
Export and import using a transportable tablespace (Further Info)
Use database links to duplicate the data using SQL
You can use the Database Configuration Assistant to generate a template from your production database. This will give you all the parameters and tablespaces, among other things. You will need to tweak the configuration somewhat; for instance the file paths may be wrong, and some parameters may need downsizing. You can then feed that template into DBCA to clone the database on you Linux machine.
To get the schemas and data you should use Data Pump (rather than the older Import / Export utlities). This can be run off the command line or from PL/SQL.
Bear in mind that using production data in a development or test environment can cause you to run foul of data protection laws and other compliance issues. It depends on what your application does and what jurisdiction you operate under. But if your production system contains citizens' personal data you need to be very careful. There are products out there which will apply masking as part of a data import process (Oracle sells one) but they tend to be expensive. Rolling your own masking product can be tricky: if this applies to your situation be sure to get your compliance staff (legal team) involved early.
I would suggest you install Oracle XE which is free to use on your local if your development is not something that is related to core database features. You can then use the methods given above to pump data into Oracle XE and compile your code on it, though for development I don't think you would need data as much as that in production.

Resources