Friday, February 3, 2012

11g data pump parameters


Export Parameters :

Parameter
Description
abort_step
Undocumented feature
access_method
Data Access Method – default is Automatic
attach
Attach to existing job – no default
cluster
Start workers across cluster; default is YES
compression
Content to export: default is METADATA_ONLY
content
Content to export: default is ALL
current_edition
Current edition: ORA$BASE is the default
data_options
Export data layer options
directory
Default directory specification
dumpfile
dumpfile names: format is (file1,…) default is expdat.dmp
encryption
Encryption type to be used: default varies
encryption_algorithm
Encryption algorithm to be used: default is AES128
encryption_mode
Encryption mode to be used: default varies
encryption_password
Encryption key to be used
estimate
Calculate size estimate: default is BLOCKS
estimate_only
Only estimate the length of the job: default is N
exclude
Export exclude option: no default
filesize
file size: the size of export dump files
flashback_time
database time to be used for flashback export: no default
flashback_scn
system change number to be used for flashback export: no default
full
indicates a full mode export
include
export include option: no default
ip_address
IP Address for PLSQL debugger
help
help: display description on export parameters, default is N
job_name
Job Name: no default
keep_master
keep_master: Retain job table upon completion
log_entry
logentry
logfile
log export messages to specified file
metrics
Enable/disable object metrics reporting
mp_enable
Enable/disable multi-processing for current session
network_link
Network mode export
nologfile
No export log file created
package_load
Specify how to load PL/SQL objects
parallel
Degree of Parallelism: default is 1
parallel_threshold
Degree of DML Parallelism
parfile
parameter file: name of file that contains parameter specifications
query
query used to select a subset of rows for a table
remap_data
Transform data in user tables
reuse_dumpfiles
reuse_dumpfiles: reuse existing dump files; default is No
sample
Specify percentage of data to be sampled
schemas
schemas to export: format is ‘(schema1, .., schemaN)’
service_name
Service name that job will charge against
silent
silent: display information, default is NONE
status
Interval between status updates
tables
Tables to export: format is ‘(table1, table2, …, tableN)’
tablespaces
tablespaces to transport/recover: format is ‘(ts1,…, tsN)’
trace
Trace option: enable sql_trace and timed_stat, default is 0
transport_full_check
TTS perform test for objects in recovery set: default is N
transportable
Use transportable data movement: default is NEVER
transport_tablespaces
Transportable tablespace option: default is N
tts_closure_check
Enable/disable transportable containment check: def is Y
userid
user/password to connect to oracle: no default
version
Job version: Compatible is the default



Import Parameters :

Parameter
Description
abort_step
Undocumented feature
access_method
Data Access Method – default is Automatic
attach
Attach to existing job – no default
cluster
Start workers across cluster; default is Y
content
Content to import: default is ALL
data_options
Import data layer options
current_edition
Applications edition to be used on local database
directory
Default directory specification
dumper_directory
Directory for stream dumper
dumpfile
import dumpfile names: format is (file1, file2…)
encryption_password
Encryption key to be used
estimate
Calculate size estimate: default is BLOCKS
exclude
Import exclude option: no default
flashback_scn
system change number to be used for flashback import: no default
flashback_time
database time to be used for flashback import: no default
full
indicates a full Mode import
help
help: display description of import parameters, default is N
include
import include option: no default
ip_address
IP Address for PLSQL debugger
job_name
Job Name: no default)’
keep_master
keep_master: Retain job table upon completion
logfile
log import messages to specified file
master_only
only import the master table associated with this job
metrics
Enable/disable object metrics reporting
mp_enable
Enable/disable multi-processing for current session
network_link
Network mode import
nologfile
No import log file created
package_load
Specify how to load PL/SQL objects
parallel
Degree of Parallelism: default is 1
parallel_threshold
Degree of DML Parallelism
parfile
parameter file: name of file that contains parameter specifications
partition_options
Determine how partitions should be handle: Default is NONE
query
query used to select a subset of rows for a table
remap_data
Transform data in user tables
remap_datafile
Change the name of the source datafile
remap_schema
Remap source schema objects to new schema
remap_table
Remap tables to a different name
remap_tablespace
Remap objects to different tablespace
reuse_datafiles
Re-initialize existing datafiles (replaces DESTROY)
schemas
schemas to import: format is ‘(schema1, …, schemaN)’
service_name
Service name that job will charge against
silent
silent: display information, default is NONE
skip_unusable_indexes
Skip indexes which are in the unsed state)
source_edition
Applications edition to be used on remote database
sqlfile
Write appropriate SQL DDL to specified file
status
Interval between status updates
streams_configuration
import streams configuration metadata
table_exists_action
Action taken if the table to import already exists
tables
Tables to import: format is ‘(table1, table2, …, tableN)
tablespaces
tablespaces to transport: format is ‘(ts1,…, tsN)’
trace
Trace option: enable sql_trace and timed_stat, default is 0
transform
Metadata_transforms
transportable
Use transportable data movement: default is NEVER
transport_datafiles
List of datafiles to be plugged into target system
transport_tablespaces
Transportable tablespace option: default is N
transport_full_check
Verify that Tablespaces to be used do not have dependencies
tts_closure_check
Enable/disable transportable containment check: def is Y
userid
user/password to connect to oracle: no default
version
Job version: Compatible is the default


The Following commands are valid while in interactive mode.


Command                                           Description
--------------------                                   ----------------------------------------------------------
ADD_FILE                                                    Add dumpfile to dumpfile set.
CONTINUE_CLIENT                              Return to logging mode. Job will be re-started if idle.
EXIT_CLIENT                                          Quit client session and leave job running.
FILESIZE                                                  Default filesize (bytes) for subsequent ADD_FILE commands.
HELP                                                          Summarize interactive commands.
KILL_JOB                                                Detach and delete job.
PARALLEL                                              Change the number of active workers for current job.
START_JOB                                            Start/resume current job.
STATUS                                                   Frequency (secs) job status is to be monitored where  the default (0) will show new status when available.

                                                                 STATUS[=interval]

STOP_JOB                                      Orderly shutdown of job execution and exits the client.
                                                             STOP_JOB=IMMEDIATE performs an immediate shutdown of the Data Pump job.

2 comments:

Post a Comment

Auto Scroll Stop Scroll