Export Parameters :
Parameter
|
Description
|
abort_step
|
Undocumented feature
|
access_method
|
Data Access Method – default is Automatic
|
attach
|
Attach to existing job – no default
|
cluster
|
Start workers across cluster; default is YES
|
compression
|
Content to export: default is METADATA_ONLY
|
content
|
Content to export: default is ALL
|
current_edition
|
Current edition: ORA$BASE is the default
|
data_options
|
Export data layer options
|
directory
|
Default directory specification
|
dumpfile
|
dumpfile names: format is (file1,…) default is expdat.dmp
|
encryption
|
Encryption type to be used: default varies
|
encryption_algorithm
|
Encryption algorithm to be used: default is AES128
|
encryption_mode
|
Encryption mode to be used: default varies
|
encryption_password
|
Encryption key to be used
|
estimate
|
Calculate size estimate: default is BLOCKS
|
estimate_only
|
Only estimate the length of the job: default is N
|
exclude
|
Export exclude option: no default
|
filesize
|
file size: the size of export dump files
|
flashback_time
|
database time to be used for flashback export: no default
|
flashback_scn
|
system change number to be used for flashback export: no default
|
full
|
indicates a full mode export
|
include
|
export include option: no default
|
ip_address
|
IP Address for PLSQL debugger
|
help
|
help: display description on export parameters, default is N
|
job_name
|
Job Name: no default
|
keep_master
|
keep_master: Retain job table upon completion
|
log_entry
|
logentry
|
logfile
|
log export messages to specified file
|
metrics
|
Enable/disable object metrics reporting
|
mp_enable
|
Enable/disable multi-processing for current session
|
network_link
|
Network mode export
|
nologfile
|
No export log file created
|
package_load
|
Specify how to load PL/SQL objects
|
parallel
|
Degree of Parallelism: default is 1
|
parallel_threshold
|
Degree of DML Parallelism
|
parfile
|
parameter file: name of file that contains parameter specifications
|
query
|
query used to select a subset of rows for a table
|
remap_data
|
Transform data in user tables
|
reuse_dumpfiles
|
reuse_dumpfiles: reuse existing dump files; default is No
|
sample
|
Specify percentage of data to be sampled
|
schemas
|
schemas to export: format is ‘(schema1, .., schemaN)’
|
service_name
|
Service name that job will charge against
|
silent
|
silent: display information, default is NONE
|
status
|
Interval between status updates
|
tables
|
Tables to export: format is ‘(table1, table2, …, tableN)’
|
tablespaces
|
tablespaces to transport/recover: format is ‘(ts1,…, tsN)’
|
trace
|
Trace option: enable sql_trace and timed_stat, default is 0
|
transport_full_check
|
TTS perform test for objects in recovery set: default is N
|
transportable
|
Use transportable data movement: default is NEVER
|
transport_tablespaces
|
Transportable tablespace option: default is N
|
tts_closure_check
|
Enable/disable transportable containment check: def is Y
|
userid
|
user/password to connect to oracle: no default
|
version
|
Job version: Compatible is the default
|
Import Parameters :
Parameter
|
Description
|
abort_step
|
Undocumented feature
|
access_method
|
Data Access Method – default is Automatic
|
attach
|
Attach to existing job – no default
|
cluster
|
Start workers across cluster; default is Y
|
content
|
Content to import: default is ALL
|
data_options
|
Import data layer options
|
current_edition
|
Applications edition to be used on local database
|
directory
|
Default directory specification
|
dumper_directory
|
Directory for stream dumper
|
dumpfile
|
import dumpfile names: format is (file1, file2…)
|
encryption_password
|
Encryption key to be used
|
estimate
|
Calculate size estimate: default is BLOCKS
|
exclude
|
Import exclude option: no default
|
flashback_scn
|
system change number to be used for flashback import: no default
|
flashback_time
|
database time to be used for flashback import: no default
|
full
|
indicates a full Mode import
|
help
|
help: display description of import parameters, default is N
|
include
|
import include option: no default
|
ip_address
|
IP Address for PLSQL debugger
|
job_name
|
Job Name: no default)’
|
keep_master
|
keep_master: Retain job table upon completion
|
logfile
|
log import messages to specified file
|
master_only
|
only import the master table associated with this job
|
metrics
|
Enable/disable object metrics reporting
|
mp_enable
|
Enable/disable multi-processing for current session
|
network_link
|
Network mode import
|
nologfile
|
No import log file created
|
package_load
|
Specify how to load PL/SQL objects
|
parallel
|
Degree of Parallelism: default is 1
|
parallel_threshold
|
Degree of DML Parallelism
|
parfile
|
parameter file: name of file that contains parameter specifications
|
partition_options
|
Determine how partitions should be handle: Default is NONE
|
query
|
query used to select a subset of rows for a table
|
remap_data
|
Transform data in user tables
|
remap_datafile
|
Change the name of the source datafile
|
remap_schema
|
Remap source schema objects to new schema
|
remap_table
|
Remap tables to a different name
|
remap_tablespace
|
Remap objects to different tablespace
|
reuse_datafiles
|
Re-initialize existing datafiles (replaces DESTROY)
|
schemas
|
schemas to import: format is ‘(schema1, …, schemaN)’
|
service_name
|
Service name that job will charge against
|
silent
|
silent: display information, default is NONE
|
skip_unusable_indexes
|
Skip indexes which are in the unsed state)
|
source_edition
|
Applications edition to be used on remote database
|
sqlfile
|
Write appropriate SQL DDL to specified file
|
status
|
Interval between status updates
|
streams_configuration
|
import streams configuration metadata
|
table_exists_action
|
Action taken if the table to import already exists
|
tables
|
Tables to import: format is ‘(table1, table2, …, tableN)
|
tablespaces
|
tablespaces to transport: format is ‘(ts1,…, tsN)’
|
trace
|
Trace option: enable sql_trace and timed_stat, default is 0
|
transform
|
Metadata_transforms
|
transportable
|
Use transportable data movement: default is NEVER
|
transport_datafiles
|
List of datafiles to be plugged into target system
|
transport_tablespaces
|
Transportable tablespace option: default is N
|
transport_full_check
|
Verify that Tablespaces to be used do not have dependencies
|
tts_closure_check
|
Enable/disable transportable containment check: def is Y
|
userid
|
user/password to connect to oracle: no default
|
version
|
Job version: Compatible is the default
|
The Following commands are valid while in interactive mode.
Command Description
-------------------- ----------------------------------------------------------
ADD_FILE Add dumpfile to dumpfile set.
CONTINUE_CLIENT Return to logging mode. Job will be re-started if idle.
EXIT_CLIENT Quit client session and leave job running.
FILESIZE Default filesize (bytes) for subsequent ADD_FILE commands.
HELP Summarize interactive commands.
KILL_JOB Detach and delete job.
PARALLEL Change the number of active workers for current job.
START_JOB Start/resume current job.
STATUS Frequency (secs) job status is to be monitored where the default (0) will show new status when available.
STATUS[=interval]
STOP_JOB Orderly shutdown of job execution and exits the client.
STOP_JOB=IMMEDIATE performs an immediate shutdown of the Data Pump job.
2 comments:
Very useful...Thanks !
Thanks Cyber
Post a Comment