Teradata stage
|
Stage type
|
Usage guidelines
|
Parallel
read
|
Teradata
utility limit
|
Teradata
Enterprise
|
Native
Parallel
|
1)Writing a large number of rows in parallel
2)Supports OPEN and CLOSE commands
3)Limited to INSERT (new table) or
APPEND (existing table)
4)Subject to the limits of Teradata FastLoad
(but also supports APPEND)
5) Locks the target table in exclusive mode
|
Yes
|
applies
|
Teradata
MultiLoad
(MultiLoad utility)
|
Plug-In
|
1)Insert, Update, Delete, Upsert of moderate
data volumes
2)Locks the target tables in exclusive mode
|
No
|
applies
|
Teradata
MultiLoad
(MultiLoad utility)
|
Plug-In
|
1)Insert, Update, Delete, Upsert of moderate
data volumes
2)Locks the target tables in exclusive mode
|
No
|
applies
|
Teradata
MultiLoad
(TPump utility)
|
Plug-In
|
1) Insert, Update, Delete, Upsert of small
volumes of data in a large database
2)Does not lock the target tables
3)Must not be run in parallel, because each
node and use counts toward system-wide
Teradata utility limit
|
No
|
applies
|
Teradata API
|
Plug-In
|
1) Insert, Update, Delete, Upsert of small
volumes of data
2)Allows concurrent writes (does not lock
target)
3)Slower than TPump for equivalent
operations
|
Yes
|
none
|
Source Teradata stages
Teradata stage
|
Stage type
|
Usage guidelines
|
Parallel
read
|
Teradata
utility limit
|
Teradata
Enterprise
|
Native
Parallel
|
1)Reading a large number of rows in
parallel
2)Supports OPEN and CLOSE
commands
3)Subject to the limits of Teradata
FastExport
|
Yes
|
applies
|
Teradata API
|
Plug-In
|
Reading a small number of rows
sequentially
|
No
|
none
|
Comparison with existing Teradata stages
Stages related to Teradata
Teradata TPUMP stage
ISTOOL for EXPORT IMPORT Datastage Components
This post help you to understand export and import of DataStage jobs and other components/assets from command line.
Use istool to export and import DataStage jobs and other Information Server components using UNIX command line.
Location of command:
UNIX: /opt/IBM/InformationServer/Clients/istools/cli
Windows: IBMInformationServerClientsistoolscli
Export Datastage Components From Command Line
—————————————————————–
Complete Project Export :
cd /opt/IBM/InformationServer/Clients/istools/cli
./istool export -dom XYZ123:9080 -u dsadm -p dsadm -ar /tmp/Test1.isx -ds XYZ123/Test1/*/*.*
JobWise Export :
cd /opt/IBM/InformationServer/Clients/istools/cli
./istool export -dom XYZ123:9080 -u dsadm -p dsadm -ar /tmp/Test1.isx -ds XYZ123/Test1/Jobs/TestJob.pjb
Syntax : ./istool -dom [domain]:9080 -u [user] -p [password] -ar [Path/ExportFileName.isx] -ds [domain/ProjectName/Jobs/JobNameName.pjb] -[Options]
Import Datastage Components From Command Line
—————————————————————–
cd /opt/IBM/InformationServer/Clients/istools/cli
./istool import -dom XYZ123:9080 -u dsadm -p dsadm -ar /tmp/Test1.isx -ds XYZ123/Test1
Syntax : ./istool -dom [domain]:9080 -u [user] -p [password] -ar [Path/ExportFileName.isx] -ds [domain/ProjectName]
Check List Of Datastage Projects from Command Line:
——————————————————————–
DSHOME=`cat /.dshome`;
echo $DSHOME
DSHOME=`cat /.dshome`
cd $DSHOME
. ./dsenv
cd ./bin
./dsjob -lprojects => { For Listing the Datastage Projects}
./dsjob -ljobs project => { For Listing the Datastage Jobs in given Project}
Additional Information’s :
————————————————
Components Extension’s
——————————-
.pjb Extension : Datastage Job’s
.tbd Extension : Datastage Table Definitions
.isx Extension : You can import Individual Jobs/Whole Project with this extention
ISTOOL Exit codes:
————————
0 – Success
1 – Warning
2 – Partial failure
3 – Export failed
4 – Export preview failed
5 – Invalid archive file
10 – Failed to connect to Information Server domain
11 – Invalid command syntax
istool export [ Options ]
Possible values For: Options
Long Name: Short Name : Description:
-help , -h : print command usage
-domain , -dom : Name of Information Server domain
-username , -u : Name of user account on Information Server domain
-password , -p : Password for Information Server account
-verbose , -v : display progress
-silent , -s : silence command output
-archive , -ar : Archive file
-preview , -pre : Preview export task without executing
-abortIfError , -abort : Abort task on N errors
-updatearchive, -up : Update any existing Package file
-datastage , -ds : DataStage Assets
Options For : Exporting the DS Jobs.
-includedependent , -incdep : Include dependent items
-nodesign , -nodes : Exclude design items
-includeexecutable , -incexec : Include executable items
-base , -base : Prefix for Paths,DataStage item paths
Datastage Projects Using Command line to Create/Delete/View
This post will help you to understand Create/Delete and View of DataStage Projects from command line[Unix].
Note : Always use DSADM user to create or delete projects.
Check List Of Datastage Projects from Command Line:
——————————————————————–
DSHOME=`cat /.dshome`;
echo $DSHOME
DSHOME=`cat /.dshome`
cd $DSHOME
. ./dsenv
cd ./bin
./dsjob -lprojects =>{ For Listing the Datastage Projects}
./dsjob -ljobs project =>{ For Listing the Datastage Jobs in given Project}
Create Projects From Command Line
———————————————–
DSHOME=`cat /.dshome`
cd $DSHOME
. ./dsenv
cd bin
./dsadmin -domain XYZ1234 -user dsadm -password dsadm -server XYZ1234 -createproject Test1 -location /tmp/Test1
Syntax : ./dsadmin -domain [domain] -user [user] -password [password] -server [server] -createproject [project name] -location [prjoect path/NewProjectName];
Delete Projects From Command Line
———————————————–
DSHOME=`cat /.dshome`
cd $DSHOME
. ./dsenv
cd bin
./dsadmin -domain XYZ1234 -user dsadm -password dsadm -server XYZ1234 -deleteproject Test1
Syntax : ./dsadmin -domain [domain] -user [user] -password [password] -server [server] –
deleteproject [project name];
Datastage Course details – Syllabus
✔purpose of Data warehouse
✔DW Architecture
✔Data warehouse vs. OLTP Applications
✔Data Mart
✔Data warehouse Lifecycle
✔Entity Relationship(ER) Model
✔Fact and Dimension Tables
✔Logical Modeling
✔Physical Modeling
✔Schemas – Star Schema & Snowflake Schemas
✔Fact less Fact Tables
✔Types of available ETL Tools
✔Key tools in the ETL market
✔Oracle
✔.NET
✔Datastage ✔5X2 & 8x
✔Designer client
✔Director client
✔Import/export manager
✔Multi client manager
✔Console for IBM Information server
✔Web console for IBM information server
✔IBM Information server Architecture
✔IBM Data Quality Architecture
✔Enterprise Information Integration
✔Web Sphere Data stage Components
✔Partitioning Methods
✔Partitioning Techniques
✔Designer Canvas
✔Central Storage
✔Job Designing
✔Creating the Jobs
✔Compiling and Run the Jobs
✔Exporting and importing the jobs
✔Parameter passing
✔System(SMP) & Cluster system(MPP)
✔Importing Method(Flat file, Txt, Xls and Database files)
✔OSH Importing Method
✔Configuration file
✔Oracle Database
✔Dynamic RDBMS
✔ODBC
✔SQL Server
✔Teradata
FILE STAGES
✔Sequential File
✔Dataset
✔Lookup File set
✔Head
✔Tail
✔Row Generator
✔Column Generator
✔Slowly changing dimensions implementation
✔Aggregator
✔Copy
✔Compress
✔Expand
✔Filter
✔Modify
✔Sort
✔Switch
✔Lookup
✔Join
✔Marge
✔Change Capture
✔Change Apply
✔Compare
✔Difference
✔Funnel
✔Remove Duplicate
✔Surrogate Key Generator
✔Pivot stage
✔Transformer
✔Local Containers
✔Validation
✔Scheduling
✔Status
✔View logs
✔Monitoring
✔Suppress and Demote the Warnings
✔Peek view
✔Delete Project
✔Protect Project
✔Environmental variables
✔Auto purge
✔RCP
✔OSH
✔Commands Execute
✔Multiple Instances
✔Job Sequence Settings
✔Job sequencer
✔Start loop Activity
✔End loop Activity
✔Notification Activity
✔Terminator Activity
✔Nested Condition Activity
✔Exception handling Activity
✔Execute Command Activity
✔Wait for file Activity
✔User variable Activity
✔Adding Check Points
✔Restartable
✔Data Quality Stages
✔Investigate Stage
✔Standardize Stage
✔Match Frequency Stage
✔Reference Match Stage
✔Unduplicated Match Stage
✔Survive Stage
✔Viewing audit details in operational DASHBOARD reports