0% found this document useful (0 votes)
11 views58 pages

Informatica Best Practices

Uploaded by

Wilson Castaneda
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
11 views58 pages

Informatica Best Practices

Uploaded by

Wilson Castaneda
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 58

TERADATA - INFORMATICA

Overview

Chris Ward
Big Data Services CoE
Senior Teradata Consultant
Agenda
• Introductions
• BDS/DI COE & Partner Alliance
• Features
> Teradata TPT and the API
> Push Down Optimization
> Metadata Manager
> Hadoop & Informatica BDE
• Tips From the Expert
> Design Considerations
> Loader Utilities / PDO Design Patterns
> Performance Tips
> Administration Tips
> Informatica Versions
• Next Steps

2 7/10/2014 Teradata Confidential


Introductions

• Chris Ward: Senior Consultant, BDS CoE Teradata

3 7/10/2014 Teradata Confidential


Agenda
• Introductions
• BDS/DI COE & Partner Alliance
• Features
> Teradata TPT and the API
> Push Down Optimization
> Metadata Manager
> Hadoop & Informatica BDE
• Tips From the Expert
> Design Considerations
> Loader Utilities / PDO Design Patterns
> Performance Tips
> Administration Tips
> Informatica Versions
• Next Steps

4 7/10/2014 Teradata Confidential


BDS EDM CoE Architecture Team Mission
Mission
Build Data Management expertise within the Big Data space and
become trusted advisors to our customers and the center of
expertise for DM within Teradata

Enterprise DI Tool Specific Architecture and Delivery


1. Tool specific expertise: INFA, IIS/DS, Talend
2. Capture, share and promote best practices
3. DM Architecture Delivery

Offer and Architecture Delivery


1. Engagements and offers focus on DM Optimization
2. Entry-point for TD Industries looking for DM/DI expertise
3. Pre-sales Assistance, Offer Development, New technology incubation

5 7/10/2014 Teradata Confidential


DI Center of Expertise
• Establish and Promote Teradata ETL Best Practices
> Incorporate 3rd Party ETL Tool Methodology into TSM (Teradata Solution
Methodology)

• Centralized Involvement on all Engagements


> Learn from other accounts - “We ran into this problem”
> Leverage innovation – “We solved it like this”
> Serve the client – “What are your ETL needs”

• Resource Management
> Centralized Talent Base
> Training (Formal, OJT, Mentoring, Cross Training)
> Right person, Right place, Right time

• 3rd Party Tool Relationship


> Direct line to Support
> Direct line to Engineering
> Quick Escalation when needed

6 7/10/2014 Teradata Confidential


Agenda
• Introductions
• BDS/DI COE & Partner Alliance
• Features
> Teradata TPT and the API
> Push Down Optimization
> Metadata Manager
> Hadoop & Informatica BDE
• Tips From the Expert
> Design Considerations
> Loader Utilities / PDO Design Patterns
> Performance Tips
> Administration Tips
> Informatica Versions
• Next Steps

7 7/10/2014 Teradata Confidential


Teradata Parallel Transporter

• Teradata’s premier load product


• Combines many load protocols (FastLoad, MultiLoad,
FastExport, TPump) into one tool
• Database load protocols are same, client code rewritten
• No learning curve for protocols – learn new features
• Most everything about load tools still applies
> Similar basic features, parameters, limitations (e.g., number of
concurrent load jobs), when to use, etc.
• Benefits: Performance and API for ETL tools
• Invoked with:
> Script interface or
> C++ API (Informatica)
• Parallel input streams for performance on load server
> Data and functional parallelism on the client box

8 7/10/2014 Teradata Confidential


Teradata Load/Unload Protocols & Products

• Load/Unload Protocols
> FastLoad Protocol – Bulk loading of empty tables
> MultiLoad Protocol – Bulk Insert, Update, Upsert, & Delete
> FastExport Protocol – Export data out of Teradata DB
> TPump – SQL application for continuous loading

• Load/Unload Products which use the load protocols


> Standalone Utilities
– FastLoad, MultiLoad, TPump, FastExport
– Separate tools & languages, script interface only
> JDBC
– FastLoad & FastExport protocols for pure Java applications
> Teradata Parallel Transporter
– Execute all load/unload protocols in one product
– Plug-in Operators: Load, Update, Stream, Export
– Provides C++ API to protocols for ISV partners

9 7/10/2014 Teradata Confidential


Teradata Parallel Transporter Architecture

Databases Files Message Queues

User Written Scripts Informatica Custom Programs

Script Parser Parallel Transporter API


TPT Infrastructure
Parallel
Data Source Operators Transporter
Components

Load Update Export Stream

Teradata Database
10 7/10/2014 Teradata Confidential
Before API – Integration with Scripting

5. Read
messages Message & Statistics File

Data –
Metadata Named
2. Write 4. Read 4. Write
Data Pipe Data Msgs.

Data Source Teradata


Informatica FastLoad
3. Invoke Utility 4. Load Database
Data
1. Build &
Write Script 4. Read Script

FastLoad Script File

1. INFA PowerCenter/PowerExchange creates Teradata utility script & writes to file.


2. INFA PowerCenter/PowerExchange reads source data & writes to intermediate
file (lowers performance to land data in intermediate file).
3. INFA invokes Teradata utility (Teradata doesn’t know INFA called it)
4. Teradata tool reads script, reads file, loads data, writes messages to file
5. INFA PowerCenter/PowerExchange reads messages file and searches for errors,
11 etc. 7/10/2014 Teradata Confidential
With API - Integration Example

Metadata Pass data & Load parameters

Load Teradata
Data Source Informatica Protocol
Functions
Load Data
Database

Return codes & error


messages passed

1. INFA PowerCenter/PowerExchange passes parameters to API


2. INFA PowerCenter/PowerExchange reads source data & passes data buffer
to API
3. Teradata tool loads data and passes return codes and messages back to caller

12 7/10/2014 Teradata Confidential


Teradata Parallel Transporter, API Interface

Data Sources
PowerCenter
Oracle, etc. Parallel Transporter
Flat file
API

Stream
Update
Export
Load

13 7/10/2014
Teradata Database
Teradata Confidential
Parallel Input Streams Using API

Source

PowerCenter
Launches Multiple
Instances that PowerCenter PowerCenter PowerCenter
Read/Transform
Data in Parallel
TPT API

Teradata Parallel

TPT Load
TPT Load

Instance
Instance
TPT Load
Instance

Transporter Reads
Parallel Streams
with Multiple
Instances Launched
Through API

14 7/10/2014
Teradata Database
Teradata Confidential
Why TPT- API
• Scalable increased performance
> Use the power of PowerCenter and TD parallel processing
> Multiple load stream instances from source through TD target
> No landing of data, just pass buffers in memory
> 64 bit instead of TTU 32 bit
• Provides Tight Integration with PowerCenter
> ETL Integration is faster and easier
> PowerCenter has control over the entire load process
– Errors and statistics returned programmatically
– Checkpoints controlled by PowerCenter

15 7/10/2014 Teradata Confidential


Connection Settings in Informatica for TPT

16 7/10/2014 Teradata Confidential


Session Source / Target Configuration

17 7/10/2014 Teradata Confidential


Sample Benchmarks (2 column, fixed width data file)

DISCALMER: All timings are relative and will vary depending on


database / Informatica / network hardware

104.8M 10M

Informatica TPT API LOAD 4 min 58 secs 34 sec

Informatica ODBC Loading 1 hr 30 mins + 9 min 14 secs

TPT Script LOAD 4 mins 28 secs 38 secs

Informatica TPT API Export 1 min 45 secs 16 sec

Informatica ODBC Exporting 7 min 24 secs 44 sec

TPT Script Export 2 min 8 secs 16 sec

18 7/10/2014 Teradata Confidential


DI CoE Best Practice

• Read Source
• Load to Stage table via TPT API
• Perform basic transformations such as data type checks only
> Ie where field data type is not a string
– Load raw value into VARCHAR
– Test and convert to true data type and load into separate field

19 7/10/2014 Teradata Confidential


Agenda
• Introductions
• BDS/DI COE & Partner Alliance
• Features
> Teradata TPT and the API
> Push Down Optimization
> Metadata Manager
> Hadoop & Informatica BDE
• Tips From the Expert
> Design Considerations
> Loader Utilities / PDO Design Patterns
> Performance Tips
> Administration Tips
> Informatica Versions
• Next Steps

20 7/10/2014 Teradata Confidential


What is Pushdown Optimization (PDO)?

Repository Repository Data Server


Server ETL/ELT processing
can be combined

ETL ETL Pushdown


Instructions Instructions Processing

Staging Warehouse

Push transformation processing from


Source Target
Database PowerCenter engine to relational database Database

21 7/10/2014 Teradata Confidential


Page 21
A Taxonomy of Data Integration Techniques

There are three main approaches:


1. ETL Approach: (1) Extract from the source systems, (2)
Transform inside the Informatica engine on integration engine
servers, and (3) Load into target tables in the data warehouse.
2. ELT Approach: (1) Extract from the source systems, (2) Load into
staging tables inside the data warehouse RDBMS servers, and (3)
Transform inside the RDBMS engine using generated SQL with a
final insert into the target tables in the data warehouse.
3. Hybrid ETLT Approach: (1) Extract from the source systems, (2)
Transform inside the Informatica engine on integration engine
servers, (3) Load into staging tables in the data warehouse, and
(4) apply further Transformations inside the RDBMS engine using
generated SQL with a final insert into the target tables in the data
warehouse.

22 7/10/2014 Teradata Confidential


ETL versus ELT

Q: Which approach is best?


A: It depends.

23 7/10/2014 Teradata Confidential


ETL versus ELT

When does ETL win?


• Ordered transformations not well suited to set processing.
• Integration of third party software tools best managed by
Informatica outside of the RDBMS (e.g., name and address
standardization utilities).
• Maximize in-memory execution for multiple step transformations
that do not require access to large volumes of historical or lookup
data (note: caching plays a role).
• Streaming data loads using message-based feeds with “real-time”
data acquisition.

24 7/10/2014 Teradata Confidential


ETL versus ELT

When does ELT win?


• Leverage of high performance DW platform for execution reduces
capacity requirements on ETL servers - this is especially useful
when peak requirements for data integration are in a different
window than peak requirements for data warehouse analytics.
• Significantly reduce data retrieval overhead for transformations
that require access to historical data or large cardinality lookup
data already in the data warehouse.
• Batch or mini-batch loads with reasonably large data sets,
especially with pre-existing indices that may be leveraged for
processing.
• Optimize performance for large scale operations that are well
suited for set operations such as complex joins and large
cardinality aggregations.

25 7/10/2014 Teradata Confidential


ETL versus ELT

The Answer: Hybrid ETLT


• Best of both worlds!
• Use the appropriate combination of ETL and ELT based on
transformation characteristics.
• Informatica creates and defines metadata to drive transformations
independent of the execution engine.
• ETL versus ELT execution is selectable within Informatica and can
be switched back-and-forth without the redefinition of
transformation rules.

Note that not all transformation rules are supported with pushdown
optimization (ELT) in the current release of Informatica 8.x.

26 7/10/2014 Teradata Confidential


Why PDO

• Cost-effectively scale by using a flexible, adaptable DI architecture


> Do the work in the “appropriate” environment

• Increase developer and team productivity


> All code in one development tool
> Developers know one tool

• Eliminate the need to write custom-coded solutions


> No longer need to develop Bteq Scripts

• Easily adapt to changes in Teradata Database


> Future versions will adapt to the new TD functionality

27 7/10/2014 Teradata Confidential


Why PDO

• Improved traceability through metadata management


> Better understanding and estimation on Impact of change
> From source systems to EDW to BI reporting
> On the contrary, Bteq does not offer metadata

• Delivered superior performance


> Data never leaves Teradata
> Heavy transformation from stage to work layers

• Meet SLA
> Well-positioned to make data available on time

• Increased efficiency in project delivery (client example)


> Templated approach for pushdown development
> Reduced maintenance headaches
> Estimated 25% time saved using pushdown with templated approach
> Test 15% more ELTs per release

28 7/10/2014 Teradata Confidential


DI CoE Delta Design Pattern
• Perform Business Rules
> FK joins
> Transformations
> Business Rules

• Perform Delta Check


> Join source to target only
> Filter None Changed Rows

• Update Existing Rows


> set effective end dt

• Insert New & Changed Rows


> Set effective start / end dt

29 7/10/2014 Teradata Confidential


Delta PDO Informatica Implementation
• Delta

• Update

• Insert

30 7/10/2014 Teradata Confidential


PDO Comparison from a Client

• Used for comparison only


• 10 Million rows copied from table to table

Connection Type Elapsed Time


ODBC 6hrs +

PDO 3 sec

31 7/10/2014 Teradata Confidential


Page 31
Database Considerations - Example

Pushdown Informatica Informatica -


[Database 1partition
partitioning-10
partitions]
Well Tuned 0:0:54 0:2:48 0:18:50
Database(Non-
Skewed PI )
With Index No 0:13:45 0:4:38 0:18:50
parallel [Skew PI]

Well Tuned 0:1:43 0:3:11 0:20:25


Database with no
Index
Un Tuned 0:28:18 0:4:50 0:39:10
Database(worst
case scenario- No
Indexes)

32 7/10/2014 Teradata Confidential


Agenda
• Introductions
• BDS/DI COE & Partner Alliance
• Features
> Teradata TPT and the API
> Push Down Optimization
> Metadata Manager
> Hadoop & Informatica BDE
• Tips From the Expert
> Design Considerations
> Loader Utilities / PDO Design Patterns
> Performance Tips
> Administration Tips
> Informatica Versions
• Next Steps

33 7/10/2014 Teradata Confidential


PowerCenter AE Metadata Manager
What problems does it solve?

 Lack confidence on the data


presented to business users
 Find it hard to relate business
concepts to technical artifacts
Analyst /Subject  Need to collaborate with developers
Matter Expert for greater productivity

 Too much time spent on impact


analysis vs. actual work
 Too much time spent trying to locate
technical artifacts and recreating
what already exists
Developer

 Costs associated with


managing disparate systems
generating metadata

Architect

34 7/10/2014 Teradata Confidential


PowerCenter Advanced Edition
Improve IT Responsiveness

Comprehensive impact analysis


delivers greater IT responsiveness
• Analyze detailed dependencies
amongst data assets and determine
where they are used
• Enables assessment and
management of change for faster
data integration project delivery
• Delivers a robust integration
metadata catalog, with direct
connectivity to enterprise
applications and mainframe systems

35 7/10/2014 Teradata Confidential


PowerCenter Advanced Edition
Deliver Trusted Data

Personalized lineage increases


insight into complex data
relationships
• Intuitive, dynamic data lineage
enables users to understand what
the data means, where it came from,
where it’s going, and how it’s
changed
• Integrated profiling results for
additional insight into data
• Highlight, filtering and
personalization capabilities improve
usability

36 7/10/2014 Teradata Confidential


Agenda
• Introductions
• BDS/DI COE & Partner Alliance
• Features
> Teradata TPT and the API
> Push Down Optimization
> Metadata Manager
> Hadoop & Informatica BDE
• Tips From the Expert
> Design Considerations
> Loader Utilities / PDO Design Patterns
> Performance Tips
> Administration Tips
> Informatica Versions
• Next Steps

37 7/10/2014 Teradata Confidential


Visual Development Environment

Hadoop & Informatica BDE

Informatica BDE

SOURCE
DATA Profile
Profile
INTERFACE
Parse
ETL
TPT API
Databases Database
Cleanse
Match TDHC
Files
HIVE (HiveQL and UDFs) JDBC

INTERFACE
Servers & MAPREDUCE
Sql-H
Mainframe LOAD Sql-H
YARN Teradata
Informatica
JMS Queue’s Batch INTERFACE
HDFS INTERFACE
HDFS API

Replicate HDFS API Profile


Profile TPT API
Social Stream
Parse
Archive
HIVE ELT/ETL ODBC
Cleanse
Sensor data AMBARI PDO
JDBC Match

38 7/10/2014 Teradata Confidential


Hadoop & Informatica BDE

• Allows the creation of Mappings that can be ‘push down’ to


be fully executed in Hadoop

• Uses Hive + Custom Extensions.

• Loads directly from Hadoop via TPT API / TDHC directly to


Teradata
• Development Benefits • Enterprise System Benefits
• Utilize ETL ease-of-use • GUI Driven design
• Leverage existing code, developers, • Maintains Metadata
and tools • Data Lineage
• Single code-base to maintain • Integration to external access points
• Limit cross-training • Consistent data types
• Automatic optimizations • Impact analysis
• Automatic support for future • Security, Governance, and Scalability
versions

39 7/10/2014 Teradata Confidential


Platform Considerations
In Data
On ETL Server In Hadoop
Warehouse

Transformation
Row-by-Row Set-based Row-by-Row
Strength

Transformation Sorting and Set-


Set-based Row-by-Row
Weakness based

Data Processing Real-time data


Batch processing Batch processing
Latency integration

Transient, non-
Main Data Storage Raw data or
persistent, non- High value data
Type infrequently used
critical

Robust security Robust security


Security of data Evolving
controls controls

Data Landing
Data Integration Reporting and Zone, Discovery
Primary Usage
Development Analytics Process,
Scalable Real-time

40 7/10/2014 Teradata Confidential


Agenda
• Introductions
• BDS/DI COE & Partner Alliance
• Features
> Teradata TPT and the API
> Push Down Optimization
> Metadata Manager
> Hadoop & Informatica BDE
• Tips From the Expert
> Design Considerations
> Loader Utilities / PDO Design Patterns
> Performance Tips
> Administration Tips
> Informatica Versions
• Next Steps

41 7/10/2014 Teradata Confidential


General Considerations
• Documentation
> Make every object self documenting. Metadata.
• Parameterization
> Use parameter files and hard code as little as possible
• Profiling
> Profile the data before you begin development
• Data Quality
> Define Data Quality Strategy early on

• Mapping Design Considerations


> Minimize the number of transformations. Don’t use 2 transformations
when 1 will do. Keep it simple.
> Avoid have multiple threads in same mapping
> Use Global parameters in parameter files.
> Where files used many times, consider staging in database first.

42 7/10/2014 Teradata Confidential


Mapping Design Considerations
• Reusable Transformations
> For transformations used more than once. I.e. Common
Dimension Lookups, String to Date expressions.
> Put in Common folder.

• Reusable Mapplets
> For groups of transformations that are commonly used together,
make them reusable. I.e. Data quality rules where data types
are converted, then lookup performed and result return in
standard format.
> Put in Common Folder.

• Source / Targets
> Use Shortcuts and put in common Folder

43 7/10/2014 Teradata Confidential


Loader Utilities
• Teradata External Loaders (Old Method)
> Informatica can write out to a named pipe or staged file and
calls ‘FASTLOAD, MLOAD, TPUMP’ utilities.
> NOTE: Partitioning and parallelization not supported when calling
external loader executable.
> Informatica treats target has to be treat as a flat file. Dates have
to written as a particular string format.
• Teradata Parallel Transporter (TPT) API
> Teradata Parallel Transporter API integrates Power Center and
Teradata Parallel Transporter API (Teradata PT API) for data
extraction and loading in parallel.
> Mappings are designed and developed exactly the same way as
any other relational data database tables.
> Can Load / Extract data from Teradata in Parallel

44 7/10/2014 Teradata Confidential


Loader Utilities (continued)

The following matrix shows under what conditions the loader


utilities can be used.

Indexes

Non Rows Requires


Referential Empty <= Viewable Loader NPI
UTILTITY Join USI NUSI Integrity Triggers Table 100000 Immediately Session Table
LOAD /
x x
FASTLOAD
UPDATE /
x x x
MLOAD
STREAM /
x x x x x x x x x
TPUMP
ODBC x x x x x x x x x
EXPORT/
FAST x x x x x x N/A x x
EXPORT

45 7/10/2014 Teradata Confidential


Teradata Specific Tips
> Load into stage table using TPTAPI.
> During load, convert and cast fields to the correct data type.
Minimize cast/convert functions later specifically for columns
which will be joined on.
> Use Pushdown optimization to move from Stage to Target.

46 7/10/2014 Teradata Confidential


Design Pattern - Parallel lookups:

–A mapping
containing parallel
lookups cannot be
pushed to the
database Convert To
–Design the
mapping and make
the lookups
sequential.

47 7/10/2014 Teradata Confidential


47
Design Pattern – Unconnected lookups:

–A mapping containing
unconnected lookups
might perform slower
than the integration
service’s DTM when
utilizing PDO.
–The Cause is SQL Convert To
generated by the
integration service can
be complex and slow as
unconnected lookups
are converted into outer
joins.
–Compensate by
connecting the lookups
whenever possible.

48 7/10/2014 Teradata Confidential


48
Design Pattern – Sorted Aggregation

–Mappings containing
an Aggregator
downstream from a
Sorter transformation
can not utilize
pushdown
- Handle this by Convert To
redesigning the
mapping to achieve full
or source-side
pushdown optimization,
configure the
Aggregator
transformation so that it
does not use sorted
input, and remove the
Sorter transformation

49 7/10/2014 Teradata Confidential


49
Use of Filter transformations

50 7/10/2014 Teradata Confidential


Design Pattern – Variable Ports Example

Pushdown
Optimization is not
possible for this
mapping because
variable ports are not
supported. Consider
replacing the
following
(NET_AMOUNT =
AMOUNT – FEE,
DOLLAR_AMT =
NET_AMOUNT *
RATE) with
(DOLLAR_AMT =
(AMOUNT – FEE) *
RATE)

51 7/10/2014 Teradata Confidential


Supported Transformations

Transformation Full Source Target


Aggregator Yes Yes No
Expression Yes1 Yes1 Yes1
Filter Yes Yes No
Joiner Yes Yes No
Lookup Yes Yes Yes2
Unconnected Lookup Yes Yes Yes
Router Yes Yes No
Sequence Generator Yes2 Yes2 Yes2
Sorter Yes Yes No
Union Yes Yes No
Update Strategy Yes No No

1 PowerCenter expressions can be pushed down only if there is an equivalent database function
2 Not all databases are supported, Refer to documentation for further details

52 7/10/2014 Teradata Confidential


Informatica-Teradata Compatibility

INFA TPT-API PDO

PC 9.6 TTU14.10 (linux/win) TTU14.10 (linux/win)


GA – Feb ’14 TTU 13.10 (others) TTU 13.10 (others)

PC 9.5.1 TTU14.0 (linux/win) TTU14.0 (linux/win)


GA – Jan ’13 TTU 13.10 TTU 13.10

PC 9.5 TTU13.10 TTU 13.10


GA – Jul ‘12

PC 9.1 TTU13.10, TTU 13.0, TTU 13.10, TTU13.0,


GA – Jun ‘11 TTU12 TTU12

PC 9.0.1 TTU13.10, TTU 13.0, TTU 13.10, TTU13.0,


GA – June ‘10 TTU12 TTU12

PC 9.0.0 TTU13.0, TTU12 TTU12


(no upgrade from
8.6.1)

GA - Nov ’09

53 7/10/2014 Teradata Confidential


Performance Tips
• Design for performance. Performance Tuning is too late!

• Tuning is only really relevant against Production Like data volumes.

• Use the right time, right place, right tool. I.e. A database is
designed to do filters, joins, sorts and aggregations. Use the
appropriate tool at the right moment.

• Find your bottlenecks and make a single change at a time and re-
test. A change can just move the bottleneck elsewhere.

• Ordered approach for Performance Tuning: Target, Source,


Mapping/Transformation, Session, System

54 7/10/2014 Teradata Confidential


Operational Reporting
> Define a methodology of monitoring the ‘batch’ earlier on.
Performance issue can be detected earlier on if there is a way to
capture and reporting ‘batch’ history
> Develop standard reusable scripts/Informatica components for
Operational reporting.
350000

300000

250000

200000

150000
Total Batch Time
100000
Total Session Elasped Time
50000

0
00

00

00
00

00

00

00

00

00

00

00

00

00
:0

:0

:0
:0
0:

0:

0:
0:

0:

0:

0:

0:

0:

0:

0:

0:

0:
00

00

00
0
0

10

10
0

0
01

1
1

01

01

01

01

01

01

01

01

01

01
20

20

20

20

20
20
2

/2

/2

/2

/2

/2

/2

/2

/2

/2

/2
1/

3/

2/

2/

1/

1/
/
11

21

13

23

12

22

12

22

11

21

11
2/

3/

4/

5/

6/

7/
2/

2/

3/

3/

4/

4/

5/

5/

6/

6/

7/
> Monitor environments all i.e. Informatica, databases etc.
Common reporting methodology to be able to look at what was
the performance like across all systems at a point in time.

55 7/10/2014 Teradata Confidential


Informatica Administration/Setup
• Use 64-bit Versions.
• Install system monitoring tools on all environments that
provide history on I/O (local disk and SAN statistics),
Memory, CPU, and Network.
• Create a Standard ‘heart beat’ sessions to benchmark
performance to all connected systems when something
changes on the system changes and run regularly.
• For connections, use Native drivers over ODBC.
• Consider Grid deployments.
• Tune network connectivity.
• Tune I/O. Consider Solid State disk or fast disks for Cache
directories.
• Add as much memory as possible. Less paging to disk, the
better.

56 7/10/2014 Teradata Confidential


Informatica TPT Install

Informatica Server
1. Install Teradata Database PowerCentre/Exchange
2. Install TTU On Informatica Server
3. Test TTU (TPT API Load / ODBC)
4. Install Informatica PowerCenter PowerCenter/Exchange TPT API
5. Install Informatica TPT API module Module

Teradata Tools & Utilities

Teradata
Database

57 7/10/2014 Teradata Confidential


Thank you
Chris Ward
Teradata Senior Consultant
[email protected]
312.543.9284

58 7/10/2014 Teradata Confidential

You might also like