0% found this document useful (0 votes)
22 views11 pages

Vishal Kumar

Vishal Kumar is a seasoned IT professional with over 17 years of experience, specializing in Data Engineering, Data Warehousing, and Cloud Data Services, particularly in the Finance and Banking domains. He has extensive expertise in ETL processes using Informatica, SQL databases, and cloud technologies like Microsoft Azure, along with a strong background in regulatory projects for banking systems. His recent roles include leading data engineering teams and developing data solutions for major clients, demonstrating a commitment to delivering high-quality data management and analytics solutions.

Uploaded by

babjimad22
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
22 views11 pages

Vishal Kumar

Vishal Kumar is a seasoned IT professional with over 17 years of experience, specializing in Data Engineering, Data Warehousing, and Cloud Data Services, particularly in the Finance and Banking domains. He has extensive expertise in ETL processes using Informatica, SQL databases, and cloud technologies like Microsoft Azure, along with a strong background in regulatory projects for banking systems. His recent roles include leading data engineering teams and developing data solutions for major clients, demonstrating a commitment to delivering high-quality data management and analytics solutions.

Uploaded by

babjimad22
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
You are on page 1/ 11

Vishal Kumar

904-436-1221 / [email protected]

Job Objective:
 Aiming senior level assignments in Data Engineering, Data Warehouse, Cloud Data Services, Analytics / Business
Intelligence with an organization of high repute by associating myself with a professionally run business organization,
that has potential for growth - both organizationally and personally and provide ample opportunity to learn, improve
and implement

Professional Summary

 Over 17+ years’ experience into IT Industry and 14+ years of Data Warehouse experience with domain knowledge of
Finance/Banking, Insurance and Title Insurance / Real Estate Services domain.
 Extensive experience in ETL and Data Integration for developing ETL mappings and scripts using Informatica Power Center
10.x/9.x/8.x using Designer, Repository Manager, Workflow Manager & Workflow Monitor, Informatics (ICS/IICS).
Exposure on Power Exchange functionality.
 Extensive experience in creating Transformations and Mappings using Informatica Designer to implement business rules
and processing tasks using Workflow Manager to move data from Multiple Sources into Targets.
 Expertise using MS SQL Server 2008/2005/2000, oracle 12c/ 11g/10g, DB2 9.0/8.0/7.0, Teradata 12, Erwin 7.0/4.0/3.5,
SQL, PL/SQL, SQL *Plus, SQL *Loader.
 Cloud Technologies (Microsoft Azure), Databricks, Blob Storage, Creating Pipelines (Azure Data Factory, Pyspark,
SparkSql, Azure DevOps, CI/CD Implementation, Octopus Deployments
 Implementation of On-Premise Application to Cloud Instance of Azure
 Working Experience on Azure Data Factory, for Orchestrating the Pipelines and Schedules.
 Working on Python for Enhancement or New Development
 Working on Go Anywhere, with Different Capabilities like Processing Emails, Emails with Attachment, Processing CSV, Flat
Files, Excel and Database objects
 Extensively used Teradata SQL Assistant and BTEQ to write SQL’s for ETL.
 Data Quality Using IDQ (Informatica Developer Tool).
 Exposure to Informatica MDM
 Exposure and Ability to Learn New technologies required for Deliverables,
 Extensive Experience in Tuning Techniques related to Database (Oracle / Sql Server).
 Hands on experience on all phases of software development life cycle (SDLC) of Data warehousing, including requirement
gathering from business users, analysis, design, creation of mapping documents, specification writing, development,
testing, and implementation.
 Expertise in working with various transformations using Aggregator, Look Up, Update Strategy, Joiner, Filter, Sequence
Generator, Normalizer, Sorter, Router, Java transformations in Informatica Powercenter Designer.
 Extensive expertise in Debugging, Performance Tuning and Optimization of the ETL Sessions and Mappings.
 Proficiency in data warehousing techniques for Data Cleansing, Slowly Changing Dimension Phenomenon, Surrogate Key
assignment, Change Data Capture (CDC).
 Extensive Experience in integration of various RDBMS like Oracle, DB2, SQL server into staging area.
 Expertise in creating Stored Procedure/Functions, Triggers and Packages in Oracle and SQL Server MS
2005/2008/2014/2016.
 Expertise in Data Analysis, Data Design, Developing and Deployment for Data Integration for BI requirements.
 Profound knowledge of the principle of Data Warehouse using Fact tables, Dimensional Table, Star Schema Modeling and
Snowflake Schema Modeling.
 Experience in Administrative activities such as creating folders, assigning security permissions to users and creating
repositories, setting up groups/users’ permissions in Informatica Repository Manager and admin console.
 Exposure knowledge of Informatica Migration and Upgrade.
 Experience working with SDLC and implementing, Waterfall and Agile methodologies.
 Experience working with Informatica in UNIX and LINUX environments.
 Coordinate ETL and database development work with developers and conduct code reviews.
 Excellent oral/written/interpersonal communication skills, quick learner, willing to adapt to the team/organizational
environment.
 Experience in Regulatory / Federal projects for Banking System, CCAR (Comprehensive Capital Analysis and Review),
DFAST, Liquidity Coverage Ratio (LCR) and Risk Weighted Assets

Skill sets

ETL Tools : Informatica 10.x/9.X/8.X, Sap Data Services (BODS), Talend, Informatica Developer (IDQ),
SSIS, Azure Data Factory
ETL Management Console : Informatica Administration Console, Sap Data Services Management Console
Database : Oracle 10g/11g/12c, DB2, Sql Server 2005/2008/2014/2016, MySQL
NoSQL : Azure Cosmos DB
Database Tools, Utilities : DB2 Visualizers Developer, Toad, SqlPlus,GoAnywhere (Managed File Transfer)
Scripting Languages : Python, Perl, UNIX, Windows Batch Scripts
Procedure Language : PL/SQL, T-SQL
Cloud Technologies : Microsoft Azure,DataBricks, PySPARK, DataLake, DeltaLake, Informatica Cloud (ICS/IICS),
Spark Sql
Operation Systems : Windows / UNIX / Linux
Scheduling Tools : Tidal, Control-M, Redwood, Informatica Inbuilt Scheduler, CronTabs
Version Controller : Perforce (P4v), Informatica Versioning
Code Promotion Tools : Azure DevOps, CI/CD, Octopus, Jenkins, Nexus and Nolio, Informatica Deployments
Project Documentation : Confluence, Jira
QA Tools : HP Quality Center
Analytics Tools : Adobe Analytics (Datawarehouse Reports)

Education:
 Bachelor of Technology in Electronics and Electrical Engineering, Year 2004, Jawaharlal Nehru Technological
University.

Projects Summary

Client: Jones Lang LaSalle Americas (JLLT) Dec2017 to Till Date


Role:Sr Data Engineer

Roles & Responsibilities:

 Operate modern data architecture approaches to meet key business objectives and provide end-to-end data solutions
 Develop good understanding of how data will flow & stored through an organization across multiple applications such
Valuation Advisory Service, Supply Chain Management and Enterprise Data Management
 Design & develop data management and data persistence solutions for application use cases leveraging relational, non-
relational databases and enhancing our data processing capabilities
 Develop POCs to influence platform architects, product managers and software engineers to validate solution proposals
and migrate
 Lead team of data engineers and Team Managers to deliver data solutions in Agile manner while mentoring Team
Members
 Translates/interprets business strategy into objectives for the team. Prioritizes and assigns business initiatives and
performs high-level review of projects to ensure conformance to business requirements and delivery timelines
 Develop a highly engaged team through communication, performance management, career development plans and
reward programs
 Hire and develop talent pool establishing a good succession planning across functional areas managed
 Promote an environment that supports diversity and reflects JLL Brand and values
 Establishes goals, success criteria and performance measures to evaluate the outcome of the projects/initiatives
managed
 Promotes a culture of teamwork and collaborates across other functional verticals within the Center of expertise in
adopting/sharing of best practices B
 Build a culture committed to delivery of outstanding customer service and open, transparent team environments
 Develop and enhance the functional & Technical capabilities of the team by providing effective coaching &
facilitating/imparting the relevant technical and functional trainings

Environment:Informatica Power Center 10.x, Sql Server, Oracle 12c/11g, Toad, Sql Developer, Control-M
Azure Data Bricks, Pyspark, Spark Sql, Python, CI/CD, Azure Dev Ops, Azure Data Factory, Azure Data Lake, Azure Sql
Server

Client: Bank oftheWest , San Ram June 2017 to December 2017


Role: Sr ETL Developer

Project: KYC/KYA Enhancements for CDD


As part of KYC and KYA process of Customer Due Diligence (CDD), the requirement was to identify the customers from
different Lines Of business if they related to one customer. The Identification process was built on several Rules and Parses
to identify the Customer with Different ID’s and Information as one. The Effort also had to clean up the data in existing
database based on survivor rules and create a golden record for each customer. Also enhanced the existing PL/SQL
Packages and Procedures which was used by ESB Team (Enterprise Service Bus) between Frontend and Database for all
activities like Search, Insert, and Update.

Roles & Responsibilities:

 Implement and Enhance ETL programs using Informatica Power Center against the Oracle 11g Data Warehouse.
 Contribute to technical architectural planning, data modeling, process flow documentation and the design and
Development of innovative solutions for the business.
 Create technical designs and mapping specifications.
 PL/SQL Enhancements and Performance Tuning
 Work with both technical staff and business constituents to translate business requirements into technical solutions.
 Perform unit and system test/validation tasks.
 Maintain technical and user documentation.
 Provide on-call support of the Data Warehouse nightly processing
 Be an active participant in both technology and business initiatives.

Environment:
Informatica Power Center 9.x, Oracle 11g, Teradata, Flat Files, XML, CSV Files, Unix Shell Scripts, Toad, Sql Developer,
Windows XP, Tidal, Sql Plus

Client Silicon Valley Bank (SVB), Santa Clara, CA May 2016 to June 2017
Role: Sr ETL Developer

Data warehouse Integrator II / CCAR and Stress Testing


Project: CCAR and Stress Testing, Comprehensive Capital Analysis and Review.

CCAR stress testing came about after the economic collapse of 2008. The Dodd-Frank Act was enacted, which requires
both the Fed and all large bank holding companies (any banking organization with consolidated Average assets of $50
billion or more) to conduct and submit the results of the annual test - then disclose those results to the public.
Stress tests are submitted according to macroeconomic scenarios provided by the Fed and are used as protection to check
for whether a bank’s system could theoretically handle a future economic shock. They test the bank’s ability to remain
“well-capitalized,” with a minimum Tier 1 common equity ratio of 5 percent, even in a severely negative economic
scenario. That hypothetical future scenario is supplied annually by the Fed to the banks so they can submit their
responsive plans under the rules of the CCAR exam, every mandated bank must submit a “capital action plan” for the
following four quarters. The Fed then assesses that bank’s financial health and gives the bank a score. This score is not
simply a “pass/fail”; banks are assessed in terms of exactly how strong they are. The Fed looks at whether banking
organization have sufficient capital available to operate soundly even under extreme economic duress, and
whether they have what the Fed calls “a robust, forward-looking capital-planning process that accounts for their
unique risks.”
SVB, is Required to Submit the CCAR Annual Report by 1st Quarter of 2017, with the Specifications and Fed Edits
provided by Federal Reserve Bank.

Project: FRDS (Financial and Risk Datastore)

To meet the data and reporting requirements of CCAR and other Regulatory reporting like Risk Weighted Assets, LCR and
Projections, there was a need on enhancing the data on existing EDW Data, FRDS is the enhanced version of EDW which is
Built to consider it as single source of truth for all reporting requirements. The FRDS is built on to use the Materialized
Views which cover different subject areas like CreditLines and Loans, Customer Information, Loan Payments, Charge off
And Recovery

Roles and Responsibilities,

 Contribute to technical architectural planning, data modeling, process flow documentation and the design and
Development of innovative solutions for the business.
 Create technical designs and mapping specifications.
 Work with both technical staff and business constituents to translate business requirements into technical solutions.
 Perform unit and system test/validation tasks.
 Maintain technical and user documentation.
 Design and Code Reviews
 Provide on-call support of the Data Warehouse nightly processing
 Be an active participant in both technology and business initiatives.
 Performance Tuning of ETL and Oracle Queries.
 Lead team of 6 People across Onsite and Offshore Resources of Tata Consultancy Services.

Environment:
SAP BODS, Oracle 11g/12C, Flat Files, XML, CSV Files, Unix Shell Scripts, Toad, Sql Developer, Windows XP, Redwood,
SqlPlus, NOLIO, Jenkins, JIRA, Perforce, PL/SQL

Client :Bank ofthe West , CA September 2014 to May 2016


Role : Sr ETL Developer

Project: DataNow (BI / Analytics)


Data Now project is a part of the RBG Daily & Centralized Data Initiative that enables an interactive reporting platform for
all key RBG metrics, KPIs & Analytics input. The platform is expected to evolve into a “one-stop and only” shop for all RBG
Branch Reporting. The Major reporting Need is the ranking of Metrics across the various level against all the different
attributes of a branches performance

Roles & Responsibilities:


 Implement and Enhance ETL programs using Informatica Power Center against the Oracle 10g Data Warehouse.
 Contribute to technical architectural planning, data modeling, process flow documentation and the design and
Development of innovative solutions for the business.
 Create technical designs and mapping specifications.
 Work with both technical staff and business constituents to translate business requirements into technical solutions.
 Perform unit and system test/validation tasks.
 Maintain technical and user documentation.
 Provide on-call support of the Data Warehouse nightly processing
 Be an active participant in both technology and business initiatives.

Environment:
Informatica Power Center 9.x,Oracle 11g, Teradata, Flat Files, XML, CSV Files, Unix Shell Scripts, Toad, Sql Developer, Windows
XP, Tidal, Sql Plus

Client:Bank ofthe West (A Subsidiary of BNP PARIBAS) , CA February 2014 to August 2014
Role: Sr ETL Developer

Project: Multi Channel Customer Relationship Management (MCCRM)


The objective of this project is to remodel the EDW in Teradata to Oracle EDW based on Business Requirements. This
project also involved another stream called MCCRM (Multi Channel Customer Relationship Management) where in the
Data from Various System of records are integrated and provided to Marketing Team with SQL Server 2008 as Database.
The data provided will used by marketing team to create DataMart and utilize for various Business needs of the Bank.

Responsibilities:
 Worked closely with Business Analysis team, Reverse Engineering Team, discussed and gathered the Analytics
requirements and created the Business Model using Star schemas.
 Worked on designing and coding the complete ETL process using Informatica for various transactions and loading data
from different sources like Flat Files and Relational Database.
 Created numerous Mappings and Mapplets using Transformations like Filters, Aggregator, Lookups, Expression, and
Sequence generator, Sorter, Joiner and Update Strategy.
 Implemented slowly changing dimensions type1 and Type 2 using date
 Created Mapplets and reusable transformations to be re-used during the development life cycle.
 Used Parameter files to define values for parameter and variable used in the mappings and sessions.
 Create parameter and variables and called them whenever required.
 Dropping and recreating Indexes on the database Target table using pre-sql and post-sql techniques while loading huge
tables.
 Created session and Worklets in Workflow manager having event tasks like e-mails.
 Worked on performance tuning by finding out the bottlenecks in various mappings.
 Unit testing self and other mappings before deploying the code for QA testing by creating scenarios based on business
requirement.
 Involved in performance tuning of the mappings and SQL statements, worked on Query Optimization.
 Handled UNIX operating system tasks by generating Pre and Post-Session UNIX Shell Scripts.
 Involved in creating Tidal Jobs which call Informatica Workflows, which run in the supply chain environment.
 Responsible for Issue Resolutions using Session Logs and Workflow Logs.
 Responsible for determining Incremental Loading mappings using Mapping Variables and Parameter Files.
 Developed shell scripts for running batch jobs and scheduling them.
 Called stored procedures in PL/SQL for certain key business requirements.
 Coordinated the UA testing with business users.

 Involved in troubleshooting data quality and integrity issues.


 Involved in Unit and Integration Testing of the Data Warehouse.
 Performed the data validations and control checks to ensure the data integrity and consistency.
 Involved in performance tuning of the mappings and SQL statements, worked on Query Optimization.
 Developed PL/SQL Procedures, Functions, Packages, SQL scripts in Oracle 9i/10g.
 Involved during User Testing and performed Unit testing.
 Writing documentation to describe program development, logic, coding, testing, changes and corrections.
 Involved in Scrum meeting on day to day activities and Task

Environment:
Informatica Power Center 9.x,Oracle 11g, Teradata, Flat Files, XML, CSV Files, Unix Shell Scripts,Toad, Sql Developer,
Windows XP, Tidal ,Sql Plus

Employer:Tata Consultancy Services, Chennai June 2013 to December 2013

Integration Lead / Assistant Consultant


The objective of this project is to remodel the cards data from existing CDW to ICDW environment and provide compliance
with ICDW Architectural standards and a ‘modeled’ environment for analytics. The Data from current state Oracle needs
to be migrated to Teradata. The aim of the project is also to have the Warehouse built to have a 360-degree view of a
customer, by integrating the data from various source of the industry.

Responsibilities:
 Worked closely with Business Analysis team, Reverse Engineering Team, discussed and gathered the Analytics
requirements and created the Business Model using Star schemas.
 Worked on designing and coding the complete ETL process using Informatica for various transactions and loading data
from different sources like Flat Files and Relational Database.
 Created numerous Mappings and Mapplets using Transformations like Filters, Aggregator, Lookups, Expression, and
Sequence generator, Sorter, Joiner and Update Strategy.
 Implemented slowly changing dimensions type1 and Type 2 using date
 Created Mapplets and reusable transformations to be re-used during the development life cycle.
 Used Parameter files to define values for parameter and variable used in the mappings and sessions.
 Create parameter and variables and called them whenever required.
 Dropping and recreating Indexes on the database Target table using pre-sql and post-sql techniques while loading huge
tables.
 Created session and Worklets in Workflow manager having event tasks like e-mails.
 Worked on performance tuning by finding out the bottlenecks in various mappings.
 Unit testing self and other mappings before deploying the code for QA testing by creating scenarios based on business
requirement.
 Involved in performance tuning of the mappings and SQL statements, worked on Query Optimization.
 Handled UNIX operating system tasks by generating Pre and Post-Session UNIX Shell Scripts.
 Involved in creating Control M Jobs which call Informatica Workflows, which run in the supply chain environment.
 Responsible for Issue Resolutions using Session Logs and Workflow Logs.
 Responsible for determining Incremental Loading mappings using Mapping Variables and Parameter Files.
 Developed shell scripts for running batch jobs and scheduling them.
 Called stored procedures in PL/SQL for certain key business requirements.
 Coordinated the UA testing with business users.
 Involved in troubleshooting data quality and integrity issues.
 Involved in Unit and Integration Testing of the Data Warehouse.
 Performed the data validations and control checks to ensure the data integrity and consistency.
 Involved in performance tuning of the mappings and SQL statements, worked on Query Optimization.
 Developed PL/SQL Procedures, Functions, Packages, SQL scripts in Oracle 9i/10g.
 Wrote UNIX shell scripts to prepress data using PMCMD command.
 Involved during User Testing and performed Unit testing.
 Writing documentation to describe program development, logic, coding, testing, changes and corrections.
 Involved in Scrum meeting on day-to-day activities and Task
Environment:
Informatica Power Center 9.x, Oracle 11g, Teradata, Flat Files, XML, CSV Files, Unix Shell Scripts, Toad, Sql Developer,
Windows XP, Tidal ,Sql Plus

Tata Consultancy Services June 2013 to December 2013


Integration Lead / Sr ETL Developer

Retirement Services Plan (RPS) is one of the Line of Business of J.P Morgan Chase .P. Morgan Retirement Plan Services
LLC.The Current platform supporting this business is in Legacy Mainframes, Client has decided the current functionality in
Mainframes to Informatica and Oracle as Database. The Major Products involved in this Process is Omni Plus and Omni Pay
by SunGard.

Responsibilities:

 Worked closely with Business Analysis team, Reverse Engineering Team, discussed and gathered the Analytics
requirements and created the Business Model using Star schemas.
 Worked on designing and coding the complete ETL process using Informatica for various transactions and loading data
from different sources like Flat Files and Relational Database.
 Created numerous Mappings and Mapplets using Transformations like Filters, Aggregator, Lookups, Expression, and
Sequence generator, Sorter, Joiner and Update Strategy.
 Created Mapplets and reusable transformations to be re-used during the development life cycle.
 Used Parameter files to define values for parameter and variable used in the mappings and sessions.
 Create parameter and variables and called them whenever required.
 Dropping and recreating Indexes on the database Target table using pre-sql and post-sql techniques while loading huge
tables.
 Created session and Worklets in Workflow manager having event tasks like e-mails.
 Worked on performance tuning by finding out the bottlenecks in various mappings.
 Unit testing self and other mappings before deploying the code for QA testing by creating scenarios based on business
requirement.
 Involved in performance tuning of the mappings and SQL statements, worked on Query Optimization.
 Handled UNIX operating system tasks by generating Pre and Post-Session UNIX Shell Scripts.
 Involved in creating Control M Jobs which call Informatica Workflows, which run in the supply chain environment.
 Responsible for Issue Resolutions using Session Logs and Workflow Logs.
 Responsible for determining Incremental Loading mappings using Mapping Variables and Parameter Files.
 Developed shell scripts for running batch jobs and scheduling them.
 Called stored procedures in PL/SQL for certain key business requirements.
 Coordinated the UA testing with business users.
 Involved in troubleshooting data quality and integrity issues.
 Involved in Unit and Integration Testing of the Data Warehouse.
 Performed the data validations and control checks to ensure the data integrity and consistency.
 Involved in performance tuning of the mappings and SQL statements, worked on Query Optimization.
 Developed PL/SQL Procedures, Functions, Packages, SQL scripts in Oracle 9i/10g.
 Wrote UNIX shell scripts to prepress data using PMCMD command.
 Involved during User Testing and performed Unit testing.
 Writing documentation to describe program development, logic, coding, testing, changes and corrections.
 Involved in Scrum meeting on day-to-day activities and Task

Environment:
Informatica Power Center 9.x,Oracle 11g, Teradata, Flat Files, XML, CSV Files, Unix Shell Scripts,Toad, Sql Developer,
Windows XP, Tidal, Sql Plus
First American India Pvt Ltd
(A Subsidiary of First American Financials, Santa Ana CA) January 2005 to June 2013
Associate Team Lead / Sr ETL Developer
Project # 1

First American provides financial services through its Title Insurance and Services segment and its Specialty Insurance
segment. The First American Family of Companies core business lines include title insurance, title plant management
services, title and other real property records, home warranty products, property and casualty insurance and banking,
trust, and investment advisory services. The project Data Trace Legacy Modernization (DTLM) project aimed at converting
all the current update process running on mainframes to be migrated and implemented the use of Informatica as ETL Tool
and SQL server as Database. The objective was to reduce the cost to the company for using Mainframe CPU Usage and to
be in pace with the latest technology in market and to serve the customers of company in better way, to build various
products as per requirement of customers. Migrate the current DB2 to a Dimensional Model in SQL Server using
Dimensional Modeling

Responsibilities:

 Worked closely with Business Analysis team, discussed and gathered the Analytics requirements and created the Business
Model using Star schemas.
 Worked on designing and coding the complete ETL process using Informatica for various transactions and loading data
from different sources like Flat Files and Relational Database.
 Created numerous Mappings and Mapplets using Transformations like Filters, Aggregator, Lookups, Expression, and
Sequence generator, Sorter, Joiner and Update Strategy.
 Implemented slowly changing dimensions type1 and Type 2 using date
 Created Mapplets and reusable transformations to be re-used during the development life cycle.
 Used Parameter files to define values for parameter and variable used in the mappings and sessions.
 Create parameter and variables and called them whenever required.
 Dropping and recreating Indexes on the database Target table using pre-sql and post-sql techniques while loading huge
tables.
 Created session and Worklets in Workflow manager having event tasks like e-mails.
 Worked on performance tuning by finding out the bottlenecks in various mappings.
 Unit testing self and other mappings before deploying the code for QA testing by creating scenarios based on business
requirement.
 Involved in performance tuning of the mappings and SQL statements, worked on Query Optimization.
 Developed Procedures, Functions in Sql Server 2005/2008.
 Handled UNIX operating system tasks by generating Pre and Post-Session UNIX Shell Scripts.
 Involved in creating Crontabs Jobs which call Informatica Workflows, which run in the chain environment.
 Played a key role in day to day production support of the Informatica batch.
 Responsible for Production Support and Issue Resolutions using Session Logs and Workflow Logs.
 Responsible for determining Incremental Loading mappings using Mapping Variables and Parameter Files.
 Developed shell scripts for running batch jobs and scheduling them.
 Coordinated the UA testing with business users.
 Worked with Informatica 8.6.1 (IDQ) toolkit, Analysis, data cleaning, data matching, data conversion, exception handling,
and monitoring capabilities of IDQ 8.6.1.
 Worked on Standard and label transformation in IDQ.
 Involved in production support and troubleshooting data quality and integrity issues.
 Involved in Unit and Integration Testing of the Data Warehouse.
 Performed the data validations and control checks to ensure the data integrity and consistency.
 Involved in performance tuning of the mappings and SQL statements, worked on Query Optimization.
 Developed PL/SQL Procedures, Functions, Packages, SQL scripts in Oracle 9i/10g.
 Wrote UNIX shell scripts to prepress data using PMCMD command.
 Involved during User Testing and performed Unit testing.
 Writing documentation to describe program development, logic, coding, testing, changes and corrections.

Environment:
Informatica Power Centre 8.6.1(Repository Manager, Designer, Source Analyzer, Workflow Manager, Workflow Monitor,
Warehouse Designer, Transformation Developer, Mapplets Designer, Mapping Designer and Task Developer, IDQ Data
Quality, SQL Server 2005/2008, Flat Files, CSV Files, COBOL Files, Unix Shell Scripts.

First American India Pvt Ltd


Module Lead / ETL Developer
Project # 2

The project NOVA aimed at transforming complex data gathered from thousands of public records from Government and
Private sources into meaningful business information used by mortgage lenders, title companies, and government
agencies. First American provided title plant and property tax information to its customers which were primarily title
insurance companies and lending institutions. Title plant system was also known as Advance Title System (ATS) or Geo
Expansion. The data was usually provided from the Land American title office. Conducted analysis of the data and
converted data into the database compatible format. The Data is basically from two different sources Data Quick System
and Core logic needs to be integrated and feed data to Applications called Doc Edge by building a warehouse and Data
Marts

Responsibilities:

 Analyzing the Functional requirements and creating mapping specification documents


 Creating Technical design documents from the functional specifications and developing Informatica Mappings, Mapplets,
Sessions and Workflow to implement the business rules and for data movement.
 Extensively worked in the areas like data staging and data validation.
 Extensively used Informatica to extract high volume datasets from various sources like Text Files.
 Installed and configured Informatica Power Center and Informatica Client tools.
 Created mappings using various Transformations like Aggregator, Expression, Filter, Router, Joiner, Lookup, Update
strategy, Source Qualifier, Sequence generator, Stored Procedure.
 Created, scheduled, and monitored workflow sessions on the basis of run on demand, run on time, using Informatica
Power Center workflow manager.
 Participated in team meetings and proposed ETL Strategy.
 Suggested performance tuning techniques in Informatica Mappings for optimum performance.
 Resolved many complex defects involving analysis of mappings and stored process.
 Mapping and Tracing data from system to system in order to solve a given business or system problem.
 Created Mapplets and reusable transformations to be re-used during the development life cycle.
 Involved in Data Warehouse Technical Design Document for High Level load Strategy and ETL design.
 Developed source to target mapping documentation for ETL process.
 Design the ETL process to Extract transform and load data
 Applied business logic in the mappings for inserting and updating records in the database.
 Worked on Field analyzing like identifying the business logic for each field.
 Developed UNIX shell scripts to scheduling Scripts by using scheduling tools.
 Responsible for Production Support and Issue Resolutions using Session Logs and Workflow Logs.
 Responsible for determining Incremental Loading mappings using Mapping Variables and Parameter Files.
 Developed shell scripts for running batch jobs and scheduling them.
 Involved in production support and troubleshooting data quality and integrity issues.
 Involved in Unit and Integration Testing of the Data Warehouse.
 Involved in performance tuning of the mappings and SQL statements, worked on Query Optimization.
 Developed Procedures, Functions, SQL Server 2005/2008
 Create parameter and variables and called them whenever required.
 Wrote UNIX shell scripts to prepress data using PMCMD command.
 Involved during User Testing and performed Unit testing.
Environment:Informatica Power Center 8.6 (Repository Manager, Designer, Workflow Manager, Workflow Monitor),
Power Exchange for Mainframes (DB2), Sql Server 2005/2008 Linux, Informatica Scheduler

Project # 3

RAPS (Release, Assignment, Pre foreclosure and subordination) an assignment of mortgage was a document which indicated
That a mortgage has been transferred from the original lender or borrower to a third party. Assignments of mortgage were
More commonly seen when lenders sell mortgages to other lenders. When someone had what was known as an assumable
Mortgage, it was possible for the borrower to transfer the mortgage to another person, in which case an assignment of
Mortgage would need to be filed to record the transaction.

Responsibilities:

 Used Informatica Powercenter 8.6.1 and its all features extensively to transform and load to Sql Server 2005/2008.
 Gathered requirements for the scope of loading data from various sources to database.
 Built Integration between various Sources.
 Extract Data from various sources, flat files to Transform and load into Staging
 Designed and developed ETL Mappings using Informatica to extract data from flat files and DB2, and to load the data into
the target database.
 Worked with Variables and Parameters in the mappings.
 Used various Transformations like Router, Filter, Joiner, Update Strategy and connected and unconnected lookup for
better data messaging and to migrate clean and consistent data using the Informatica Powercenter Designer.
 Involved in the Development of Informatica mappings and Mapplets and also tuned them for Optimum performance,
Dependencies and Batch Design.
 Worked on troubleshooting the Mappings to improving Performance by identifying bottlenecks.
 Performance tuning of the process at the mapping level, session level, source level, and the target level.
 Week on Week Status report to manager on the process and Timeline
 Used Informatica Scheduler and Crontabs to schedule Informatica mappings

Environment:
Informatica Power Center 9.x,Oracle 11g, Teradata, Flat Files, XML, CSV Files, Unix Shell Scripts,Toad, Sql Developer,
Windows XP, Tidal, Sql Plus

Project # 4

Geo counties conversion was the project where data from the Counties recorders offices in United States which thecompany
Got through various vendors needs to be converted as per business unit requirement by applying various rules for Data
Cleansing and formatting the data and loading it to database. The data could be from any source mostly the data arrived
Through text files, Excel files and Access. The conversion process is to load the data into DB2 table which can be accessed
By user though front-end applications. The Data mainly provides the Historical data of the property which allows Individual
Andfinancial institutions to make a discoing on particular property

Responsibilities:
 Initial analysis of Raw Data received from business Users, Data differs from County to county and State to state
 Coding of pre-process steps using COBOL and JCL
 Prepare the reports likes fields, records data types involved and send it to Business Unit
 Designed and developed COBOL programs
 Build the Jobs in JCL in a Sequential Order to execute the COBOL Process
 Build the load Jobs in JCL using Load Utilities to load data in DB2
 Provide post live support once the county goes live into Production.
 Maintenance of Counties, Fixing issues if any in post-production
 Week on Week Status report to manager on the process and Timeline
 Once Data goes live to production, County should be scheduled for Daily Updates, which happens on daily, weekly basis.

Environment: COBOL, JCL, DB2, CICS, DYL 280, Xpeditor.

You might also like