Vishal Kumar
Vishal Kumar
904-436-1221 / [email protected]
Job Objective:
Aiming senior level assignments in Data Engineering, Data Warehouse, Cloud Data Services, Analytics / Business
Intelligence with an organization of high repute by associating myself with a professionally run business organization,
that has potential for growth - both organizationally and personally and provide ample opportunity to learn, improve
and implement
Professional Summary
Over 17+ years’ experience into IT Industry and 14+ years of Data Warehouse experience with domain knowledge of
Finance/Banking, Insurance and Title Insurance / Real Estate Services domain.
Extensive experience in ETL and Data Integration for developing ETL mappings and scripts using Informatica Power Center
10.x/9.x/8.x using Designer, Repository Manager, Workflow Manager & Workflow Monitor, Informatics (ICS/IICS).
Exposure on Power Exchange functionality.
Extensive experience in creating Transformations and Mappings using Informatica Designer to implement business rules
and processing tasks using Workflow Manager to move data from Multiple Sources into Targets.
Expertise using MS SQL Server 2008/2005/2000, oracle 12c/ 11g/10g, DB2 9.0/8.0/7.0, Teradata 12, Erwin 7.0/4.0/3.5,
SQL, PL/SQL, SQL *Plus, SQL *Loader.
Cloud Technologies (Microsoft Azure), Databricks, Blob Storage, Creating Pipelines (Azure Data Factory, Pyspark,
SparkSql, Azure DevOps, CI/CD Implementation, Octopus Deployments
Implementation of On-Premise Application to Cloud Instance of Azure
Working Experience on Azure Data Factory, for Orchestrating the Pipelines and Schedules.
Working on Python for Enhancement or New Development
Working on Go Anywhere, with Different Capabilities like Processing Emails, Emails with Attachment, Processing CSV, Flat
Files, Excel and Database objects
Extensively used Teradata SQL Assistant and BTEQ to write SQL’s for ETL.
Data Quality Using IDQ (Informatica Developer Tool).
Exposure to Informatica MDM
Exposure and Ability to Learn New technologies required for Deliverables,
Extensive Experience in Tuning Techniques related to Database (Oracle / Sql Server).
Hands on experience on all phases of software development life cycle (SDLC) of Data warehousing, including requirement
gathering from business users, analysis, design, creation of mapping documents, specification writing, development,
testing, and implementation.
Expertise in working with various transformations using Aggregator, Look Up, Update Strategy, Joiner, Filter, Sequence
Generator, Normalizer, Sorter, Router, Java transformations in Informatica Powercenter Designer.
Extensive expertise in Debugging, Performance Tuning and Optimization of the ETL Sessions and Mappings.
Proficiency in data warehousing techniques for Data Cleansing, Slowly Changing Dimension Phenomenon, Surrogate Key
assignment, Change Data Capture (CDC).
Extensive Experience in integration of various RDBMS like Oracle, DB2, SQL server into staging area.
Expertise in creating Stored Procedure/Functions, Triggers and Packages in Oracle and SQL Server MS
2005/2008/2014/2016.
Expertise in Data Analysis, Data Design, Developing and Deployment for Data Integration for BI requirements.
Profound knowledge of the principle of Data Warehouse using Fact tables, Dimensional Table, Star Schema Modeling and
Snowflake Schema Modeling.
Experience in Administrative activities such as creating folders, assigning security permissions to users and creating
repositories, setting up groups/users’ permissions in Informatica Repository Manager and admin console.
Exposure knowledge of Informatica Migration and Upgrade.
Experience working with SDLC and implementing, Waterfall and Agile methodologies.
Experience working with Informatica in UNIX and LINUX environments.
Coordinate ETL and database development work with developers and conduct code reviews.
Excellent oral/written/interpersonal communication skills, quick learner, willing to adapt to the team/organizational
environment.
Experience in Regulatory / Federal projects for Banking System, CCAR (Comprehensive Capital Analysis and Review),
DFAST, Liquidity Coverage Ratio (LCR) and Risk Weighted Assets
Skill sets
ETL Tools : Informatica 10.x/9.X/8.X, Sap Data Services (BODS), Talend, Informatica Developer (IDQ),
SSIS, Azure Data Factory
ETL Management Console : Informatica Administration Console, Sap Data Services Management Console
Database : Oracle 10g/11g/12c, DB2, Sql Server 2005/2008/2014/2016, MySQL
NoSQL : Azure Cosmos DB
Database Tools, Utilities : DB2 Visualizers Developer, Toad, SqlPlus,GoAnywhere (Managed File Transfer)
Scripting Languages : Python, Perl, UNIX, Windows Batch Scripts
Procedure Language : PL/SQL, T-SQL
Cloud Technologies : Microsoft Azure,DataBricks, PySPARK, DataLake, DeltaLake, Informatica Cloud (ICS/IICS),
Spark Sql
Operation Systems : Windows / UNIX / Linux
Scheduling Tools : Tidal, Control-M, Redwood, Informatica Inbuilt Scheduler, CronTabs
Version Controller : Perforce (P4v), Informatica Versioning
Code Promotion Tools : Azure DevOps, CI/CD, Octopus, Jenkins, Nexus and Nolio, Informatica Deployments
Project Documentation : Confluence, Jira
QA Tools : HP Quality Center
Analytics Tools : Adobe Analytics (Datawarehouse Reports)
Education:
Bachelor of Technology in Electronics and Electrical Engineering, Year 2004, Jawaharlal Nehru Technological
University.
Projects Summary
Operate modern data architecture approaches to meet key business objectives and provide end-to-end data solutions
Develop good understanding of how data will flow & stored through an organization across multiple applications such
Valuation Advisory Service, Supply Chain Management and Enterprise Data Management
Design & develop data management and data persistence solutions for application use cases leveraging relational, non-
relational databases and enhancing our data processing capabilities
Develop POCs to influence platform architects, product managers and software engineers to validate solution proposals
and migrate
Lead team of data engineers and Team Managers to deliver data solutions in Agile manner while mentoring Team
Members
Translates/interprets business strategy into objectives for the team. Prioritizes and assigns business initiatives and
performs high-level review of projects to ensure conformance to business requirements and delivery timelines
Develop a highly engaged team through communication, performance management, career development plans and
reward programs
Hire and develop talent pool establishing a good succession planning across functional areas managed
Promote an environment that supports diversity and reflects JLL Brand and values
Establishes goals, success criteria and performance measures to evaluate the outcome of the projects/initiatives
managed
Promotes a culture of teamwork and collaborates across other functional verticals within the Center of expertise in
adopting/sharing of best practices B
Build a culture committed to delivery of outstanding customer service and open, transparent team environments
Develop and enhance the functional & Technical capabilities of the team by providing effective coaching &
facilitating/imparting the relevant technical and functional trainings
Environment:Informatica Power Center 10.x, Sql Server, Oracle 12c/11g, Toad, Sql Developer, Control-M
Azure Data Bricks, Pyspark, Spark Sql, Python, CI/CD, Azure Dev Ops, Azure Data Factory, Azure Data Lake, Azure Sql
Server
Implement and Enhance ETL programs using Informatica Power Center against the Oracle 11g Data Warehouse.
Contribute to technical architectural planning, data modeling, process flow documentation and the design and
Development of innovative solutions for the business.
Create technical designs and mapping specifications.
PL/SQL Enhancements and Performance Tuning
Work with both technical staff and business constituents to translate business requirements into technical solutions.
Perform unit and system test/validation tasks.
Maintain technical and user documentation.
Provide on-call support of the Data Warehouse nightly processing
Be an active participant in both technology and business initiatives.
Environment:
Informatica Power Center 9.x, Oracle 11g, Teradata, Flat Files, XML, CSV Files, Unix Shell Scripts, Toad, Sql Developer,
Windows XP, Tidal, Sql Plus
Client Silicon Valley Bank (SVB), Santa Clara, CA May 2016 to June 2017
Role: Sr ETL Developer
CCAR stress testing came about after the economic collapse of 2008. The Dodd-Frank Act was enacted, which requires
both the Fed and all large bank holding companies (any banking organization with consolidated Average assets of $50
billion or more) to conduct and submit the results of the annual test - then disclose those results to the public.
Stress tests are submitted according to macroeconomic scenarios provided by the Fed and are used as protection to check
for whether a bank’s system could theoretically handle a future economic shock. They test the bank’s ability to remain
“well-capitalized,” with a minimum Tier 1 common equity ratio of 5 percent, even in a severely negative economic
scenario. That hypothetical future scenario is supplied annually by the Fed to the banks so they can submit their
responsive plans under the rules of the CCAR exam, every mandated bank must submit a “capital action plan” for the
following four quarters. The Fed then assesses that bank’s financial health and gives the bank a score. This score is not
simply a “pass/fail”; banks are assessed in terms of exactly how strong they are. The Fed looks at whether banking
organization have sufficient capital available to operate soundly even under extreme economic duress, and
whether they have what the Fed calls “a robust, forward-looking capital-planning process that accounts for their
unique risks.”
SVB, is Required to Submit the CCAR Annual Report by 1st Quarter of 2017, with the Specifications and Fed Edits
provided by Federal Reserve Bank.
To meet the data and reporting requirements of CCAR and other Regulatory reporting like Risk Weighted Assets, LCR and
Projections, there was a need on enhancing the data on existing EDW Data, FRDS is the enhanced version of EDW which is
Built to consider it as single source of truth for all reporting requirements. The FRDS is built on to use the Materialized
Views which cover different subject areas like CreditLines and Loans, Customer Information, Loan Payments, Charge off
And Recovery
Contribute to technical architectural planning, data modeling, process flow documentation and the design and
Development of innovative solutions for the business.
Create technical designs and mapping specifications.
Work with both technical staff and business constituents to translate business requirements into technical solutions.
Perform unit and system test/validation tasks.
Maintain technical and user documentation.
Design and Code Reviews
Provide on-call support of the Data Warehouse nightly processing
Be an active participant in both technology and business initiatives.
Performance Tuning of ETL and Oracle Queries.
Lead team of 6 People across Onsite and Offshore Resources of Tata Consultancy Services.
Environment:
SAP BODS, Oracle 11g/12C, Flat Files, XML, CSV Files, Unix Shell Scripts, Toad, Sql Developer, Windows XP, Redwood,
SqlPlus, NOLIO, Jenkins, JIRA, Perforce, PL/SQL
Environment:
Informatica Power Center 9.x,Oracle 11g, Teradata, Flat Files, XML, CSV Files, Unix Shell Scripts, Toad, Sql Developer, Windows
XP, Tidal, Sql Plus
Client:Bank ofthe West (A Subsidiary of BNP PARIBAS) , CA February 2014 to August 2014
Role: Sr ETL Developer
Responsibilities:
Worked closely with Business Analysis team, Reverse Engineering Team, discussed and gathered the Analytics
requirements and created the Business Model using Star schemas.
Worked on designing and coding the complete ETL process using Informatica for various transactions and loading data
from different sources like Flat Files and Relational Database.
Created numerous Mappings and Mapplets using Transformations like Filters, Aggregator, Lookups, Expression, and
Sequence generator, Sorter, Joiner and Update Strategy.
Implemented slowly changing dimensions type1 and Type 2 using date
Created Mapplets and reusable transformations to be re-used during the development life cycle.
Used Parameter files to define values for parameter and variable used in the mappings and sessions.
Create parameter and variables and called them whenever required.
Dropping and recreating Indexes on the database Target table using pre-sql and post-sql techniques while loading huge
tables.
Created session and Worklets in Workflow manager having event tasks like e-mails.
Worked on performance tuning by finding out the bottlenecks in various mappings.
Unit testing self and other mappings before deploying the code for QA testing by creating scenarios based on business
requirement.
Involved in performance tuning of the mappings and SQL statements, worked on Query Optimization.
Handled UNIX operating system tasks by generating Pre and Post-Session UNIX Shell Scripts.
Involved in creating Tidal Jobs which call Informatica Workflows, which run in the supply chain environment.
Responsible for Issue Resolutions using Session Logs and Workflow Logs.
Responsible for determining Incremental Loading mappings using Mapping Variables and Parameter Files.
Developed shell scripts for running batch jobs and scheduling them.
Called stored procedures in PL/SQL for certain key business requirements.
Coordinated the UA testing with business users.
Environment:
Informatica Power Center 9.x,Oracle 11g, Teradata, Flat Files, XML, CSV Files, Unix Shell Scripts,Toad, Sql Developer,
Windows XP, Tidal ,Sql Plus
Responsibilities:
Worked closely with Business Analysis team, Reverse Engineering Team, discussed and gathered the Analytics
requirements and created the Business Model using Star schemas.
Worked on designing and coding the complete ETL process using Informatica for various transactions and loading data
from different sources like Flat Files and Relational Database.
Created numerous Mappings and Mapplets using Transformations like Filters, Aggregator, Lookups, Expression, and
Sequence generator, Sorter, Joiner and Update Strategy.
Implemented slowly changing dimensions type1 and Type 2 using date
Created Mapplets and reusable transformations to be re-used during the development life cycle.
Used Parameter files to define values for parameter and variable used in the mappings and sessions.
Create parameter and variables and called them whenever required.
Dropping and recreating Indexes on the database Target table using pre-sql and post-sql techniques while loading huge
tables.
Created session and Worklets in Workflow manager having event tasks like e-mails.
Worked on performance tuning by finding out the bottlenecks in various mappings.
Unit testing self and other mappings before deploying the code for QA testing by creating scenarios based on business
requirement.
Involved in performance tuning of the mappings and SQL statements, worked on Query Optimization.
Handled UNIX operating system tasks by generating Pre and Post-Session UNIX Shell Scripts.
Involved in creating Control M Jobs which call Informatica Workflows, which run in the supply chain environment.
Responsible for Issue Resolutions using Session Logs and Workflow Logs.
Responsible for determining Incremental Loading mappings using Mapping Variables and Parameter Files.
Developed shell scripts for running batch jobs and scheduling them.
Called stored procedures in PL/SQL for certain key business requirements.
Coordinated the UA testing with business users.
Involved in troubleshooting data quality and integrity issues.
Involved in Unit and Integration Testing of the Data Warehouse.
Performed the data validations and control checks to ensure the data integrity and consistency.
Involved in performance tuning of the mappings and SQL statements, worked on Query Optimization.
Developed PL/SQL Procedures, Functions, Packages, SQL scripts in Oracle 9i/10g.
Wrote UNIX shell scripts to prepress data using PMCMD command.
Involved during User Testing and performed Unit testing.
Writing documentation to describe program development, logic, coding, testing, changes and corrections.
Involved in Scrum meeting on day-to-day activities and Task
Environment:
Informatica Power Center 9.x, Oracle 11g, Teradata, Flat Files, XML, CSV Files, Unix Shell Scripts, Toad, Sql Developer,
Windows XP, Tidal ,Sql Plus
Retirement Services Plan (RPS) is one of the Line of Business of J.P Morgan Chase .P. Morgan Retirement Plan Services
LLC.The Current platform supporting this business is in Legacy Mainframes, Client has decided the current functionality in
Mainframes to Informatica and Oracle as Database. The Major Products involved in this Process is Omni Plus and Omni Pay
by SunGard.
Responsibilities:
Worked closely with Business Analysis team, Reverse Engineering Team, discussed and gathered the Analytics
requirements and created the Business Model using Star schemas.
Worked on designing and coding the complete ETL process using Informatica for various transactions and loading data
from different sources like Flat Files and Relational Database.
Created numerous Mappings and Mapplets using Transformations like Filters, Aggregator, Lookups, Expression, and
Sequence generator, Sorter, Joiner and Update Strategy.
Created Mapplets and reusable transformations to be re-used during the development life cycle.
Used Parameter files to define values for parameter and variable used in the mappings and sessions.
Create parameter and variables and called them whenever required.
Dropping and recreating Indexes on the database Target table using pre-sql and post-sql techniques while loading huge
tables.
Created session and Worklets in Workflow manager having event tasks like e-mails.
Worked on performance tuning by finding out the bottlenecks in various mappings.
Unit testing self and other mappings before deploying the code for QA testing by creating scenarios based on business
requirement.
Involved in performance tuning of the mappings and SQL statements, worked on Query Optimization.
Handled UNIX operating system tasks by generating Pre and Post-Session UNIX Shell Scripts.
Involved in creating Control M Jobs which call Informatica Workflows, which run in the supply chain environment.
Responsible for Issue Resolutions using Session Logs and Workflow Logs.
Responsible for determining Incremental Loading mappings using Mapping Variables and Parameter Files.
Developed shell scripts for running batch jobs and scheduling them.
Called stored procedures in PL/SQL for certain key business requirements.
Coordinated the UA testing with business users.
Involved in troubleshooting data quality and integrity issues.
Involved in Unit and Integration Testing of the Data Warehouse.
Performed the data validations and control checks to ensure the data integrity and consistency.
Involved in performance tuning of the mappings and SQL statements, worked on Query Optimization.
Developed PL/SQL Procedures, Functions, Packages, SQL scripts in Oracle 9i/10g.
Wrote UNIX shell scripts to prepress data using PMCMD command.
Involved during User Testing and performed Unit testing.
Writing documentation to describe program development, logic, coding, testing, changes and corrections.
Involved in Scrum meeting on day-to-day activities and Task
Environment:
Informatica Power Center 9.x,Oracle 11g, Teradata, Flat Files, XML, CSV Files, Unix Shell Scripts,Toad, Sql Developer,
Windows XP, Tidal, Sql Plus
First American India Pvt Ltd
(A Subsidiary of First American Financials, Santa Ana CA) January 2005 to June 2013
Associate Team Lead / Sr ETL Developer
Project # 1
First American provides financial services through its Title Insurance and Services segment and its Specialty Insurance
segment. The First American Family of Companies core business lines include title insurance, title plant management
services, title and other real property records, home warranty products, property and casualty insurance and banking,
trust, and investment advisory services. The project Data Trace Legacy Modernization (DTLM) project aimed at converting
all the current update process running on mainframes to be migrated and implemented the use of Informatica as ETL Tool
and SQL server as Database. The objective was to reduce the cost to the company for using Mainframe CPU Usage and to
be in pace with the latest technology in market and to serve the customers of company in better way, to build various
products as per requirement of customers. Migrate the current DB2 to a Dimensional Model in SQL Server using
Dimensional Modeling
Responsibilities:
Worked closely with Business Analysis team, discussed and gathered the Analytics requirements and created the Business
Model using Star schemas.
Worked on designing and coding the complete ETL process using Informatica for various transactions and loading data
from different sources like Flat Files and Relational Database.
Created numerous Mappings and Mapplets using Transformations like Filters, Aggregator, Lookups, Expression, and
Sequence generator, Sorter, Joiner and Update Strategy.
Implemented slowly changing dimensions type1 and Type 2 using date
Created Mapplets and reusable transformations to be re-used during the development life cycle.
Used Parameter files to define values for parameter and variable used in the mappings and sessions.
Create parameter and variables and called them whenever required.
Dropping and recreating Indexes on the database Target table using pre-sql and post-sql techniques while loading huge
tables.
Created session and Worklets in Workflow manager having event tasks like e-mails.
Worked on performance tuning by finding out the bottlenecks in various mappings.
Unit testing self and other mappings before deploying the code for QA testing by creating scenarios based on business
requirement.
Involved in performance tuning of the mappings and SQL statements, worked on Query Optimization.
Developed Procedures, Functions in Sql Server 2005/2008.
Handled UNIX operating system tasks by generating Pre and Post-Session UNIX Shell Scripts.
Involved in creating Crontabs Jobs which call Informatica Workflows, which run in the chain environment.
Played a key role in day to day production support of the Informatica batch.
Responsible for Production Support and Issue Resolutions using Session Logs and Workflow Logs.
Responsible for determining Incremental Loading mappings using Mapping Variables and Parameter Files.
Developed shell scripts for running batch jobs and scheduling them.
Coordinated the UA testing with business users.
Worked with Informatica 8.6.1 (IDQ) toolkit, Analysis, data cleaning, data matching, data conversion, exception handling,
and monitoring capabilities of IDQ 8.6.1.
Worked on Standard and label transformation in IDQ.
Involved in production support and troubleshooting data quality and integrity issues.
Involved in Unit and Integration Testing of the Data Warehouse.
Performed the data validations and control checks to ensure the data integrity and consistency.
Involved in performance tuning of the mappings and SQL statements, worked on Query Optimization.
Developed PL/SQL Procedures, Functions, Packages, SQL scripts in Oracle 9i/10g.
Wrote UNIX shell scripts to prepress data using PMCMD command.
Involved during User Testing and performed Unit testing.
Writing documentation to describe program development, logic, coding, testing, changes and corrections.
Environment:
Informatica Power Centre 8.6.1(Repository Manager, Designer, Source Analyzer, Workflow Manager, Workflow Monitor,
Warehouse Designer, Transformation Developer, Mapplets Designer, Mapping Designer and Task Developer, IDQ Data
Quality, SQL Server 2005/2008, Flat Files, CSV Files, COBOL Files, Unix Shell Scripts.
The project NOVA aimed at transforming complex data gathered from thousands of public records from Government and
Private sources into meaningful business information used by mortgage lenders, title companies, and government
agencies. First American provided title plant and property tax information to its customers which were primarily title
insurance companies and lending institutions. Title plant system was also known as Advance Title System (ATS) or Geo
Expansion. The data was usually provided from the Land American title office. Conducted analysis of the data and
converted data into the database compatible format. The Data is basically from two different sources Data Quick System
and Core logic needs to be integrated and feed data to Applications called Doc Edge by building a warehouse and Data
Marts
Responsibilities:
Project # 3
RAPS (Release, Assignment, Pre foreclosure and subordination) an assignment of mortgage was a document which indicated
That a mortgage has been transferred from the original lender or borrower to a third party. Assignments of mortgage were
More commonly seen when lenders sell mortgages to other lenders. When someone had what was known as an assumable
Mortgage, it was possible for the borrower to transfer the mortgage to another person, in which case an assignment of
Mortgage would need to be filed to record the transaction.
Responsibilities:
Used Informatica Powercenter 8.6.1 and its all features extensively to transform and load to Sql Server 2005/2008.
Gathered requirements for the scope of loading data from various sources to database.
Built Integration between various Sources.
Extract Data from various sources, flat files to Transform and load into Staging
Designed and developed ETL Mappings using Informatica to extract data from flat files and DB2, and to load the data into
the target database.
Worked with Variables and Parameters in the mappings.
Used various Transformations like Router, Filter, Joiner, Update Strategy and connected and unconnected lookup for
better data messaging and to migrate clean and consistent data using the Informatica Powercenter Designer.
Involved in the Development of Informatica mappings and Mapplets and also tuned them for Optimum performance,
Dependencies and Batch Design.
Worked on troubleshooting the Mappings to improving Performance by identifying bottlenecks.
Performance tuning of the process at the mapping level, session level, source level, and the target level.
Week on Week Status report to manager on the process and Timeline
Used Informatica Scheduler and Crontabs to schedule Informatica mappings
Environment:
Informatica Power Center 9.x,Oracle 11g, Teradata, Flat Files, XML, CSV Files, Unix Shell Scripts,Toad, Sql Developer,
Windows XP, Tidal, Sql Plus
Project # 4
Geo counties conversion was the project where data from the Counties recorders offices in United States which thecompany
Got through various vendors needs to be converted as per business unit requirement by applying various rules for Data
Cleansing and formatting the data and loading it to database. The data could be from any source mostly the data arrived
Through text files, Excel files and Access. The conversion process is to load the data into DB2 table which can be accessed
By user though front-end applications. The Data mainly provides the Historical data of the property which allows Individual
Andfinancial institutions to make a discoing on particular property
Responsibilities:
Initial analysis of Raw Data received from business Users, Data differs from County to county and State to state
Coding of pre-process steps using COBOL and JCL
Prepare the reports likes fields, records data types involved and send it to Business Unit
Designed and developed COBOL programs
Build the Jobs in JCL in a Sequential Order to execute the COBOL Process
Build the load Jobs in JCL using Load Utilities to load data in DB2
Provide post live support once the county goes live into Production.
Maintenance of Counties, Fixing issues if any in post-production
Week on Week Status report to manager on the process and Timeline
Once Data goes live to production, County should be scheduled for Daily Updates, which happens on daily, weekly basis.