Sample
Sample
PROFILE SUMMARY:
● 25+ years of experience in the Defense, Banking & Finance, Transportation, Healthcare industries with
spanning of technologies from Oracle, Teradata, Db2, Greenplum, Hadoop, Hadoop, AWS, Snowflake
● Certified Project Management Professional by PMI organization over 20 years and managing a team.
● Strong Experience with Snowflake implementation with AWS Technology for Real-time Data
Warehouse Implementation
● Experience with several AWS tools such as DMS, Datasync, Lambda, Glue, Lake Formation, AWS
Sagemaker, AWS Glue Studio, Cloudwatch, AWS Glue DataBrew, others.
● Automated Data Integration, Security Authentication, Data Ingestion using Python.
● Experience in Design Methodology (Relational and Dimensional ModelingNO) & Data Modeling tools
like ERWIN, ERStudio Converting Conceptual, logical design into a physical design, adhering to
organization standards on both logical and physical design.
● IBM Certified Data Modeler and Teradata Certified Professional
● Expertise in Big Data Architecture design, planning, installation, deployment, application development,
and migration of traditional Data Warehouse solutions to Hadoop-based Integrated Data Lakes and
Enterprise Data Hub (EDH).
● Strong Experience in Metadata integration with tools such as Enterprise Data Catalog (EDC) by creating
a holistic view of all the data assets in the organization and moving towards a single source of truth and
trusted data asset.
● Extensive expertise with monitoring, debugging, benchmarking, and performance tuning of tools of the
Data Platform ecosystem and external tools supporting the architecture.
TECHNICAL SKILLS
Data Warehouses: Snowflake, Greenplum, Teradata, Hadoop, Oracle, Sql Server, Db2
Databases: Oracle, Db2, Natural Adabas, DB2, IMS DB/DC, IBM DB2/400
Machine Learning: Python, Sagemaker
Cloud: AWS, Snowflake
Data Management//Data Lineage: Enterprise Data Catalog (EDC), AWS Lake Formation/Glue
Data Ingestion/Integration: AWS DMS, Informatica Tools Batch and CDC
Data Visualization: Tableau
ETL: AWS DMS, AWS Batch, Informatica Developer, AWS Glue
Languages: Python, SAS
Build & Configuration/Code
Repositories (CI): Jenkins, SVN/Bitbucket
PROFESSIONAL EXPERIENCE
Responsibilities:
● Collaborated with business teams, and EA providing a strategic and tactical roadmap for migration of
current enterprise technology, product, and service stack to Big Data based ecosystem that builds value
Working as a Data Warehouse Architect/Tech Lead assigned to build, guide the team and maintain processes
written in Teradata, SAS, and Informatica.
Responsibilities
● Gathering requirements, preparing business and technical specifications
● Leading developers, coding, preparing implementation documentation, and deploying projects into
production.
● Led implementation of several Data Warehouse and Analytical initiatives at AAR
● Designed a logical model from business requirements.
● Responsible for the setup of the Teradata platform for Prod, Dev, and Quality Assurance (QA).
● Led team in designing, developing, and implementing ETL process using Informatica and Oracle Data
loader for sourcing data from multiple sources applying transformations per business rules, and loading
the data in the target Teradata DW.
● Designed, developed, and implemented Data marts used in business reporting (SAS), metrics, and end-
user querying.
● Provided data analysis in support of business intelligence and information management using SAS and
Embarcadero.
● Responsible for data integration and implementation and maintenance; problem resolution; product
enhancements; testing; review of end-user materials; reviewing and coordinating the release of product
and status reports.
● Build and designed reference architectures, models, etc. for new services that leverage Teradata/ as the
core underlying platform for Data-Warehouse, DataMart's, ETL Vendor Integrations, Data –Ingestion
● Worked with infrastructure and admin teams for Disaster Recovery strategy for Teradata on-premise
POCs of ETL, vendors like Informatica, Talent, Data Stage, etc.
Environment: ER Studio, Unix Shell Scripting, Mainframe, Oracle, Teradata, SAS, Informatica, Tableau ,
Enterprise Data Catalog(Informatica)
At Wells Fargo as a Data Architect focused on providing an end-to-end data architecture for the HELOC line
of business.
Responsibilities:
● Led successful implementation and maintenance of an enterprise-level Application
● Demonstrated the ability to relate business problems to the underlying data and the efficient utilization
of Data Warehouse technology to solve business problems.
● Designed logical model from business requirements for and converted logical model to physical model.
● Built data marts for transactional data for all subject areas.
● Assisted application developers with performance tuning of application queries and SQL.
Environment: SQL Server, Sybase, ER Studio.
This was a new initiative by Johnson Controls-ASG to revamp their existing Data Warehouse. The legacy
warehouse consists of several unstructured data objects. Brought into the organization to lead a team of 10
members towards a structured Data Warehouse initiative and introduce standards and Best Business
Practices to the organization. He was nominated by JCI to attend the Ralph Kimball Training in San Jose, CA.
Responsibilities:
● Worked extensively with financial folks in Asia, Europe, and North America for gathering requirements
for the BBP project. The project focused on building a repository with crucial cost metrics of each plant
and building reports on top of it, which would be helpful to analyze the performance of each plant.
Based on their performance they are ranked. Low-ranked plants would adopt the strategy from the best
plant. The project saved an estimated 3 million Dollars.
● Designed/Developed a Star Schema for BBP Project. Also, involved closely with the ETL team in designing
the Staging area database tables for ETL scripts. Actively involved in ETL design.
● Extracted Data from Hyperion using Autopilot script and FTP as flat files to EDW staging. This is further
loaded data using SQL Loader into the staging tables.
● Implemented Partition Exchange to reduce the existing impact of unavailability of reports to the
business team
● Improved overall performance of End-to-End job process from 24hrs down to 8 hrs by revamping the
load strategy by introducing parallel runs in Autosys job scheduler, introducing new indexes, sub-
partitioning of data, memory issue associated with Oracle Warehouse builder when sorting and grouping
of data by introducing temp tables and ensuring the use of partition keys for access and retrieval
Environment: Oracle 9.2.2, Erwin 4.1, Business Object 6.5 and Oracle Warehouse Builder 10g, HP 3000
Unix Server
Blue Cross Blue Shield of Michigan, Detroit, MI Sep 2002 to Sep 2004
Team Lead/Data Warehouse Architect
This warehouse encompasses the extraction of data from several legacy systems and ERP applications. Data
was extracted from IBM Mainframe, Client Server Application into the Enterprise Data Warehouse. Data
extract consists of Blue Exchange data in which transactions such as Membership Eligibility, Provider,
Subscriber, Data, Claims, Pre- authorization data is captured into the warehouse. These data are sent to the
NASCO for transmission to the Blue Exchange group of Entities. Both the transmitted data and the response
data are captured into the warehouse and measured for the duration. Blue Exchange charges a substantial
fee for any delay in response to the provider or the requesting Blue Exchange entity. The purpose of the
warehouse was to keep track of the charges paid and measure the response time trend.
BCBSM has a subsidiary known as Blue Care Network which deals with HMO products. Here the Data
Warehouse extract consists of extracting claims details, membership details, product details, provider details
from the Facets application into the integrated EDW environment. Here the measure consists of the Claim
payment ratio, Provider Payment Ratio, Rejected Claim Ratio, and other measures. All source data were sent
Involved in design, analysis, and technical architecture of North America, Europe, Middle East, Africa (EMEA)
MQ Series Adapter for IBM’s Worldwide Customer Database (CDb) using a combination of relational and
object-oriented methodologies. CDb is the database that stores the unique representation of the customer.
CDb captures different views of the customer from the various customer information systems within IBM
and will determine and maintain the unique representation of the customer. Currently, the unique
representation of the customer is interfaced with the North America Analytical Database (NADb) every
quarter. The architecture involves providing customer data Adapters in the standard Db2 format and using
Unix Shell Scripts, Db2 Stored Procedures, and Db2 XML Extender; an XML Business Object Document is
created which will be transported by MQ Series into the Unique Reference Number Engine. The Customer
Database is the master repository for level 1 customer data. It contains all variations and cross-references to
customers between systems.
Responsibilities:
● As an Architect for the Worldwide Customer Database,
● Analyzed the source data arriving from the different source systems and mapping them to the
Worldwide Customer Database
● Implemented performance improvement measures like parallel processing, indexing strategy, parallel
loading, and aggregation
● Responsible for designing multi-thread extraction, load, transformation, and aggregation, which
significantly improved the overall performance throughput.
● Involved in Changing and recreating Data Models as per our client requirements using Computer
Associate
● Involved in designing objects required in interfacing MQSeries with source and target databases
Converted business requirements into a system specification
● Data Gap Analysis for any changes to the source system against the Worldwide Customer Database.
● Business Requirements Study/ Migration Data Analysis/ Migration Data Dictionary & Design
Documentation,
● Responsible for the Customer database design using Computer Associate.
● Provided support for unit and integration tests.
Environment: Computer Associate 3.5.2 Data Modeler, AXIO Data Analysis Tool, Sagent Data Mart Tool,
CMVC (Change Management), Brio Query-EIS 6.0, IBM AIX Unix Server, Unix Shell Scripting, Db2 7.2.2 EEE,
Db2 Stored Procedure, C++, C program, File Aid, Lotus Notes, Lotus Word Pro, Db2 XML Extender, Windows
NT Server, Windows 2000
Involved in Operational Data Warehousing project for Maybank (Commercial Bank), Singapore, which has
over 300 branches in Singapore and Malaysia. Senior management uses the Data Warehouse for strategic
planning and budgeting.
Data was extracted from PeopleSoft 7.5, AS/400, IBM Mainframe, Windows NT Server, Unix using
FEATURED SPEAKER
Teradata Conference – Mileage at AAR
TDWI Conference – Best Practices for Deploying and Adopting a Data Catalog
RRE Conference@University of Delaware – Automation Framework in Hadoop@AAR
Informatica World – Best Practices for Deploying and Adopting a Data Catalog
Award Submission Papers
TDWI Conference – Car Hire Project Implementation