Sandhya Pochamreddy
Sandhya Pochamreddy
Senior ETL Talend developer with 7+ years of experience in Development and Production environments. Experience in
developing ETL for enterprise data warehouse and BI reports.
Experience in Talend Open Studio and Talend Integration Suite.
Experience working with Data Warehousing Concepts like OLAP, OLTP, Star Schema, Snow Flake Schema,
Logical/Physical/ Dimensional Data Modeling.
Experience in Implementation of Microsoft Business Intelligence (BI) platforms including SQL Server Integration Services
(SSIS), SQL Server Reporting Services (SSRS) in SQL Server.
Extensively used ETL methodology for performing Data Profiling, Data Migration, Extraction, Transformation and
Loading using Talend and designed data conversions from wide variety of source systems including Netezza, SQL Server,
Oracle, DB2, Teradata, Hive, Hana and non-relational sources like flat files, XML and Mainframe Files.
Extracted data from multiple operational sources for loading staging area, Data warehouse, Data Marts using SCDs
(Type 1/Type 2/ Type 3) loads.
Extensively created mappings in TALEND using tMap, tJoin, tReplicate, tConvertType, tFlowMeter, tLogCatcher,
tNormalize, tDenormalize, tJava, tAggregateRow, tWarn, tLogCatcher, tMysqlScd, tFilter, tGlobalmap, tDie etc.
Experienced in Implementation of ETL engine using Java to handle incremental loads.
Excellent experience with Talend ETL and used features of Context Variables, Database components like tMSSQLInput,
tOracleOutput, tmap, tFileCopy, tFileCompare, tSalesforceOutput, tSalesforceBulkExec, tSalesforceInput tFileExists file
components, ELT Components etc.
Capable of processing large sets of structured, semi-structured and unstructured data and supporting systems
application architecture.
Worked with different tHDFSInput, tHDFSOutput, tPigLoad, tPigFilterRow, tPigFilterColumn, tPigStoreResult, tHiveLoad,
tHiveInput, tHbaseInput, tHbaseOutput, tSqoopImport and tSqoopExport.
Able to assess business rules, collaborate with stakeholders and perform source-to-target data mapping, design and
review.
Experience in monitoring and scheduling using AutoSys, Control M& Job Conductor (Talend Admin Console) and using
UNIX (Korn& Bourn Shell) Scripting.
Experienced in creating Triggers on TAC server to schedule Talend jobs to run on server.
Strong experience in Extraction, Transformation, loading (ETL) data from various sources into Data Warehouses and
Data Marts using Informatica Power Center (Designer, Workflow Manager, Workflow Monitor, Repository Manger).
Experience in developing Informatica mappings using transformations like Source Qualifier, Connected and
Unconnected Lookup, Normalizer, Router, Filter, Expression, Aggregator, Stored Procedure, Sequence Generator,
Sorter, Joiner, Update Strategy, Union Transformations.
Experience on developing analytic reports and dashboard using business objects and Tableau.
Able to perform independently in complex troubleshooting, root-cause analysis and solution development.
Proven team player, good communication skills and quick learner.
Professional Experience
Responsibilities:
Responsible for designing and implementing ETL process to load data from different sources, perform data mining and
analyze data using visualization/reporting tools to leverage the performance of OpenStack.
Created job lets in Talend for the processes which can be used in most of the jobs in a project like to Start job and
Commit job.
Used several features of Talend such as tmap, treplicate, tfilterrow, tsort, tWaitforFile, tSalesforceOutput,
tSalesforceBulkExec, tSalesforceInput etc for ETL process
Involved in design and development of complex ETL mapping.
Utilized Big Data components like tHDFSInput, tHDFSOutput, tPigLoad, tPigFilterRow, tPigFilterColumn, tPigStoreResult,
tHiveLoad, tHiveInput, tHbaseInput, tHbaseOutput, tSqoopImport and tSqoopExport.
Implemented error handling in Talend to validate the data integrity and data completeness for the data from flat file.
To populate the data into dimensions and fact tables, efficiently involved in creating Talend Mappings.
Used Talend Admin Console Job conductor to schedule ETL Jobs on daily, weekly, monthly and yearly basis.
Extensively used ETL to load data from flat files, XML, SQL Server, Oracle database, MySql from different sources to
Data Warehouse database.
Efficient in writing complex Java code using tJava, tJavarow, tJavaFlex and handled Heap Space Issues and memory
related issues in Talend.
Created Hive queries that helped market analysts spot emerging trends by comparing fresh data with EDW reference
tables and historical metrics.
Created HBase tables to load large sets of structured, semi-structured and unstructured data coming from UNIX, NoSQL
Develop reports, dashboards using Tableau for quick reviews to be presented to Business and IT users.
Developed Ad-hoc reports using Tableau Desktop, Excel.
Developed visualizations using sets, Parameters, Calculated Fields, Dynamic sorting, Filtering, Parameter driven
analysis.
Environment: Talend Open Studio and Talend Integration Suite, tmap, treplicate, tfilterrow, tsort, tWaitforFile,
tSalesforceOutput, tSalesforceBulkExec, tSalesforceInput, SQL Server, API, Oracle, SQL, Tableau and Hadoop.
Responsibilities:
Interacted with the Business users to identify the process metrics and various key dimensions and measures. Involved
in the complete life cycle of the project.
Worked on integration projects from gathering requirements phase to implementation phase which enabled Customers
to see the data daily basis or weekly basis with very minimal manual intervention rather than spending quite amount of
manual hours.
Developed talend jobs to push data into consolidated stage (for all feeds), which is sauce for MDM process.
Used Talend most used components (tMap, tDie,tConvertType, tSOAP, tLogCatcher, tRowGenerator, tSetGlobalVar,
tHashInput & tHashOutput, tFilterRow, tAggregateRow, tFileExist, tFileCopy, tFileList, tDie and many more).
Worked on various Talend components such as tMap, tFilterRow, tAggregateRow, tFileExist, tFileCopy, tFileList, tDie
etc.
Implemented different matching rules, data validation rule to get the golden record.
Created many complex ETL jobs for data exchange from and to Database Server and various other systems Including
RDBMS, XML, CSV, and Flat file structures.
Responsible for developing, support and maintenance for the ETL (Extract, Transform and Load) processes using Talend
Integration Suite.
Conducted JAD sessions with business users and SME's for better understanding of the reporting requirements.
Developed Talend jobs to populate the claims data to data warehouse - star schema.
Used Talend Admin Console Job conductor to schedule ETL Jobs on daily, weekly, monthly and yearly basis.
Worked Extensively on Talend Admin Console and Schedule Jobs in Job Conductor.
Experienced with Java transformations for calling Hive views to extract data from Hadoop systems.
Expertized in developing shell scripts in UNIX. Created Talend mappings using the Transformations.
Involved in analyzing and extracting mongo application collections into ODS by using hive views.
Prepared ETL mapping Documents for every mapping and Data Migration document for smooth transfer of project
from development to testing environment and then to production environment.
Responsible for prioritizing the issues and assign them to the production support team and planning the deployment of
fixes for the same.
Environment: Talend Components (tMap, tDie,tConvertType, tSOAP, tLogCatcher, tRowGenerator, tSetGlobalVar, tHashInput
& tHashOutput, tFilterRow, tAggregateRow, tFileExist, tFileCopy, tFileList, tDie, SQL Server, API and many more), Oracle.
Environment: IBM WebSphere Data Stage (Designer, Director), Microsoft SQL, Netezza, Oracle, Toad, Java, MS-Access, shell
scripts, Mercury Quality Center, IBM Rational Rose.
Responsibilities:
Designed the ETL jobs using IBM Info sphere Data Stage 9.1 to Extract, Transform and load the data into staging, ODS
and EDW.
Designed and developed the ETL jobs using Parallel Edition which distributed the incoming data concurrently across all
the processors, to achieve the best performance.
Designed parallel jobs using stages such as Join, Merge, Lookup, Remove Duplicates, Copy, Filter, Funnel, Dataset,
Lookup, Pivot, and Sort, Surrogate key Generator, Change Data Capture (CDC), Modify, Row Generator and Aggregator.
Responsible for generation of DDL statements which are executed for database creation.
Responsible for preparing Physical/logical data models.
Responsible for data analysis, requirements gathering, report analysis, source-to-target mapping, frequency analysis,
process flow diagrams, and documentation. Handled Performance Tuning of Jobs to ensure faster Data Loads.
Designed sequence jobs using the activities such as Job Activity, Nested Condition, Notification Activity, Sequencer
Activity, Terminator Activity and Execute Command.
Performed the Integration and System testing on the ETL jobs. Responsible for preparing ad hoc jobs.
Translated business rules and functionality requirements into ETL derivations.
Scheduled jobs using Autosys scheduler utility based on the requirements and monitored the production processes
closely for any possible errors.
Imported the required Metadata from heterogeneous sources at the process level.
Created Job Parameters and Environment variables to run the same job for different sources and targets.
Used the Director to schedule running the job, testing and debugging its components, and monitoring.
Created Shared Containers for Re-using the Business functionality.
Collaborated with BO team to design Crystal reporting and reports for enterprise reporting applications.
Worked with Developers to troubleshoot and resolve issues in job logic as well as performance.
Responsibilities:
Responsible for requirement gathering from various groups. Followed Iterative Waterfall model for Software
Development Life Cycle Process (SDLC).
Designed and developed Informatica mapping codes to build business rules to load data. Extensively worked on
Informatica Lookup, stored procedure and update transformations to implement complex rules and business logic.
Analyzed and created facts and dimension tables.
Analyzed business requirements and worked closely with the various application teams and business teams to develop
ETL procedures that are consistent across all applications and systems.
Developed the functionality at the database level PL/SQL using Toad tool as well at the Unix OS level using shell
scripting.
Experience with dimensional modelling using star schema and snowflake models.
Experienced in identifying and documenting data integration issues, challenges such as duplicate data, nonconfirmed
data, and unclean data
Imported Source/Target Tables from the respective databases and created Reusable Transformations (Joiner, Routers,
Lookups, Rank, Filter, Expression and Aggregator), Mapplets and Mappings using Designer module of Informatica.
Used Informatica power center for (ETL) extraction, transformation and loading data from heterogeneous source
systems. Worked extensively on SQL coding to check the data quality coming from the respective parties.
Worked cooperatively with the team members to identify and resolve various issues relating to Informatica and
databases.
Applied performance tuning techniques for cubes to reduce calculate time and partitioned cubes.