ETL Testing Validation and Check List
ETL Testing Validation and Check List
Check the mapping of fields that support data staging and in data marts.
Check for duplication of values generated using sequence generators.
Check the correctness of surrogate keys that uniquely identify rows of data.
Check for data-type constraints of the fields present in staging and core levels.
Check the data loading status and error messages after ETLs (extracts, transformations,
loads).
Look for string columns that are incorrectly left or right trimmed.
Make sure all tables and specified fields were loaded from source to staging.
Verify that not-null fields were populated.
Verify that no data truncation occurred in each field.
Make sure data types and formats are as specified during database design.
Make sure there are no duplicate records in target tables.
Make sure data transformations are correctly based on business rules.
Verify that numeric fields are populated precisely.
Make sure every ETL session completed with only planned exceptions.
Verify all data cleansing, transformation, and error and exception handling.
Verify stored procedure calculations and data mappings.
Load the database with peak expected production volumes to help ensure that the
volume of data can be loaded by the ETL process within the agreed-on window.
Compare ETL loading times to loads performed with a smaller amount of data to
anticipate scalability issues. Compare the ETL processing times component by
component to pinpoint any areas of weakness.
Monitor the timing of the reject process, and consider how large volumes of rejected
data will be handled.
Perform simple and multiple join queries to validate query performance on large
database volumes. Work with business users to develop sample queries and acceptable
performance criteria for each query.
System testing checklist
One of the objectives of data warehouse testing is to help ensure that the required business
functions are implemented correctly. This phase includes data verification, which tests the
quality of data populated into target tables. A system-testing checklist can help with this
process.
Make sure the functionality of the system meets the business specifications.
Look for the count of records in source tables and compare them with counts in target
tables, followed by analysis of rejected records.
Check for end-to-end integration of systems and connectivity of the infrastructure—for
example, make sure hardware and network configurations are correct.
Check all transactions, database updates, and data-flow functions for accuracy.
Validate the functionality of the business reports.
Trouble Shooting:
Check the records in the corresponding table and delete the violated record.
Check the run date mentioned in the transfer control table and try incrementing it.
Trouble Shooting:
Issue Details: While running a job if job gets failed and in session log we got an error related
to integration service.
Trouble Shooting:
Check the profile file for the corresponding Autosys job.
Check for the correct repository of Power center corresponding to job.
Verify that the Integration service should be updated correctly.
Open the Power Center Monitor and verify that the workflow folder should be available
under the same Integration service.
Issue Details: While running a job if job gets failed and in logs, Batch launched id access issue
is occurred.
Trouble Shooting:
Issue Details: If during record counts check in source and target if it’s mismatched in large
account.
Trouble Shooting:
Check the load type, if its full load this is an issue and raise the defect.
If its incremental load, check the source extract time and stop extract time, Change the
timestamp in the parameter table and re run the job.
Check the no of processed rows and the no of loaded rows in the load summary of the
session log.
Issue Details: If Autosys job got failed and no workflow or session logs got generated.
Trouble Shooting:
Check the JIL source of the job: In Command line of JIL source, Hyphen ‘–‘anddot ‘.’
Should be placed at appropriate position.
Check if the profile files and the connections strings are pointing to the right
databases/servers.
Issue Details: If after running Autosys job data is not loaded at target side and in logs in the
load summary section there is 0 extraction and transformation for Full load.
Trouble Shooting:
Check in the source table, there should be data in the source table. There is also a
possibility that the source is having the older data than the cut of data in the control
table or the last processed time stamp.
If data is loading from STG to DW side, in the main STG table data should be present.
Check the max load date in the DW table and process date in the stage table; if it is
already matching then increment the process date.
Issue Details: If after running AutoSys job data is not loaded at target side and in logs in the
load summary section there is 0 extraction and transformation for Incremental load.
Trouble Shooting:
Check the transfer parameter entries for source and stop extract time and then check
the same in the logs. Time period for extraction of data load to till data load should be
corrected in the transfer parameter table.
Trouble Shooting:
Check the transfer parameter table and check the parameter values corresponding to
the incremental Job.
10. Autosys Job got failed which don’t have work flow.
Issue Details: If the AutoSys Job which don’t have workflow got failed.
Trouble Shooting:
Check the AutoSys logs in .err file. If the job got failed file size of .err becomes none
zero byte and if the job got succeeded vice versa.
Issue Details: If the job got failed when loading data from flat files to table.
Trouble Shooting:
Check the AutoSys logs and catch the error from .err file.
If the issue is related with files for example: invalid file or any other issues found, run
the gunzip –t filename command where the file is placed. It will return the exact error
for that file.
Issue Details: During data comparison for source and target if large number of differences are
found in DB comparator result.
Trouble Shooting:
Check the metadata columns at the target end and remove those columns from the
target end query of DB comparator.
Check the order by in the both queries and modify the queries with proper order by
clause by using primary or unique key in both source and target end.
Remove the timestamps from the comparison rule as they are intercepted differently by
the Sybase and the oracle database.
Trouble Shooting:
Issue Details: If the box jobs keep running for long time.
Resolution:
Verify that there should not be any job under Box job should be on hold status.
Change the status of the on hold sub job to off hold and trigger the box job.
Put the failed sub-jobs on ice if it’s not a mandatory/critical dependent job
Issue Details: If not able to see the workflow status after running job in workflow monitor,
getting error while opening.
Trouble shooting:
There is a network down reason issue and testing team need to contact to support team for
Informatica Power Center.