Infa Practice Test 1
Infa Practice Test 1
For the below scenario which type of Reference Table should be created.
Business has requested to load Product reference data which is currently in a flat file ( 6000 records) to a table, the reference table data will be
updated by using IDQ jobs and the users will check the updates through Informatica Analyst and may edit the data if any correction is needed?
UnManaged Editable
(Correct)
Managed
UnManaged
Managed Editable
(Incorrect)
Explanation
All the data which is currently exists in the table can be viewed and edited in the Informatica Analyst and Audit trail (who has inserted/updated
the records) is maintained in the Analyst.
UnManaged Editable: -
Consider the below scenarios,
1. When you have a reference table which is updated by multiple process/ users
2. You have planned to perform a full load to the reference table using external tool and Incremental updates to the reference table is handled by
Informatica.
In the above scenarios Informatica allows users to work with reference data from external database sources. i.e. if you created a reference table in
ODS_Reference Database, you can update the data from ETL tools or manual sqls etc, when you login to Informatica Analyst you can see the
update data which you have performed using externally, and if you update any record from Informatica Analyst GUI, the data updates will reflect
in the table.
When you use UnManaged Editable, data is not copied to staging database, instead managed directly by source system.
UnManaged: -
Consider the scenario in the above example, after you perform the data load from external sources, Business users are reviewing the data in
Informatica Analyst and they have identified few data discrepancies, but they cannot update the data directly from Informatica Analyst, they need
to communicate back to the corresponding team about the data issues.
(If the Reference tables are created with UnManaged Editable option, then only the data can be edited by Business users in this scenario)
Managed: -
All the Reference data which Informatica has provided are by default Managed, Informatica Manages the Metadata for table names and table
structure in the repository. And If you Import csv files to create Reference Table, it will automatically create as MANAGED.
Here are more details on this topic--> https://idwbi.teachable.com/courses/idq-interview-questions/lectures/5503666
Question 2: Incorrect
Probabilistic matching techniques can be used in
Note: There can be multiple correct answers to this question.
Standardizer - Remove/Replace custom strings
Classifier
(Incorrect)
Labeler - Token execution mode
(Correct)
Parser - Token Parser Mode
(Correct)
Explanation
Probabilistic matching techniques can be used in Labeler - Token Execution mode and Parser - Token Parser Mode.
Parsing or standardizing works correctly only if a column has a clear identifier such as an address or a company name. It requires an exact match
to the reference table entries to make a positive match. For a given input, if these identifiers are not very clear, you need numerous
transformations to parse and label the data.
In addition, you need reference tables to identify the known values, token sets for different token types (word, number, and so on), regular
expressions for custom data structures and patterns to split data by known patterns and their frequency of occurrence.
The probabilistic model predicts the relationships between words that allow the labeling of ambiguous data and to reduce the complexity of Data
Quality mappings. A probabilistic model also improves performance and gives better results.
Here are more details on this topic--> https://idwbi.teachable.com/courses/idq-interview-questions/lectures/3707049
Question 3: Incorrect
Select all the Data Quality transformations which supports parameterized reference tables?
Note: There can be multiple correct answers to this question
Parser Transformation - Pattern-based parsing
(Incorrect)
Address Validator
(Incorrect)
Standardizer
(Correct)
Parser Transformation - Token Parsing Mode
(Correct)
Case
(Correct)
Explanation
Informatica Data Quality support parameterized reference table, you can use parameterized reference tables in Standardizer,Labeler, Parser
Transformation - Token Parsing Mode, Case (Title Case and Sentence Case) transformations. Here are more details on this topic :-
https://idwbi.teachable.com/courses/idq-interview-questions/lectures/6229418
Question 4: Incorrect
Using Data Quality for Excel, users can
Write data quality rules in excel and export the rules as .xml which can be imported in to data quality as mapplet objects.
Use the data quality rules created in Informatica developer and process the source data to find data quality issues, duplicates,
and the Survivor records.
Use the data quality rules created in Informatica developer and apply the rules directly from excel
(Correct)
Use excel spreadsheets directly as Reference data and Lookup Objects, instead of using .dic files
(Incorrect)
Explanation
DQ for excel helps non Informatica users to use the data quality rules created in IDQ and validate data in excel.
DQ for excel helps non Informatica users to use the data quality rules created in IDQ and validate data in excel. DQ for excel helps users to reuse
the data quality rules and perform data validation from their PC and check how the source data will get transformed in the target data objects.
In IDQ, developer has created a mapplet which will standardize data based on the reference data and parse address or customer names.
Developers can save the mapplet as web service and give the web service ( after deploying the web service to application). Users can use the DQ
for excel and create a new service(Add) using the web service url and validate the data in excel.
Users can create a batch of 100 records and validate the data using DQ for Excel. The web service Maximum Occurrences should be configured
to unbounded to allow the web service to process multiple records in one batch.
Here are more details on this topic :- https://idwbi.teachable.com/courses/idq-interview-questions/lectures/6282430
Question 5: Incorrect
You have MRS,DIS,CMS in a Informatica domain, Customer has requested you to create a new CMS and reuse the existing MRS and DIS and
parse the address data, based on the results, the final approach will be planned.
The Jobs in the current environment are critical and should process the data in real time, customer does not want the new jobs to cause any
Impact to the existing jobs?
select all the statements which are incorrect.
A domain cannot have more than one Content Management Service
(Correct)
Multiple Content Management Services (CMS)s can be associated to One Data Integration Service (DIS).
(Correct)
Multiple Content Management Services (CMS)s cannot be associated to One Data Integration Service (DIS).
Reduce the MaxAddressObjectCount so that the new Address Validation jobs will not Impact the existing Address validation
Jobs
(Incorrect)
Explanation
A domain can have more than one Content Management Service and Multiple Content Management Services (CMS)s cannot be associated
to One Data Integration Service (DIS).
Question 6: Correct
Scorecards can be created on _____
Profile Definitions
(Correct)
Logical Data Objects (LDO)
Reference Tables
Physical Database Objects (PDO)
Explanation
Scorecards can only be created on the profile definitions.
A graphical representation of valid values for a column or the output of a rule in profile results. Use scorecards to measure data quality during
different points in the data quality life cycle.
When you create a scorecard, you can assign a weight to each metric(A metric is a column of a data source or output of a rule that is part of a
scorecard). Create a metric group to categorize related metrics in a scorecard into a set.
Scorecard results are stored in the Profile Warehouse.
Here are more details on this topic--> https://idwbi.teachable.com/courses/idq-interview-questions/lectures/3706720
Question 7: Incorrect
Which of the following is not a valid status to correct exception records in Informatica Analyst?
Reprocess Record
(Incorrect)
Accept Record
Bad Record
(Correct)
Reject Record
Explanation
Users cannot update the status to Bad Record to correct exceptions.
Below are the status which are available in Informatica Analyst to correct exceptions.
Accept:- You determine that the current data is acceptable to the business.
Reject:- You determine that the current data is unacceptable to the business.
Reprocess:- You cannot determine the status of the current data
Here are more details on this topic :- https://idwbi.teachable.com/courses/idq-interview-questions/lectures/6288367
Question 8: Incorrect
Human Task, tasks can be distributed by
By number of records in the groups
By number of records
By number of Items in the groups
(Incorrect)
By number of items
(Correct)
Explanation
Tasks can be distributed by Number of Items or by data value.
From v 10.x The human task metadata will be stored in the workflow connection specified in DIS properties. Here are more details on this
topic :- https://idwbi.teachable.com/courses/idq-interview-questions/lectures/6286776
Question 9: Correct
During the PowerCenter upgrade process, the PowerCenter Repository database contents should be converted in to a format which can be
accessed by Model Repository Service?
True
False
(Correct)
Explanation
PowerCenter and Model Repository are configured in two separate databases. As per your requirement you can upgrade both PowerCenter and
Model Repository together or one at each time. If the PowerCenter mappings have any IDQ mapplets those are treated as PowerCenter objects
and will be handled in the PowerCenter upgrade process.
Here are more details on this topic--> https://idwbi.teachable.com/courses/idq-interview-questions/lectures/6270034
Question 10: Incorrect
Regular expressions can be used in
Note: There can be multiple correct answers to this question.
Labeler - Character Mode
(Incorrect)
Parser - Pattern Based Parser
(Incorrect)
Labeler - Token Mode
(Correct)
Parser - Token Mode
(Correct)
Explanation
A regular expression is a special sequence of characters that helps to match an input pattern, using a specialized syntax held in a pattern. Below
are few examples
\d A digit: [0-9]
\D A non-digit: [^0-9]
\w A word character: [a-zA-Z_0-9]
\W A non-word character: [^\w]
IDQ follows Java standards, any Java regular expression can be used in IDQ.
Here are more details on this topic :- https://idwbi.teachable.com/courses/idq-interview-questions/lectures/3707023
Question 11: Incorrect
Select all the tasks which are performed by MRS
Stores metadata created by Informatica Developer, Informatica Analyst and Data Integration Service in a relational
database.
(Correct)
Runs mappings in a deployed application
Generates data previews in Informatica Developer
(Incorrect)
Runs mappings in a deployed application
Explanation
DIS executes the mapping and generates data previews.
Informatica Architecture is a Service Oriented Architecture, it will have a Service Consumer (sends a request) and Service Provider (receives a
response). Services communicates with each other by passing data and coordinating activities.
The Informatica domain is a collection of Nodes (logical representation of the machine on which Informatica server is Installed) and Application
services.
Example of Application Services are
Model Repository Service
Data Integration Service
Analyst Service
Content Management Service
Application services (Configured in the Informatica Administrator) provides the functionalities of the Informatica platform. External clients
Interact with these services via nodes.
Each Application Service (mentioned above) have Internal dependency and requires to be configure in a specific order.
Model Repository Service should be configured first followed by Data Integration Service, Analysts Service and Content Management Service.
A Model Repository Service database (refereed as Model Repository) must be created where all Model Repository Service metadata is stored
(Profiles, Reference data, Mapplets, Mapping, Workflows etc), Model Repository is also used by Informatica Analyst and Data Integration
Service to store its Metadata. MRS manages the request received for Data Integration Service, Informatica Analyst and Informatica Developer
Data Integration Service (is like the Integration service in PowerCenter) receives requests from Model Repository Service, Analysts Service and
Content Management Service, which executes the jobs (profiles, Workflows, Scorecards etc) use the Data Integration Service as a run time
engine. In Data Integration Service, Profile Warehouse can be configured to store the results of Profiles and Scorecards.
Analyst Service allows Business users to perform Adhoc profiling and score carding of the data, maintain the contents of reference data and
handling the exception data.
Content Management Service manages the Reference data like the Address Doctor files, Reference Files, Data required for Identity Matching and
Probabilistic and Classification models. Content Management Service can only be associated with a specific Model Repository Service and Data
Integration Service.
Services should be created in the below order.
Model Repository Service
Data Integration Service
Content Management Service
Analyst Service
Here are more details on this topic :- https://idwbi.teachable.com/courses/idq-interview-questions/lectures/3706748
Question 12: Incorrect
For mapping task, only one one parameter set can be added and deployed in the same application
True
(Correct)
False
(Incorrect)
Explanation
Parameter Sets are similar to Parameter files. In Parameter files you define the required parameters and tag the parameters to the mappings or
workflows.
The Issue with Parameter files is they have to be placed on a server and the same file has to be updated when you deploy to different
environments. Using Parameter Sets you don't have to create any additional files, the data is stored in the Model Repository.
Use the addParameterSetEntries command to add additional parameters or use the UpdateParameterSetEntries to update Parameter values for the
existing parameters, when you deploy the IDQ objects to different environments you should execute the UpdateParameterSetEntries command to
update the parameters according to the deployed environment.
Also for each mapping task in the workflow, you can use only one parameter set.
Multiple Parameter sets can be deployed in one application, and using infacmd command with -ps you can choose which parameter set to be
used.
Here are more details on this topic :- https://idwbi.teachable.com/courses/idq-interview-questions/lectures/6316748
Question 13: Incorrect
Select all the statements which are correct about the Merge Transformation?
Note: There can be multiple correct answers to this question.
Merge Transformation retains the datatypes of all the Input data ports and combine the fields. The output data type can only
be defined as string
(Incorrect)
Same Input port can not be included in multiple strategies in Merge Transformation.
Merge transformation is similar to the Concatenate function (available in PowerCenter and Data Quality).
(Correct)
Merge Transformation converts all the Input data ports (from the upstream transformation) to string before combining the
fields.
(Correct)
Explanation
Merge transformation is similar to the Concatenate function (available in PowerCenter and Data Quality).
In Merge Transformation String datatype is created by default for every port, users can’t change the datatypes for any ports (input and output).
Any non-string port from the upstream transformations will also be converted to string because the merge functions coverts the data to string
before merging the data from different fields.
Here are more details on this topic--> https://idwbi.teachable.com/courses/idq-interview-questions/lectures/3707505
Question 14: Correct
If source data has address data in the below format, which Address Validation Template should be used?
Hybrid Address Validation
Multi line Address Validation
(Correct)
Discrete Address Validation
Address Rows Validation
Explanation
Use the Multi line Template If source data is split across multiple lines and the data flows through different attributes.
Multi line Address Validation Input Template accepts source data containing multiple address elements. Use the Multi line Template If source
data is split across multiple lines and the data flows through different attributes.
Source file has the Customer full name as Mr Joe Smith ( sample source record :- 1001|Issued|Mr Joe Smith|100.34|New|UY76200T)
From the Labeler, the LabeledOutput is as below.
Pattern Based parser will provide full flexibility to customize each individual pattern output
Here are more details on this topic--> https://idwbi.teachable.com/courses/idq-interview-questions/lectures/3709850
Question 33: Incorrect
From Informatica Developer 10.x, objects can only be exported to PowerCenter 10.x?
False
(Correct)
True
(Incorrect)
Explanation
From Informatica Developer 10.x, users can export objects to PowerCenter 9.5.1, PowerCenter 9.6.1 & PowerCenter 10.x. Here are more details
on this topic--> https://idwbi.teachable.com/courses/idq-interview-questions/lectures/6270087
Question 34: Incorrect
Which of the following is not a valid Task status (highlighted in blue color box) in Informatica Analyst?
On Schedule
Elapsed
(Correct)
Assigned
(Incorrect)
OverDue
Explanation
Elapsed is not a valid Task status
Question 35: Correct
All the statements are correct w.r.t Profile and Scorecard, except
Note: There can be multiple correct answers to this question.
Profile can be deleted without Impacting Scorecard
If Profile is deleted, the dependent Scorecard will be Invalidated
(Correct)
From Informatica v 10.x, users can define decimal values in Scorecard Metric Thresholds
(Correct)
If data source object is deleted then both the profile and the Scorecard will be Invalidated
Explanation
Scorecard Metrics cannot have decimal values and if profile is deleted the scorecards will be invalidated.Here are more details on this topic :-
https://idwbi.teachable.com/courses/idq-interview-questions/lectures/3706722
Question 36: Incorrect
Which IDQ transformations can generate Score
Note: There can be multiple correct answers to this question.
Labeler
(Correct)
Comparative
(Correct)
Classifier
(Correct)
Parser
(Correct)
Explanation
Below transformations can generate scores in IDQ.
Classifier
Comparative
Exception
Parser
Labeler
Match
Weighted Average
Here are more details on this topic :- https://idwbi.teachable.com/courses/idq-interview-questions/lectures/6273788
Question 37: Incorrect
____ is the service which is introduced from Informatica version 10.x?
System Services
(Correct)
Search Service
Queue Service
(Incorrect)
Reporting and Dashboard Service
Explanation
In IDQ 10.x, the domain includes system services. A system service is an application service that can have a single instance in the domain.
Below are the new features in IDX 10.x
Generate and Execute DDL: - You can create tables in a database by generating and executing a DDL script. (Similar to the options available in
Target designer in Informatica PowerCenter). DDL scripts can be generated for one or more relational objects, and if a relation object already
exists in the database, you can drop the target and re-create it.
Configure the mappings as Dynamic: - You can configure dynamic mappings, i.e. you can create mappings with dynamic sources and targets
that allow metadata changes to the data sources. When you configure a source or target to be dynamic, the Data Integration Service can interpret
metadata changes to relational and flat file data sources at run time.
Support to the PowerCenter Transformations: - You can import the following PowerCenter transformations into the Developer
tool.Normalizer transformationSequence Generator transformationUpdate Strategy transformation
Object Versioning: - In IDQ 10.x when the Model repository is integrated with a version control system, the version control system protects
objects from being overwritten by other members of the development team. You can check objects out and in, undo the checkout of objects, and
view and restore historical versions of objects.
Terminate Event Task :- From Informatica v 10.1.1, users can configure Terminate Event task to Abort the workflow, if any mapping tasks fails
in the workflow.
System Services: - From IDQ 10.x, the domain includes system services. A system service is an application service that can have a single
instance in the domain. The domain includes the following system services:Email ServiceResource Manager ServiceScheduler Service
Resource Manager Service: - The Resource Manager Service manages computing resources in the domain and dispatches jobs to achieve
optimal performance and scalability. The Resource Manager Service collects information about nodes with the compute role. The service matches
job requirements with resource availability to identify the best compute node to run the job.
Schedule Workflows: - You can schedule the profiles, scorecards, deployed mappings and deployed workflows using a Scheduler Service.
Target Load Order Constraints: - You can configure constraints to control the order in which rows are loaded and committed across target
instances in a mapping.
Inclusive Gateway: - From IDQ v 10.1.1, You can configure to run multiple tasks in parallel using Inclusive Gateway task. For example, you
can configure an Inclusive gateway with multiple branches and add a Mapping task to each branch. When the workflow runs, the Data Integration
Service concurrently runs the mappings that the tasks identify.
New Monitoring Views in Monitoring Tool: - Below are the new features in the monitoring toolSummary Statistics view - Displays resource
usage, object distribution, and object states for a selected time range.Detailed Statistics View: - Displays graphs of throughput and resource usage
information for the source and target.Create customized views on the Execution summary views
Address Validator Transformation:- Effective in version 10.0, you can define parameters to set the following transformation
properties:Geocode data typeGlobal Max Field LengthMax Result CountOptimization LevelStandardize Invalid Address
For more details check --> https://idwbi.teachable.com/courses/idq-interview-questions/lectures/3706749
Question 38: Incorrect
Data has to be extracted from various sources and compared against the master data for each load, the existing process is delaying the load
process due to the source data comparison with master data for every load. Customer is insisting to follow the same load approach and asked your
recommendation.
Select the best approach
Standardize the data and perform dual source matching, as data is standardized by removing the noise words, and validating
against the reference tables, the match process will be efficient and load time will be considerably less as compared to
performing the match with out standardization
Use Identity Match with Persistent RecordID which uses Universal ID which allows users to store matching keys to make the
subsequent matching process efficient. It will significantly improve the performance when source data is matched against
master data regularly and the speed of the match operation is critical
(Correct)
Reading the data from two pipelines and performing a dual source is not recommended, when time is a constraint. Use Union
transformation to combined the data from both the sources and perform single source Field matching and in consolidation
transformation use row based strategy with modal exact option to prioritize the data from master source.
All the other approaches mentioned will not reduce the load window, as match process is a complex process. Instead of using
match transformation use Comparison transformation to compare strings between source data and master data and based on
the score generated create a ETL process to load as per your requirement.
(Incorrect)
Explanation
Use Identity Match with Persistent RecordID which uses Universal ID which allows users to store matching keys to make the subsequent
matching process efficient. It will significantly improve the performance when source data is matched against master data regularly and the speed
of the match operation is critical
Question 39: Incorrect
________contain key building algorithms that have been developed and optimized for specific countries, language and for each class of data.
Address reference data
(Incorrect)
Identity populations
(Correct)
Classification models
Probabilistic models
Explanation
The Content Management Service manages the following types of reference data:
Address reference data: -You use address reference data when you want to validate the postal accuracy of an address or fix errors in an address.
Use the Address Validator transformation to perform address validation.
Identity populations: -You use identity population data when you want to perform duplicate analysis on identity data. An identity is a set of
values within a record that collectively identify a person or business. Use a Match transformation or Comparison transformation to perform
identity duplicate analysis.
Probabilistic models and classifier models: - You use probabilistic or classifier model data when you want to identify the type of information
that a string contains. Use a probabilistic model in a Parser or Labeler transformation. Use a classifier model in a Classifier transformation.
Probabilistic models and classifier models use probabilistic logic to identify or infer the type of information in the string. Use a Classifier
transformation when each input string contains a significant amount of data.
Reference tables: - You use reference tables to verify the accuracy or structure of input data values in data quality transformations.
In addition to the above, Informatica also provides below Content to customers based on the Subscription.
Data Domains Accelerator
Australia/New Zealand Accelerator
Brazil Accelerator
Financial Services Accelerator
France Accelerator
Germany Accelerator
Portugal Accelerator
Spain Accelerator
United Kingdom Accelerator
U.S./Canada Accelerator
Core Accelerator
Core Accelerator :- The Core accelerator contains mapplets, rules and reference data objects that other accelerators can reuse.
Address data cleansing
Contact data cleansing
Corporate data cleansing
General data cleansing
Matching and deduplication data cleansing
Product data cleansing
Here are more details on this topic :- https://idwbi.teachable.com/courses/idq-interview-questions/lectures/3706824
Question 40: Incorrect
Matching process cannot be disabled for which of the following Persistence Methods in Identity Match?
Remove IDs from the database
Update the database with new IDs
(Incorrect)
Update the current IDs in the database
Do not update the database
(Correct)
Explanation
If you want to perform match against the store with out updating, you can select Do not update the database as Matching Process. By default the
matching will be enabled and you cannot disable the match process
Question 41: Incorrect
______input is mandatory for the match process?
ClusterId
(Incorrect)
GroupKey
SequenceId
(Correct)
RowId
Explanation
Yes, SequenceId is mandatory for the match process. Every record in an input data set must include a unique sequence identifier. If a data set
contains duplicate sequence identifiers, the Match transformation cannot identify duplicate records correctly. Use the Key Generator
transformation to create unique identifiers if none exist in the data.
When you create an index data store for identity data, the Match transformation adds the sequence identifier for each record to the data store.
When you configure the transformation to compare a data source with the index data store, the transformation might find a common sequence
identifier in both data sets. The transformation can analyze the sequence identifiers if they are unique in the respective data sets.
Note: - The Match Mapping will be Invalid without connecting the SequenceId port from the upstream mapping.
Here are more details on this topic --> https://idwbi.teachable.com/courses/idq-interview-questions/lectures/3717453
Question 42: Incorrect
What is the best approach to validate the Address data?
Note: There can be multiple correct answers to this question.
Use Standardizer transformation to remove noise words and standardize data before sending the data to Address Validator.
(Correct)
For each country create a separate Address Validator transformation to validate the address data
For highly complex and discrete attributes use Basic template model with Discrete Address Validation Input Template
(Incorrect)
For Australian address, Informatica Address Doctor uses AMAS (Address Matching Approval System) rules for validation of
address when the Input mode is configured as ‘Certified’.
(Correct)
Explanation
For Australian address, Informatica Address Doctor uses AMAS (Address Matching Approval System) rules for validation of address when the
Input mode is configured as ‘Certified’.
The certification mode required additional database files (AUS5C1.MD and AUS5C2.MD, which Informatica will provide and renew based on
the subscription) and SetConfig.xml should be configured Section <DataBase> with Type=” CERTIFIED” for ISO=” AUS” or “ALL”.
The databases contain Postal Address File (PAF) data which includes Australia Post’s Delivery Point Identifiers (DPIDs).
Here are more details on this topic--> https://idwbi.teachable.com/courses/idq-interview-questions/lectures/3707393
Question 43: Incorrect
Which of the following is not a valid Human Task step?
Review step
Exception step
Publish step
(Correct)
Cluster step
(Incorrect)
Explanation
Below steps can be created in Human Task
Cluster step:- Create a Cluster step when you want a user to examine duplicate record clusters and create a preferred record from the values in the
duplicate records.
Exception step:- Create an Exception step when you want a user to examine and fix errors in records.
Review step:- Create a Review step when you want a user to review the work done in an Exception or Cluster step. Review steps are not
mandatory.
Here are more details on this topic :- https://idwbi.teachable.com/courses/idq-interview-questions/lectures/6289031
Question 44: Incorrect
Select all the IDQ objects which cannot be exported to PowerCenter?
Note: There can be multiple correct answers to this question.
Human Tasks
(Correct)
Workflows
(Correct)
Rules created in Developer
Logical Data Objects
(Incorrect)
Explanation
Workflows and Human Tasks cannot be exported to PowerCenter.
No, not all objects can be exported from IDQ to PowerCenter.
For example :- Workflows ,Human Tasks, Connections and Lookups with Customized SQls created in Mapplets cannot be exported to
PowerCenter.
Also, the parameter files which are used in IDQ are not exported to PowerCenter, because the format of parameter files are different between
IDQ and PowerCenter. The parameters resolve to their default values when IDQ mappings/mapplets with parameters are imported to
PowerCenter.
Here are more details on this topic :- https://idwbi.teachable.com/courses/idq-interview-questions/lectures/6279663
Question 45: Incorrect
Select the correct statement w.r.t Scorecards?
Scorecard results are stored in Model Repository as part of Profile definition.
(Incorrect)
Scorecard results are displayed only on GUI and are not stored in any database.
Scorecard results are stored in the Staging Database.
Scorecard results are stored in the Profile Warehouse.
(Correct)
Explanation
Scorecard results are stored in the Profile Warehouse.
A graphical representation of valid values for a column or the output of a rule in profile results. Use scorecards to measure data quality during
different points in the data quality life cycle.
When you create a scorecard, you can assign a weight to each metric(A metric is a column of a data source or output of a rule that is part of a
scorecard). Create a metric group to categorize related metrics in a scorecard into a set.
Scorecard results are stored in the Profile Warehouse.
Here are more details on this topic :- https://idwbi.teachable.com/courses/idq-interview-questions/lectures/3706720
Question 46: Incorrect
Which transformation creates links between records that share duplicate characteristics across more than one data field
Merge Transformation
Consolidation Transformation
Match Transformation
(Incorrect)
Association Transformation
(Correct)
Explanation
Association Transformation is an extension of the data matching process which process output data from a match transformation.
The Association transformation creates links between records that share duplicate characteristics across more than one data field so that they are
treated as members of a single set in data consolidation operations.
Here are more details on this topic--> https://idwbi.teachable.com/courses/idq-interview-questions/lectures/3717535
Question 47: Incorrect
Select all the Validation modes which can be configured using AddressDoctor Webservice?
Note: There can be multiple correct answers to this question.
Certified
Batch
(Correct)
Parse
(Incorrect)
Interactive
(Correct)
Explanation
Yes,You can develop a Web Service to validate addresses. This Web Service could be accessed by third party applications.You can also validate
address using AddressDoctor Cloud and AddressDoctor Software Library.
Below URL will have the wsdl file details.
http://validator2.addressdoctor.com/addBatch/Index.html
Using AddressDoctor Webservice, you can configure Batch,FastCompletion and Interactive validation modes
Here are more details on this topic--> https://idwbi.teachable.com/courses/idq-interview-questions/lectures/6270656
Question 48: Incorrect
Select all the options which are required to use DQ for Excel?
Data Transformation Excel License option
(Incorrect)
Web Services option
(Correct)
DQ for Excel installation on PC
(Correct)
MS Excel version 2007 or 2010
(Correct)
Explanation
Below are required to use DQ for Excel.
DQ for Excel installation on PC
MS Excel version 2007 or 2010
Web Services option
Here are more details on this topic :- https://idwbi.teachable.com/courses/idq-interview-questions/lectures/6288332
Question 49: Correct
Below objects can be Imported from PowerCenter to MRS, expect
Relational Connections
(Correct)
Mappings
Mapplets
Reusable Transformations
Explanation
Connections should be manually configured. Connections are not exported when objects are exported from PowerCenter to IDQ and vice
versa.Here are more details on this topic :- https://idwbi.teachable.com/courses/idq-interview-questions/lectures/6276462
Question 50: Incorrect
As per the requirement, the Customer full name has to be parsed to Prefix, FirstName and LastName.
In the Labeler transformation, below LabeledOutput and TokenizedData are generated.
But the actual output is as below, the Customer Name is not parsed and the completed string is shown as Overflow
Pattern Based parser will provide full flexibility to customize each individual pattern output
Here are more details on this topic :- https://idwbi.teachable.com/courses/idq-interview-questions/lectures/3709850
Question 51: Incorrect
Records with score =1 will flow through the Unique Records group and records with 0.82 will flow in the Cluster Output
group
(Incorrect)
All the records will flow through the Standard output group
All the four records are duplicate and will flow through the Cluster Output group. These records should be manually
reviewed in the Analyst tool.
(Correct)
Generate duplicate record table and validate if the records are duplicate, if yes, configure the records to be reviewed by
Business Analyst by setting up Unmanaged Editable Reference table in Informatica Analyst.
Explanation
Duplicate records will flow through the Cluster Output group and Business users have to analyse the records in the Analyst tool.
Question 52: Incorrect
Are Probabilistic models part of Informatica Base Content set?
True
(Incorrect)
False
(Correct)
Explanation
Informatica doesn’t provide Probabilistic models as part of the base content set.
A probabilistic model is a reference data object. Use a probabilistic model to understand the contents of a data string that contains multiple data
values. A probabilistic model identifies the types of information in each value in the string.
Probabilistic models use Natural Language Processes (NLP) to identify the type of information in a string. NLP detect relevant terms in the input
string and disregard terms that are not relevant.
Here are more details on this topic --> https://idwbi.teachable.com/courses/idq-interview-questions/lectures/3707014
Question 53: Incorrect
Select the attribute which is not required in Exception task processing?
IS_MASTER
(Correct)
REVIEW_STATUS
WORKFLOW_ID
USER_COMMENT
(Incorrect)
Explanation
IS_MASTER is required for Duplicate exception task.
When you correct exceptions task or review exceptions task, below attributes should be present in the table. If below attributes are not present,
users cannot open the created exception task in Informatica Analyst,Informatica will display error stating the required column is missing.
To correct exception, below attributes should exist in the table.
ROW_IDENTIFIER
REVIEW_STATUS
WORKFLOW_ID
USER_COMMENT
UPDATED_STATUS
RECORD_STATUS
Here are more details on this topic :- https://idwbi.teachable.com/courses/idq-interview-questions/lectures/6285055
Question 54: Incorrect
Customer has requested to verify the address attributes of Singapore, India, Belgium and United States the address information is formatted on a
single line (line1) with comma separator and country name is on different line (line2), based on the Input data Customer is expecting 5-7
suggestions per address. select all the statements which are incorrect?
It is recommended to enter the single line address elements in the correct sequence (example :- House Number , Street No,
Postal Code, Locality)
Country Name is optional when you pass complete address info in Suggestion List mode
Single line address validation is not available in Suggestion List mode.
(Correct)
Single line address validation is available for United States, Singapore and Belgium
(Correct)
Explanation
Single line address validation is available for United States, Singapore and Belgium
Question 55: Incorrect
Select the transformations which support parameterizing the Cache directory?
Address Validator
Key Generator
(Correct)
Decision
(Incorrect)
Association
(Correct)
Explanation
Below are some of the transformations which support cache directory parameterizing
Association
Consolidation
Duplicate Record Exception
Key Generator
Match
Here are more details on this topic :- https://idwbi.teachable.com/courses/idq-interview-questions/lectures/6272878
Question 56: Incorrect
Select the expression functions which are not supported by the Decision Transformation?
Note: There can be multiple correct answers to this question.
XOR
(Incorrect)
MonthCompare
AES_DECRYPT
(Correct)
DEC_BASE64
(Correct)
Explanation
DEC_BASE64 and AES_DECRYPT are not supported by the Decision Transformation.
The following expressions are not supported by the Decision Transformation.
ABORT
AES_DECRYPT
AES_ENCRYPT
COMPRESS
DEC_BASE64
DECODE
DECOMPRESS
ERROR
IIF
TO_BIGINT
You cannot remove the association between a tag and an object after the tag is associated with object.
Tags associated with an object in the Developer tool appear as tags for the same objects in the Analyst tool after users
associate the tag with one or more objects in Analyst
(Incorrect)
Tags are specific to Informatica Developer objects.
Tags associated with an object in the Developer tool appear as tags for the same objects in the Analyst tool.
(Correct)
Explanation
Tag is metadata that defines an object in the Model repository based on business usage. Create tags to group objects according to their business
usage.
After you create a tag, you can associate the tag with one or more objects. You can remove the association between a tag and an object. You can
use a tag to search for objects associated with the tag in the Model repository. The Developer tool displays a glossary of all tags.
For example, you create a tag named XYZCorp_CustomerOrders and assign it to tables that contain information for the customer orders from the
XYZ Corporation. Users can search by the XYZCorp_CustomerOrders tag to identify the tables associated with the tag.
Note: Tags associated with an object in the Developer tool appear as tags for the same objects in the Analyst tool.
There are no matching records between Orders and Customers
There might be some Orphan records in Orders or Customer table, due to which all the rows cannot be joined between
Orders and Customers
(Correct)
One of the table is empty, due to which Informatica developer didn't generate any join Analysis, the venn diagram in the
picture represents one of the table is empty
All the key column values exists between Orders and Customers, Orders table should be joined to Customers using a left join
because as per the picture the Order table is shown as driving table.
(Incorrect)
Explanation
Based on the picture, the Order has more number of rows compared to customers and there might be some orphan records in one of the table.
Join analysis describes the degree of potential joins between two data columns. Use a join profile to analyze column joins in a data source or
across multiple data sources. A join profile displays results as a Venn diagram and as numerical and percentage values. You create and run a join
profile from an enterprise discovery profile.
Address reference data: -You use address reference data when you want to validate the postal accuracy of an address or fix errors in an address.
Use the Address Validator transformation to perform address validation.
Identity populations: -You use identity population data when you want to perform duplicate analysis on identity data. An identity is a set of
values within a record that collectively identify a person or business. Use a Match transformation or Comparison transformation to perform
identity duplicate analysis.
Probabilistic models and classifier models: - You use probabilistic or classifier model data when you want to identify the type of information
that a string contains. Use a probabilistic model in a Parser or Labeler transformation. Use a classifier model in a Classifier transformation.
Probabilistic models and classifier models use probabilistic logic to identify or infer the type of information in the string. Use a Classifier
transformation when each input string contains a significant amount of data.
Reference tables: - You use reference tables to verify the accuracy or structure of input data values in data quality transformations.
In addition to the above, Informatica also provides below Content to customers based on the Subscription.
Data Domains Accelerator
Australia/New Zealand Accelerator
Brazil Accelerator
Financial Services Accelerator
France Accelerator
Germany Accelerator
Portugal Accelerator
Spain Accelerator
United Kingdom Accelerator
U.S./Canada Accelerator
Core Accelerator
Core Accelerator :- The Core accelerator contains mapplets, rules and reference data objects that other accelerators can reuse.
Address data cleansing
Contact data cleansing
Corporate data cleansing
General data cleansing
Matching and deduplication data cleansing
Product data cleansing
Here are more details on this topic :- https://idwbi.teachable.com/courses/idq-interview-questions/lectures/3706824
1 A
2 d
3 a
4 b
5 a
6 b
7 c
8 b
9 c
10 d
11 d
12 TRUE
13 c
14 b
15 a
16 a
17 a
18 c
19 b
20 b
21 d
22 ab
23 b
24 b
25 TRUE
26 cde
27 c
28 c
29 d
30 c
31 d
32 c
33 d
34 a
35 b
36 c
37 bd
38 d
39 bd
40 a
41 b
42 A or D
43 a
44 bd
45 b
46 b
47 a
48 a
49 c
50 FALSE
51 d
52 a
53 a
54 d
55 a
56 b
57 c
58 c
59 b
60 d
61 b
62 a
63 c
64 b
65 d
66 b
67 1245
68 c
69 a