0% found this document useful (0 votes)
146 views47 pages

Infa Practice Test 1

Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
146 views47 pages

Infa Practice Test 1

Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
You are on page 1/ 47

Question 1: Incorrect

For the below scenario which type of Reference Table should be created.
Business has requested to load Product reference data which is currently in a flat file ( 6000 records) to a table, the reference table data will be
updated by using IDQ jobs and the users will check the updates through Informatica Analyst and may edit the data if any correction is needed?

UnManaged Editable
(Correct)

Managed

UnManaged

Managed Editable
(Incorrect)
Explanation
All the data which is currently exists in the table can be viewed and edited in the Informatica Analyst and Audit trail (who has inserted/updated
the records) is maintained in the Analyst.
UnManaged Editable: -
Consider the below scenarios,
1. When you have a reference table which is updated by multiple process/ users
2. You have planned to perform a full load to the reference table using external tool and Incremental updates to the reference table is handled by
Informatica.
In the above scenarios Informatica allows users to work with reference data from external database sources. i.e. if you created a reference table in
ODS_Reference Database, you can update the data from ETL tools or manual sqls etc, when you login to Informatica Analyst you can see the
update data which you have performed using externally, and if you update any record from Informatica Analyst GUI, the data updates will reflect
in the table.
When you use UnManaged Editable, data is not copied to staging database, instead managed directly by source system.
UnManaged: -
Consider the scenario in the above example, after you perform the data load from external sources, Business users are reviewing the data in
Informatica Analyst and they have identified few data discrepancies, but they cannot update the data directly from Informatica Analyst, they need
to communicate back to the corresponding team about the data issues.
(If the Reference tables are created with UnManaged Editable option, then only the data can be edited by Business users in this scenario)
Managed: -
All the Reference data which Informatica has provided are by default Managed, Informatica Manages the Metadata for table names and table
structure in the repository. And If you Import csv files to create Reference Table, it will automatically create as MANAGED.
Here are more details on this topic--> https://idwbi.teachable.com/courses/idq-interview-questions/lectures/5503666
Question 2: Incorrect
Probabilistic matching techniques can be used in
Note: There can be multiple correct answers to this question.

Standardizer - Remove/Replace custom strings

Classifier
(Incorrect)

Labeler - Token execution mode
(Correct)

Parser - Token Parser Mode
(Correct)
Explanation
Probabilistic matching techniques can be used in Labeler - Token Execution mode and Parser - Token Parser Mode.
Parsing or standardizing works correctly only if a column has a clear identifier such as an address or a company name. It requires an exact match
to the reference table entries to make a positive match. For a given input, if these identifiers are not very clear, you need numerous
transformations to parse and label the data.
In addition, you need reference tables to identify the known values, token sets for different token types (word, number, and so on), regular
expressions for custom data structures and patterns to split data by known patterns and their frequency of occurrence.
The probabilistic model predicts the relationships between words that allow the labeling of ambiguous data and to reduce the complexity of Data
Quality mappings. A probabilistic model also improves performance and gives better results.
Here are more details on this topic--> https://idwbi.teachable.com/courses/idq-interview-questions/lectures/3707049
Question 3: Incorrect
Select all the Data Quality transformations which supports parameterized reference tables?
Note: There can be multiple correct answers to this question

Parser Transformation - Pattern-based parsing
(Incorrect)

Address Validator
(Incorrect)

Standardizer
(Correct)

Parser Transformation - Token Parsing Mode
(Correct)

Case
(Correct)
Explanation
Informatica Data Quality support parameterized reference table, you can use parameterized reference tables in Standardizer,Labeler, Parser
Transformation - Token Parsing Mode, Case (Title Case and Sentence Case) transformations. Here are more details on this topic :-
https://idwbi.teachable.com/courses/idq-interview-questions/lectures/6229418
Question 4: Incorrect
Using Data Quality for Excel, users can

Write data quality rules in excel and export the rules as .xml which can be imported in to data quality as mapplet objects.

Use the data quality rules created in Informatica developer and process the source data to find data quality issues, duplicates,
and the Survivor records.

Use the data quality rules created in Informatica developer and apply the rules directly from excel
(Correct)

Use excel spreadsheets directly as Reference data and Lookup Objects, instead of using .dic files
(Incorrect)
Explanation
DQ for excel helps non Informatica users to use the data quality rules created in IDQ and validate data in excel.
DQ for excel helps non Informatica users to use the data quality rules created in IDQ and validate data in excel. DQ for excel helps users to reuse
the data quality rules and perform data validation from their PC and check how the source data will get transformed in the target data objects.
In IDQ, developer has created a mapplet which will standardize data based on the reference data and parse address or customer names.
Developers can save the mapplet as web service and give the web service ( after deploying the web service to application). Users can use the DQ
for excel and create a new service(Add) using the web service url and validate the data in excel.
Users can create a batch of 100 records and validate the data using DQ for Excel. The web service Maximum Occurrences should be configured
to unbounded to allow the web service to process multiple records in one batch.
Here are more details on this topic :- https://idwbi.teachable.com/courses/idq-interview-questions/lectures/6282430
Question 5: Incorrect
You have MRS,DIS,CMS in a Informatica domain, Customer has requested you to create a new CMS and reuse the existing MRS and DIS and
parse the address data, based on the results, the final approach will be planned.
The Jobs in the current environment are critical and should process the data in real time, customer does not want the new jobs to cause any
Impact to the existing jobs?
select all the statements which are incorrect.

A domain cannot have more than one Content Management Service
(Correct)

Multiple Content Management Services (CMS)s can be associated to One Data Integration Service (DIS).
(Correct)

Multiple Content Management Services (CMS)s cannot be associated to One Data Integration Service (DIS).

Reduce the MaxAddressObjectCount so that the new Address Validation jobs will not Impact the existing Address validation
Jobs
(Incorrect)
Explanation
A domain can have more than one Content Management Service and Multiple Content Management Services (CMS)s cannot be associated
to One Data Integration Service (DIS).
Question 6: Correct
Scorecards can be created on _____

Profile Definitions
(Correct)

Logical Data Objects (LDO)

Reference Tables

Physical Database Objects (PDO)
Explanation
Scorecards can only be created on the profile definitions.
A graphical representation of valid values for a column or the output of a rule in profile results. Use scorecards to measure data quality during
different points in the data quality life cycle.

When you create a scorecard, you can assign a weight to each metric(A metric is a column of a data source or output of a rule that is part of a
scorecard). Create a metric group to categorize related metrics in a scorecard into a set.
Scorecard results are stored in the Profile Warehouse.
Here are more details on this topic--> https://idwbi.teachable.com/courses/idq-interview-questions/lectures/3706720
Question 7: Incorrect
Which of the following is not a valid status to correct exception records in Informatica Analyst?

Reprocess Record
(Incorrect)

Accept Record

Bad Record
(Correct)

Reject Record
Explanation
Users cannot update the status to Bad Record to correct exceptions.
Below are the status which are available in Informatica Analyst to correct exceptions.
Accept:- You determine that the current data is acceptable to the business.
Reject:- You determine that the current data is unacceptable to the business.
Reprocess:- You cannot determine the status of the current data
Here are more details on this topic :- https://idwbi.teachable.com/courses/idq-interview-questions/lectures/6288367
Question 8: Incorrect
Human Task, tasks can be distributed by

By number of records in the groups

By number of records

By number of Items in the groups
(Incorrect)

By number of items
(Correct)
Explanation
Tasks can be distributed by Number of Items or by data value.

From v 10.x The human task metadata will be stored in the workflow connection specified in DIS properties. Here are more details on this
topic :- https://idwbi.teachable.com/courses/idq-interview-questions/lectures/6286776
Question 9: Correct
During the PowerCenter upgrade process, the PowerCenter Repository database contents should be converted in to a format which can be
accessed by Model Repository Service?

True

False
(Correct)
Explanation
PowerCenter and Model Repository are configured in two separate databases. As per your requirement you can upgrade both PowerCenter and
Model Repository together or one at each time. If the PowerCenter mappings have any IDQ mapplets those are treated as PowerCenter objects
and will be handled in the PowerCenter upgrade process.
Here are more details on this topic--> https://idwbi.teachable.com/courses/idq-interview-questions/lectures/6270034
Question 10: Incorrect
Regular expressions can be used in
Note: There can be multiple correct answers to this question.

Labeler - Character Mode
(Incorrect)

Parser - Pattern Based Parser
(Incorrect)

Labeler - Token Mode
(Correct)

Parser - Token Mode
(Correct)
Explanation
A regular expression is a special sequence of characters that helps to match an input pattern, using a specialized syntax held in a pattern. Below
are few examples
\d A digit: [0-9]
\D A non-digit: [^0-9]
\w A word character: [a-zA-Z_0-9]
\W A non-word character: [^\w]
IDQ follows Java standards, any Java regular expression can be used in IDQ.
Here are more details on this topic :- https://idwbi.teachable.com/courses/idq-interview-questions/lectures/3707023
Question 11: Incorrect
Select all the tasks which are performed by MRS

Stores metadata created by Informatica Developer, Informatica Analyst and Data Integration Service in a relational
database.
(Correct)

Runs mappings in a deployed application

Generates data previews in Informatica Developer
(Incorrect)

Runs mappings in a deployed application
Explanation
DIS executes the mapping and generates data previews.
Informatica Architecture is a Service Oriented Architecture, it will have a Service Consumer (sends a request) and Service Provider (receives a
response). Services communicates with each other by passing data and coordinating activities.
The Informatica domain is a collection of Nodes (logical representation of the machine on which Informatica server is Installed) and Application
services.
Example of Application Services are
Model Repository Service
Data Integration Service
Analyst Service
Content Management Service
Application services (Configured in the Informatica Administrator) provides the functionalities of the Informatica platform. External clients
Interact with these services via nodes.
Each Application Service (mentioned above) have Internal dependency and requires to be configure in a specific order.
Model Repository Service should be configured first followed by Data Integration Service, Analysts Service and Content Management Service.
A Model Repository Service database (refereed as Model Repository) must be created where all Model Repository Service metadata is stored
(Profiles, Reference data, Mapplets, Mapping, Workflows etc), Model Repository is also used by Informatica Analyst and Data Integration
Service to store its Metadata. MRS manages the request received for Data Integration Service, Informatica Analyst and Informatica Developer
Data Integration Service (is like the Integration service in PowerCenter) receives requests from Model Repository Service, Analysts Service and
Content Management Service, which executes the jobs (profiles, Workflows, Scorecards etc) use the Data Integration Service as a run time
engine. In Data Integration Service, Profile Warehouse can be configured to store the results of Profiles and Scorecards.
Analyst Service allows Business users to perform Adhoc profiling and score carding of the data, maintain the contents of reference data and
handling the exception data.
Content Management Service manages the Reference data like the Address Doctor files, Reference Files, Data required for Identity Matching and
Probabilistic and Classification models. Content Management Service can only be associated with a specific Model Repository Service and Data
Integration Service.
Services should be created in the below order.
Model Repository Service
Data Integration Service
Content Management Service
Analyst Service
Here are more details on this topic :- https://idwbi.teachable.com/courses/idq-interview-questions/lectures/3706748
Question 12: Incorrect
For mapping task, only one one parameter set can be added and deployed in the same application

True
(Correct)

False
(Incorrect)
Explanation
Parameter Sets are similar to Parameter files. In Parameter files you define the required parameters and tag the parameters to the mappings or
workflows.
The Issue with Parameter files is they have to be placed on a server and the same file has to be updated when you deploy to different
environments. Using Parameter Sets you don't have to create any additional files, the data is stored in the Model Repository.
Use the addParameterSetEntries command to add additional parameters or use the UpdateParameterSetEntries to update Parameter values for the
existing parameters, when you deploy the IDQ objects to different environments you should execute the UpdateParameterSetEntries command to
update the parameters according to the deployed environment.
Also for each mapping task in the workflow, you can use only one parameter set.
Multiple Parameter sets can be deployed in one application, and using infacmd command with -ps you can choose which parameter set to be
used.
Here are more details on this topic :- https://idwbi.teachable.com/courses/idq-interview-questions/lectures/6316748
Question 13: Incorrect
Select all the statements which are correct about the Merge Transformation?
Note: There can be multiple correct answers to this question.

Merge Transformation retains the datatypes of all the Input data ports and combine the fields. The output data type can only
be defined as string
(Incorrect)

Same Input port can not be included in multiple strategies in Merge Transformation.

Merge transformation is similar to the Concatenate function (available in PowerCenter and Data Quality).
(Correct)

Merge Transformation converts all the Input data ports (from the upstream transformation) to string before combining the
fields.
(Correct)
Explanation
Merge transformation is similar to the Concatenate function (available in PowerCenter and Data Quality).
In Merge Transformation String datatype is created by default for every port, users can’t change the datatypes for any ports (input and output).
Any non-string port from the upstream transformations will also be converted to string because the merge functions coverts the data to string
before merging the data from different fields.
Here are more details on this topic--> https://idwbi.teachable.com/courses/idq-interview-questions/lectures/3707505
Question 14: Correct
If source data has address data in the below format, which Address Validation Template should be used?


Hybrid Address Validation

Multi line Address Validation
(Correct)

Discrete Address Validation

Address Rows Validation
Explanation
Use the Multi line Template If source data is split across multiple lines and the data flows through different attributes.
Multi line Address Validation Input Template accepts source data containing multiple address elements. Use the Multi line Template If source
data is split across multiple lines and the data flows through different attributes.

Here are more details on this topic :- https://idwbi.teachable.com/courses/idq-interview-questions/lectures/3707076


Question 15: Incorrect
In Case transformation, in which of the following Case Reference tables can be used?

Lower Case

Upper Case

Merge Case
(Incorrect)

Title Case
(Correct)
Explanation
Reference tables can be used when the case conversion type is Title Case and Sentence case. Here are more details on this topic-->
http://idwbi.teachable.com/courses/idq-interview-questions/lectures/3707693
Question 16: Incorrect
Parameters can not be used in ______

Rules

Mapping specification
(Correct)

Transformations

Scorecards
(Incorrect)
Explanation
Parameters cannot be used in Mapping specification
Question 17: Correct
Can the Input Request and Output SOAP Response be viewed in IDQ,with out executing the mapping?

There is no option in IDQ to view the SOAP message with out executing the mapping.

Data viewer option is disabled for web service transformation

In data viewer IDQ displays the Input SOAP Request, user has to enter the values for the required fields and click Run, the
IDQ GUI will display the Output SOAP Response
(Correct)

IDQ displays the Input SOAP message in data viewer for http based SOAP services, for https data cannot be viewed with out
executing the mapping.
Explanation
Users can click on the output web service transformation and pass the required values in the input operation and execute data viewer to see the
SOAP response.
Question 18: Incorrect
Which Address Validation Modes use Reference Data?
Note: There can be multiple correct answers to this question.

Batch
(Correct)

Certified
(Correct)

Parse
(Incorrect)

Address Code Lookup
(Correct)

Country Recognition
(Incorrect)
Explanation
Address Code Lookup, Batch, Certified, Interactive, Suggestion List. Country Recognition and Parse Mode doesn’t use Reference Data for
address validation. For more detailed explanation - https://idwbi.teachable.com/courses/idq-interview-questions/lectures/3707262
Question 19: Correct
Which transformation is used to remove noise words?
Note: There can be multiple correct answers to this question.

Match Transformation - Identity Match
(Correct)

Standardizer Transformation
(Correct)

Consolidation

Match Transformation - Classic Match
Explanation
Noise words appear before strings like name, address, phone numbers etc. Noise words are considered irrelevant. Common English noise words
are THE, OF, FOR etc.
Noise Words generally have the following characteristics:
They are common
They rarely appear on their own in a name
They are not stably present in the search or file data
They are generally ‘qualification’ words rather than ’naming’ words
They are unlikely to be used on their own for searching
Standardizer and Match (Identity Match) transformations can be used to remove Noise words.
Here are more details on this topic--> https://idwbi.teachable.com/courses/idq-interview-questions/lectures/3709681
Question 20: Incorrect
Which of the following Address Doctor Subscription File can be used in Batch Mode?

xxx5En.MD
(Incorrect)

xxx5GC.MD

xxx5FC.MD

xxx5BI.MD
(Correct)
Explanation
xxx5BI.MD: - Batch Interactive. Used in Batch /Interactive/Certified Mode.
Address Doctor supplies below subscription files:
xxx5BI.MD: - Batch Interactive. Used in Batch /Interactive/Certified Mode
xxx5FC.MD: - Fast Complete. Used in Suggestion List mode
xxx5C1.MD – xxx5cnn.MD: - Certified. Used in Certified Mode (will typically be multiple files)
xxx5GC.MD: - Geocoding
xxx5CA.MD: - Cameo
xxx5En.MD: -Extended/Supplementary (applies to some countries only)
Note: -
1)Informatica Address Doctor refers to suggestion list mode as fast completion mode.
2)xxx in the file names indicates the ISO country code.
3)n - indicates a number.
Here are more details on this topic --> https://idwbi.teachable.com/courses/idq-interview-questions/lectures/3707289
Question 21: Incorrect
In which Data Quality transformations strategies are not created?
Note: There can be multiple correct answers to this question.

Address Validator
(Correct)

Parser

Exception
(Correct)

Weighted Average
(Correct)
Explanation
In Address Validator, Association, Exception and Weighted Average, you cannot create any strategies.
No, not all Data Quality transformations use strategies, some of the transformations have pre-defined templates and some transformation
generates scores based on the Input strings.
For example :- In Address Validator, Association, Exception and Weighted Average, you cannot create any strategies. You configure the
transformation based on the available templates within the transformation.
For more details - https://idwbi.teachable.com/courses/idq-interview-questions/lectures/6267062
Question 22: Correct
Which of the following is not true?
Note: There can be multiple correct answers to this question.

In PowerCenter folders can be created as shared and non shared and in Informatica Developer projects can be created as
shared and non shared

In PowerCenter and Informatica Developer, folders can be created as shared and non shared
(Correct)

In PowerCenter you can drag and drop objects and in Informatica Developer you need to right click and select, to open any
objects from navigator in to the work space

In PowerCenter and Informatica Developer you can drag and drop the objects from navigator in to the work space
(Correct)
Explanation
Here are more details on this topic --> https://idwbi.teachable.com/courses/idq-interview-questions/lectures/4662329
Question 23: Correct
In which Validation Mode,the Address Validator transformation can be executed with default Address Doctor license key?

Country Recognition

Certified

Parse
(Correct)

Suggestion List
Explanation
You can use Address Validator in Parse Validation Mode only with default license.
When Content Management Service (CMS) is created, it creates a default License Key. This key cannot be used for the actual data validation.
This key is used only to unlock the AD engine and to enable Parse Only mode.
Here are more details on this topic--> https://idwbi.teachable.com/courses/idq-interview-questions/lectures/6270820
Question 24: Incorrect
Mapping Specification can be exported to
Note: There can be multiple correct answers to this question.

Excel
(Correct)

IDQ Mapplet
(Incorrect)

Virtual Table
(Correct)

PowerCenter
(Correct)
Explanation
Mapping specifications can be exported to PowerCenter from Informatica Analyst.
you can also export the rules in to a excel document and share with non Informatica users on the business rules used and you can load the data to
a flat file or table after applying the defined rules in the Mapping specifications.

Here are more details on this topic :-


https://idwbi.teachable.com/courses/idq-interview-questions/lectures/6316572
Question 25: Correct
You have received a flat file which has 1000 records, some of the records have both data and label and some have only data info. You need to
create Classifier Model?

Classifier Model can be Imported with data values only, and labels can be assigned manually after the data is Imported from
Developer GUI.
(Correct)

Classifier Model can not be created by importing data from flat file, load the data to a relational table and then create the
Classifier Model.

Classifier Model can not be Imported if input data doesn't have both data and label values.

Classifier Model can not be edited after the model is created and compiled, edit the file which will be created on Informatica
server when Classifier model is created and sync the changes in developed to update the label info.
Explanation
Classifier Model can be Imported with data values only, and labels can be assigned manually after the data is Imported from Developer GUI.
A Classifier model is a reference data object that you can use to identify the type of information in a data record.
When you run a mapping that contains a Classifier transformation, the transformation searches for common data values between the classifier
model and each input record. The transformation uses the common values to classify the information in each record and classifies to which of a
set of categories or sub-populations a new.
Classifier Models can be trained to classify the below
Language Identification
Product Classification
Issue Classification
Security Classification
Identify potential spam
Here are more details on this topic :- https://idwbi.teachable.com/courses/idq-interview-questions/lectures/3707016
Question 26: Incorrect
____ is not a valid Data Quality Dimension?

Conformity
(Incorrect)

Consistency

Completeness

Compliance
(Correct)
Explanation
Compliance is a regulatory requirement.
In order to ensure data is trustworthy, it is important to understand the key dimensions of data quality to assess how the data are “bad” in the first
place.
Completeness: - Is all the requisite information available. Do any data values have missing elements or are they in an unusable state
Consistency: - Are data values the same across the data sets. Are there any distinct occurrences of the same data instances that provide conflicting
information
Conformity: - Do data values comply with the specified formats. If so, do all the data values comply with those formats
Accuracy: - Do data objects accurately represent the “real world” values they are expected to model. What data is Incorrect or out of date
Integrity: - Is there are any data missing or is there any data with missing references.The inability to join related records together may actually
introduce duplication across your systems
Timeliness: - Is the required data information is available when it is expected and needed. Timeliness of data is very important.
Duplicates: - What data records or attributes are repeated.
Here are more details on this topic :- https://idwbi.teachable.com/courses/idq-interview-questions/lectures/3830032
Question 27: Correct
Should the data be sorted before passing the data into the Match Transformation?

Yes, only for Dual Source Match process

No, IDQ will internally sort the data based on the created clusters.

Yes, data has to be sorted on the SequenceID on which match process is performed.

Yes, sorting is required for field Match Process only.
(Correct)
Explanation
For classic matching we need to sort the data on GroupKey field. Sorting is not required for Identity matching. Here are more details on this topic
:- https://idwbi.teachable.com/courses/idq-interview-questions/lectures/3717416
Question 28: Incorrect
Select the datatypes which are not supported in Association Transformation?

String

Double
(Correct)

Numeric

BigInt
(Correct)
Explanation
The Association transformation accepts string and numerical data values on input ports. If you add an input port of another data type, the
transformation converts the port data values to strings.
For more details https://idwbi.teachable.com/courses/234642/lectures/3717556
Question 29: Incorrect
Source data has Customer full name, the data has to be checked against the Prefix.dic, FirstName.dic and LastName.dic
Below strategies are created in Labeler transformation.

Source file has the Customer full name as Mr Joe Smith ( sample source record :- 1001|Issued|Mr Joe Smith|100.34|New|UY76200T)
From the Labeler, the LabeledOutput is as below.

But the customer is expecting the LabeledOutput as below.


select the correct response.

Update the Labeler transformation with Exclusive mode and identify the values which are present in source and missing
from .dic files. Send the list to Informatica and get updated .dic files

Source data and the Strategies in Labeler are using different delimiters, update the Strategies in Labeler as per source data.
(Correct)

.dic file does not have the values which are present in the source, due to which Lables are not created.

Incorrect Execution mode is selected. select the execution mode as Character to get the data as expected by the Customer.
(Incorrect)
Explanation
Source data and the Strategies in Labeler are using different delimiters, update the Strategies in Labeler as per source data.
Input Data will be validated based on the overall Labeler configuration, but the input data has to be send in a way that the input strings are
recognized properly by the labeler.
Delimiters help Labeler to search for part of the strings (first name or last Name) without delimiters we can't have a granular control on how we
want Labeler to recognize and tokenized the data.
If incorrect delimiter is specified, the input data will not be recognized properly
If no delimiter is specified, Data Quality recognizes it as a Word Symbol
Example: - If the input string is Mr Joe Smith and if you need to separate the strings as First name and Last name using token labeler and if you
have not used space as a delimiter in this scenario, Labeler will search for the entire string (Mr Joe Smith) to be present in the dictionary
files.Proper delimiter (in this case space delimiter) has to be used for the labeler to recognize the string as name name.
Here are more details on this topic :- https://idwbi.teachable.com/courses/idq-interview-questions/lectures/3709823
Question 30: Incorrect
In which Execution Mode in Labeler probabilistic matching techniques can be used?

Token
(Correct)

Probabilistic matching techniques are used in Parser transformation only

String

Character
(Incorrect)
Explanation
A probabilistic model is a reference data object. Use a probabilistic model to understand the contents of a data string that contains multiple data
values. A probabilistic model identifies the types of information in each value in the string.
Probabilistic models use Natural Language Processes (NLP) to identify the type of information in a string. NLP detect relevant terms in the input
string and disregard terms that are not relevant.
In Token Execution mode you can use probabilistic matching techniques.
Here are more details on this topic --> https://idwbi.teachable.com/courses/idq-interview-questions/lectures/3707014
Question 31: Incorrect
Select all the data integration service plugins that can be defined in DIS
Note: There can be multiple correct answers to this question.

Profile Service
(Correct)

Workflow Service
(Correct)

SQL Service
(Correct)

Mapping Service
(Correct)
Explanation
All the options are valid DIS plugins.
Below Data Integration service plugins can be defined in DIS

Users can enable(true) or disable(false) these plugins as per their requirement.


Here are more details on this topic :- https://idwbi.teachable.com/courses/idq-interview-questions/lectures/6285447
Question 32: Incorrect
Pattern Based Parser requires the output from _______ transformation as Input?

Association

Labeler - Character Mode
(Incorrect)

Standardizer

Labeler - Token Mode
(Correct)
Explanation
It is used to parse the patterns made of multiple strings, if data from multiple strings have to be parsed apart or sorted. The patterns should be
recognizable by the upstream ports LabelData port and Tokenized data port from the Labeler transformation. For example, if the source contains
multiple string value with SSN, EMPNO, DateofBirth, if these values have to be parsed in to three separate attributes, based on the system token
sets each multi string can be separated.
Pattern Based Parser requires the output from Labeler - Token Mode transformation as Input

Pattern Based parser will provide full flexibility to customize each individual pattern output
Here are more details on this topic--> https://idwbi.teachable.com/courses/idq-interview-questions/lectures/3709850
Question 33: Incorrect
From Informatica Developer 10.x, objects can only be exported to PowerCenter 10.x?

False
(Correct)

True
(Incorrect)
Explanation
From Informatica Developer 10.x, users can export objects to PowerCenter 9.5.1, PowerCenter 9.6.1 & PowerCenter 10.x. Here are more details
on this topic--> https://idwbi.teachable.com/courses/idq-interview-questions/lectures/6270087
Question 34: Incorrect
Which of the following is not a valid Task status (highlighted in blue color box) in Informatica Analyst?

On Schedule

Elapsed
(Correct)

Assigned
(Incorrect)

OverDue
Explanation
Elapsed is not a valid Task status
Question 35: Correct
All the statements are correct w.r.t Profile and Scorecard, except
Note: There can be multiple correct answers to this question.

Profile can be deleted without Impacting Scorecard

If Profile is deleted, the dependent Scorecard will be Invalidated
(Correct)

From Informatica v 10.x, users can define decimal values in Scorecard Metric Thresholds
(Correct)

If data source object is deleted then both the profile and the Scorecard will be Invalidated
Explanation
Scorecard Metrics cannot have decimal values and if profile is deleted the scorecards will be invalidated.Here are more details on this topic :-
https://idwbi.teachable.com/courses/idq-interview-questions/lectures/3706722
Question 36: Incorrect
Which IDQ transformations can generate Score
Note: There can be multiple correct answers to this question.

Labeler
(Correct)

Comparative
(Correct)

Classifier
(Correct)

Parser
(Correct)
Explanation
Below transformations can generate scores in IDQ.
Classifier
Comparative
Exception
Parser
Labeler
Match
Weighted Average
Here are more details on this topic :- https://idwbi.teachable.com/courses/idq-interview-questions/lectures/6273788
Question 37: Incorrect
____ is the service which is introduced from Informatica version 10.x?

System Services
(Correct)

Search Service

Queue Service
(Incorrect)

Reporting and Dashboard Service
Explanation
In IDQ 10.x, the domain includes system services. A system service is an application service that can have a single instance in the domain.
Below are the new features in IDX 10.x
Generate and Execute DDL: - You can create tables in a database by generating and executing a DDL script. (Similar to the options available in
Target designer in Informatica PowerCenter). DDL scripts can be generated for one or more relational objects, and if a relation object already
exists in the database, you can drop the target and re-create it.
Configure the mappings as Dynamic: - You can configure dynamic mappings, i.e. you can create mappings with dynamic sources and targets
that allow metadata changes to the data sources. When you configure a source or target to be dynamic, the Data Integration Service can interpret
metadata changes to relational and flat file data sources at run time.
Support to the PowerCenter Transformations: - You can import the following PowerCenter transformations into the Developer
tool.Normalizer transformationSequence Generator transformationUpdate Strategy transformation
Object Versioning: - In IDQ 10.x when the Model repository is integrated with a version control system, the version control system protects
objects from being overwritten by other members of the development team. You can check objects out and in, undo the checkout of objects, and
view and restore historical versions of objects.
Terminate Event Task :- From Informatica v 10.1.1, users can configure Terminate Event task to Abort the workflow, if any mapping tasks fails
in the workflow.
System Services: - From IDQ 10.x, the domain includes system services. A system service is an application service that can have a single
instance in the domain. The domain includes the following system services:Email ServiceResource Manager ServiceScheduler Service
Resource Manager Service: - The Resource Manager Service manages computing resources in the domain and dispatches jobs to achieve
optimal performance and scalability. The Resource Manager Service collects information about nodes with the compute role. The service matches
job requirements with resource availability to identify the best compute node to run the job.
Schedule Workflows: - You can schedule the profiles, scorecards, deployed mappings and deployed workflows using a Scheduler Service.
Target Load Order Constraints: - You can configure constraints to control the order in which rows are loaded and committed across target
instances in a mapping.
Inclusive Gateway: - From IDQ v 10.1.1, You can configure to run multiple tasks in parallel using Inclusive Gateway task. For example, you
can configure an Inclusive gateway with multiple branches and add a Mapping task to each branch. When the workflow runs, the Data Integration
Service concurrently runs the mappings that the tasks identify.
New Monitoring Views in Monitoring Tool: - Below are the new features in the monitoring toolSummary Statistics view - Displays resource
usage, object distribution, and object states for a selected time range.Detailed Statistics View: - Displays graphs of throughput and resource usage
information for the source and target.Create customized views on the Execution summary views
Address Validator Transformation:- Effective in version 10.0, you can define parameters to set the following transformation
properties:Geocode data typeGlobal Max Field LengthMax Result CountOptimization LevelStandardize Invalid Address
For more details check --> https://idwbi.teachable.com/courses/idq-interview-questions/lectures/3706749
Question 38: Incorrect
Data has to be extracted from various sources and compared against the master data for each load, the existing process is delaying the load
process due to the source data comparison with master data for every load. Customer is insisting to follow the same load approach and asked your
recommendation.
Select the best approach

Standardize the data and perform dual source matching, as data is standardized by removing the noise words, and validating
against the reference tables, the match process will be efficient and load time will be considerably less as compared to
performing the match with out standardization

Use Identity Match with Persistent RecordID which uses Universal ID which allows users to store matching keys to make the
subsequent matching process efficient. It will significantly improve the performance when source data is matched against
master data regularly and the speed of the match operation is critical
(Correct)

Reading the data from two pipelines and performing a dual source is not recommended, when time is a constraint. Use Union
transformation to combined the data from both the sources and perform single source Field matching and in consolidation
transformation use row based strategy with modal exact option to prioritize the data from master source.

All the other approaches mentioned will not reduce the load window, as match process is a complex process. Instead of using
match transformation use Comparison transformation to compare strings between source data and master data and based on
the score generated create a ETL process to load as per your requirement.
(Incorrect)
Explanation
Use Identity Match with Persistent RecordID which uses Universal ID which allows users to store matching keys to make the subsequent
matching process efficient. It will significantly improve the performance when source data is matched against master data regularly and the speed
of the match operation is critical
Question 39: Incorrect
________contain key building algorithms that have been developed and optimized for specific countries, language and for each class of data.

Address reference data
(Incorrect)

Identity populations
(Correct)

Classification models

Probabilistic models
Explanation
The Content Management Service manages the following types of reference data:

Address reference data: -You use address reference data when you want to validate the postal accuracy of an address or fix errors in an address.
Use the Address Validator transformation to perform address validation.
Identity populations: -You use identity population data when you want to perform duplicate analysis on identity data. An identity is a set of
values within a record that collectively identify a person or business. Use a Match transformation or Comparison transformation to perform
identity duplicate analysis.
Probabilistic models and classifier models: - You use probabilistic or classifier model data when you want to identify the type of information
that a string contains. Use a probabilistic model in a Parser or Labeler transformation. Use a classifier model in a Classifier transformation.
Probabilistic models and classifier models use probabilistic logic to identify or infer the type of information in the string. Use a Classifier
transformation when each input string contains a significant amount of data.
Reference tables: - You use reference tables to verify the accuracy or structure of input data values in data quality transformations.
In addition to the above, Informatica also provides below Content to customers based on the Subscription.
Data Domains Accelerator
Australia/New Zealand Accelerator
Brazil Accelerator
Financial Services Accelerator
France Accelerator
Germany Accelerator
Portugal Accelerator
Spain Accelerator
United Kingdom Accelerator
U.S./Canada Accelerator
Core Accelerator
Core Accelerator :- The Core accelerator contains mapplets, rules and reference data objects that other accelerators can reuse.
Address data cleansing
Contact data cleansing
Corporate data cleansing
General data cleansing
Matching and deduplication data cleansing
Product data cleansing
Here are more details on this topic :- https://idwbi.teachable.com/courses/idq-interview-questions/lectures/3706824
Question 40: Incorrect
Matching process cannot be disabled for which of the following Persistence Methods in Identity Match?

Remove IDs from the database

Update the database with new IDs
(Incorrect)

Update the current IDs in the database

Do not update the database
(Correct)
Explanation
If you want to perform match against the store with out updating, you can select Do not update the database as Matching Process. By default the
matching will be enabled and you cannot disable the match process
Question 41: Incorrect
______input is mandatory for the match process?

ClusterId
(Incorrect)

GroupKey

SequenceId
(Correct)

RowId
Explanation
Yes, SequenceId is mandatory for the match process. Every record in an input data set must include a unique sequence identifier. If a data set
contains duplicate sequence identifiers, the Match transformation cannot identify duplicate records correctly. Use the Key Generator
transformation to create unique identifiers if none exist in the data.
When you create an index data store for identity data, the Match transformation adds the sequence identifier for each record to the data store.
When you configure the transformation to compare a data source with the index data store, the transformation might find a common sequence
identifier in both data sets. The transformation can analyze the sequence identifiers if they are unique in the respective data sets.
Note: - The Match Mapping will be Invalid without connecting the SequenceId port from the upstream mapping.
Here are more details on this topic --> https://idwbi.teachable.com/courses/idq-interview-questions/lectures/3717453
Question 42: Incorrect
What is the best approach to validate the Address data?
Note: There can be multiple correct answers to this question.

Use Standardizer transformation to remove noise words and standardize data before sending the data to Address Validator.
(Correct)

For each country create a separate Address Validator transformation to validate the address data

For highly complex and discrete attributes use Basic template model with Discrete Address Validation Input Template
(Incorrect)

For Australian address, Informatica Address Doctor uses AMAS (Address Matching Approval System) rules for validation of
address when the Input mode is configured as ‘Certified’.
(Correct)
Explanation
For Australian address, Informatica Address Doctor uses AMAS (Address Matching Approval System) rules for validation of address when the
Input mode is configured as ‘Certified’.
The certification mode required additional database files (AUS5C1.MD and AUS5C2.MD, which Informatica will provide and renew based on
the subscription) and SetConfig.xml should be configured Section <DataBase> with Type=” CERTIFIED” for ISO=” AUS” or “ALL”.
The databases contain Postal Address File (PAF) data which includes Australia Post’s Delivery Point Identifiers (DPIDs).
Here are more details on this topic--> https://idwbi.teachable.com/courses/idq-interview-questions/lectures/3707393
Question 43: Incorrect
Which of the following is not a valid Human Task step?

Review step

Exception step

Publish step
(Correct)

Cluster step
(Incorrect)
Explanation
Below steps can be created in Human Task
Cluster step:- Create a Cluster step when you want a user to examine duplicate record clusters and create a preferred record from the values in the
duplicate records.
Exception step:- Create an Exception step when you want a user to examine and fix errors in records.
Review step:- Create a Review step when you want a user to review the work done in an Exception or Cluster step. Review steps are not
mandatory.
Here are more details on this topic :- https://idwbi.teachable.com/courses/idq-interview-questions/lectures/6289031
Question 44: Incorrect
Select all the IDQ objects which cannot be exported to PowerCenter?
Note: There can be multiple correct answers to this question.

Human Tasks
(Correct)

Workflows
(Correct)

Rules created in Developer

Logical Data Objects
(Incorrect)
Explanation
Workflows and Human Tasks cannot be exported to PowerCenter.
No, not all objects can be exported from IDQ to PowerCenter.
For example :- Workflows ,Human Tasks, Connections and Lookups with Customized SQls created in Mapplets cannot be exported to
PowerCenter.
Also, the parameter files which are used in IDQ are not exported to PowerCenter, because the format of parameter files are different between
IDQ and PowerCenter. The parameters resolve to their default values when IDQ mappings/mapplets with parameters are imported to
PowerCenter.
Here are more details on this topic :- https://idwbi.teachable.com/courses/idq-interview-questions/lectures/6279663
Question 45: Incorrect
Select the correct statement w.r.t Scorecards?

Scorecard results are stored in Model Repository as part of Profile definition.
(Incorrect)

Scorecard results are displayed only on GUI and are not stored in any database.

Scorecard results are stored in the Staging Database.

Scorecard results are stored in the Profile Warehouse.
(Correct)
Explanation
Scorecard results are stored in the Profile Warehouse.
A graphical representation of valid values for a column or the output of a rule in profile results. Use scorecards to measure data quality during
different points in the data quality life cycle.

When you create a scorecard, you can assign a weight to each metric(A metric is a column of a data source or output of a rule that is part of a
scorecard). Create a metric group to categorize related metrics in a scorecard into a set.
Scorecard results are stored in the Profile Warehouse.
Here are more details on this topic :- https://idwbi.teachable.com/courses/idq-interview-questions/lectures/3706720
Question 46: Incorrect
Which transformation creates links between records that share duplicate characteristics across more than one data field

Merge Transformation

Consolidation Transformation

Match Transformation
(Incorrect)

Association Transformation
(Correct)
Explanation
Association Transformation is an extension of the data matching process which process output data from a match transformation.
The Association transformation creates links between records that share duplicate characteristics across more than one data field so that they are
treated as members of a single set in data consolidation operations.
Here are more details on this topic--> https://idwbi.teachable.com/courses/idq-interview-questions/lectures/3717535
Question 47: Incorrect
Select all the Validation modes which can be configured using AddressDoctor Webservice?
Note: There can be multiple correct answers to this question.

Certified

Batch
(Correct)

Parse
(Incorrect)

Interactive
(Correct)
Explanation
Yes,You can develop a Web Service to validate addresses. This Web Service could be accessed by third party applications.You can also validate
address using AddressDoctor Cloud and AddressDoctor Software Library.
Below URL will have the wsdl file details.
http://validator2.addressdoctor.com/addBatch/Index.html
Using AddressDoctor Webservice, you can configure Batch,FastCompletion and Interactive validation modes
Here are more details on this topic--> https://idwbi.teachable.com/courses/idq-interview-questions/lectures/6270656
Question 48: Incorrect
Select all the options which are required to use DQ for Excel?

Data Transformation Excel License option
(Incorrect)

Web Services option
(Correct)

DQ for Excel installation on PC
(Correct)

MS Excel version 2007 or 2010
(Correct)
Explanation
Below are required to use DQ for Excel.
DQ for Excel installation on PC
MS Excel version 2007 or 2010
Web Services option
Here are more details on this topic :- https://idwbi.teachable.com/courses/idq-interview-questions/lectures/6288332
Question 49: Correct
Below objects can be Imported from PowerCenter to MRS, expect

Relational Connections
(Correct)

Mappings

Mapplets

Reusable Transformations
Explanation
Connections should be manually configured. Connections are not exported when objects are exported from PowerCenter to IDQ and vice
versa.Here are more details on this topic :- https://idwbi.teachable.com/courses/idq-interview-questions/lectures/6276462
Question 50: Incorrect
As per the requirement, the Customer full name has to be parsed to Prefix, FirstName and LastName.
In the Labeler transformation, below LabeledOutput and TokenizedData are generated.

Mapping to the Pattern Based Parser is as below

Below are the possible patterns defined in the Parser Transformation

The expected output from Pattern based parser is as below

But the actual output is as below, the Customer Name is not parsed and the completed string is shown as Overflow

Select the correct response



The Patterns are incorrectly defined in the Parser Transformation.
(Correct)

The delimiter defined in the Labeler is not matching with the delimiter defined in the Pattern based Parser Transformation.

The Pattern based transformation is searching for the Customer Full Name in the .dic files, Informatica provides the
Prefix.dic , FirstName.dic & SurName.dic files separately, create three strategies with with each of the .dic files and parse the
Name .
(Incorrect)

The Patterns defined in the Patterns section (in Parser transformation) might have spaces due to which the Customer Name is
not parsed correctly.
Explanation
The Patterns are incorrectly defined in the Parser Transformation.
It is used to parse the patterns made of multiple strings, if data from multiple strings have to be parsed apart or sorted. The patterns should be
recognizable by the upstream ports LabelData port and Tokenized data port from the Labeler transformation. For example, if the source contains
multiple string value with SSN, EMPNO, DateofBirth, if these values have to be parsed in to three separate attributes, based on the system token
sets each multi string can be separated.
Pattern Based Parser requires the output from Labeler - Token Mode transformation as Input

Pattern Based parser will provide full flexibility to customize each individual pattern output
Here are more details on this topic :- https://idwbi.teachable.com/courses/idq-interview-questions/lectures/3709850
Question 51: Incorrect


Records with score =1 will flow through the Unique Records group and records with 0.82 will flow in the Cluster Output
group
(Incorrect)

All the records will flow through the Standard output group

All the four records are duplicate and will flow through the Cluster Output group. These records should be manually
reviewed in the Analyst tool.
(Correct)

Generate duplicate record table and validate if the records are duplicate, if yes, configure the records to be reviewed by
Business Analyst by setting up Unmanaged Editable Reference table in Informatica Analyst.
Explanation
Duplicate records will flow through the Cluster Output group and Business users have to analyse the records in the Analyst tool.
Question 52: Incorrect
Are Probabilistic models part of Informatica Base Content set?

True
(Incorrect)

False
(Correct)
Explanation
Informatica doesn’t provide Probabilistic models as part of the base content set.
A probabilistic model is a reference data object. Use a probabilistic model to understand the contents of a data string that contains multiple data
values. A probabilistic model identifies the types of information in each value in the string.
Probabilistic models use Natural Language Processes (NLP) to identify the type of information in a string. NLP detect relevant terms in the input
string and disregard terms that are not relevant.
Here are more details on this topic --> https://idwbi.teachable.com/courses/idq-interview-questions/lectures/3707014
Question 53: Incorrect
Select the attribute which is not required in Exception task processing?

IS_MASTER
(Correct)

REVIEW_STATUS

WORKFLOW_ID

USER_COMMENT
(Incorrect)
Explanation
IS_MASTER is required for Duplicate exception task.
When you correct exceptions task or review exceptions task, below attributes should be present in the table. If below attributes are not present,
users cannot open the created exception task in Informatica Analyst,Informatica will display error stating the required column is missing.
To correct exception, below attributes should exist in the table.
ROW_IDENTIFIER
REVIEW_STATUS
WORKFLOW_ID
USER_COMMENT
UPDATED_STATUS
RECORD_STATUS
Here are more details on this topic :- https://idwbi.teachable.com/courses/idq-interview-questions/lectures/6285055
Question 54: Incorrect
Customer has requested to verify the address attributes of Singapore, India, Belgium and United States the address information is formatted on a
single line (line1) with comma separator and country name is on different line (line2), based on the Input data Customer is expecting 5-7
suggestions per address. select all the statements which are incorrect?

It is recommended to enter the single line address elements in the correct sequence (example :- House Number , Street No,
Postal Code, Locality)

Country Name is optional when you pass complete address info in Suggestion List mode

Single line address validation is not available in Suggestion List mode.
(Correct)

Single line address validation is available for United States, Singapore and Belgium
(Correct)
Explanation
Single line address validation is available for United States, Singapore and Belgium
Question 55: Incorrect
Select the transformations which support parameterizing the Cache directory?

Address Validator

Key Generator
(Correct)

Decision
(Incorrect)

Association
(Correct)
Explanation
Below are some of the transformations which support cache directory parameterizing
Association
Consolidation
Duplicate Record Exception
Key Generator
Match
Here are more details on this topic :- https://idwbi.teachable.com/courses/idq-interview-questions/lectures/6272878
Question 56: Incorrect
Select the expression functions which are not supported by the Decision Transformation?
Note: There can be multiple correct answers to this question.

XOR
(Incorrect)

MonthCompare

AES_DECRYPT
(Correct)

DEC_BASE64
(Correct)
Explanation
DEC_BASE64 and AES_DECRYPT are not supported by the Decision Transformation.
The following expressions are not supported by the Decision Transformation.
ABORT
AES_DECRYPT
AES_ENCRYPT
COMPRESS
DEC_BASE64
DECODE
DECOMPRESS
ERROR
IIF
TO_BIGINT

Here are more details on this topic--> https://idwbi.teachable.com/courses/idq-interview-questions/lectures/3707512


Question 57: Incorrect
Which of the following are incorrect about Rule Specifications?
Note: There can be multiple correct answers to this question.

Rules which are created in developer can be used in the Rules Specifications
(Incorrect)

Rules Specification will be Invalid if it contains more than one Input and one output
(Correct)

When rules are saved in Analyst, the rules are automatically generated as mapplets in Developer
(Correct)

Reference tables can be used in Rule Specification
Explanation
Reference tables can be used in Rule Specification and also rules created in Developer and Analyst can be used in Rules Specifications
Question 58: Correct
To schedule IDQ workflows,third party scheduler has to be used as IDQ doesn't have scheduling capability

True

False
(Correct)
Explanation
From Informatica v 10.x You can schedule the profiles, scorecards, deployed mappings and deployed workflows using a Scheduler Service.
Below are the new features of IDQ 10.x
Generate and Execute DDL: - You can create tables in a database by generating and executing a DDL script. (Similar to the options available in
Target designer in Informatica PowerCenter). DDL scripts can be generated for one or more relational objects, and if a relation object already
exists in the database, you can drop the target and re-create it.
Configure the mappings as Dynamic: - You can configure dynamic mappings, i.e. you can create mappings with dynamic sources and targets
that allow metadata changes to the data sources. When you configure a source or target to be dynamic, the Data Integration Service can interpret
metadata changes to relational and flat file data sources at run time.
Support to the PowerCenter Transformations: - You can import the following PowerCenter transformations into the Developer
tool.Normalizer transformationSequence Generator transformationUpdate Strategy transformation
Object Versioning: - In IDQ 10.x when the Model repository is integrated with a version control system, the version control system protects
objects from being overwritten by other members of the development team. You can check objects out and in, undo the checkout of objects, and
view and restore historical versions of objects.
Terminate Event Task :- From Informatica v 10.1.1, users can configure Terminate Event task to Abort the workflow, if any mapping tasks fails
in the workflow.
System Services: - From IDQ 10.x, the domain includes system services. A system service is an application service that can have a single
instance in the domain. The domain includes the following system services:Email ServiceResource Manager ServiceScheduler Service
Resource Manager Service: - The Resource Manager Service manages computing resources in the domain and dispatches jobs to achieve
optimal performance and scalability. The Resource Manager Service collects information about nodes with the compute role. The service matches
job requirements with resource availability to identify the best compute node to run the job.
Schedule Workflows: - You can schedule the profiles, scorecards, deployed mappings and deployed workflows using a Scheduler Service.
Target Load Order Constraints: - You can configure constraints to control the order in which rows are loaded and committed across target
instances in a mapping.
Inclusive Gateway: - From IDQ v 10.1.1, You can configure to run multiple tasks in parallel using Inclusive Gateway task. For example, you
can configure an Inclusive gateway with multiple branches and add a Mapping task to each branch. When the workflow runs, the Data Integration
Service concurrently runs the mappings that the tasks identify.
New Monitoring Views in Monitoring Tool: - Below are the new features in the monitoring toolSummary Statistics view - Displays resource
usage, object distribution, and object states for a selected time range.Detailed Statistics View: - Displays graphs of throughput and resource usage
information for the source and target.Create customized views on the Execution summary views
Address Validator Transformation:- Effective in version 10.0, you can define parameters to set the following transformation
properties:Geocode data typeGlobal Max Field LengthMax Result CountOptimization LevelStandardize Invalid Address
Here are more details on this topic :- https://idwbi.teachable.com/courses/idq-interview-questions/lectures/3706749
Question 59: Incorrect
In DQ for excel, how many maximum number of records can be configured in a batch

100
(Correct)

1000

There is no limit on the number of records, based on the requirement the any number of records can be configured
(Incorrect)

10
Explanation
DQ for excel helps non Informatica users to use the data quality rules created in IDQ and validate data in excel. DQ for excel helps users to reuse
the data quality rules and perform data validation from their PC and check how the source data will get transformed in the target data objects.
In IDQ, developer has created a mapplet which will standardize data based on the reference data and parse address or customer names.
Developers can save the mapplet as web service and give the web service ( after deploying the web service to application). Users can use the DQ
for excel and create a new service(Add) using the web service url and validate the data in excel.
Users can create a batch of 100 records and validate the data using DQ for Excel. The web service Maximum Occurrences should be configured
to unbounded to allow the web service to process multiple records in one batch.
Here are more details on this topic :- https://idwbi.teachable.com/courses/idq-interview-questions/lectures/6282430
Question 60: Incorrect
Select the Objects which can be exported from Informatica Analyst to PowerCenter?

Mapping Specifications
(Correct)

From Informatica Analyst, no objects can be directly exported to PowerCenter.
(Incorrect)

Rule Specifications

Data Specifications
Explanation
You cannot export Mapping Specifications from Informatica Analyst to PowerCenter
Question 61: Incorrect
A customer file has 5 million records and 500 records are duplicates, the customer data has to be analyzed based on CustomerName, DateofBirth
and SSN. Filter Exact Match option is enabled in Match transformation, how the mapping performance will be impacted?

It is recommend to group the similar records and send the grouped data to the Match Process instead of using Filter Exact
Match option,similar records will be created on the same cluster, which will be (similar records) consolidated and removed by
Consolidator transformation.

It is recommended to use the Filter Exact Match option and allow the 500 duplicate records go through the matching process,
as every row of data goes through the matching process irrespective of selecting Filter Exact Match option.
(Incorrect)

Filter Exact Match option can only be used in Dual Source matching, and in this scenario the option is disabled by default.

Enabling Filter Exact Match option helps improve performance by removing duplicate records out of the matching
process. However, there is additional logic (internally added) to remove the duplicate records that adds little overhead to
internal mapping logic, hence its not recommended to enable Filter Exact Match option for this scenario.
(Correct)
Explanation
For this scenario, its not recommended to enable Filter Exact Match option.
Filter Exact Match can be configured in clustered output match type only. When Filter Exact Match is enabled it Improves the performance of the
Match Transformation by eliminating the identical records from the matching process. The identical records are directly written to the clustered
output.
If you have a source dataset with large number of identical records enabling Filter Exact Match will improve performance by removing identical
records out of the matching process.
Based on the source data and the number of exact matches in the dataset, it has to be checked if Match Mapping can be executed with or without
enabling the Filter Exact Match option.
If the data set has a lot of exact matches, there is a lot of overhead to process these exact matches and Match mapping can be very resource
intensive, to Improve the Match performance in this scenario we can enable the Filter Exact Match Option.
Here are more details on this topic--> https://idwbi.teachable.com/courses/idq-interview-questions/lectures/3717467
Question 62: Incorrect
Select the correct statement about Tags?


You cannot remove the association between a tag and an object after the tag is associated with object.

Tags associated with an object in the Developer tool appear as tags for the same objects in the Analyst tool after users
associate the tag with one or more objects in Analyst
(Incorrect)

Tags are specific to Informatica Developer objects.

Tags associated with an object in the Developer tool appear as tags for the same objects in the Analyst tool.
(Correct)
Explanation
Tag is metadata that defines an object in the Model repository based on business usage. Create tags to group objects according to their business
usage.
After you create a tag, you can associate the tag with one or more objects. You can remove the association between a tag and an object. You can
use a tag to search for objects associated with the tag in the Model repository. The Developer tool displays a glossary of all tags.
For example, you create a tag named XYZCorp_CustomerOrders and assign it to tables that contain information for the customer orders from the
XYZ Corporation. Users can search by the XYZCorp_CustomerOrders tag to identify the tables associated with the tag.
Note: Tags associated with an object in the Developer tool appear as tags for the same objects in the Analyst tool.

Here are more details on this topic :- https://idwbi.teachable.com/courses/idq-interview-questions/lectures/3706828


Question 63: Correct
Select the correct statement about WebServices?

Address Validator web services can only be executed in Batch Mode, Fast Completion,Country Recognization
& Interactive modes
(Correct)

Similar to PowerCenter in IDQ also, workflows/mappings can be invoked by web service API.

IDQ Mappings and Mapplets can be deployed as WebServices in Data Quality.

Using REST Web Service Consumer transformation, you can configure the HTTP method as GET only
Explanation
Address Validator web services can only be executed in Batch Mode, Fast Completion,Country Recognization & Interactive modes
Question 64: Incorrect
A source file has CustomerInfo (5000 records), matching has to be performed based on individual data such as Name, DOB, SSN within the same
data set and generate match scores.
Select the best mapping approach for the this scenario.

Standardize the data, and filter duplicate records before sending the data to match transformation, so that the match cluster
size will be small which will speed up the match process.

Standardize the data and create Groupkeys based on Name, DOB, SSN and Include City or State, so that small groups
are created and match process can speed up processing the data with differing City or State by ensuring unnecessary
comparisons are avoided.

Standardize the data and send the data to Match transformation and generate match scores. As the source data set is small,
GroupKey is not required.
(Correct)

Standardize the data and generate a group key using NYSIIS or Soundex strategy in Key Generator transformation,
as Irrespective of the data volume, its mandatory to pass GroupKey to the Match process.
(Incorrect)
Explanation
Matching rules are provided by business, i.e. if you are working on Customer data, should the data be matched on Name, DOB and SSN or
should you include any address attributes or should you individually match and creates links between records that share duplicate characteristics,
these kind of rules are provided by the business.
Grouping depends on the source volume, which developer has to profile the data and come up with a key which can create correct no of records
in the group. (Informatica recommends 5000 records in each group).
Example
Let’s say, a source file has CustomerInfo (5000 records), matching has to be performed based on individual data such as Name, DOB, SSN
within the same data set and generate match scores.
If matching on very large data sets, using a GroupKey can speed up processing by ensuring unnecessary comparisons are avoided. For instance, if
matching based on Address fields only, it does not make sense to compare addresses where the City or State differ. So, a GroupKey based on
either of these could be beneficial. However, if you are matching based on individual data such as Name, DOB, SSN within the same data set, a
GroupKey based on City or State could exclude potential matches where an individual has duplicate records with differing City or State data.
For the above scenario, GroupKey is not required as the source data set is small.Standardize the data and send the data to Match transformation
Here are more details on this topic --> https://idwbi.teachable.com/courses/idq-interview-questions/lectures/6270548
Question 65: Incorrect
Join Analysis is performed on Party and PartyRelation table, based on the below picture select the correct answer


There are no matching records between Orders and Customers

There might be some Orphan records in Orders or Customer table, due to which all the rows cannot be joined between
Orders and Customers
(Correct)

One of the table is empty, due to which Informatica developer didn't generate any join Analysis, the venn diagram in the
picture represents one of the table is empty

All the key column values exists between Orders and Customers, Orders table should be joined to Customers using a left join
because as per the picture the Order table is shown as driving table.
(Incorrect)
Explanation
Based on the picture, the Order has more number of rows compared to customers and there might be some orphan records in one of the table.
Join analysis describes the degree of potential joins between two data columns. Use a join profile to analyze column joins in a data source or
across multiple data sources. A join profile displays results as a Venn diagram and as numerical and percentage values. You create and run a join
profile from an enterprise discovery profile.

Here are more details on this topic :- https://idwbi.teachable.com/courses/idq-interview-questions/lectures/3706702


Question 66: Incorrect

Select the correct statement w.r.t the above picture?



Using the Step_13A_Cust_Claims step, user should examine and fix errors in records.
(Correct)

Using the Step_13A_Cust_Claims step, user should review the records.
(Incorrect)

Using the Step_13A_Cust_Claims step, user should create a preferred record from the list of records.

Using the Step_13A_Cust_Claims step, user should examine duplicate record clusters in the the records.
Explanation
The icon used in Step_13A_Cust_Claims represents a ExceptionStep. Users check the Issues in the data and fix them using Informatica Analyst.
Question 67: Incorrect
______ is the file format of Local Population File ?

.ysp
(Incorrect)

.ycp

.ylp
(Correct)

.ylf
Explanation
Three basic types of population files exist. They are as follows:
ysp (Standard Population File)
ylp (Local Population File)
ycp (Custom Population File)
Here are more details on this topic :- https://idwbi.teachable.com/courses/idq-interview-questions/lectures/3717498
Question 68: Incorrect
Select the recommend group size for classic matching?

5000
(Correct)

50000

can be derived based on the total no of records processed by match transformation divided by the no of match strategies
(Incorrect)

500
Explanation
1. Match Performance Analysis: - Before executing the Match mapping, run the Match Performance Analysis to view the number of
comparisons that will have to take place. Match Performance Analysis provides the below stats, cleansing and standardizing the data before
passing to the Match transformation will significantly improve the below stats.
A large number of NULLs in a group or in your columns being matched may take a longer time to match.
2. Limiting the Cache file size: - The port size for all fields especially the fields used as search keys and group keys should be set to the
minimum required precision so that the size of cache files will be reduced which in turns improves the performance.
Below formulae can be used to provide a rough calculation of how much disk space (in MB) would be needed to run matching on a particular
dataset:
Classic Matching: - d * n * 0.0000025
Identity Matching: - d * n * 0.000005
where d=size of the data records in chars and n=number of records.
3. Limiting the number of ports: - Pass the ports which are used within the matching to the Match Transformation.
4. Group Sizes: - Informatica recommends the group sizes for Classic Matching should not be more than 5000-6000. The number of pairs a
dataset with N records will generate is calculated by the formula (((N*N)-N)/2)
Here are more details on this topic :- https://idwbi.teachable.com/courses/idq-interview-questions/lectures/3717461
Question 69: Incorrect
select all the correct statements w.r.t Import in Developer and Analyst?

Rule specifications get imported when exporting/importing the Business Glossary from one environment to another
(Incorrect)

PowerCenter mappings with VSAM sources can not be imported into IDQ
(Correct)

Similar to PowerCenter you can choose different target Projects to import multiple objects in IDQ.

Mapping Specifications can be exported only after validating the object as valid.

Data Domains can only be imported using Advanced Import Wizard
Explanation
PowerCenter mappings with VSAM sources can not be imported into IDQ
Question 70: Incorrect
Informatica distributes the following types of content
Note: There can be multiple correct answers to this question

Data Domains
(Correct)

Identity population files
(Correct)

Address reference data files
(Correct)

Mapping Specifications
(Incorrect)
Explanation
The Content Management Service manages the following types of reference data:

Address reference data: -You use address reference data when you want to validate the postal accuracy of an address or fix errors in an address.
Use the Address Validator transformation to perform address validation.
Identity populations: -You use identity population data when you want to perform duplicate analysis on identity data. An identity is a set of
values within a record that collectively identify a person or business. Use a Match transformation or Comparison transformation to perform
identity duplicate analysis.
Probabilistic models and classifier models: - You use probabilistic or classifier model data when you want to identify the type of information
that a string contains. Use a probabilistic model in a Parser or Labeler transformation. Use a classifier model in a Classifier transformation.
Probabilistic models and classifier models use probabilistic logic to identify or infer the type of information in the string. Use a Classifier
transformation when each input string contains a significant amount of data.
Reference tables: - You use reference tables to verify the accuracy or structure of input data values in data quality transformations.
In addition to the above, Informatica also provides below Content to customers based on the Subscription.
Data Domains Accelerator
Australia/New Zealand Accelerator
Brazil Accelerator
Financial Services Accelerator
France Accelerator
Germany Accelerator
Portugal Accelerator
Spain Accelerator
United Kingdom Accelerator
U.S./Canada Accelerator
Core Accelerator
Core Accelerator :- The Core accelerator contains mapplets, rules and reference data objects that other accelerators can reuse.
Address data cleansing
Contact data cleansing
Corporate data cleansing
General data cleansing
Matching and deduplication data cleansing
Product data cleansing
Here are more details on this topic :- https://idwbi.teachable.com/courses/idq-interview-questions/lectures/3706824
1 A
2 d
3 a
4 b
5 a
6 b
7 c
8 b
9 c
10 d
11 d
12 TRUE
13 c
14 b
15 a
16 a
17 a
18 c
19 b
20 b
21 d
22 ab
23 b
24 b
25 TRUE
26 cde
27 c
28 c
29 d
30 c
31 d
32 c
33 d
34 a
35 b
36 c
37 bd
38 d
39 bd
40 a
41 b
42 A or D
43 a
44 bd
45 b
46 b
47 a
48 a
49 c
50 FALSE
51 d
52 a
53 a
54 d
55 a
56 b
57 c
58 c
59 b
60 d
61 b
62 a
63 c
64 b
65 d
66 b
67 1245
68 c
69 a

You might also like