A Interview Faq's - 2
A Interview Faq's - 2
A Interview Faq's - 2
By PenchalaRaju.Yanamala
1. Slowly changing dimension by default it will take flag among flag, date, and version.
In scd we will do only insert and update.
3. using update strategy t/r at mapping level and session level which scenarios do you
use this.
7. client daggara vunna server unix lo vuntundi, nuvvu ikkada nunchi akkada vunna
data unix command dwara ela techukuntavu.
Ex: 4,5,6....
12. Surrogate key generally system generated key but how does it created? Is
there any option for it?
ANS: U mean we will edit in transformation developer or in mapping designer ani tirigi
manamu question veyyali. But the answer is NO.
16. Peer reviews or code reviews both means same. Get full data regarding this.
where do u store our bugs? If your team member did some mistake and
you found it then what you do?
17. unit test cases and test case documents who will write and how will you do
unit testing? what is the way you approach to do testing? who will approve
your test cases?
21. Inline view or inline query means IF ANY QUERY WHICH FOLLOW WITHIN THE
FROM CLAUSE OF ANOTHER QUERY.
23. How many repositories are there and what are the folders in your
repository?
24. Did you did any mappings? if yes then how did u did that and what is the
difficulty you faced in creating it.
26. Excel sheets are first " file - save as" to CSV files i.e save the excel sheet with an
extension ".csv" and bring into source qualifier. This will be created in the form of
flat files.
27. project architecture must know. what u will do in ur staging area. what u will
do after staging area.
Q) Nuvvu oka target ki both nextval and currval ports attach chesavu then does the
currval works or not?
ANS: Yes it works. nextval, currval both will generate keys. But
currval=nextval+increment by value.
29. Index: b-tree index is preferred for OLTP. bitmap index is used on static data
[unchanged data] on DWH. bitmap is only of non unique index b-tree is of two
types unique and non unique. Composite index means one index created on more
ports.
Ex: create index indexname on table name
<port1, port2, port3, port4>; We can increase performance of index using
partitioning.
30. Difference between oracle 8i and 9i: merge, case, 999 columns we can create,
grouping sets [rollup,cube], list partitioning, nvl2, coalesce, creating user defined
datatypes, multi table inserts.
31. sql: only once we can execute but in plsql we can execute many times.
sql we cannot stop the flow but in plsql we can stop the flow.
sql we cannot manage errors but in plsql we can manage errors {raise_....}
ANS: It is ACTIVE t/r. because if we use primary key then it sorts some records
eventhough it acts as unionall.
33. we can use worklet with in a worklet. open the worklet designer and in tasks
menu click on insert worklet.
ANS: confirmed, degenerated, scd, junk, causal, role playing, status, audit
dimensions. Degenerated: these dimensions identify operations transaction control
number such as invoice [bill]. This exists only in fact table. Role playiing dimension: it
occurs when a physical dimension appears several times in a fact table, each
represented as a separate logical table with unique column with views.
audit: a physical dimension that tags each fact table row with operational metadata
when the row is created.
ANS: additive, semi additive, non additive, psuedo, textual, desired, year to date facts.
ANS: A table with multi part key capturing a many to many relationship that cannotbe
accomodate by the natural granularity of single fact or dimension table. Servers bridge
between the fact table and dimension table to allow many values of dimensions are
ragged hierarchies.
38. U can use server grid to automate the distribution of session. it is a server
object that distributes sessions in workflow to servers based on server
availability. the grid maintains connections to multiple servers.
44. What are the dimension tables and fact tables in ur project?
47. Have u done unit test plan [UTP] OR unit test specifications [UTS]?
ANS: There are 5 phases of test. unit testing, peer review testing, team lead
testing, system intergration, UAT or client testing. but we involve in unit testing and
peer review testing. Unit testing contains of two phases and they are test cases and
test script. In test cases we write columnid,source name,condition,target name, result,
expected result, remarks. In test script we write number of records in the
source = number of records in target.
ORACLE:
49. What is the order of execution of ports that informatica server process?
ANS:
ANS: For example for daily update values we go for mapping variable so that we can
just run the session and change some values in tha mapping parameter instead of
changing all the values.
53. can we insert records into child table without the primarykey of the parent
table?
ANS: YES.
DELETE FROM EMP WHERE ROWID NOT IN (SELECT MAX(ROWID) FROM EMP
GROUP BY EMPNO);
56. Nuvvu chesina mappings ela pamputavu client ki?
ANS:
64. ONE TO MANY RELATION WE USED IN DATAWARE HOUSE. That is one fact
table to many dimension tables.
ANS: 50. I dont deal with all the sources as other team also use other sources. I
didnot remember them.
ANS: 1Terabyte.
ANS: No i wont do a mapping daily. It may take one week or so to complete it.
Q) We have 10 sources. Using joiner T/R how many joins you should use to join
them?
Q) There are 3 sessions, even though 2 sessions fails, my 3rd session should
execute. How do you build this?
Q) How many records you loaded into your project? How much time it takes?
ANS:
ANS: Normal join. because normal join gives records which match the condition.
Where as in case of detail join it gives records which match the condition plus master
table records, so performance decreases. Similarly for master join it gives records
which match the condition plus detail table records, so performance decreeases.
ANS: Yes.
Q) Can you write SQL query in source qualifier transformation for flat files?
ANS: No
Q) What is the use of sorted input option? In which T/R can you useit?
ANS: Input, Output, Lookup, Rank, Variable, return, groupby, master, in-out, key.
ANS:
ANS:
Q) How can you call a stored procedure in a lookup T/R?
Q) If you are using flatfile sources, how do you locate your flatfile?
ANS:
Q) What is the rank data cache size and rank index cache size?
ANS: Data cache is twice the Index cache. Data cache = 2GB Index cache = 1GB.
Q) Expression condition?
ANS: Yes.
Q) From oracle database, we are importing 2 source tables and at that time what
T/R we should use?
ANS:
ANS: It is the directory where all the cache details are stored. It is $PMCachedir
ANS: Exactly we cannot say. But it will contain Lookup, Update strategy, Router,
Stored procedure transformations in more number.
Q) How many ways you can update relational sources and what are they?
Q) Where should you place the flatfile to import the flatfile to the designer?
ANS: It should be placed local to the server or if you can give exact path then you can
place anywhere according to your convinience.
Q) To provide support for mainframes source data which files are used of
source definitions?
ANS: COBOL.
Q) Which T/R should you need while using the gloabal sources as a source
definition?
ANS: In 2 ways.
1. Drag the port from another transforamtion
2. Click the add buttion on the ports tab.
Q) Can you use mappings parameter or variables created in one mapping into
another mapping or mapplet?
ANS: No
Q) What are the types of data that passes between informatica server & stored
procedure?
ANS: Input and output parameters, Return value & Status code.
ANS: you specify the target load order based on source qualifiers in a maping. If u
have the multiple source qualifiers connected to the multiple targets, you can
designate the order in which informatica server loads data into the targets.
Q) What are the basic needs to join two sources in source qualifier T/R?
ANS: The sources must be homogeneous. Primary key and Foreign key relationship
must exists between them.
ANS: This transformation is used to maintain the history data or just most recent
changes in to target table.
Q) What are two levels in update strategy T/R?
ANS: Insert, Delete, Update, Update as update, Update as insert, Update esle insert,
Truncate table.
ANS: Yes. But first we have to copy mappings then we can copy sessions.
ANS: No.
Q) In sequential batch can you run the session if previous session fails?
ANS: Yes.By setting the option always runs the session.
ANS: We can start our required session only in case of sequential batch. But in case
of concurrent batch we cant do like this.
ANS: We cannot. If you want to start batch that resides in a batch,create a new
independent batch and copy the necessary sessions into the new batch.
Q) After dragging the ports of 3 sources to a single source qualifier can you
map three ports directly to the target?
ANS: No. Without having any common ports and unless we join them, we cannot
directly map the ports from source qualifier to target.
ANS: Yes
ANS: Yes.
ANS: No.
ANS: 1. Inner join gives rows which match the condition where as outer join gives all
data.
2. Inner join uses direct table name or alias name. But in outer join we will
mandatorily give alias name for table name.
Q) How to use sequence created in oracle in informatica?
Q) What are minimum and maximum values for index and data cache?
ANS:
Q) Can you update the target at session level also? If yes then why do we use
update strategy T/R?
ANS:
Q) How separate 1:30 to 3:00 'o clock we have maintainance the internal time?
ANS: Email task send email when the workflow is success or fail.
ANS: In our cobol sources we are having single source records but we want
them in multiple records so how many records we want that number we will give
in the 'occurs' option.
Q) Differences between connected and unconnected lookup?
ANS: In the normal mode the data loading into the target in row by row basis and we
can see the information in session log file in row by row. If it fails the session,
then use the session recovery option in the session properties. Then again we
can run the session. Also the target table contains indexes, check constraints,
primary key, foreign keys then you go for normal loading. In the bulk load,
loading the data into the target in group of rows and we can't see the
information where the session fail. In the bulk mode session recovery is not
possible. We can improve the session performance. Whenever you perform the
data loading in bulk mode the target database table [oracle,sybase,SQL server]
doesnot contain indexes, primary key, check constraints and foreign keys.
Q) In the update strategy T/R how to use delete option in session properties?
Q) How to get particular records [Ex: 3rd, 5th, 8th} from a flatfile to load into
target?
ANS: 1) External loading is done by SQL loader. SQL loader is a tool used to perform
bulk loading. 2)
Q) Explain the differences of the following DWH & OLAP, DWH & ODS, ODS &
OLTP, OLTP & DWH, OLTP & DSS.
Q) How do you Import the flat file that is on the Informatica Unix server? I mean
in the source analyzer.
ANS: go to the cmd prompt and ftp the path of the unix box. Then use lcd command
to go to the local current directory where u want to save the file.then use the put
command to copy the file to the local directory.
Q) What are the out put files that the informatica server creates during the
session running?
ANS: Informatica server log, Session log, session detail file, performance detail file,
reject file, control file, post session email, indicator file, output file, cache file.
ANS: Join data originating from same source data base.Filter records when the
informatica server reads source data. Specify an outer join rather than the
default inner join specify sorted records.Select only distinct values from the
source. Creating custom query to issue a special SELECT statement for the
informatica server to read source data.
Q) How many ways you can update a relational source defintion and what are
they?
ANS: NO. Informatica is not at all concern with back end database.It displays you all
the information that is to be stored in repository.If want to reflect back end
changes to informatica screens, again you have to import from back end to
informatica by valid connection.And you have to replace the existing files with
imported files.
Q) How can you recognise whether or not the newly added rows in the source r
gets insert in the target ?
ANS: In the Type2 maping we have three options to recognise the newly added rows
1. Version number 2. Flagvalue 3. Effective date Range.
ANS: source will gets inserted in target along with a new version number. And newly
added dimension in source will inserted into target with a primary key.
Type2 Dimension/Flag current Maping: This maping is also used for slowly
changing dimensions.In addition it creates a flag value for changed or new
dimension. Flag indiactes the dimension is new or newlyupdated.Recent
dimensions will gets saved with cuurent flag value 1. And updated dimensions r
saved with the value 0.
ANS: If one of session is configured to "run if previous completes" and that previous
session fails.
ANS: The Designer automatically creates a RANKINDEX port for each Rank
transformation.The Informatica Server uses the Rank Index port to store the
ranking position for each record in a group. For example, if you create
a Rank transformation that ranks the top 5 salespersons for each quarter, the
rank index numbers the salespeople from 1 to 5.
Q) How can you create or import flat file definition in to the warehouse
designer?
ANS: NO.Unless and until u join those three ports in source qualifier you cannot map
them directly.
ANS: The informatica server builds a cache in memory when it processes the first
row af a data in a cached look up transformation. It allocates memory for the
cache based on the amount you configure in the transformation or session
properties. The informatica server stores condition values in the index cache
and output values in the data cache.
Q) What are the different options used to configure the sequential batches?
ANS: When you add a relational or a flat file source definition to a maping, you need
to connect it to a source qualifier transformation. The source qualifier
transformation represnets the records that the informatica server reads when it
runs a session.
ANS: If you configure a session in a sequential batch to stop on failure, you can run
recovery starting with the failed session. The Informatica Server completes the
session and then runs the rest of the batch. Use the Perform Recovery session
property To recover sessions in sequential batches configured to stop on
failure:
1.In the Server Manager, open the session property sheet.
2.On the Log Files tab, select Perform Recovery, and click OK.
3.Run the session.
4.After the batch completes, open the session property sheet.
5.Clear Perform Recovery, and click OK.
If you do not clear Perform Recovery, the next time you run the session, the
Informatica Server attempts to recover the previous session.
If you do not configure a session in a sequential batch to stop on failure, and the
remaining sessions in the batch complete, recover the failed session as a
standalone session.
Q) What are the new features of the server manager in the informatica 5.0?
ANS: you can use command line arguments for a session or batch.
This allows you to change the values of session parameters,and mapping
parameters and maping variables. Parallel data processig: This feature is
available for powercenter only.If we use the informatica server on a SMP
system,yoU can use multiple CPU's to process a session concurently.
Process session data using threads: Informatica server runs the session in two
processes.
Q) What is transformation?
ANS: If you partition a session with a relational target informatica server creates
multiple connections to the target database to write target data concurently.If
you partition a session with a file target the informatica server creates one target file
for each partition. you can configure session properties to merge these target files.
Q) What are the different types of schemas?
ANS: two types of schemas r there: Star schema and snow flake Schema.
Q) How the informatica server sorts the string values in Rank transformation?
ANS: When the informatica server runs in the ASCII data movement mode it sorts
session data using Binary sortorder. If you configure the seeion to use a binary sort
order, the informatica server caluculates the binary value of each string and returns
the specified number of rows with the higest binary values for the string.
ANS: Status code provides error handling for the informatica server during the
session. The stored procedure issues a status code that notifies whether or not stored
procedure completed sucessfully. This value can not seen by the user.It only used by
the informatica server to determine whether to continue running the session or stop.
ANS: Under certain circumstances, when a session does not complete, you need to
truncate the target tables and run the session from the beginning. Run the
session from the beginning when the Informatica Server cannot run recovery or
when running recovery might result in inconsistent data.
ANS: When using incremental aggregation, you apply captured changes in the source
to aggregate calculations in a session. If the source changes only incrementally
and you can capture changes, you can configure the session to process only
those changes.
This allows the Informatica Server to update your target incrementally, rather than
forcing it to process the entire source and recalculate the same calculations each time
you run the session.
Q) What are the diffrences between joiner transformation and source qualifier
transformation?
ANS: You can join hetrogenious data sources in joiner transformation which we can
not achieve in source qualifier transformation. yoU need matching keys to join
two relational sources in source qualifier transformation. Where as you doesn't
need matching keys to join two sources. Two relational sources should come
from same datasource in source qualifier. you can join relatinal sources which
are coming from different sources also.
Q) If a session fails after loading of 10,000 records in to the target. How can you
load the records from 10001 th record when u run the session next time?
ANS: Data movement modes determines how informatcia server handles the
charector data.You choose the data movement in the informatica server configuration
settings.Two types of data movement modes avialable in informatica.ASCII mode and
Uni code mode.
ANS: For a relational sources informatica server creates multiple connections for each
parttion of a single source and extracts seperate range of data for each
connection. Informatica server reads multiple partitions of a single source
concurently. Similarly for loading also informatica server creates multiple
connections to the target and loads partitions of data concurently. For XML and
file sources,informatica server reads multiple files concurently. For loading the
data informatica server creates a seperate file for each partition(of a source
file).U can choose to merge the targets.
ANS:
Master thread: Creates and manages all other threads
Mapping thread: One maping thread will be creates for each session.Fectchs
session and maping information.
Pre and post session threads: This will be created to perform pre and post
session operations. Reader thread: One thread will be created for each
partition of a source.It reads data from source. Writer thread: It will be created
to load data to the target.Transformation thread: It will be created to tranform
data.
ANS: Maping parameter represents a constant value that you can define before
running a session. A mapping parameter retains the same value throughout the
entire session. When you use the maping parameter , yoU declare and use the
parameter in a maping or maplet. Then define the value of parameter in a
parameter file for the session. Unlike a mapping parameter, a maping variable
represents a value that can change throughout the session. The informatica
server saves the value of maping variable to the repository at the end of session
run and uses that value next time yoU run the session.
ANS: Use lookup transformation in u'r mapping to lookup data in a relational table,
view, synonym. Informatica server queries the look up table based on the
lookup ports in the transformation. It compares the lookup transformation port
values to lookup table column values based on the look up condition. The
default T/R for lookup T/R is Target.
Q) What is polling?
ANS: It displays the updated information about the session in the monitor window.
The monitor window displays the status of each session when U poll the
informatica server.
ANS: you can shedule a session to run at a given time or intervel,or u can manually
run the session. Different options of scheduling Run only on demand: server
runs the session only when user starts session explicitly
Run once: Informatica server runs the session only once at a specified date
and time.
Run every: Informatica server runs the session at regular intervels as u
configured.
Customized repeat: Informatica server runs the session at the dats and times
secified in the repeat dialog box.
Q) Can you copy the session to a different folder or repository?
ANS: Yes. By using copy session wizard u can copy a session in a different folder
Or repository. But that target folder or repository should consists of mapping of
that session. If target folder or repository is not having the maping of copying
session ,you should have to copy that maping first before u copy the session.
Q) What are two types of processes that informatica runs the session?
ANS: Load manager Process: Starts the session, creates the DTM process,
and sends post-session email when the session completes.
DTM process: Creates threads to initialize the session, read, write, and
transform data, and handle pre- and post-session operations.
ANS: A dimension which links with more than one fact table is called as Confirmed
dimesion.
ANS: we have bitmap index, b-tree index, function based index, reverse key index
and composit index. we will use bitmap index in DWH.
ANS: This is the data base used to captur daily business activites and this is
normalized databse.
Q) To achieve the session partition what r the necessary tasks u have to do?
ANS: Configure the session to partition source data. Install the informatica server on a
machine with multiple CPU's.
Q) Which tool you use to create and manage sessions and batches and to
monitor and stop the informaticaserver?
Q) How can you access the remote source into your session?
ANS: When you edit, schedule the sesion each time, informatica server directly
communicates the repository to check whether or not the session and users are
valid. All the metadata of sessions and mappings will be stored in repository.
ANS: After the loadmanger performs validations for session,it creates the DTM
process.DTM is to create and manage the threads that carry out the session
tasks.I creates the master thread.Master thread creates and manges all the
other threads.
ANS: Both pipelines begin with the same original data source.
Both input pipelines originate from the same Source Qualifier transformation.
Both input pipelines originate from the same Normalizer transformation.
Both input pipelines originate from the same Joiner transformation.
Either input pipelines contains an Update Strategy transformation.
Either input pipelines contains a connected or unconnected Sequence
Generator transformation.
ANS: Manages the session and batch scheduling: When you start the informatica
server the load maneger launches and queries the repository for a list of
sessions configured to run on the informatica server. When u configure the
session the loadmanager maintains list of list of sessions and session start
times. When u sart a session loadmanger fetches the session information from
the repository to perform the validations and verifications prior to starting DTM
process.
Locking and reading the session: When the informatica server starts a
session lodamaager locks the session from the repository. Locking prevents U
starting the session again and again.Reading the parameter file: If the session uses a
parameter files,loadmanager reads the parameter file and verifies that the session
level parematers are declared in the file.
Verifies permission and privelleges: When the sesson starts load manger checks
whether or not the user have privelleges to run the session.
Creating log files: Loadmanger creates logfile contains the status of session.
Q) what are the settiings that u use to cofigure the joiner transformation?
ANS: 1.Master and detail source 2.Type of join 3.Condition of the join.
Q) What are the basic needs to join two sources in a source qualifier?
ANS: 1.Two sources should have primary and Foreign key relation ships.
2.Two sources should have matching data types.
Q) How can u work with remote database in informatica? Did you work directly
by using remote connections?
ANS: To work with remote datasource u need to connect it with remote connections.
But it is not preferable to work with that remote source directly by using remote
connections.Instead u bring that source into U r local machine where
informatica server resides. If u work directly with remote source the session
performance will decreases by passing less amount of data across the network
in a particular time.
ANS: Within a session: When you configure a session, you can instruct the
Informatica Server to either treat all records in the same way (for example, treat
all records as inserts), or use instructions coded into the session mapping to flag
records for different database operations. Within a mapping: Within a mapping,
you use the Update Strategy transformation to flag records.
Q) What is Datadriven?
ANS: The informatica server follows instructions coded into update strategy
transformations with in the session mapping determine how to flag records for
insert,update, delete or reject If you do not choose data driven option setting,
the informatica server ignores all update strategy transformations in the
mapping.
Q) What are the basic needs to join two sources in a source qualifier?
ANS: View - store the SQL statement in the database and let you use it as a table.
Every time you access the view, the SQL statement executes.
Materialized view - Materialized view can be used to precalculate the
expensive joins and aggregates prior to execution and the result is stored in a
table in database and can refer to it in future. the adv of this is increse in
performance. it works as snapshot. View is nothing but an sql query stored. it
will not store data in tables.
Q) Can you use the maping parameters or variables created in one maping into
any other reusable transformation?
ANS: Yes. Because reusable tranformation is not contained with any maplet or
maping.
ANS: The PowerCenter repository allows you to share metadata across repositories
to create a data mart domain. In a data mart domain, you can create a single
global repository to store metadata used across an enterprise, and a number of
local repositories to share the global metadata as needed.
ANS: It is a web based application that enables you to run reports againist repository
metadata. with a meta data reporter,you can access information about your
repository with out having knowledge of sql,transformation language or
underlying tables in the repository.
Q) What is the filename which you need to configure in UNIX while installing
infromatica?
ANS: When a Joiner transformation occurs in a session, the Informatica Server reads
all the records from the master source and builds index and data caches based
on the master rows.After building the caches, the Joiner transformation reads
records from the detail source and perform joins.
ANS: A surrogate key is a substitution for the natural primary key. It is a unique
identifier or number (normally created by a database sequence generator ) for
each record of a dimension table that can be used for the primary key to the
table. A surrogate key is useful because natural keys may change.
Q) What is a Cube?
ANS: DRILL DOWN, DRILL ACROSS and TIME HANDLING. To be able to drill
down/drill across is the most basic requirement of an end user in a
datawarehouse. Drilling down most directly addresses the natural end-user
need to see more detail in an result. Drill down should be as generic as
possible becuase there is absolutely no good way to predict users drill-down
path.
ANS: Unlike most dimensions "Time dimension" do not change. You can populate it
once and use for years. So the easiest way is to use spread-sheet.
ANS: Real Time Data warehous is an analytic component of an enterprise level data
stream that supports continuous, asynchronous, multi-point delivery of data. In
a RTDW data moves straight from the source systems to decision makers
without any form for staging.
ANS: When a non-key attribute identifys the value of another non-key atribute then
the table is set to contain transitive dependecncy.
Q) What oracle tools are available to design and build a data warehosue/data
mart?
ANS: Data Warehouse Builder, Oracle Designer, Oracle Express, Express Objects
etc.
ANS: You can instruct the PowerCenter Server to rebuild the lookup cache if you
need to by checking "Re-cache from lookup source option" in the lookup
transformation properties tab.