Ab Initio Scenario Based Interview Questions and Answers
Ab Initio Scenario Based Interview Questions and Answers
And Answers
1. Mention what information does a .dbc file extension provides to connect to the
database?
Answer: The .dbc extension provides the GDE with the information to connect with the database
are
Name and version number of the database to which you want to connect
Name of the computer on which the database instance or server to which you want to
connect runs, or on which the database remote access software is installed
Name of the server, database instance or provider to which you want to link.
3. Suppose we assign you a new project. What would be your initial point and the key steps
that you follow?
Answer: The first thing that largely matters is defining the objective of the task and then engages
the team in it. This provides a solid direction for the accomplishment of the task. This is
important when one is working on a set of data which is completely unique or fresh. After this,
the next big thing that needs attention is effective data modeling. This includes finding the
missing values and data validation. The last thing is to track the results.
4. What do you mean by the term data warehousing? Is it different from Data Mining?
Answer: Many times there is a need to have data retrieval, warehousing can simply be
considered to assure the same without affecting the efficiency of operational systems. It simply
supports decision support and always works in addition to the business applications and
Customer Relationship Management and warehouse architecture. Data mining is closely related
to this approach. It assures simple findings of required operators from the warehouse.
6. What is a cursor? Within a cursor, how would you update fields on the row just fetched?
Answer: The oracle engine uses work areas for internal processing in order to the execute sql
statement is called cursor. There are two types of cursors like Implicit cursor and an explicit
cursor. An implicit cursor is using for internal processing and Explicit cursor is using for user
open for data required.
10. Do you think effective communication is necessary for data processing? What is your
strength in terms of same?
Answer: The biggest ability that one could have in this domain is the ability to rely on the data
or the information. Of course, communication matters a lot in accomplishing several important
tasks such as the representation of the information. There are many departments in an
organization and communication make sure things are good and reliable for everyone.
19. What would be the next step after collecting the data?
Answer: Once the data is collected, the next important task is to enter it in the concerned
machine or system. Well, gone are those days when storage depends on papers. In the present
time, data size is very large and it needs to be performed in a reliable manner. A digital approach
is a good option for this as it simply lets users perform this task easily and in fact without
compromising with anything. A large set of operations then need to be performed for the
meaningful analysis. In many cases, the conversion also largely matters and users are always free
to consider the outcomes which best meet their expectations.
20. Suppose you find the term Validation mentioned with a set of data, what does that
simply represent?
Answer: It represents that the concerned data is clean, correct and can thus be used reliably
without worrying about anything. Data validation is widely regarded as the key points in the
processing system.
21. How scientific data processing is different from commercial data processing?
Answer: Scientific data processing simply means data with a great amount of computation i.e.
arithmetic operations. In this, a limited amount of data is provided as input and a bulk data is
there at the outcome. On the other hand, commercial data processing is different. In this, the
outcome is limited as compared to the input data. The computational operations are limited in the
commercial data processing.
22. Name any two stages of the data processing cycle and provide your answer in terms of a
comparative study of them?
Answer: The first is Collection and the second one is the preparation of data. Of course, the
collection is the first stage and preparation is the second in a cycle dealing with data processing.
The first stage provides a baseline to the second and the success and simplicity of the first
depends on how accurately the first has been accomplished. Preparation is mainly the
manipulation of important data. The collection breaks data sets while Preparation joins them
together.
23. What do you mean by a transaction file and how it is different from that of a Sort file?
Answer: The Transaction file is generally considered to hold input data and that is for the time
when a transaction is under process. All the master files can be updated with it simply. Sorting is
done to assign a fixed location to the data files on the other hand.
27. What are the different forms of output that can be obtained after processing of data?
Answer: These are
1. Tables
2. Plain Text files
3. Image files
4. Maps
5. Charts
6. Vectors
7. Raw files
Sometimes data is required to be produced in more than one format and therefore the software
accomplishing this task must-have features available in it to keep up the pace in this matter.
28. What exactly do you know about the typical data analysis?
Answer: It generally involves the organization as well as the collection of important files in the
form of important files. The main aim is to know the exact relationship among the industrial data
or the full data and the one which is analyzed. Some experts also call it one of the best available
approaches to find errors. It entails the ability to spot problems and enable the operator to find
out the root causes of the errors.
In case of reformat if the destination field names are same or a subset of the source fields then no
need to write anything in the reformat xfr unless you won’t want to use any real transform other
than reducing the set of fields or split the flow into a number of flows to achieve the
functionality.
REVOKE means cancel the grant (permissions). So, Grant or Revoke both commands depend
upon D.B.A..
32. How would you find out whether a SQL query is using the indices you expect?
Answer: Explain plan can be reviewed to check the execution plan of the query. This would
guide if the expected indexes are used or not.
36. What is the difference between a Scan component and a RollUp component?
Answer: Rollup is for the group by and Scan is for the successive total. Basically, when we need
to produce a summary then we use scan. Rollup is used to aggregate data.
37. What is the Difference between DML Expression and XFR Expression?
Answer: The main difference b/w DML & xfr is that
DML represents the format of the metadata.
XFR represent the transform functions.which will contain business
rules
1. Real-Time processing
2. Multiprocessing
3. Time-Sharing
4. Batch processing
5. Adequate Processing
42. What is the diff b/w look-up file and look-up, with a relevant example?
Answer: Generally, the Lookup file represents one or more serial files (Flat files). The amount
of data is small enough to be held in memory. This allows transform functions to retrieve records
much more quickly than it could retrieve from Disk.
Set AB_AIR_ROOT
Login to EME web interface- http://serverhost:[serverport]/abinitio
Through GDE, you can connect to EME data-store
Through air-command
45. How can you force the optimizer to use a particular index?
Answer:
Use hints /*+ */, these acts as directives to the optimizer
46. What are the operations that support avoiding duplicate record?
Answer: Duplicate records can be avoided by using the following:
54. What are the facts that can compromise data integrity?
Answer: There are several errors that can cause this issue and can transform many other
problems. These are:
57. Can sorting and storing be done through single software or you need different for these
approaches?
Answer: Well, it actually depends on the type and nature of the data. Although it is possible to
accomplish both these tasks through the same software, many software has their own
specialization and it would be good if one adopts such an approach to get the quality outcomes.
There are also some pre-defined set of modules and operations that largely matters. If the
conditions imposed by them are met, users can perform multiple tasks with similar software. The
output file is provided in the various formats.