0% found this document useful (0 votes)
7 views

Datastage_Interview1

Uploaded by

sai Charan
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as RTF, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
7 views

Datastage_Interview1

Uploaded by

sai Charan
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as RTF, PDF, TXT or read online on Scribd
You are on page 1/ 3

Read the String functions in DS

A: Functions like [] -> sub-string function and ':' -> concatenation operator

Syntax: string [ [ start, ] length ]

string [ delimiter, instance, repeats ]

What are Sequencers?

A: Sequencers are job control programs that execute other jobs with preset Job parameters.

2. What are other Performance tunings you have done in your last project to increase the performance
of slowly running jobs?

A:

1. Staged the data coming from ODBC/OCI/DB2UDB stages or any database on the server using
Hash/Sequential files for optimum performance also for data recovery in case job aborts.

2. Tuned the OCI stage for 'Array Size' and 'Rows per Transaction' numerical values for faster inserts,
updates and selects.

3. Tuned the 'Project Tunables' in Administrator for better performance.

4. Used sorted data for Aggregator.

5. Sorted the data as much as possible in DB and reduced the use of DS-Sort for better performance of
jobs

6. Removed the data not used from the source as early as possible in the job.

7. Worked with DB-admin to create appropriate Indexes on tables for better performance of DS queries

8. Converted some of the complex joins/business in DS to Stored Procedures on DS for faster execution
of the jobs.

9. If an input file has an excessive number of rows and can be split-up then use standard logic to run jobs
in parallel.

10. Before writing a routine or a transform, make sure that there is not the functionality required in one
of the standard routines supplied in the sdk or ds utilities categories.

Constraints are generally CPU intensive and take a significant amount of time to process. This may be the
case if the constraint calls routines or external macros but if it is inline code then the overhead will be
minimal.

11. Try to have the constraints in the 'Selection' criteria of the jobs itself. This will eliminate the
unnecessary records even getting in before joins are made.

12. Tuning should occur on a job-by-job basis.

13. Use the power of DBMS.

14. Try not to use a sort stage when you can use an ORDER BY clause in the database.

15. Using a constraint to filter a record set is much slower than performing a SELECT … WHERE….

16. Make every attempt to use the bulk loader for your particular database. Bulk loaders are generally
faster than using ODBC or OLE.

Do u know about METASTAGE?

A: MetaStage is used to handle the Metadata which will be very useful for data lineage and data analysis
later on. Meta Data defines the type of data we are handling. This Data Definitions are stored in
repository and can be accessed with the use of MetaStage.

Explain the differences between Oracle8i/9i?

Oracle 8i does not support pseudo column sysdate but 9i supports

Oracle 8i we can create 256 columns in a table but in 9i we can upto 1000 columns(fields)

. How do you merge two files in DS?

A: Either use Copy command as a Before-job subroutine if the metadata of the 2 files are same or create
a job to concatenate the 2 files into one if the metadata is different.

28. What are Static Hash files and Dynamic Hash files?

A: As the names itself suggest what they mean. In general we use Type-30 dynamic Hash files. The Data
file has a default size of 2Gb and the overflow file is used if the data exceeds the 2GB size.

what is Hash file stage and what is it used for?

A: Used for Look-ups. It is like a reference table. It is also used in-place of ODBC, OCI tables for better
performance.
30. Have you ever involved in updating the DS versions like DS 5.X, if so tell us some the steps you
have taken in doing so?

A: Yes. The following are some of the steps; I have taken in doing so:

1) Definitely take a back up of the whole project(s) by exporting the project as a .dsx file

2) See that you are using the same parent folder for the new version also for your old jobs using the
hard-coded file path to work.

3) After installing the new version import the old project(s) and you have to compile them all again. You
can use 'Compile All' tool for this.

4) Make sure that all your DB DSN's are created with the same name as old one's. This step is for moving
DS from one machine to another.

5) In case if you are just upgrading your DB from Oracle 8i to Oracle 9i there is tool on DS CD that can do
this for you.

6) Do not stop the 6.0 server before the upgrade, version 7.0

install process collects project information during the upgrade. There is NO rework (recompilation of
existing jobs/routines) needed after the upgrade.

31. Did you Parameterize the job or hard-coded the values in the jobs?

A: Always parameterized the job. Either the values are coming from Job Properties or from a ‘Parameter
Manager’ – a third part tool. There is no way you will hard–code some parameters in your jobs. The
often Parameterized variables in a job are: DB DSN name, username, password, dates W.R.T for the data
to be looked against at.

What is Difference between E-R Modeling and Dimentional Modeling

Basic diff is E-R modeling will have logical and physical model. Dimensional model will have only physical
model.

E-R modeling is used for normalizing the OLTP database design.

Dimensional modeling is used for de-normalizing the ROLAP/MOLAP design.

You might also like