Exam A
Exam A
QUESTION 1
In your ETL application design you have found several areas of common processing requirements in the mapping specifications. These common logic
areas found include: code validation lookups and name formatting. The common logic areas have the same logic, but the jobs using them would have
different column metadata. Choose the action that gives you the best reusability design to effectively implement these common logic areas in your ETL
application?
A. Create parallel routines for each of the common logic areas and for each of the unique column metadata formats.
B. Create separate jobs for each layout and choose the appropriate job to run within a job sequencer.
C. Create parallel shared containers and define columns combining all data formats.
D. Create parallel shared containers with Runtime Column Propagation (RCP) ON and define only necessary common columns needed for the
logic.
Answer: D
Explanation/Reference:
Explanation:
QUESTION 2
When optimizing a job, Balanced Optimization will NOT search the job for what pattern?
A. Links
B. Stages
C. Sequencers
D. Property Settings
Answer: C
Explanation/Reference:
Explanation:
QUESTION 3
You are asked to optimize the fork join job design in the exhibit. This job uses the sort aggregator and a left outer join on the ZIP code column.
Currently all partitioning is set to "Auto" and automatic sort insertion is allowed.
Which change will reduce the cost of partitioning that occurs in this job?
A. Use Entire partitioning on the input links to the Aggregator and Join stages.
B. Hash partition and sort on ZIP code column on the input links to the Aggregator and Join stages.
C. Hash partition and sort on ZIP code column prior to the Copy stage and use entire partitioning on the Aggregator and Join stages.
D. Hash partition and sort on ZIP code column prior to the Copy stage, and use same partitioning on the Aggregator and Join stages.
Answer: D
Explanation/Reference:
Explanation:
QUESTION 4
You have a parallel job that based on operational recoverability requirements needs to be broken up into two separate parallel jobs. You have decided
to use the Data Set stage to support this job design change. What two characteristics of Data Sets make them a good design consideration in your jobs
design change? (Choose two.)
Answer: CD
Explanation/Reference:
Explanation:
QUESTION 5
What two binding types are supported by Information Services Director (ISD) for a parallel job that is designed to be used as a service? (Choose two.)
A. EJB
B. SQL
C. HDFS
D. SOAP
E. STREAMS
Answer: AD
Explanation/Reference:
Explanation:
QUESTION 6
You are assigned to correct a job from another developer. The job contains 20 stages sourcing data from two Data Sets and many sequential files. The
annotation in the job indicates who wrote the job and when, not the objective of the job. All link and stage names use the default names. One of the
output columns has an incorrect value which should have been obtained using a lookup. What could the original developer have done to make this task
easier for maintenance purposes?
Answer: B
Explanation/Reference:
Explanation:
QUESTION 7
You are asked by management to document all jobs written to make future maintenance easier.
Which statement is true about annotations?
A. The short job description can be identified within the Description Annotation stage.
B. The Description Annotation stage contains both the short and full descriptions for the job.
C. The background for the Description Annotation stage can be changed for each unique stage.
D. The Description Annotation stage can be added several times at different locations to identify business logic.
Answer: A
Explanation/Reference:
Explanation:
QUESTION 8
A job design consists of an input Row Generator stage, a Filter stage, followed by a Transformer stage and an output Sequential File stage. The job is
run on an SMP machine with a configuration file defined with three nodes. The $APT_DISABLE_COMBINATION variable is set to True. How many
player processes will this job generate?
A. 8
B. 10
C. 12
D. 16
Answer: A
Explanation/Reference:
Explanation:
QUESTION 9
A. Same
B. Modulus
C. Sort Merge
D. Round Robin
Answer: B
Explanation/Reference:
Explanation:
QUESTION 10
A job design consists of an input Row Generator stage, a Sort stage, followed by a Transformer stage and an output Data Set stage. The job is run on
an SMP machine with a configuration file defined with four nodes. The $APT_DISABLE_COMBINATION variable is set to True. How many player
processes will this job generate?
A. 7
B. 16
C. 13
D. 16
Answer: C
Explanation/Reference:
Explanation:
QUESTION 11
The data going into the target Sequential Files stage is range-partitioned and sorted. Which technique method would be the most efficient to create a
globally sorted target sequential file?
Answer: B
Explanation/Reference:
Explanation:
QUESTION 12
In the exhibit, a Funnel stage has two input links. Input 1 (Seq_File) comes from a Sequential File stage with "Readers per Node" set to "2". Input 2
(Dataset) comes from a dataset created with 3 partitions. In the Funnel stage, the funnel type is set to "Sequence".
The parallel configuration file contains 4 nodes. How many instances of the Funnel stage run in parallel?
A. 1
B. 2
C. 4
D. 6
Answer: C
Explanation/Reference:
Explanation:
QUESTION 13
Your job sequence must be restartable. It runs Job1, Job2, and Job3 serially. It has been compiled with "Add checkpoints so sequence is restartable".
Job1 must execute every run even after a failure. Which two properties must be selected to ensure that Job1 is run each time, even after a failure?
(Choose two.)
Answer: AD
Explanation/Reference:
Explanation:
QUESTION 14
Which two actions are available when editing a message handler? (Choose two.)
A. Abort job
B. Demote to warning
C. Suppress from job log
D. Demote to informational
E. Suppress from the project
Answer: CD
Explanation/Reference:
Explanation:
QUESTION 15
What is the result of running the following command: dsjob -report DSProject ProcData
A. Generates a report about the ProcData job, including information about its stages and links.
B. Returns a report of the last run of the ProcData job in a DataStage project named DSProject.
C. Runs the DataStage job named ProcData and returns performance information, including the number of rows processed.
D. Runs the DataStage job named ProcData and returns job status information, including whether the job aborted or ran without warnings.
Answer: B
Explanation/Reference:
Explanation:
QUESTION 16
You would like to pass values into parameters that will be used in a variety of downstream activity stages within a job sequence. What are two valid
ways to do this? (Choose two.)
Answer: AE
Explanation/Reference:
Explanation:
QUESTION 17
On the DataStage development server, you have been making enhancements to a copy of a DataStage job running on the production server. You have
been asked to document the changes you have made to the job. What tool in DataStage Designer would you use?
A. Compare Against
B. diffapicmdline.exe
C. DSMakeJobReport
D. Cross Project Compare
Answer: D
Explanation/Reference:
Explanation:
QUESTION 18
You are working on a project that contains a large number of jobs contained in many folders. You would like to review the jobs created by a former
developer of the project. How can you find these jobs?
Answer: B
Explanation/Reference:
Explanation:
QUESTION 19
Your customer is using Source Code Control Integration for Information server and have tagged artifacts for version 1. You must create a deployment
package from the version 1. Before you create the package you will have to ensure the project is up to date with version 1. What two things must you
do to update the meta-data repository with the artifacts tagged as version 1? (Choose two.)
Answer: DE
Explanation/Reference:
Explanation:
QUESTION 20
You want to find out which table definitions have been loaded into a job, and specifically which stages of the job they has been loaded into? How will
you determine this?
A. Select the job, right-click, then click the Find where used (deep) command.
B. Select the job, right-click, then click the Find dependencies (deep) command.
C. Select the job, right-click, then click the Find where used command. Then right-click and select "Show the dependency path from the job".
D. Select the job, right-click, then click the Find dependencies command. Then right-click and select "Show the dependency path from the job".
Answer: D
Explanation/Reference:
Explanation:
QUESTION 21
You are responsible for deploying objects into your customers production environment. To ensure the stability of the production system the customer
does not permit compilers on production machines. They have also protected the project and only development machines have the required compiler.
What option will enable jobs with a parallel transformer to execute in the customers production machines?
A. Add $APT_COMPILE_OPT=-portable
B. Set $APT_COPY_TRANSFORM_OPERATOR
C. Use protected projects in the production environment.
D. Create a package with Information Server Manager and select the option to include executables.
Answer: D
Explanation/Reference:
Explanation:
QUESTION 22
What two features distinguish the Operations Console from the Director job log? (Choose two.)
A. Jobs can be started and stopped in Director, but not in the Operations Console.
B. The Operations Console can monitor jobs running on only one DataStage engine.
C. Workload management is supported within Director, but not in the Operations Console.
D. The Operations Console can monitor jobs running on more than one DataStage engine.
E. The Operations Console can run on systems where the DataStage clients are not installed.
Answer: DE
Explanation/Reference:
Explanation:
QUESTION 23
A. Stages
B. File sets
C. Schemas
D. Data sets
E. Operators
Answer: DE
Explanation/Reference:
Explanation:
QUESTION 24
Which two environment variables add additional reporting information in the job log for DataStage jobs? (Choose two.)
A. $APT_IO_MAP
B. $OSH_EXPLAIN
C. $APT_STARTUP_STATUS
D. $APT_EXPORT_FLUSH_COUNT
E. $APT_PM_STARTUP_CONCURRENCY
Answer: BC
Explanation/Reference:
Explanation:
QUESTION 25
A job validates account numbers with a reference file using a Join stage, which is hash partitioned by account number. Runtime monitoring reveals that
some partitions process many more rows than others. Assuming adequate hardware resources, which action can be used to improve the performance
of the job?
Answer: B
Explanation/Reference:
Explanation:
QUESTION 26
You are asked by your customer to collect partition level runtime metadata for DataStage parallel jobs. You must collect this data after each job
completes. What two options allow you to automatically save row counts and CPU time for each instance of an operator? (Choose two.)
A. $APT_CPU_ROWCOUNT
B. $APT_PERFORMANCE_DATA
C. Enable the job property "Record job performance data".
D. Open up the job in Metadata Workbench and select the "Data Lineage" option.
E. Click the Performance Analysis icon in the toolbar to open the Performance Analyzer utility.
Answer: BC
Explanation/Reference:
Explanation:
QUESTION 27
Which option is required to identify a particular job player processes?Which option is required to identify a particular job? player processes?
Answer: B
Explanation/Reference:
Explanation:
QUESTION 28
A. Import another copy of the table metadata into the project where it is required.
B. Use the "Shared Table Creation Wizard" to place the table in the shared repository.
C. Export DataStage table definitions from one project and importing them into another project.
D. Use the InfoSphere Metadata Asset Manager (IMAM) to move the DataStage table definition to the projects where it is needed.
Answer: B
Explanation/Reference:
Explanation:
QUESTION 29
Which two parallel job stages allow you to use partial schemas? (Choose two.)
A. Peek stage
B. File Set stage
C. Data Set stage
D. Column Export stage
E. External Target stage
Answer: BE
Explanation/Reference:
Explanation:
QUESTION 30
In addition to the table and schema names, what two element names must be specified when you create a shared table definition in DataStage
Designer? (Choose two.)
A. Database
B. Host system
C. Project name
D. Database instance
E. DataStage server system name
Answer: AB
Explanation/Reference:
Explanation:
QUESTION 31
When using Runtime Column Propagation, which two stages require a schema file? (Choose two.)
A. Peek stage
B. Pivot stage
C. Column Import stage
D. DB2 Connector stage
E. Sequential File stage
Answer: CE
Explanation/Reference:
Explanation:
QUESTION 32
What are the two Transfer Protocol Transfer Mode property options for the FTP Enterprise stage? (Choose two.)
A. FTP
B. EFTP
C. TFTP
D. SFTP
E. RFTP
Answer: AD
Explanation/Reference:
Explanation:
QUESTION 33
Your job will write its output to a fixed length data file. When configuring the sequential file stage as a target what format and column tab properties
need to be considered for this type of file output?
A. On the Output Link format tab, change the 'Delimiter' property to whitespace.
B. On the Output Link format tab, add the 'Record Type' property to the tree and set its value to be 'F'.
C. On the Output Link column tab, insure that all the defined column data types are fixed length types.
D. On the Output Link column tab, specify the record size total based on all of the columns defined
Answer: C
Explanation/Reference:
Explanation:
QUESTION 34
Identify the two statements that are true about the functionality of the XML Pack 3.0. (Choose two.)
Answer: CE
Explanation/Reference:
Explanation:
QUESTION 35
Identify the two delimiter areas available to be configured in the Sequential File format tab properties? (Choose two.)
A. File delimiter
B. Null delimiter
C. Final delimiter
D. Field delimiter
E. End of group delimiter
Answer: CD
Explanation/Reference:
Explanation:
QUESTION 36
When using a Sequential File stage as a source what are the two reject mode property options? (Choose two.)
A. Set
B. Fail
C. Save
D. Convert
E. Continue
Answer: BE
Explanation/Reference:
Explanation:
QUESTION 37
Which two statements are true about Data Sets? (Choose two.)
Answer: BD
Explanation/Reference:
Explanation:
QUESTION 38
What is the correct method to process a file containing multiple record types using a Complex Flat File stage?
Answer: D
Explanation/Reference:
Explanation:
QUESTION 39
When using the Column Export stage, what are two export column type property values allowed for the combined single output column result? (Choose
two.)
A. Vector
B. Binary
C. Integer
D. Decimal
E. VarChar
Answer: BE
Explanation/Reference:
Explanation:
QUESTION 40
Which two file stages allow you to configure rejecting data to a reject link? (Choose two.)
Answer: CE
Explanation/Reference:
Explanation:
QUESTION 41
Identify two items that are created as a result of running a Balanced Optimization on a job that accesses a Hadoop distributed file system as a source?
(Choose two.)
Answer: AC
Explanation/Reference:
Explanation:
QUESTION 42
A customer must compare a date column with a job parameter date to determine which output links the row belongs on. What stage should be used for
this requirement?
A. Filter stage
B. Switch stage
C. Compare stage
D. Transformer stage
Answer: D
Explanation/Reference:
Explanation:
QUESTION 43
Rows of data going into a Transformer stage are sorted and hash partitioned by the Input.Product column. Using stage variables, how can you
determine when a new row is the first of a new group of Product rows?
A. Create a stage variable named sv_IsNewProduct and follow it by a second stage variable named sv_Product. Map the Input.Product column to
sv_Product. The derivation for sv_IsNewProduct is: IF Input.Product = sv_Product THEN "YES" ELSE "NO".
B. Create a stage variable named sv_IsNewProduct and follow it by a second stage variable named sv_Product. Map the Input.Product column to
sv_Product. The derivation for sv_IsNewProduct is: IF Input.Product <> sv_Product THEN "YES" ELSE "NO".
C. Create a stage variable named sv_Product and follow it by a second stage variable named sv_IsNewProduct. Map the Input.Product column to
sv_Product. The derivation for sv_IsNewProduct is: IF Input.Product = sv_Product THEN "YES" ELSE "NO".
D. Create a stage variable named sv_Product and follow it by a second stage variable named sv_IsNewProduct. Map the Input.Product column to
sv_Product. The derivation for sv_IsNewProduct is: IF Input.Product <> sv_Product THEN "YES" ELSE "NO".
Answer: B
Explanation/Reference:
Explanation:
QUESTION 44
Which statement describes what happens when Runtime Column Propagation is disabled for a parallel job?
A. An input column value flows into a target column only if it matches it by name.
B. An input column value flows into a target column only if it is explicitly mapped to it.
C. You must set APT_AUTO_MAP project environment to true to allow output link mapping to occur.
D. An input column value flows into a target column based on its position in the input row. For example, first column in the input row goes into the first
target column.
Answer: B
Explanation/Reference:
Explanation:
QUESTION 45
Which statement is true when using the SaveInputRecord() function in a Transformer stage.
A. You can only use the SaveInputRecord() function in Loop variable derivations.
B. You can access the saved queue records using Vector referencing in Stage variable derivations.
C. You must retrieve all saved queue records using the GetSavedInputRecord() function within Loop variable derivations.
D. You must retrieve all saved queue records using the GetSavedInputRecord() function within Stage variable derivations.
Answer: C
Explanation/Reference:
Explanation:
QUESTION 46
In the Slowly Changing Dimension stage, a dimension columns Purpose code property can trigger which two actions. (Choose two.)
Answer: BC
Explanation/Reference:
Explanation:
QUESTION 47
Answer: C
Explanation/Reference:
Explanation:
QUESTION 48
In a Transformer, which two mappings can be handled by default type conversions. (Choose two.)
Answer: DE
Explanation/Reference:
Explanation:
QUESTION 49
Identify two different types of custom stages you can create to extend the Parallel job syntax? (Choose two.)
A. Input stage
B. Basic stage
C. Group stage
D. Custom stage
E. Wrapped stage
Answer: DE
Explanation/Reference:
Explanation:
QUESTION 50
Which two statements are true about stage variables in a Transformer Stage? (Choose two.)
Answer: AC
Explanation/Reference:
Explanation:
QUESTION 51
Answer: D
Explanation/Reference:
Explanation:
QUESTION 52
Suppose a user ID has been created with DataStage and QualityStage component authorization. Which client application would be used to give that
user ID DataStage Developer permission?
Answer: C
Explanation/Reference:
Explanation:
QUESTION 53
Which two data repositories can be used for user authentication within the Information Server Suite? (Choose two.)
Answer: CE
Explanation/Reference:
Explanation:
QUESTION 54
Which two statements are true about the use of named node pools? (Choose two.)
A. Grid environments must have named node pools for data processing.
B. Named node pools can allow separation of buffering from sorting disks.
C. When named node pools are used, DataStage uses named pipes between stages.
D. Named node pools limit the total number of partitions that can be specified in the configuration file.
E. Named node pools constraints will limit stages to be executed only on the nodes defined in the node pools.
Answer: BE
Explanation/Reference:
Explanation:
QUESTION 55
Which step is required to change from a normal lookup to a sparse lookup in an ODBC Connector stage?
Answer: C
Explanation/Reference:
Explanation:
QUESTION 56
Which method is used to specify when to stop a job because of too many rejected rows with an ODBC Connector?
Answer: B
Explanation/Reference:
Explanation:
QUESTION 57
Which two pieces of information are required to be specified for the input link on a Netezza Connector stage? (Choose two.)
A. Partitioning
B. Server name
C. Table definitions
D. Buffering settings
E. Error log directory
Answer: AD
Explanation/Reference:
Explanation:
QUESTION 58
Which requirement must be met to read from a database in parallel using the ODBC connector?
Answer: B
Explanation/Reference:
Explanation:
QUESTION 59
Which two statements about the Additional Connections Options property in the Teradata Connector stage to specify details about the number of
connections to Teradata are true? (Choose two.)
Answer: BD
Explanation/Reference:
Explanation:
QUESTION 60
Configuring the weighting column of an Aggregator stage affects which two options. (Choose two.)
A. Sum
B. Maximum Value
C. Average of Weights
D. Coefficient of Variation
E. Uncorrected Sum of Squares
Answer: AE
Explanation/Reference:
Explanation:
QUESTION 61
The parallel framework was extended for real-time applications. Identify two of these aspects.
(Choose two.)
A. XML stage.
B. End-of-wave.
C. Real-time stage types that re-run jobs.
D. Real-time stage types that keep jobs always up and running.
E. Support for transactions within source database connector stages.
Answer: BD
Explanation/Reference:
Explanation:
QUESTION 62
How must the input data set be organized for input into the Join stage? (Choose two.)
A. Unsorted
B. Key partitioned
C. Hash partitioned
D. Entire partitioned
E. Sorted by Join key
Answer: BE
Explanation/Reference:
Explanation:
QUESTION 63
The Change Apply stage produces a change Data Set with a new column representing the code for the type of change. What are two change values
identified by these code values? (Choose two.)
A. Edit
B. Final
C. Copy
D. Deleted
E. Remove Duplicates
Answer: CD
Explanation/Reference:
Explanation:
QUESTION 64
A. Join stage
B. Merge stage
C. Lookup stage
D. Funnel stage
Answer: B
Explanation/Reference:
Explanation:
QUESTION 65
A 100MB input dataset has even distribution across 400 unique key values. When you run with a 4-node configuration file, which two changes could
improve sort performance in this scenario? (Choose two.)
Answer: BE
Explanation/Reference:
Explanation:
QUESTION 66
A. The Data Rules stage works with rule definitions only; not executable rules.
B. As a best practice, you should create and publish new rules from the Data Rules stage.
C. If you have Rule Creator role in InfoSphere Information Analyzer, you can create and publish rule definitions and rule set definitions directly
from the stage itself.
D. When a job that uses the Data Rules stage runs, the output of the stage is passed to the downstream stages and results are stored in the
Analysis Results database (IADB).
Answer: A
Explanation/Reference:
Explanation:
QUESTION 67
Which job design technique can be used to give unique names to sequential output files that are used in multi-instance jobs?
Topic 2, Volume B
QUESTION 68
The ODBC stage can handle which two SQL Server data types? (Choose two.)
A. Date
B. Time
C. GUID
D. Datetime
E. SmallDateTime
Answer: CE
Explanation/Reference:
Explanation:
QUESTION 69
Which DB2 to InfoSphere DataStage data type conversion is correct when reading data with the DB2 Connector stage?
A. XML to SQL_WVARCHAR
B. BIGINT to SQL_BIGINT (INT32)
C. VARCHAR, 32768 to SQL_VARCHAR
D. CHAR FOR BIT DATA to SQL_VARBINARY
Answer: A
Explanation/Reference:
Explanation:
QUESTION 70
Answer: B
Explanation/Reference:
Explanation:
QUESTION 71
Which two statements about using a Load write method in an Oracle Connector stage to tables that have indexes on them are true? (Choose two.)
Answer: CE
Explanation/Reference:
Explanation:
QUESTION 72
Which Oracle Connector stage property can be set to tune job performance?
A. Array size
B. Memory size
C. Partition size
D. Transaction size
Answer: A
Explanation/Reference:
Explanation:
QUESTION 73
In a Transformer expression for a stage variable, there is a nullable input column. Assume the legacy NULL processing option is turned off. What
happens when a row is processed that contains NULL in that input column?
Answer: C
Explanation/Reference:
Explanation:
QUESTION 74
Identify two different types of custom stages you can create to extend the Parallel job syntax? (Choose two.)
A. Input stage
B. Basic stage
C. Group stage
D. Custom stage
E. Wrapped stage
Answer: DE
Explanation/Reference:
Explanation:
QUESTION 75
When using the loop functionality in a transformer, which statement is true regarding Transformer processing.
QUESTION 76
In your job you are using a Sequential File stage as a source. You have decided to use a schema file as an alternative way of specifying column
definitions and record format properties. What two items are true in configuring this stage in your parallel job? (Choose two.)
Answer: AD
Explanation/Reference:
Explanation:
QUESTION 77
Which stage classifies data rows from a single input into groups and computes totals?
A. Modify stage
B. Compare stage
C. Aggregator stage
D. Transformer stage
Answer: C
Explanation/Reference:
Explanation:
QUESTION 78
Which statement describes a SCD Type One update in the Slowly Changing Dimension stage?
A. Adds a new row to the fact table.
B. Adds a new row to a dimension table.
C. Overwrites an attribute in the fact table.
D. Overwrites an attribute in a dimension table.
Answer: D
Explanation/Reference:
Explanation:
QUESTION 79
Answer: C
Explanation/Reference:
Explanation:
QUESTION 80
You are processing groups of rows in a Transformer. The first row in each group contains "1" in the Flag column and "0" in the remaining rows of the
group. At the end of each group you want to sum and output the QTY column values. Which technique will enable you to retrieve the sum of the last
group?
A. Output the sum that you generated each time you process a row for which the LastRow() function returns True.
B. Output the sum that you generated up to the previous row each time you process a row with a "1" in the Flag column.
C. Output a running total for each group for each row. Follow the Transformer stage by an Aggregator stage. Take the MAX of the QTY column
for each group.
D. Output the sum that you generated up to the previous row each time you process a row with a "1" in the Flag column. Use the FirstRow()
function to determine when the group is done.
Answer: C
Explanation/Reference:
Explanation:
QUESTION 81
The derivation for a stage variable is: Upcase(input_column1) : ' ' : Upcase(input_column2). Suppose that input_column1 contains a NULL value.
Assume the legacy NULL processing option is turned off. Which behavior is expected?
Answer: B
Explanation/Reference:
Explanation:
QUESTION 82
A. By default, table definitions created in DataStage Designer are visible to other Information Server products.
B. Table definitions created in DataStage Designer are local to DataStage and cannot be shared with other Information Server products.
C. When a table definition is created in one DataStage project, it is automatically available in other DataStage projects, but not outside of
DataStage.
D. Table definitions created in DataStage Designer are not by default available to other Information Server products, but they can be shared with other
Information Server products.
Answer: D
Explanation/Reference:
Explanation:
QUESTION 83
What are two advantages of using Runtime Column Propagation (RCP)? (Choose two.)
Answer: BE
Explanation/Reference:
Explanation:
QUESTION 84
In a file defining a partial schema, which two properties define the record? (Choose two.)
A. intact
B. delim
C. position
D. length_of_record
E. record_delim_string
Answer: AE
Explanation/Reference:
Explanation:
QUESTION 85
What role must a user have to delete shared metadata assets from the repository?
A. DataStage Administrator
B. Business Glossary Administrator
C. Common Metadata Administrator
D. Information Analyzer Administrator
Answer: C
Explanation/Reference:
Explanation:
QUESTION 86
Your job is using the Sequential File stage as a source. Identify two stage property configurations that will allow you to accomplish parallel reads?
(Choose two.)
A. Read Method property set to 'File Pattern' and the environment variable $APT_IMPORT_PATTERN_USES_FILESET is set to 'FALSE'.
B. Read Method property set to 'File Pattern' and the 'Readers per Node' property value set to greater than 1.
C. Read Method property is set to 'Parallel' with the Key property set to a Input column that has the file names.
D. Read Method property set to 'Specific File(s)' with only one file specified and the 'Readers Per Node' property value set to greater than 1.
E. Read Method property set to 'Specific File(s)' with more than one file specified and each file specified has the same format/column properties.
Answer: DE
Explanation/Reference:
Explanation:
QUESTION 87
Identify the two statements that are true about the functionality of the XML Pack 3.0. (Choose two.)
Answer: CE
Explanation/Reference:
Explanation:
QUESTION 88
When using the Sequential File stage as a source, what two property options allow you to add extra columns about the file(s) you are reading onto the
output link? (Choose two.)
Answer: AC
Explanation/Reference:
Explanation:
QUESTION 89
The parallel framework supports standard and complex data types in the SQL type column tab property. Identify the two complex data types? (Choose
two.)
A. Vector
B. VarChar
C. Linked list
D. Subrecord
E. Binary trees
Answer: AD
Explanation/Reference:
Explanation:
QUESTION 90
How does the Complex Flat File stage (CFF) support the use of OCCURS clauses within COBOL files?
Answer: D
Explanation/Reference:
Explanation:
QUESTION 91
Identify two areas that DataStage can integrate with a Hadoop environment. (Choose two.)
A. Use the Pig Activity stage in a sequencer to invoke a Pig MapReduce program.
B. Use the Hive Activity stage in a sequences to invoke a Hive MapReduce program.
C. Use the Big Data File stage to access files on the Hadoop Distributed File System.
D. Use the Oozie Workflow Activity stage in a sequencer job to invoke Oozie work flows.
E. Use the MapReduce Transformer stage to convert Hadoop Distributed File System data.
Answer: CD
Explanation/Reference:
Explanation:
QUESTION 92
You are using the Complex Flat File stage as a source in your job. What are two types of data specifically supported by the Complex Flat File stage for
your job? (Choose two.)
A. XML files.
B. Mainframe data sets with VSAM files.
C. Mainframe data files with DB2 tables.
D. Data from files that contain Hadoop record types.
E. Data from flat files that contain multiple record types.
Answer: BE
Explanation/Reference:
Explanation:
QUESTION 93
The number of File Set data files created depends upon what two items? (Choose two.)
A. Amount of memory.
B. Schema definition of the file.
C. Operating system limitations.
D. Number of logical processing nodes.
E. Number of disks in the export or default disk pool connected to each processing node in the default node pool.
Answer: DE
Explanation/Reference:
Explanation:
QUESTION 94
What are the two Transfer Protocol Transfer Mode property options for the FTP Enterprise stage? (Choose two.)
A. FTP
B. EFTP
C. TFTP
D. SFTP
E. RFTP
Answer: AD
Explanation/Reference:
Explanation:
QUESTION 95
Identify two restructure stages that allow you to create or organize vectors in the output link results? (Choose two.)
A. Split Vector
B. Column Import
C. Merge records
D. Split Subrecord
E. Make Subrecord
Answer: DE
Explanation/Reference:
Explanation:
QUESTION 96
You are using the Change Capture stage in your job design to identify changes made to the input link (before image) to obtain the results in the output
link. This job will be using a multi-node configuration file when executed. What are two requirements on the input link data for the Change Capture
stage? (Choose two.)
A. sorted
B. merged links
C. parallel mode
D. key partitioned
E. sequential mode
Answer: AD
Explanation/Reference:
Explanation:
QUESTION 97
What are two statistics or attributes that can be added to the output of a Data Rule stage? (Choose two.)
A. Record ID
B. System timestamp
C. Number of violations
D. Job execution owner
E. Percentage of violations
Answer: AB
Explanation/Reference:
Explanation:
QUESTION 98
What are two properties dependent of both Column for Calculation and Summary Column for Recalculation? (Choose two.)
A. Summary
B. Missing Value
C. Sum of Squares
D. Mean Deviation
E. Coefficient of Variation
Answer: AB
Explanation/Reference:
Explanation:
QUESTION 99
A DataStage job uses an Inner Join to combine data from two source parallel datasets that were written to disk in sort order based on the join key
columns. Which two methods could be used to dramatically improve performance of this job? (Choose two.)
Answer: BE
Explanation/Reference:
Explanation:
QUESTION 100
Indicate two possible job design topologies that Information Services Director (ISD) parallel jobs can conform to. (Choose two.)
Answer: BC
Explanation/Reference:
Explanation:
QUESTION 101
How must the input data set be organized for input into the Join stage? (Choose two.)
A. Unsorted
B. Key partitioned
C. Hash partitioned
D. Entire partitioned
E. Sorted by Join key
Answer: BE
Explanation/Reference:
Explanation:
QUESTION 102
What are the valid join operations for the Join stage? (Choose two.)
A. Inner join
B. Where join
C. Top outer join
D. Right outer join
E. Bottom inner join
Answer: AD
Explanation/Reference:
Explanation:
QUESTION 103
QUESTION 104
You are editing the configuration file. Which statement is true about file system efficiency?
Answer: C
Explanation/Reference:
Explanation:
QUESTION 105
Answer: B
Explanation/Reference:
Explanation:
QUESTION 106
A. DataStage Administrator
B. The DSODBCConfig.cfg file
C. The DSAppWatcher.sh script
D. Information Server Web Console
Answer: B
Explanation/Reference:
Explanation:
QUESTION 107
What two project environment variables can be considered in your parallel jobs to support your optimization strategy of partitioning and sorting?
(Choose two.)
A. $APT_NO_PART_INSERTION
B. $APT_OPT_SORT_INSERTION
C. $APT_RESTRICT_SORT_USAGE
D. $APT_PARTITION_FLUSH_COUNT
E. $APT_TSORT_STRESS_BLOCKSIZE
Answer: AE
Explanation/Reference:
Explanation:
QUESTION 108
The effective use of naming conventions means that objects need to be spaced appropriately on the DataStage Designer canvas. For stages with
multiple links,expanding the icon border can significantly improve readability. This approach takes extra effort at first, so a pattern of work needs to be
identified and adopted to help development. Which feature of Designer can improve development speed?
A. Palette Layout
B. Repository Layout
C. Snap to Grid Feature
D. Show Performance Statistics Feature
Answer: C
Explanation/Reference:
Explanation:
QUESTION 109
You have created three parallel jobs (Job A, B and C) in which the output of one job is the input to the other job. You are required to create processing
that manages this data relationship of the jobs and provide job level restart-ability. What two tasks will accomplish these objectives? (Choose two.)
A. Enable runtime column propagation for the job sequence running the jobs.
B. Set the 'Add checkpoints so sequence is restartable' option in the Sequencer job.
C. Enable the 'Add checkpoints so container is restartable' option for each of the parallel shared containers.
D. Create two parallel shared containers that has the logic of Job B and Job C and include those shared containers in the correct flow order within
the design of Job A.
E. Create a Sequencer job that has triggered events configured allowing Job A to run first, then Job B to run when A completes successfully, and
then Job C to run when Job B completes successfully.
Answer: BE
Explanation/Reference:
Explanation:
QUESTION 110
Which statement is true about improving job performance when using Balanced Optimization?
A. Convert a job to use bulk staging tables for Big Data File stages.
B. Balance optimization attempts to balance the work between the source server, target sever, and the job.
C. If the job contains an Aggregator stage, data reduction stages will be pushed into a target data server by default.
D. To ensure that a particular stage can only be pushed into a source or target connector, you can set the Stage Affinity property to source or
target.
Answer: D
Explanation/Reference:
Explanation:
QUESTION 111
In your parallel job design you have selected a parallel shared container to be included. Which area of your job design is required to be configured to
use the parallel shared container?
A. List the jobs that will be allowed to use the container.
B. Configure your job properties General tab "Allow Container" option to be True.
C. Configure the job name reference the parallel shared container originates from.
D. Configure the number of input and/or output links to support the parallel shared container.
Answer: D
Explanation/Reference:
Explanation:
QUESTION 112
A job design reads from a complex flat file, performs some transformations on the data, and outputs the results to a ISD output stage. How can the
parameter values can be passed to this job at run-time?
Answer: C
Explanation/Reference:
Explanation:
QUESTION 113
What two binding types are supported by Information Services Director (ISD) for a parallel job that is designed to be used as a service? (Choose two.)
A. EJB
B. SQL
C. HDFS
D. SOAP
E. STREAMS
Answer: AD
Explanation/Reference:
Explanation:
QUESTION 114
Modular development techniques should be used to maximize re-use of DataStage jobs and components. Which statement represents Modular
Development techniques?
Answer: D
Explanation/Reference:
Explanation:
QUESTION 115
A. 2
B. 3
C. 4
D. 5
Answer: B
Explanation/Reference:
Explanation:
QUESTION 116
You would like to run a particular processing job within a job sequence for each weekday. What two methods could be used? (Choose two.)
Answer: CE
Explanation/Reference:
Explanation:
QUESTION 117
When you run a parallel job, any error messages and warnings are written to the job log and can be viewed from the Director client. What message
handler level is available?
A. stage level
B. record level
C. project level
D. parameter level
Answer: C
Explanation/Reference:
Explanation:
QUESTION 118
Which two commands start a DataStage job named procRows in a project named DSProject. It also passes the value 200 to the numRows parameter.
(Choose two.)
Answer: BE
Explanation/Reference:
Explanation:
QUESTION 119
What two repositories does the Source Code Control Integration for DataStage and QualityStage serve? (Choose two.)
A. WebSphere Repository
B. Source Code Repository
C. Universe Project Repository
D. Information Analyzer Repository
E. Information Server metadata Repository
Answer: BE
Explanation/Reference:
Explanation:
QUESTION 120
You have finished changes to many jobs and shared containers. You must export all of your changes and integrate them into a test project with other
objects. What is a way to select the objects you changed for the export?
Answer: D
Explanation/Reference:
Explanation:
QUESTION 121
You have made a copy of your job and made major changes to a job in your project. You now want to identify all the changes that have been made.
What task will allow you to identify these changes?
Answer: C
Explanation/Reference:
Explanation:
QUESTION 122
What two Information Server tools can be used to create deployment packages for DataStage jobs? (Choose two.)
A. istool
B. Information Server Console
C. Information Server Manager
D. InfoSphere Metadata Assset Manager
E. Information Server Web Console - Administration tab
Answer: AC
Explanation/Reference:
Explanation:
QUESTION 123
Two new columns have been deleted from a sequential file that a number of DataStage jobs write to. You have imported a table definition for the
changed file. How can you determine what jobs might be affected by this change?
A. Select the new table definition, right-click, then click on the Where used command.
B. Select the original table definition, right-click, then click on the Where used command.
C. Select the original table definition, right-click, then click on the Find dependencies command.
D. Select the new table definition, right-click, click on the Compare against command, then select original table definition.
Answer: B
Explanation/Reference:
Explanation:
QUESTION 124
A job using a one-node configuration file writes to a target Sequential File stage. The target Sequential File stage has been set to write to three different
sequential files. How many instances of the Sequential File stage will run in parallel?
A. One.
B. Two.
C. Three.
D. The job will abort, because a configuration file with at least three nodes is required.
Answer: C
Explanation/Reference:
Explanation:
QUESTION 125
The data going into the target Sequential File stage is sorted in each partition by the Date field. You want the data in the sequential file to be sorted by
Date. Which collection algorithm should you choose in the Sequential File stage?
A. Auto
B. Ordered
C. Sort Merge
D. Round Robin
Answer: C
Explanation/Reference:
Explanation:
QUESTION 126
A job design consists of an input Sequential Files stage, a Sort stage, followed by a Transformer stage and an output Sequential File stage. In the input
Sequential Files stage, the "Number of readers per node" property has been set to 3. The job is run on an SMP machine with a configuration file
defined with 3 nodes. The $APT_DISABLE_COMBINATION variable is set to True. How many player processes will this job generate?
A. 6
B. 8
C. 10
D. 12
Answer: B
Explanation/Reference:
Explanation:
QUESTION 127
A job using a three-node configuration file writes to a target Sequential File stage. The target Sequential File stage has been set to write to two different
sequential files. How many instances of the Sequential File stage will run?
A. One instance.
B. Two instances.
C. Three instances.
D. Four instances.
Answer: B
Explanation/Reference:
Explanation:
QUESTION 128
Each row contains the CustID, ProductID, ProductType and TotalAmount columns. You want to retain the record with the largest value in the
TotalAmount column for each CustID and ProductType using a RemoveDuplicates stage. How can this be accomplished?
Answer: A
Explanation/Reference:
Explanation:
QUESTION 129
What two computer system resources on the DataStage engine are monitored in the Operations Console? (Choose two.)
A. Disk
B. CPU
C. Memory
D. Scratch disk
E. XMETA database
Answer: BC
Explanation/Reference:
Explanation:
QUESTION 130
You want to better understand the tsort operators that are inserted into the job flow at runtime. Which environment variable can provide this
information?
A. $OSH_DUMP
B. $APT_DUMP_SCORE
C. $APT_STARTUP_STATUS
D. $APT_PM_PLAYER_MEMORY
Answer: B
Explanation/Reference:
Explanation:
QUESTION 131
Which job log message heading identifies the score in the job log?
Answer: D
Explanation/Reference:
Explanation:
QUESTION 132
You are experiencing performance issues for a given job. You are assigned the task of understanding what is happening at run time for that job. What
step should you take to understand the job performance issues?
Answer: C
Explanation/Reference:
Explanation:
QUESTION 133
Answer: AC
Explanation/Reference:
Explanation:
QUESTION 134
Which option is required to identify a particular job player processes?Which option is required to identify a particular job? player processes?