Data Stage Expert Quiz
Data Stage Expert Quiz contains set of 75 Datastage MCQ Questions With Answers which will help you to clear Expert level quiz.
1) What is APT_CONFIG in datastage?
- APT_CONFIG is just an environment variable used to identify the configuration (*.apt) file.
- APT_CONFIG is a system variable used to idetify the temporary memory.
- APT_CONFIG is just an environment variable used to idetify the Configuration of SMP/MMP server
- A&C are true
2) What is a conductor node?
- A node where execution starts.
- A node where data stage server is installed
- Both (a) and (b)
- None of the above
3) How to send notifications from Datastage as a text message (sms) to a Mobile
- create a batch script which will take care of sending messages and invoke it from Datastage using ExecDOS or ExecSh subroutines passing the required parameters
- DSSendMail Before-job or After-job subroutine is used to send sms messages
- configure an SMTP (email) server as a mobile phone gateway.a Notification Activity can be used to send message with a job log and any desired details
- All of the above
4) What is a command to analyze hashed file?
- STAT command
- FILE command
- Both A&B
- Only A
5) What are the command line functions that import and export the DS jobs?
- dsimport.exe- imports the DataStage components.
- dsexport.exe- exports the DataStage components
- dsimpt.exe- imports the DataStage components.
- dsexpt.exe- exports the DataStage components
- dsimprt.exe- imports the DataStage components.
- dsexprt.exe- exports the DataStage components
- None of the above
6) What is Meta Stage?
- Basically this is processing stage exists in Parallel job.
- Basically this is processing stage exists in Server job.
- Basically this is Meta data repository.
- None of above.
7) OCI stage can connect and execute?
- SQL Server
- Oracle procedure
- DB2 table
- Files
8) Functions of an Orchestrate operators are
- Orchestrate Operators read records from input data sets, perform actions on the input records, and write results to output data sets.
- An operator may perform an action as simple as copying records from an input data set to an output data set without modification.
- Alternatively, an operator may modify a record by adding, removing, or modifying fields during execution.
- All
9) Import Orchestrate Schema wizard allows to import meta data from
- Orchestrate schema file
- from a file set
- from a data set
- All of the Above
10) Which of the following is True?
- @INROWNUM counts incoming rows to a transformer in a datastage job
- ‘@OUTROWNUM counts outcoming rows to a transformer in a datastage job
- Used to generate sequences, primary keys, id’s, numbering rows and also for debugging and error tracing
- Gives the row numbers of the input link and output link correspondingly to a Modify stage
- Only A
11) How do you remove duplicates without using remove duplicate stage?
- Using sort stage duplicates can be removed by making options allow duplicates= false
- It’s not possible
- Using Sort stage will by default remove the duplicates
12) Merging of two files in Datastage is carried out by
- Merge Stage
- Funnel Stage
- Both A & B
- None of the Above
13) In case of processing of a job with surrogate key generator, on multiple nodes; generated key values can
- Be in a series
- Skip few values in between of the series
- Both A & B
- None of the above
14) Select the correct statement. A.Lookup stage reference link will allow duplicate records by default. B.Lookup stage reference link should not allow duplicate records by default.
- A and C are correct
- B only correct
- B and C are correct
- All are correct
15) Select the correct statement on lookup . A. Paging fault will occur when reference dataset size is greater than physical(RAM) memory. B. Will give the reject from primary or driving source.
- A and B are correct
- A only
- B and C are correct
- All
16) Select the correct statement of Write Range Map stage A.Can have One Input link and one Output link. B.Only one input link C.Only run in sequential mode
- A and D are correct
- B and C are correct
- Only B
- Only C
17) Select the correct statement A.LookUp allow duplicates in Primary and Will give warnings if duplicates available in secondary. B.Merge allow duplicates in secondary and Will give warn
- A and B are correct
- All
- A,B and C are correct
- None of the statements
18) DSJobWaveNo key word is called as
- System Variable
- DS Parameters
- DSRoutine
- DataStage Macro
19) Default percentage for Buffer Free Run for a Stage is
- 30
- 75
- 50
- 60
20) Which Stage reads the EBCDIC data file
- Complex Flat Set
- Complex File Format
- Complex Flat File
- External Source
21) COBOL File Definitions recognizes
- Occurs Depending On
- Occurs Redefines
- Redefines
- Signal
22) Which type of stages allow to specify an orchestrate as a Datastage “Stage”
- Custom
- Wrapped
- Buid
- Build and Customer
23) DataStage generates and compiles number of files to build an operator which the Build stage executes. Which one of the following is not generated by Datastage for the Build Stage
- Header File
- Object Files
- Source File
- Control Files
24) Orchestrate automatically does not perform default conversions for which of the following
date type
- int8 to int32
- sfloat to dfloat
- string to ustring
- None
25) The Advantages of Argument-List Processing
- It automatically produces a usage string.
- It provides automatic run-time validation of your options
- It creates a property list you can traverse to examine the input arguments and
- their values
- ALL
26) Which feature allows to process records whose size exceeds the 32 kilobyte limit imposed by Orchestrate
- Combinable Operators
- Unit-of-Work Batch Processing
- None
- Both
27) To encourage user to update the short description for a job, how can you make the short description visible and updateable on the main canvas.
- Click the show job short description option in the job properties
- Add an annotation stage to the job canvas and copy and paste in the short description
- Right click on the job canvas and choose show job short description in the submenu
- Add a description annotation field to the job canvas and select the short description property.
28) In a Lookup Operators, a Key column in the stream link is VARCHAR and has Unicode set in the extended attribute while the corresponding column in a reference link also Varchar without Unicode set.
What will allow correct comparison of the data?
- Convert both columns to CHAR, pad with spaces and remove Unicode from the extended attribute in Transformer operators prior to the lookup operators
- Convert the column in the reference link to the UTF-8 code page using the stringToUstring function in a transformer operator to the lookup operator and set the extended attribute of the column
- Remove the Unicode from the extended attribute of the column from the beginning of the job to the lookup operator and then set the extended attribute of the column in the output mapping section of the lookup operator
- Convert the column in the stream link to the default code page using the UstringToString Function in Transformer operator prior to the Lookup Operator
29) Which of the stage does not support reject link
- ODBC Stage
- Oracle Enterprise Stage
- Sequential File Stage
- None of the above
30) We can call a database stored procedure using _______ stage in job
- ODBC
- Dynamic RDBMS
- Stored procedure
- All above
31) Select odd stage
- SQL Server Enterprise
- Sequential File
- Stored Procedure
- Surrogate Key Generator
32) Select odd combination of stages
- Change capture & Row Generator
- Sequential File & Dataset
- Compare & Switch
- SQL Server Enterprise & Dynamic RDBMS
33) Select correct statement
- In a Project Job name can be duplicate
- In a Job Stage name can be duplicate
- In a Job Link name can’t be duplicate
- In a Project Routine name can be duplicate
34) A rejected link is used for:
- Capture rejected data from transformer stage.
- Capture rejected link from lookup stage.
- Capture rejected link from sequential file.
- All of the above
35) The __________ is a passive stage which provides a communication channel between Data Stage processes running simultaneously in the same job
- Link Collector Stage
- Inter-Process Stage
- Lookup Stage
- Transformer Stage.
36) What is the limit to the number of instance of a job that can be run simultaneously for a multi-instance job?
- 10
- 100
- 1000
- There is no limit defined by Data Stage.
37) Transformer in Server jobs will support
- 1 reference link and 1 stream link
- n stream links and 1 reference link
- n reference links and 1 stream link
- none
38) One of the below stage is a native parallel database stage
- Oracle OCI Load
- Informix Enterprise
- Dynamic RDBMS
- Teradata API
39) Which one of below difference between Merge stage and Lookup stage in Parallel jobs is true?
- Lookup supports single reference link and Merge supports multiple reference links
- Lookup supports multiple reject links and Merge supports only one reject link
- Lookup supports single stream link and Merge supports multiple stream links
- Lookup supports single output link and Merge supports multiple output links
40) Among the following, which stage can be used to remove duplicate records in parallel jobs?
- Sort stage
- Remove duplicate stage
- Hashed file stage
- A & B
41) How many max number of input links support by Link collector stage in server jobs?
- 64
- 128
- N
- Depends on OS
42) Which one is an invalid trigger expression in an activity in a Job Sequence?
- Equality(Conditional)
- Unconditional
- ReturnValue(Conditional)
- Custom(Conditional)
43) Which two system variables must be used in a parallel Transformer derivation to generate a unique sequence of integers across the input records?
- @PARTITIONNUM, @DATE
- @INROWNUM, @PARTITIONNUM
- @DATE, @NUMPARTITIONS
- A or C
44) A job has two input sources that need to be combined. Each input source exceeds
available physical memory. The files are in the same format and must be combined using
a key value. It is guaranteed that there will be at least one match.For the above scenar
- Funnel
- Merge
- Lookup
- Transformer
45) In a Transformer you add a new column to an output link named JobName that is to contain the name of the job that is running. What can be used to derive values for this column?
- DataStage function
- Link Variable
- System variable
- DataStage Macro
46) Which feature in DataStage will allow you determine all the jobs that are using a Custom Routine?
- Extended Job view
- Reporting Assistant
- Usage Analysis
- Impact Analysis
47) In which situation should a BASIC Transformer stage be used in a Parallel job?
- in a job containing complex routines migrated from DataStage Server Edition
- in a job requiring lookups to hashed files
- in a large-volume job flow
- in a job requiring complex, reusable logic
48) Which statement about job parameter usage is true?
- You can change job parameters while a job is running and the changes will
- immediately be applied mid-joB
- You can use environment variables to set parameter values linked to the Job Sequence
- You can change the parameter values in an initialization file linked to a Job Sequence.ini file
- Changes to the job parameters in the Designer do not require a recompile to be applied to the joB
49) Which one of the below would determine the number of records in a Dataset?
- Dataset Management utility
- DataStage Administrator
- Dataset stage properties
- All of the above
50) Which of these is not a field level data type property?
- Pad character
- Calculated value of the field
- Nullable
- None of the above
51) What is the below way is not used delete a persistent parallel Dataset?
- standard UNIX command rm
- orchadmin command rm
- delete the dataset using Dataset Management utility
- None of the above
52) Which is not a before/after job/stage sub routine
- DSWaitForFile
- DSSendMail
- UtilityRunJob
- ExecSHSilent
53) To de-duplicate data in source data in Data Stage server job which of the following stage is used?
- Hash File Stage
- Remove Duplicate Stage
- Lookup File Stage.
- None of the above
54) Which of the following value helps setting up the value of a job parameter as default value as defined in the Data Stage environment?
- $PROJDEF
- $ENV
- $VAR
- None of the above
55) There is way by which server job functionality can be incorporated in data stage parallel job. This can be done by:
- By converting parallel job into server job.
- By running parallel job under environment where only server job can run.
- Creating server container and using them in parallel job.
- We can not incorporate server job functionality in parallel job.
56) Which of the following is supported complex data type in the parallel job?
- Sub-records
- Tagged Sub records
- Vectors
- All of the above
57) The operator using the data set as input performs no repartitioning and takes as input the partitions output by the preceding stage. With this partitioning method, records stay on the same processing node; that is, they are not redistributed, this partiti
- Hash
- Round robin
- Same
- Auto
58) Which of the collecting technique in Data Stage will eagerly read any row from any input partition as it becomes available, but if it detects that, for example, the data needs sorting as it is collected, it will do that.
- Auto
- Round robin.
- Ordered
- Sorted Merge
59) Multi Instance Jobs are nothing but
- Jobs running in Parallel
- Jobs running in sequence
- Both
- None of the Above
60) You are reading customer data using a Sequential File stage and transforming it using the Transformer stage. The Transformer is used to cleanse the data by trimming spaces from character fields in the input. The cleansed data is to be written to a target
- Hash on the customer ID field
- Round Robin
- . Random
- Entire
61) An Aggregator stage using a Hash technique processes a very large number of rows during month end processing. The job occasionally aborts during these large runs with an obscure memory error. When the job is rerun, processing the data in smaller amounts c
- Set the Combinability option on the Stage Advanced tab to Combinable allowing the Aggregator to use the memory associated with other operators.
- Change the partitioning keys to produce more data partitions.
- Add a Sort stage prior to the Aggregator and change to a sort technique on the Stage Properties tab of the Aggregator stage
- Set the environment variable APT_AGG_MAXMEMORY to a larger value.
62) A customer is interested in selecting the right RDBMS environment to run DataStage Enterprise Edition to solve a multi-file and relational database data merge. The customer realizes the value of running in parallel and is interested in knowing which RDBMS
- DB2/UDB Enterprise
- Oracle Enterprise
- ODBC Enterprise
- Sybase Enterprise
63) A job has two input sources that need to be combined. Each input source
exceeds available physical memory. The files are in the same format and must be combined using a key
value. It is guaranteed that there will be at least one match
- Merge
- Funnel
- Lookup
- Transformer
64) Which two steps are required to change from a normal lookup to a sparse lookup in an ODBC
Enterprise stage?
- Change the lookup option in the stage properties to “Sparse”.
- Replace columns at the beginning of a SELECT statement with a wildcard asterisk (*).
- Establish a relationship between the key field column in the source stage with the database table field.
- Both A and C
65) Which three statements are true about File Sets?
- File sets are partitioned
- File sets are stored as header file and data files
- File sets are readable by external applications.
- All of the above
66) You need to update three dimension tables and one fact table in a star schema data warehouse
using the Slowly Changing Dimensions (SCD) stage. How many SCD stages will you need in
your jobs?
- One
- Two
- Three
- Four
67) Which three statements are true about stage variables?
- Stage variables retain their value between rows in the stage.
- Stage variables are executed before output link derivations.
- Stage variables can be used to control execution of constraints.
- All of the above
68) Which two statements are true about DataStage parallel routines?
- Parallel routines can be called from any stage type.
- Parallel routines are coded in C/C++.
- Parallel routines are coded outside of DataStage.
- Both B & C
69) Which two methods should be used to delete a persistent parallel Data Set?
- Delete the data set in the Data Set stage.
- Delete the Data Set in the Data Set Management utility.
- Delete the Data Set using the orchadmin command.
- Both B & C
70) Which statements are true about Transformer stage variables? (Choose two)
- A stage variable can evaluate to a null result.
- A stage variable can be used in derivations.
- A stage variable derivation can use other stage variables
- Both B & C
71) How to handle Date convertions in Datastage – Convert a mm/dd/yyyy format to yyyy-dd-mm
- Iconv(Filedname,”D/MDY[2,2,4]”)
- Oconv(Iconv(Filedname,”D/MDY[2,2,4]”),”D-YDM[4,2,2]”)
- Oconv(Filedname,”D/MDY[4,2,2]”)
- Any of the above
72) Which of the following is valid using a Trim function in DataStage?
- Trim(” a b c d “) results “a b c d”
- Trim(” a b c d “,” “,”B”) results “a b c d”
- Both A&B
- Only B
73) Horizontal Pivot type functionality
- Horizontal pivoting (columns into rows)
- Vertical pivoting (rows into columns)
- A & B
- None
74) A job reads from a dataset using a DataSet stage. This data goes to a Transformer stage and then is written to a sequential file using a Sequential File stage. The default configuration file has 3 nodes. The job creating the dataset and the current job bo
- 3
- 1
- 7
- 9