Anda di halaman 1dari 151

How to remove reverse duplicates in datastage ?

Abhisek

DataStage Interview Questions

Scenario:
Objective: The interface aims to consolidate the round-trip ticket cost of passengers.
Description: The source file is a .txt file, train_route_src.txt, consolidated with trains from a particular
source to destination. There are many repetitions in the source file (Reverse Duplication: eg. BLR --> BBS
and BBS --> BLR). Remove the reverse duplicates and maintain the target in...
D Hemakumar
o

Apr 24th, 2015

Source---->columngenarator--------->transformer---->target
generate the two columns col1 and col2 and generate the sequence numbers for both
columns in transformer write constraint col1>col2
take the source destination and distancefare columns in the output link
i hope this may work
can anyone pls solve in datastage.........
source column is..
A
A
B
B
B
C
C
D
........want output columns(2 cols) as.........
A1
A2
B1
B2
B3
C1
C2
D1
THANX IN ADVANCE.
D Hemakumar
o

Apr 24th, 2015

Take three stage variables in the transformer


StageVar=inputcol
StageVar1=Stagevar
StageVar2=if StageVar=StageVar1 then StageVar2+1 else 1
map the StageVar2 to the output link column
jallu
o

Apr 11th, 2015

We can use stage variable in transformer stage,


u will take 2 stage variable 1 is stg1 , 2 is stg2,
u have to write if stg2=col_n then stg1+1 else ---> stg1
col_n--->stg2 , so u can get required output

Single column with multiple rows should be converted into single


row with commas in datastage
o

sai3689

DataStage Interview Questions

Can any one help....


I have single column in Source like.....
a
b
c
d
want target as.....
a,b,c,d
thanks in advance
D Hema
o

Apr 23rd, 2015

Use the stag variable in the transformer


StageVar=StageVar:,: input column name
go to the constraints :-Lastrow()
map that stage variable to the output column
jallu
o

Apr 11th, 2015

By using pivot stage in datastage


Answer Question Select Best Answer

FEB

132014

01:50 PM

7131
Views

6
Ans

Datastage job scenario question


o

premox5

DataStage Interview Questions

My input has a unique column-id with the values 10,20,30.....how can i get first record in one o/p file,last
record in another o/p file and rest of the records in 3rd o/p file?
hari
o

Apr 23rd, 2015

Can you explain more about this explanation


chandrasekhar
o

Apr 22nd, 2015

Take three stage variables as stgA,stgB,stgC . stgA --columnname . stgB --If columnname =
then A else if columnname (inrouwnum >0 and currentdate=sysdate then A esle B stgC
--stgB --If columnname = ...
Answer Question Select Best Answer

MAR

172015

07:18 AM

766
Views

1
Ans

How to convert multiple rows into single row in datastage


o

sai3689

DataStage Interview Questions

Hi,
Can you please solve this in datastage..
my input is
name,city
johnson,newyork
johnson,mumbai
johnson,delhi
want output as
name,city1,city2,city3
johnson,newyork,mumbai,delhi
pls explain with steps
thanking in advance...
sudip
o

Apr 23rd, 2015

Seq stg------>Sort stg--------> Trnsfrm stg-------->Tgt 1. read data in seq file stage. 2. in sort
stage enable key column change as true. 3. in Trnsfrm define stage var, logic -> if key
change =1 thn...
Answer Question Select Best Answer

APR

162008

09:58 AM

3695
Views

11
Ans

DataStage Technical Questions


o

mohanreddy

DataStage Interview Questions

1. Is the Hash file is active or passive? If we take as source ?2. can u take seq file as look up?3. In hash
file dynamic 30, there r two types:1)genaric2)specific what is the meaning?4. how to connect MERGE
STAGE while source as two tables?5. what is the purpose of MERGE?6. how can DS job scheduled in
Unix?7. how do u know how many rows rejected?8. use of universe stage?9. what is SEQ file buffer?10.
diff...
Pravin Patil
o

Apr 19th, 2015

For Question 15
how do u decide when go for join or lookup?
If Large amount of data is comming from I/P source then use Join.
If Small amount of data is comming from I/P source then use Lookup.
Pravin Patil
o

Apr 19th, 2015

For Question 47
Environment variable is predefined variable.
An user defined environment variable is a place holder to store a value that can used in the
entire project.
Answer Question Select Best Answer

NOV

082011

10:01 AM

1938
Views

3
Ans

SRC records
o

swapna

DataStage Interview Questions

SRC has 1 record I want 10records in target how is it possible ? plz explain me ?
Murali
o

Apr 19th, 2015

Use the copy stage, between source and target.


Take 10 nodes in the configuration file.
Select "Partitioning methodology" as "Entire" in copy stage.
You will 10 Output records.
ramesh
o

Mar 25th, 2015

Goto transformer stage ,in that stage select system variables ,select iteration and specify
the condition in loopig is @iteration=10
Answer Question Select Best Answer

APR

062015

08:11 AM

396
Views

2
Ans

Single column with multiple rows should be converted into single


row with commas
o

sai3689

DataStage Interview Questions

Can any one help....


Source is.....
a
b
c
d
want target as.....
a,b,c,d
thanks in advance
amulya kumar panda
o

Apr 17th, 2015

add one extra column that value 1 or any thing but value same Use a vertical Pivot option
available in Pivot Enterprise stage(Stage tab-->Properties-->Pivot Type=Vertical) Again in
stage-->Pivot ...
vasuu
o

Apr 13th, 2015

Uses pivot stage or u can tx by stage variables


Answer Question Select Best Answer

OCT

192013

11:18 PM

10821
Views

6
Ans

What is the architecture of your Datastage project


o

Sam Geek

DataStage Interview Questions

I came across this question many times in interview, In specific what can i answer..... Please help..

rstrainings
o

Apr 15th, 2015

we can say like there is a staging area to store the data in the form of tables, and then it
transforms to ETL stage (Here we will do all the conversions,remove
duplicates,joining,merging etc) then di...
shiv
o

Jun 11th, 2014

The above answer, is to architecture of Datastage, its not the architecture of project. Project
architecture would be like: ***************************** You have: 1 Source--------> 1 Staging
Area----...
Answer Question Select Best Answer

MAR

172015

08:20 AM

745
Views

1
Ans

Convert multiple rows into columns in datastage


o

sai3689

DataStage Interview Questions

hi
i have sourse like
name,city
johny,newyork
johny,mumbai
johny,del
want output like
name,city1,city2,city3

johny,nework,mumbai,delhi
thanks in advance..
Ruchir
o

Apr 7th, 2015

Use a vertical Pivot option available in Pivot Enterprise stage(Stage tab-->Properties->Pivot Type=Vertical)
Again in stage-->Pivot properties tab-->group by on Name column and Pivot using City
column.
Then select array size=3.
Answer Question Select Best Answer

JUN

302006

01:03 AM

5690
Views

5
Ans

What is the difference between Symetrically parallel


processing,Massively parallel processing?
o

balu

DataStage Interview Questions

Himanshu Maheshwari
o

Apr 2nd, 2015

The environment in which you run your parallel jobs is defined by your system s
architecture and hardware resources. All parallel processing environments are categorized
as one of: SMP (...
Saurabh Sinha
o

Aug 1st, 2014

In SMP every processor share a single copy of the operating system (OS)
In MPP each processor use its own operating system (OS) and memory.
Answer Question Select Best Answer

OCT

202011

02:29 PM

11586
Views

8
Ans

Datastage Real time scenario


o

amulas

DataStage Interview Questions

source table
name
A
A
B
B
B
C
C
D
In source table data like this
but I want traget table like this
name count
A1
A2
B1
B2
B3

C1
C2
D1

pls any one one solve this........


satish
o

Apr 2nd, 2015

Use sort stage in sort cluster key change coloum true.


then transformer stage create stage variable
If DSLink5.keyChange=1 Then DSLink5.keyChange Else StageVar+1
then map the sagevar to CountRow column
Purna Yalamanchili

Mar 7th, 2015

Use if condition in transformer


name=stagevar
if name=stagevar then stagevar1+1 else 1=stagevar1
stagevar1=count
Answer Question Select Best Answer

FEB

262015

11:23 PM

1098
Views

7
Ans

Scenario in datastage
o

sravanthi

DataStage Interview Questions

input is
cola

1
2
3
this should be populated at the output as
cola
1
22
333
satish
o

Apr 2nd, 2015

In Transformer use stage variable Str(%Input_column%,%Input_column%)


sai3689

Mar 17th, 2015

@iteration
Answer Question Select Best Answer

MAR

192015

01:55 AM

791
Views

2
Ans

Convert multiple rows into single row


o

sai3689

DataStage Interview Questions

hi...
i have sourse like
name,city

johny,newyork
johny,mumbai
johny,del
want output like.....
name,city1,city2,city3
johny,nework,mumbai,delhi
----------------------------------------thanks in advance...
sai3689

Apr 1st, 2015

Thank u abishek
Abhishek Surkar
o

Mar 27th, 2015

Use Pivot Stage-->open Pivot stage select Properties tab select Pivot Type = Vertical
-->select Pivot Properties tab click Check Group by for Name --> click Check Pivot for city
-->Make Array size =4 -->
Map Output columns to output link and click Ok -->
Compile --> Run
Answer Question Select Best Answer

MAR

242015

09:32 AM

286
Views

0
Ans

I would like to have output as like this 101 90 next line 102 65
next line 102 55 etc...
o

Anjaneya Gupta

DataStage Interview Questions

11. Find The Output With out PIVOTE stage


Source File: Target
SNO, MATHS, PHYSICS, CHEMISTRY SNO, MARKS
101, 70, 80, 90 101, 70
102, 65, 55, 75 101, 80
101, 90
102, 65
102, 55
102, 75
Answer Question

MAR

222015

08:59 AM

429
Views

0
Ans

How do you implement SCD type 2 in datastage ?


o

teja

DataStage Interview Questions

Answer Question

NOV

042014

09:51 PM

2873
Views

4
Ans

Datastage scenario - Flat file


o

sravanthi

DataStage Interview Questions

Source is a flat file and has 200 records . These have to split across 4 outputs equally . 50 records in
each .
The total number of records in the source may vary everyday ,according to the count records are to split
equally at 4 outputs.
Could someone post an answer for this question.
Thanks
Vinay
o

Mar 19th, 2015

We can use the split command in the Filter section of the Sequential
file stage:
Split -l 50 File.txt Segment
O/P: will give 4 files namely Segmentaa,Segmentab,Segmentac and Segmented each of 50
records.
Gandhi
o

Jan 13th, 2015

Hi all,
Keep four files as output.Simply give modulus(col_name,4)=0 constraint for first file
,modulus(col_name,4)=1 second file and modulus(col_name,4)=2 for third file and finally
modulus(col_name,4)=3
Answer Question Select Best Answer

MAR

04:31 AM

132015

542
Views

1
Ans

How is Cartesian join performed using a join stage?


o

sravanthi

DataStage Interview Questions

I have
source
10
20
20
reference 10
20
20
can someone tell me how a cartesian join performed using a join stage in datastage.
Thanks
Sravanthi
satjay
o

Mar 17th, 2015

Create a dummy key column in both the links (eg with value 1) and join with this key
column. This is the approach to join two tables without keys.
Answer Question Select Best Answer

OCT

022014

08:01 AM

2549
Views

4
Ans

Dataset stage and sequential file stage


o

taruna.arya

DataStage Interview Questions

what is the difference between dataset stage and sequential file stage? even one more to be added here,
What is the difference between dataset stage and fileset stage?
kritika
o

Feb 23rd, 2015

Dataset: It stores data in the ASCII format so it takes less time to read data for datastage IT
can accommodate large amount of data Sequential file stage: It contains data in the
readable format,...
ravikumar a v
o

Jan 18th, 2015

Dataset stage is used for reference in put link in other jobs design and where as sequential
file stage is not possible.
Answer Question Select Best Answer

DEC

192013

08:51 AM

3211
Views

2
Ans

What are the uses of using parameters in datastage?


o

premox5

DataStage Interview Questions

Ishu
o

Feb 20th, 2015

Use of parameters: 1. avoid hardcoding. 2. Incase a value for some parameter is changed
in future, we can change it at a single point instead of making changes everywhere in every
job wherever it is ...
Lubna Khan
o

Jan 27th, 2014

By Using Parameters we can avoid hardcoding and assign values at runtime

Split Data From Flat File

sekintrance

DataStage Interview Questions

How do you split the data from a flat file in to columns?


Maximus
o

Feb 19th, 2015

You can use split command in Unix.


Answer Question Select Best Answer

OCT

162014

07:59 AM

2236
Views

2
Ans

Sequential File Stage

Alekhya ch

DataStage Interview Questions

Explain some scenarios where a sequential file stage runs in parallel?


Chandra
o

Jan 22nd, 2015

If you have multiple files ,need to change Read method to FILE PATTERN ,Then its run
parallel mode
Balu
o

Nov 25th, 2014

Changing the "no.of readers per node" to more than one . you can read a sequential file in
parallel mode
Answer Question Select Best Answer

NOV

042014

09:48 PM

2190
Views

3
Ans

Datastage scenario - two columns in the source

sravanthi

DataStage Interview Questions

I have two columns in the source , COl A and Col B .


Input is like
Cola Colb
100 ABCDEF
I should achieve the output as
Cola Colb
100 A
100 B
100 C
100 D
100 E
100 F
Note : Colb is not of fixed length .
Yash
o

Jan 7th, 2015

Its pretty much easy using loop.


With below loop conditions in transformer.You can easily achieve it
Loop Condition: @ITERATION
praveen
o

Nov 25th, 2014

Use pivot stage


Answer Question Select Best Answer

OCT

132011

11:17 AM

6380
Views

6
Ans

Job sequencing

amulas

DataStage Interview Questions

I have 3 jobs A,B & C, which are dependent each other, I want to run A & C jobs daily and B job run only
on Sunday. How can I do it?
Sreelesh
o

Jan 6th, 2015

Create a single sequence which will run on all days. Add a stage to call a routine which will
check whether the date is a sunday. If sunday call the job B. Else the sequence will call only
A &C.
Kartik Dharia
o

Jan 5th, 2015

You can create a new job to test whether the current day is Sunday or not. If it is true then
create a file else create a 0KB file. Then you can create only 1 sequence with jobs A and C
one after the...
Answer Question Select Best Answer

NOV

202014

09:22 AM

2288
Views

2
Ans

How to convert varchar to date ?

saikumar

DataStage Interview Questions

i have one source file which contains the below data


input column: DOB varchar(8)
20000303
20000409
1999043
1999047
validate the Date if valid date else pass the default-1999-12-31,convert varchar to date
Sam
o

Dec 17th, 2014

Hi, Check for valid date as below: 1.> Len of the string =8 then Date < sysdate else default
date 2.> We use a) "Iconv" function - Internal Convertion. b) "Oconv" function - Exter...
sonali
o

Dec 6th, 2014

In transformer stage use stringtodate(datecol,format)


Answer Question Select Best Answer

132010

JUL

12:32 AM

7192
Views

4
Ans

Data Granularity

manju_thanneeru

DataStage Interview Questions

What is data granularity? Explain.


Nutan Raut
o

Dec 15th, 2014

High level of detail-low level of granularity.& low level of detail-high level of granularity.

pari
o

Apr 26th, 2014

Explain data granularity and how it is applicable to data ware house?


Answer Question Select Best Answer

DEC

032014

03:52 AM

987
Views

1
Ans

Copy Stage in Processing

vidyasagarvuna

DataStage Interview Questions

Having single input in the source wants the repetition of the input in the column of the target using only
copy stage.
@sheesh
o

Dec 4th, 2014

Use entire partition and run on multiple nodes


Answer Question Select Best Answer

OCT

022014

07:54 AM

2733
Views

5
Ans

Having single input source want three different outputs.

taruna.arya

DataStage Interview Questions

I have a source file having data like:


10

10
10
20
20
20
30
30
40
40
50
60
70
i want three output from the above input file, these output would be:
1) having only unique records no duplicates should be there. Like:
10
20
30
40
50
60
70
2) having only duplicate records,...
Mihir
o

Nov 27th, 2014

Use Sort Stage-->Define the key field, In Property, Key Change column is TRUE.
Then use a Transformer, In constraint , KeyChange=1 for Unique record O/P and
KeyChange=0 for Duplicate O/P.
Dheerendra
o

Nov 11th, 2014

Use aggregator...take count of rows.....then add filter...specify conditions as row count = 1


and >1
Answer Question Select Best Answer

NOV

142014

02:41 AM

1388
Views

1
Ans

What are the Datastage native operators

Gopan.P

DataStage Interview Questions

What are the Datastage native operators??


C
C++
BASIC
OSH
Could you brief this please
srinivas
o

Nov 26th, 2014

OSH (orchestrate ) is the answer , when we develop any parallel job it converts into OSH
script in the back ground , of course C++ because when we use transformer stage the code
generated in the background is C++.
Answer Question Select Best Answer

152011

JUL

02:02 AM

7036
Views

7
Ans

What is the use of node in data stage ?

nareshketepalli

DataStage Interview Questions

If we increase the nodes wt will happens?


barak
o

Nov 21st, 2014

Increasing node will not always increase speed it might create lock for example in mass
update .
Indudhara V
o

Oct 8th, 2014

Node is just a process. Its a logical thing used to increase the efficiency of the jobs by
running them in parallel just like multi processing in operating system. Each node processes
may run on the same processor or different processors.
Answer Question Select Best Answer

OCT

022014

08:05 AM

2069
Views

2
Ans

What is range look up and how it is used in datastage.

taruna.arya

DataStage Interview Questions

Say I have 5 rows in a source table and for each rows 10 rows matching in a lookup table and my range
is for lookup is 9 to 99. what will be the row count in output table?
barak
o

Nov 21st, 2014

That depends on , if you set up "more than one row match" you will get you "Cartesian
Product"
ramireddy
o

Oct 17th, 2014

You will get same no of records in the source even there are duplicates in look up data
Answer Question Select Best Answer

OCT

162014

08:02 AM

1750
Views

1
Ans

Design a job

Alekhya ch

DataStage Interview Questions

I have table(Emp) with the columns Eid,Ename,Sal,month(sal),year(sal) and DOB (say 15th-Jan-1981).
Design a job such that the output displays Ename,year(sal),tot(sal) and current age i.e. For Ex: 18 yrs
barak
o

Nov 21st, 2014

Using Transformer Days Between and divided by 365 - we can check for year with 366 if
that's a big deal .
Answer Question Select Best Answer

OCT

022014

07:58 AM

1673
Views

3
Ans

Source file and 3 output tables records problem

taruna.arya

DataStage Interview Questions

Suppose i am having a source file and 3 output tables and I want first row to be written to first table
second row to second table, third row to third table likewise how can we achieve this using datastage
without using partitioning?
Dheerendra
o

Nov 14th, 2014

Use SEQ---> Transformer stage ----> 3 SEQ Files In transformer stage add constraints as
Mod(@INROWNUM,3) =1, Mod(@INROWNUM,3) =2, Mod(@INROWNUM,3) =0
Pratheeksha
o

Nov 12th, 2014

For this, one option could be :


Get the row number column from the source seq file, in the transformer use stage variable
which is defined as below:
if mod(i/p col,3)=0 then
if mod(i/p col,2)=0 then
"2"
else
"3"
else "1"
Answer Question Select Best Answer

OCT

022014

08:10 AM

2098
Views

1
Ans

How to identify updated records in datastage.

taruna.arya

DataStage Interview Questions

How can we identify updated records in datastage? Only updated records without having any row-id and
date column available.
Vikram Singh
o

Nov 11th, 2014

I believe you can use the change capture stage which will show you the data before any
update change made and after the update change was made. Should work, if there are no
other constraints.
Answer Question Select Best Answer

JAN

032013

12:26 AM

14325
Views

9
Ans

How to get top five rows in DataStage?

naveen.chinthala

DataStage Interview Questions

How to get top five rows in DataStage? I tried to use @INROWNUM,@OUTROWNUM system variables
in transformer..but they are not giving unique sequential numbers for every row...please help!
Thanks in advance!!

Vikram Singh
o

Nov 11th, 2014

You can use Head Stage, that would be the most convenient way of getting top N rows from
the dataset.
Poorna
o

Aug 7th, 2014

You can restrict data it at source stage level its self @ using filter option.
Apply in filter : head -5
Answer Question Select Best Answer

OCT

022014

08:06 AM

1578
Views

2
Ans

What is slowly changing dimension(SCD)

taruna.arya

DataStage Interview Questions

what is slowly changing dimension(SCD)? How do we rectify it in datastage.


sabz R

Nov 7th, 2014

We can use surrogate key generator to rectify this problem in datastage..


sabz
o

Nov 5th, 2014

SCD is a common problem particular to database.This applies to cases where the attributes
for a record varies over time.This can be solved by 3 types.type1,type2,type3
Answer Question Select Best Answer

OCT

022014

08:03 AM

1680
Views

2
Ans

How to read a comma delimiter file in sequential file stage?

taruna.arya

DataStage Interview Questions

How we read comma delimiter file in sequential file stage and how we can remove the header and footer
from a comma delimiter file?
deepakatsit

Oct 29th, 2014

For Removing Header file you can use SED Command of UNIX to remove Header.

deepakatsit

Oct 23rd, 2014

For Sequential file there is an option "First Row Column Name" set to true so Header is
removed and File End Type You need to select through which Footer is Removed.
Now For Comma Delimiter Use : :"Field Defaults -> Delimiter = Comma"
Answer Question Select Best Answer

112014

JUL

05:03 PM

3076
Views

7
Ans

Datastage job

hema123

DataStage Interview Questions

i have a sequence of job in DataStage which is taking more than 4 hrs which is supposed to complete in
less than 1 hr
so what could be the possibilities to take much longer time than expected?
RamyaSujith
o

Oct 6th, 2014

Check if any stage is reading/processing data sequentially which could have been done in
parallel.

Raveena Mittal
o

Sep 11th, 2014

When sometimes ,some DML statements are executed in the db and they are not
committed,then also it may keep the job running for long hours
Answer Question Select Best Answer

182011

MAY

11:23 AM

15961
Views

19
Ans

Datastage Job Scenario question

Boopathy Srinivasan

DataStage Interview Questions

input file A contains12345678910input file B contains6789101112131415Output file X


contains12345Output file y contains678910Output file z contains1112131415How can we do in this in a
single ds job in px ?....could you please give me the logic to implement ???
Read Best Answer

Editorial / Best Answer


Answered by: vinod chowdary
o

Jul 27th, 2011

Hello guy's, I would like to solve this by using the Change capture stage. First, i am going to
use source as A and refrerence as B both of them are connected to Change capture stage.
From, change capture stage it connected to filter stage and then targets X,Y and Z. In the
filter stage: keychange column=2 it goes to X [1,2,3,4,5] Keychange column=0 it goes to Y
[6,7,8,9,10] Keychange column=1 it goes to Z [11,12,13,14,15] Revert me PLz
Jithin
o

Sep 11th, 2014

Do a full outer join between two files and from transformer draw three output links
1st link-->wherever left side is null
2nd link->wherever right side is null
3rd link->wherever match is there
ghost
o

May 17th, 2014

create one px job. src file= seq1 (1,2,3,4,5,6,7,8,9,10) 1st lkp = seq2
(6,7,8,9,10,11,12,13,14,15) o/p - matching recs - o/p 1 (6,7,8,9,10) not-matching records o/p 2 (1,2,3,4,5) 2nd lkp: s...
Improve Answer

FEB

082006

07:05 AM

12732
Views

10
Ans

What is the difference between datastage and informatica

anunath

DataStage Interview Questions

pundlik
o

Sep 5th, 2014

Compared to Informatica the Data Stage is more User friendly


sunrays_arun

Oct 26th, 2007

SAS DI studio is best when compared to Informatica and Datastage as it generates SAS
code at the back end .SAS is highly flexible compared to other BI solution.

Getting files In datastage


o

Kiranchandra

DataStage Interview Questions

How to get a files from different servers to one server in datastage by using unix command?
Murali
o
Code

Aug 19th, 2014

1. scp test.ksh dsadm@10.87.130.111:/home/dsadm/sys/


Answer Question Select Best Answer

2 AUG

182014

06:14 AM

2495
Views

1
Ans

Display files date wise like aug 18th,19th,29th data files by using
Unix cmd?
o

Kiranchandra

DataStage Interview Questions

Ashok
o

Aug 19th, 2014

You can display files date wise by doing normal ls -latr cmd
Answer Question Select Best Answer

2 AUG

052014

09:29 PM

2643
Views

1
Ans

Flatfiles(.txt,.csv,xml) to laod sequential files?


o

Kiranchandra

DataStage Interview Questions

How can we load three different flat files(1 file .txt,2 file .csv,3 file xml) to sequential file at a time?
Devesh Ojha
o

Aug 11th, 2014

If Metadata is the same then we can load by doing the Union operation and if metadata is
diffrent then first sync the metadata and then load them .
Answer Question Select Best Answer

3 SEP

122013

05:21 AM

4517
Views

3
Ans

How to get the last day of the current month?


o

naveen.chinthala

DataStage Interview Questions

I have explored all the available functions in the transformer stage, but could not find the exact function to
get the last day of the current month. Can you please show me which function is available for this logic.
Arunjith B Indivar

Jul 27th, 2014

DaysInMonth(CurrentDate())
venueksh

Apr 8th, 2014

Oracle

has LAST_DAY function:

SELECT LAST_DAY(to_date(07/04/2014,MM/DD/YYYY)) from dual;


SELECT LAST_DAY(SYSDATE) from dual;
Answer Question Select Best Answer

242005

4 MAY

05:54 PM

10322
Views

3
Ans

What are Routines and where/how are they written and have you
written any routines before?
o
DataStage Interview Questions

Routines are stored in the Routines branch of the DataStage Repository, where you can create, view or
edit. The following are different types of routines: 1) Transform functions 2) Before-after job
subroutines 3) Job Control routines
Chalapathirao Maddali
o
Jul 11th, 2014

Datasatge has 2 types of routines ,Below are the 2 types. 1.Before/After Subroutine.
2.Transformer routines/Functions. Before/After Subroutines : These are built-in
routines.which can be called in...
bvrp

Nov 1st, 2005

RoutinesRoutines are stored in the Routines branch of the DataStage Repository,where you
can create, view, or edit them using the Routine dialog box. Thefollowing program
components are classified as ...
Answer Question Select Best Answer

5 JUN

202005

05:59 AM

4130
Views

6
Ans

Whats difference betweeen operational data stage (ODS) & data


warehouse?
o

DataStage Interview Questions

Ramyapriya Sudhakar
o
Jul 9th, 2014

operational data store : It is unlike real EDW ,data is refreshed near real time and used for
routine business activity. It is used as an interim logical area for data warehouse. This is the
pla...
Dharmendra
o

Sep 22nd, 2006

An operational data store (or "ODS") is a database designed to integrate data from multiple
sources to facilitate operations, analysis and reporting. Because the data originates from
multiple sources,...
Answer Question Select Best Answer

072014

6 JUL

02:07 AM

2933
Views

0
Ans

Write the real time scenario of different stages in datastage?


o

manojkitty

DataStage Interview Questions

Answer Question

7 DEC

282012

12:49 AM

2078
Views

1
Ans

Purpose of using user defined environment variables and


parameter sets
o

google_yahoo

DataStage Interview Questions

What is the purpose of using user defined environment variables and parameter sets. I m little bit
confused. Could any one explain me in detail.?
Charmi
o

Jul 1st, 2014

Hi,
Parameter Set is used when you want a set of user defined variables to be used many
times in a project.
For example, Variables like Server Name, User Id, Password can be added in a parameter
set and that can be used across the jobs , instead of including the three variables
everytime.
Answer Question Select Best Answer

8 APR

112012

12:13 AM

2762
Views

1
Ans

What are Hierarchies? Examples?

upendarkm

DataStage Interview Questions

Rajesh B
o

Jun 8th, 2014

Hi The hierarchy is nothing but parent and child relationship....


lets say the country is parent -->state is child to it---> city is child to it--->house no is child to
it
here if anyone needs hierarchy you can say from country-->state-->city-->house is the
hierarchy relationship
Answer Question Select Best Answer

9 OCT

272011

05:15 AM

7775
Views

4
Ans

How to extract job parameters from a file?


o

ramamulas

DataStage Interview Questions

karthick
o

May 30th, 2014

Parameter file will have comma delimiter.


use cat file1.txt |cut -d, -f1| tr -d "
" command to extract the first field ...
use execute command activity for extracting all the parameter. then finally pass the value to
actual job.
Mallikarjuna_G

Aug 18th, 2013

Write a server job routine that takes input as the file and reads the parameters from it. If the
file contains more than one parameter each in a separate line, the your routine should
concatenate them...
Answer Question Select Best Answer

10 OCT

242013

09:58 PM

36097
Views

1
Ans

Scenario based question


o

Sam Geek

DataStage Interview Questions

How to find if the next value in a column is incrementing or not


for ex
100
200
300
400
If the curval greater than previous val then print greater if lesser print lesser
For ex
100
200
150
400, Here 150
karthick
o

May 30th, 2014

Previous value in one stag_v1 and presend value in stg_v2 compare the two , if greater
then stg_v1=stg_v2 and move to next value. else loop it.
Answer Question Select Best Answer

11 NOV

032007

11:52 AM

4245
Views

3
Ans

DataStage - delete header and footer on the source sequential


o

srinivas

DataStage Interview Questions

How do you you delete header and footer on the source sequential file and how do you create header and
footer on target sequential file using datastage?
Kalai
o

May 7th, 2014

"Output --> Properties --> Option --> Filter --> add sed command here" to delete header and
footer records
leelasankar.pr

Jul 22nd, 2008

By using UNIX sed command we can delete header and footeri.e; for header
p'and footer
sed -n '$|p'
Answer Question Select Best Answer

12 FEB

132014

02:26 PM

sed -n '1|

10476
Views

5
Ans

Datastage scenario Question


o

NaveenKrish

DataStage Interview Questions

A Sequences is calling activity 1, activity 2 and activity 3.while running, activity 1 and 2 got finished but 3
got aborted. How can I design a sequence such that the sequence has to run from activity 2 when I
restart the sequences?
Mallikarjuna_G

May 7th, 2014

To make the job re-run from activity 3, we need to introduce restartability in the sequence
job. For this below points have to be taken care of in Job Sequence Adding Checkpoints:
Checkpoints have t...
Ritwik
o

Apr 21st, 2014

You have to check the " Do not checkpoint run " checkbox for activity 2. If you set the
checkbox for a job that job will be run if any of the job later in the sequence fails and the
sequence is restarted.
Answer Question Select Best Answer

13 JAN

212014

03:59 PM

5154
Views

2
Ans

How to seperate two diff datatypes and load it into two files?
o

premox5

DataStage Interview Questions

i have a source file1 consist of two datatypes


file1:
no(integer)
1
2
3
&
dept(char)
cs
it
ie
and i want to seperate these two datatypes and load it into target files
file2 & file3.
how can i do this in datastage and by using which stage?
dileep Janga

Jan 31st, 2014

I think this Question is to confuse the Job Aspirant by using Datatypes and all... Its very
simple... File1-->2 Columns. 1.NO(Integer) 2.DEPT(Char). Target1: NO(Integer), Target2:
DEPT(Char). Take ...
Lubna Khan
o

Jan 27th, 2014

In Transformer stage there is one function IsInteger and IsChar , We can identify If IsInteger
(column name) then file1 else file2
Answer Question Select Best Answer

14 JAN

202006

09:16 AM

14939
Views

12
Ans

What is the exact difference betwwen Join,Merge and Lookup


Stage??
o

Phantom

DataStage Interview Questions

vij
o

Jan 6th, 2014

Default partition technique is Auto in all please check once


mallika_chaithu

Jun 8th, 2011

Hope the below one helps you.Join Stage: 1.) It has n input links(one being primary and
remaining being secondary links), one output link and there is no reject link2.) It has 4 join
operations: inner...
Answer Question Select Best Answer

15 DEC

192010

10:53 PM

11435
Views

7
Ans

Remove Leading Zero

goodfriendsri

DataStage Interview Questions

How to remove leading zeros in data and transform data to target?


Sunitha
o

Dec 2nd, 2013

The main differences between 8.1 and 8.5 are 8.5 has Input looping ,Output looping. 8.5
has saving ,editing and compling is 40% faster. 8.5 has functions like LastRow,LastInGroup
and Iteration system...
Rupesh Agrawal
o
Nov 8th, 2013

Convert data to integer and convert back to string.


Answer Question Select Best Answer

16 JUN

052006

01:15 PM

4420
Views

2
Ans

How we can create rank using datastge like in informatica.


o

kishore

DataStage Interview Questions

Itishree
o

Dec 2nd, 2013

You can use either of the below options: Seq File->Sort->Remove Duplicate ->O/P in
Remove duplicate stage, choose the key column and duplicate to retain last properties Seq
File->Remove Duplicate ->O/...
Murat Nur
o

Dec 28th, 2006

Hi,if ranking means that belowprop_id rank1

11

21

32&nbs...

Answer Question Select Best Answer

17 OCT

132011

11:20 AM

5314
Views

4
Ans

Dataset utility question


o

amulas

DataStage Interview Questions

How can we retrieve the particular rows in dataset by using orchadmin command?
sunitha
o

Nov 29th, 2013

Orchadmin dump -part 0 -n 17 -field name input.ds


Sushils13

Aug 20th, 2013

$orchadmin dump [options] Sample.ds


options : -p period(N) : Lists every Nth record from each partition starting from first record.
Answer Question Select Best Answer

18 SEP

022013

11:49 PM

6089
Views

7
Ans

How to Seperate Repetating & non-Repetating data.


o

Sushils13

DataStage Interview Questions

I have data in Table as


ID
1
2
3
1
4
3
3
5
6
6
7
Now I want to have Output as repeatating records in one table & non-repeatating in another table
Table 1 :
Id
2
4
5
7
Table 2:
ID
1
1
3
3

3
6
6
<br...< p=""></br...<>
bhargav
o

Nov 28th, 2013

First take copy stage after take the 2 stages from copy at a time lookup and aggregater in
aggregater take count -> id after filter1 count =1 out put to lookup as a reference link and
stream link is copy stage again take filter2-> count=1 target1 and count1 target 2
Rohit K A
o

Nov 28th, 2013

Define 3 Stage Variables in Transformer Stage. StageVar1 will hold the I/P field ID values.
StageVar2 hold StageVar1 Value and then write a condition in StageVar3 -> If
StageVar1=Stagevar2 Then "Repea...
Answer Question Select Best Answer

19 MAR

062013

02:27 AM

1559
Views

1
Ans

Is it possible to use server components in a parallel job? If Yes,


How?
o

pamidisanthi

DataStage Interview Questions

Rupesh Agrawal
o
Nov 8th, 2013

No, It is not possible. Both component uses different code base.

What are the major differences between DataStage 8.1 and


DataStage 8.5?

naveen.chinthala

DataStage Interview Questions

rajesh
o

Oct 26th, 2013

Implement looping concepts in transformer


Seshadri B S
o

Jun 19th, 2013

Last row Handling functions in transformer are additional features


Answer Question Select Best Answer

APR

182011

04:18 AM

2967
Views

1
Ans

What is data quality and why it's needed ?

singh6

DataStage Interview Questions

vinay
o

Oct 24th, 2013

Name cleaning
Address cleaning
Answer Question Select Best Answer

SEP

032013

03:25 AM

3907
Views

1
Ans

Increment Values based on column (Looping)

Ram_1104

DataStage Interview Questions

I have source like this


Num, SeqNo,Ln,Qty
101, 1 ,1,5
I wanna target following below
Num,SeqNo,Ln,Qty
101, 1 , 1, 1
101, 1 , 2, 1
101, 1 , 3, 1
101, 1 , 4, 1

101, 1 , 5, 1
Based on Qty value records will be incremented.If qty value is 4 then o/p will be like below
Num,SeqNo,Ln,Qty
101, 1...
akhilesh.shivhare

Sep 13th, 2013

Yours i/p data


Num,SeqNum,Ln,Qty
101,1,1,5
and o/p data
Num,SeqNum,Ln,Qty
101,1,1,1
101,1,2,1
101,1,3,1
101,1,4,1
101,1,5,1
steps- use transformer
-define stage variable stgvar_Qty with value 1 means stgvar_Qty=1
-define loop with condition @ITERATION
Answer Question Select Best Answer

JUN

242012

10:46 AM

34869
Views

1
Ans

Most Difficult Jobs you have created

dwhnovice

DataStage Interview Questions

Hi Guys,
I have 3 yrs for experiences in datastage , though not much practical experience due to various reasons,
now I have been asked this question in 4 of my interviews and I am always flounder at it . Have tried
different approaches , telling the truth , telling a real situation I faced which actually was not that difficult ,
but I always seems to flounder at this question...
Kamalakar Kalidindi
o

Aug 27th, 2013

I too fed up with this question..i gave answer like this, every new job is difficult..when we are
building that job for first time, it will be difficult only..among those implementing SCD type -2
(ins...
Answer Question Select Best Answer

AUG

182013

02:14 AM

6679
Views

1
Ans

Explain Datastage configuration File and its importance

Mallikarjuna_G

DataStage Interview Questions

Sushils13

Aug 22nd, 2013

The Datastage configuration file is specified at runtime by a $APT_CONFIG_FILE variable.


Configuration file structure Datastage EE configuration file defines number of nodes,
assigns resources to ea...
Answer Question Select Best Answer

JUN

272005

06:21 AM

25493
Views

11
Ans

How to handle Date convertions in Datastage? Convert a mm/dd/yyyy


format to yyyy-dd-mm?

o
DataStage Interview Questions

We use a) "Iconv" function - Internal Convertion. b) "Oconv" function - External Convertion. Function to
convert mm/dd/yyyy format to yyyy-dd-mm is Oconv(Iconv(Filedname,"D/MDY[2,2,4]"),"D-MDY[2,2,4]")
vipul choudhary

Aug 22nd, 2013

There are two types of hash file


1. modulus hash file- file size increases randomly
2. splitting hash file-file size decreases randomly
Muralidhar
o

Jul 29th, 2013

By default datastage date format is :yyyy-mm-dd.The input is mm/dd/yyyy.


String_to_date(inputdate,"%mm%dd%yyyy")
Answer Question Select Best Answer

SEP

042011

02:56 AM

4712
Views

4
Ans

What is the meaning by OCI stage ?

ARJUN REDDY

DataStage Interview Questions

Sushils13

Aug 20th, 2013

Call Interface (OCI), it is a set of low-level APIs used to interact with Oracle
databases. It allows one to use operations like logon, execute, parss etc. using a C or C++
program.
Oracle

Ramon
o

Sep 22nd, 2011

It is your connectivity stage object to your database. Think of it as your db stage.


Answer Question Select Best Answer

NOV

212011

11:15 AM

6167
Views

2
Ans

What are the Environmental variables in Datastage?

yaoliang

DataStage Interview Questions

Sushil
o

Aug 19th, 2013

Basically Environment variable is predefined variable those we can use while creating DS
job.We can set either as Project level or Job level.Once we set specific variable that
variable will be available into the project/job
glaciya

Dec 12th, 2011

The Environmental variables in datastage are some pathes which can support system can
use as shortcuts to fulfill the program running instead of doing nonsense activity. In most
time, environmental variables are defined when the software have been installed or being
installed.
Answer Question Select Best Answer

JUN

072011

04:10 AM

6303
Views

2
Ans

How to perform loj nd ij in lookup stage

G_G_GOUD

DataStage Interview Questions

How to perform left outer join and right outer join in lookup stage
sivaksa

Aug 13th, 2013

You need right outer join in lookup..please change the link order like left to right, Right to left
put lookup failure condition ..continue...
mallika_chaithu

Jun 9th, 2011

In Lookup stage properties, you will have constraints option. If you click on constraints
button- you will get options like continue, drop, fail and rejectIf you select the option
continue: it means l...
Answer Question Select Best Answer

AUG

182011

08:25 AM

6213
Views

2
Ans

Job based scenario

devanand

DataStage Interview Questions

TABLE T1 with c1 and Table T2 with c1


13
24
25
3
4
5
...
sivaksa

Aug 13th, 2013

Yes A is the correct answer

Dinesh
o

Aug 19th, 2011

Use lookup stage...and choose reject option !


you will get the matched records in master output file(3,4,5,5) and unmatched in reject
file(1,2,2,6,7).
Answer Question Select Best Answer

OCT

132011

11:26 AM

16856
Views

8
Ans

Datastage real time scenario

amulas

DataStage Interview Questions

I have source like this


a,b,c,1,2,3 ( All this in one column)
I wanna target following below
a,b,c,1 ( Ist row)
a,b,c,2 (2nd row)
a,b,c,3 (3rd row)
Thanks in advance.

sivaksa
o

Aug 13th, 2013

Guys try simple way.


Source.....trans.....pivot...target
trans...we can use filed function and create new column
col1:filed(inputcol,,,4,1)
col2:filed(inputcol,,,5,1)
Col3:filed(inputcol,,,6,1)
And pivot put output col :col1,col2,col3
Then you can get output
Prabhakar Achyuta
o

Jul 23rd, 2013

Very simple..please go through below. Input: CUST_ID ---------- a,b,c,1,2,3 Output:


CUST_ID: ------------- a,b,c,1 a,b,c,2 a,b,c,3 Solution: We can solve it from below solution.
JOB flow: Transfo...
Answer Question Select Best Answer

242005

MAY

05:58 PM

7141
Views

3
Ans

What are Sequencers?

o
DataStage Interview Questions

Sequencers are job control programs that execute other jobs with preset Job parameters.
Read Best Answer

Editorial / Best Answer


jagannimmagadda

Member Since Jan-2006 | Jan 10th, 2006

A sequencer allows you to synchronize the control flow of multiple activities in a job
sequence. It can have multiple input triggers as well as multiple output triggers.The
sequencer operates in two modes:ALL mode. In this mode all of the inputs to the sequencer
must be TRUE for any of the sequencer outputs to fire.ANY mode. In this mode, output
triggers can be fired if any of the sequencer inputs are TRUE
Mallikarjuna_G

Aug 10th, 2013

Sequencer Activity Stage is stage that is used to control and coordinate the flow of
execution in a Job Sequence. It can accept multiple input triggers and multiple output
trigger. It has two modes...
Bimal Pradhan

May 17th, 2011

As the name suggests, it is basically to execute the jobs in sequence.For example, if you
have 5 jobs which hare interdependent , then you would call them from a sequencer.the
execution of 2nd job will be dependent and will be triggered only after execution of the 1st
job.
Improve Answer

242005

MAY

05:57 PM

8581
Views

6
Ans

How did you handle an 'Aborted' sequencer?

o
DataStage Interview Questions

In almost all cases we have to delete the data inserted by this from DB manually and fix the job and then
run the job again.
Read Best Answer

Editorial / Best Answer


ritu singhai

Member Since Jul-2008 | Jul 5th, 2008

Have you set the compilation options for the sequence so that in case job aborts, you need
not to run it from from the first job. By selecting that compilation option you can run that
aborted sequence from the point the sequence was aborted.
Like for example, you have 10 jobs(job1, job2, job3 etc.) in a sequence and the job 5 aborts,
then by checking "Add checkpoints so sequence is restartable on failure" and "Automatically
handle activities that fail" you can restart this sequence from job 5 only. it will not run the
jobs 1,2,3 and 4.
Please check these options in your sequence.
Hope this helps.
Mallikarjuna_G

Aug 10th, 2013

Two things needs to be handled when a Job Sequence aborts. 1. We must have an
exception handling code that notifies us about the failure 2. When we re-run the sequence
after fixing the Sequence, it ...
Riten
o

May 30th, 2012

To Handle Aborted Sequence These are the steps to be taken: First Add Terminator to Job
Sequence and Choose Terminator with Other option(Trigger in Job sequence) ( if job fails it
will go to Other lin...
Improve Answer

AUG

282007

01:01 AM

3718
Views

7
Ans

How to connect two stages which do not have any common columns
between them?

Praveen

DataStage Interview Questions

Mallikarjuna_G

Aug 10th, 2013

If those two stages are sources and if it is valid that though the names of columns from two
sources differ but they are same, then use a copy stage infront of one source stage and
rename column as pe...

muralidhar
o

May 30th, 2013

This is little tricky question.we can do in two ways.


1)we can connect using lookup stage if both columns having same datatype.
2)We can add one extra column that is common for two files then we can connect.
Answer Question Select Best Answer

JUN

172008

07:51 AM

5043
Views

4
Ans

Fetching last row from a particular column of sequential file

sharmilas

DataStage Interview Questions

how to fetch the last row from a particular column.. Input file may be sequential file...
Mallikarjuna_G

Aug 10th, 2013

There are multiple ways to do this, given that input is a sequential file. Two of them are
1. Use "Filter" option available in Seq file stage. Specify a unix command like: tail -1
(recommended)

2. Use a tail stage next to Seq filestage and then specify.


Akil
o

Apr 1st, 2013

There are 2 ways to fetch the last row of any file if the number of records in the file is not
known before run time:
1) Use the tail stage and run the tail stage in sequential mode
2) Use the lastrow() function in the transformer stage
Answer Question Select Best Answer

JUN

262013

08:37 AM

8432
Views

3
Ans

Datastage Performance Tuning

rameshkk

DataStage Interview Questions

How to improve the Performance Tuning any 5 points


Mallikarjuna_G

Aug 10th, 2013

Use Copy stage instead of a Transformer for simple operations like : placeholder
between stages renaming Columns dropping Columns implicit (default) type
Conversions USe Stage variab...
Prabhakar Achyuta
o

Jul 23rd, 2013

1. First filter then extract. But dont extract and filter. Use SQL instead of table method when
extracting. Say 1 million records are coming from input table but there is a filter condition
(Acct_Type...
Answer Question Select Best Answer

132011

MAY

05:33 AM

5450
Views

6
Ans

What is the purpose of the copy stage:

siva3me

DataStage Interview Questions

What is the uses of the copy stage with out copies input link to output link datasetsand it can have any
purposes pls send me with example
Mallikarjuna_G

Aug 10th, 2013

Use Copy stage for simple operations like :


to multicast the same input data among multiple output links.
placeholder between stages to avoid buffer issues
renaming Columns
dropping Columns
implicit (default) type Conversions
Akila
o

Apr 1st, 2013

Besides used for making copies of the input, copy stage is also helpful for achieving 1)
dropping columns between stages 2) to change column names 3) if the scenario is to end
the job flow directly f...
Answer Question Select Best Answer

APR

112012

12:24 AM

3499
Views

3
Ans

How do you use Surrogatekey key in reporting?

upendarkm

DataStage Interview Questions

Mallikarjuna_G

Aug 10th, 2013

In a typical data warehouse environment, we normally have a requirement to keep history.


So, we would end up having multiple rows for a given Primary key. So, we define a new
column that doesn't have ...
djkool4u

Jul 11th, 2013

Is Slowly Changing Dimension (SCD) Implementation we use Surrogate key as the primary
key is being duplicated for the sake of keeping history data for the records with the same
pk.
Answer Question Select Best Answer

AUG

282012

03:14 AM

5463
Views

2
Ans

Datastage partition

sreereddi

DataStage Interview Questions

Which partition followed by joiner,merge. lookup,remove duplicate


Mallikarjuna_G

Aug 10th, 2013

For Join, Merge and Remove duplicates, have data on links hash key partitioned an sorted
on Key columns specified. For lookup - primary link needs to be hash key partitioned and
sorted and reference link has to use entire partition method.
Shaik
o

Aug 31st, 2012

Key partition is required and data should be sorted before all these stage while using
Answer Question Select Best Answer

AUG

312012

01:42 PM

5182
Views

4
Ans

Which partition is best in performance level?

yasodha krishnasamy

DataStage Interview Questions

Mallikarjuna_G

Aug 10th, 2013

It depends on the job design and requrement. When you specify SAME partitioning,
datastage uses the partioing method defined in the previous stage and will not perform any
partitioning in the current...
Muralidhar

May 30th, 2013

Same partition.
The reason is it keeps the previous partition and send to output as it is.

No of processors in DataStage
o

naveen.chinthala

DataStage Interview Questions

How do we know the no of processors in a job? Is there any specific calculation for this.
Mallikarjuna_G

Aug 10th, 2013

There will not be any processors in the job. Your question could be - find the number of
nodes/processors on which the job is running. So the answer to this - go to Director and
open the log for the ...
Kumaresh
o

Jul 19th, 2013

Run the job. Go to director log. Check for the apt config data displayed over there. That will
show the number of processors/nodes.
Answer Question Select Best Answer

JUN

022013

01:04 AM

3602
Views

2
Ans

Job running slow


o

sunitha.gummudu

DataStage Interview Questions

Hi,
Job1 is running for 10 minutes for the first time and the same job1 is running for 15 minutes for the
second time ,the load being the same for both.Could someone explain..
Mallikarjuna_G

Aug 10th, 2013

One reason could be that - for the first time the job is just loading/inserting data into target.
When you run the same job again, it would take more time as it tries to update from second
time. Upda...
Raju Nath
o

Jun 27th, 2013

Hi you can use the delete and then load command during loading then it will be the same
time. First time when you are loading that time the dataset/table is empty that is why its
taking less time bu...
Answer Question Select Best Answer

JUN

262013

08:34 AM

4286
Views

3
Ans

Count in Dataset
o

rameshkk

DataStage Interview Questions

How to get the dataset record count? with out using orchadmin command
Mallikarjuna_G

Aug 10th, 2013

There are only two ways to this -

1. Use datastage dataset management utility[GUI].


2. Use orchadmin utility from commandline
shakeel
o

Jul 12th, 2013

Browse the path where data set loaded and run wc -l filename
Answer Question Select Best Answer

MAR

282005

06:18 PM

4951
Views

4
Ans

Dimension Modelling types along with their significance


o
o

DataStage Interview Questions

Data Modelling is Broadly classified into 2 types. a) E-R Diagrams (Entity - Relatioships). b) Dimensional
Modelling.
BHANU
o

Jul 9th, 2013

Dimensional modeling is a technique for conceptualizing and visualizing the data there are
two types of dimensional modeling
1.Snowflake schema
2.Star schema
cyberdiptikanta

Jun 26th, 2010

Dimensional Modelling:It is a designing methodlogy for designing datawarehouse with


dimensions& facts.It is of three types 1) Conceptual Modelling 2) Logical Modelling 3)
Physical Modelling1) Conc...

Answer Question Select Best Answer

OCT

042008

05:25 AM

9032
Views

3
Ans

Downstream & Upstream


o

Vanshika248

DataStage Interview Questions

What are these terms used for in Datastage - Downstream and Upstream?
mohan
o

May 17th, 2013

Upstream means data channel from source system to ETL.


there are verity of source system (RDBMS, Flat Files, XML, ODBC).this data is integrated
by ETL.this is input to ETL.
same as upstream.
Downstream means data generated from ETL.
teena
o

Feb 21st, 2013

Could you plz site examples for upstream and downstream applications
Answer Question Select Best Answer

DEC

05:28 PM

062005

6977
Views

8
Ans

If your running 4 ways parallel and you have 10 stages on the


canvas, how many processes does datastage create?
o

vishut

DataStage Interview Questions

ss
o

Mar 28th, 2013

JJ, you are wrong.


Apt_ Dump_Score
By using this only we can find.
jj
o

Nov 26th, 2012

1 conductor process
3 Section leader process
3 player process ( modify and filter are combined)
1 source sequential file
1 target sequential file
Total : 9
Answer Question Select Best Answer

MAR

042013

11:47 AM

4566
Views

2
Ans

What is the size of your database ?


o

goutam421

DataStage Interview Questions

Please answer this question


santosh varma
o

Mar 21st, 2013

Your database size is the sum of all datafiles,tempfiles,redologs. so check the sum from
dba_data_files,dba_temp_files and v$logfile
Harikrishna Chidrala
o
Mar 15th, 2013

It Depends about the size,Example My Development Database size is 6TB and UAT Server
is 12 TB and Production is ~25TB.
Answer Question Select Best Answer

MAR

062013

02:17 AM

2456
Views

1
Ans

What is host name in datastage?


o

skjilani29

DataStage Interview Questions

ysubba
o

Mar 21st, 2013

Nothing but server name

Answer Question Select Best Answer

OCT

202011

11:05 PM

4970
Views

1
Ans

Datastage 8.1
o

rachel797.ds

DataStage Interview Questions

In datastage 8.1 What is the limit of the file size? Is there limit from # of rows and # of field perspective for
file extract to be fed in profile stage
varun khare

Mar 8th, 2013

In 8.0/8.1 it uses parser which requires the entire XML document to be loaded into memory,
so that youre limited by the amount of available memory. In 8.5 it uses a really clever
streaming app
Answer Question Select Best Answer

JAN

252008

06:47 AM

4746
Views

4
Ans

Dataset in UNIX
o

manoharkolukula

DataStage Interview Questions

How to see the data in the Dataset in UNIX. What command we have to use to see the data in Dataset in
UNIX?
karn khera
o

Feb 23rd, 2013

The command which must be used is orchadmin dump datasetname(without quotes). But
before that run this command "cd `cat/.dshome`dsenv" without quotes. The reason is: cd
`cat/.dshome`-> This will cha...
Saravanan Mani
o
Jun 20th, 2012

Orchadmin rm datasetname
Answer Question Select Best Answer

DEC

282012

12:08 AM

4479
Views

1
Ans

Diff between userdefined environment variables and parameter


sets
o

google_yahoo

DataStage Interview Questions

Hi all, pls let me know the purpose of using User defined environment variables and parameters sets
MallikarjunaG

Jan 7th, 2013

An user defined environment variable is a place holder to store a value that can used in the
entire project. Parameter set is new time saving feature added in DataStage 8x. Parameter
set is set of jo...
Answer Question Select Best Answer

MAR

132009

07:37 AM

7896
Views

11
Ans

Sequential file with Duplicate Records


o

rajivkumar23us

DataStage Interview Questions

A sequential file has 8 records with one column, below are the values in the column separated by space,1
1 2 2 3 4 5 6In a parallel job after reading the sequential file 2 more sequential files should be created,
one with duplicate records and the other without duplicates.File 1 records separated by space: 1 1 2 2File
2 records separated by space: 3 4 5 6How will you do it
hussy
o

Dec 5th, 2012

Its very simple: 1. Introduce a sort stage very next to sequential file, 2. Select a property
(key change column) in sort stage and you can assign 0-Unique or 1- duplicate or viceversa
as you wish. ...
Hemant Kanthed
o
Feb 19th, 2012

After source sequential we can use sort stage with dump_key in which 0 is assigned to
duplicate record and 1 is assigned to non duplicate record after sort stage we can use
transformer stage in whic...
Answer Question Select Best Answer

OCT

302009

12:16 AM

5127
Views

3
Ans

Sequential File Stage


o

srikanth.ds

DataStage Interview Questions

If are given a list of .txt files and asked to read only the first 3 files using seq file stage. How will you do it?
matan
o

Nov 16th, 2012

In sequential file we can take a single file by using the file as Specified file.
But we can take the more than one file use File Pattern with different file names.
Metadata must be same.
narra satish

Mar 25th, 2011

HI, In sequential file we can take a single file byusing the file as Specifyed file.But we can
take the more than one file use File Pattern with different file names.Metadata must be
same.
Answer Question Select Best Answer

AUG

112010

04:34 AM

14215
Views

9
Ans

Transformer Stage Functions


o

greek143

DataStage Interview Questions

If you have Numerical+Characters data in the source, how will you load only Character data to the target?
Which functions will you use in Transformer stage?
raj
o

Nov 8th, 2012

Example : raje123ndh456ar
Code
1. convert(0123456789,,raje123ndh456ar) = rajendhar
now it is converted to character string and we can load only character.
Vinay Sharma
o

Oct 4th, 2012

Use alpha function.


Exp-->alpha(123vinay567)
Ans-->vinay
i thing this is your requeriment..
Answer Question Select Best Answer

2 AUG

212008

01:44 AM

4095
Views

8
Ans

Transformer Order of Execution


o

rajeshdhannawat1

DataStage Interview Questions

What is the order of execution in the transformer?


amarnath
o

Oct 15th, 2012

What is
1.Stage variables
2.Constraints
3.Derivations
AMIT
o

Jun 21st, 2012

Order of execution
1. Stage variable
2. Constraints
3. Derivations
Answer Question Select Best Answer

292012

3 MAY

08:14 AM

7777
Views

2
Ans

Why copy stage is better than transformer?


o

Jothi D

DataStage Interview Questions

Incase of only dropping columns(without any transformations or business rules), we can go for copy stage
instead of transformer.But can anyone tell me exactly why copy stage is better in performance than
transformer?
Atangel

Oct 12th, 2012

Transformer is a heavier stage as apart from the orchestrate we do have the C++ operators
for various derivations. Right from compilation to execution a Transformer will always take
more time than a copy stage which is a simple straight forward passive(if we can call) stage
kkreddy
o

Jul 1st, 2012

Copy stage is a passive stage and Transformer is Active and it involves the process usage
and copy does not . So its good to go with copy since it takes less time then transformer to
propagate the columns over the link
Answer Question Select Best Answer

4 AUG

242011

07:58 AM

7595
Views

3
Ans

Conditional load of records in Datastage


o

dronadula

DataStage Interview Questions

Source has sequential file stage in 10 records and move to transformer stage it has one output link 2
records and reject link has 5 records ? But i want remaining 3 records how to capture
Shirisha
o

Oct 10th, 2012

You can choose O.W option in constraints, so remaining records you will get.
anil_k_nayaka

Jan 25th, 2012

In the transformer Stage Constraints we can define the constraints where this records can
get to which link so that the missing records can get catched.
Answer Question Select Best Answer

042008

5 JUL

04:33 AM

6642
Views

13
Ans

Sequencer Scenario
o

pavan.daddanalla

DataStage Interview Questions

Scenario- if suppose we have 3 jobs in sequencer, while running if job1 is failed then we have to run job2
and job 3 ,how we can run? plz ans this thanks in advance
Rekha Ramakrishnan

Aug 23rd, 2012

In the first sequence trigger condition there would be option like condition, unconditional,
failed etc. You can select Unconditional so that even if first sequence fails or succeed it
automatically redirect to next sequence.
Gopi N
o

Jul 11th, 2012

If the scenario we have like 1st is abort and then trigger 2nd and 3rd do the below, give the
trigger condition like, 1> If job aborted give the link trigger condition which is going to 2nd
and 3rd j...
Answer Question Select Best Answer

6 FEB

232008

08:45 AM

7263
Views

7
Ans

Join and Lookup Stage


o

datastage8

DataStage Interview Questions

If you have a huge volume of data to be referenced, which stage will you use? Join or Lookup stage?
Why
Gopi N
o

Jul 11th, 2012

If we have a Huge data at reference defiantly we should go to Join stage because Look up
takes much time to process hence we will use entire partition but in Join stage we will give
the sorted data and it will simplify better than Look-up
arjunreddy
o

Aug 24th, 2011

Look up stage
Answer Question Select Best Answer

7 APR

112012

12:07 AM

3314
Views

1
Ans

Can you filter data in hashed file? (Y/N why)


o

upendarkm

DataStage Interview Questions

Gopi N
o

Jul 11th, 2012

We can filter the data in hash file based on the key column, If we have a key duplicate we
reject in to files as well by writing a SQL statement

How to use Excel file as input in DataStage??


o

Sowmya

DataStage Interview Questions

joel
o

Jun 20th, 2012

The above answer is wrong. Please follow the steps below to load the excel in dataastage :first open excel sheet and open->saveas-> then save the file as .csv extension while
importing u have to s...
srinu5077
Profile Answers by srinu5077 Questions by srinu5077

Jan 4th, 2012

I did same as step1, but when i import the metadata it asking username and password,
when press ok button it shows NO MATCHES FOUND, pls help me
Thank you
Regards
Srinivas
Answer Question Select Best Answer

JAN

082007

10:48 AM

2509
Views

1
Ans

What is the diffrence between the Dynamic RDBMS Stage &


Static RDBMS Stage ?
o

izack

DataStage Interview Questions

Shiv
o

Jun 15th, 2012

1. Dynamic RDBMS is the only stage that supports N inputs and N outputs.
2. Using Dynamic RDBMS we can read multiple tables independently.
Answer Question Select Best Answer

AUG

292005

08:10 AM

3771
Views

3
Ans

How do we do the automation of dsjobs?


o
o

DataStage Interview Questions

mcrao1

Jun 4th, 2012

DataStage Job run from Unix Command Line I am running DataStage Job from Unix
Command Line with job level parameters the job is getting abort , can someone correct if
there is any syntax problem in t...

Pavan
o

Aug 29th, 2005

We can call Datastage Batch Job from Command prompt using 'dsjob'. We can also pass all
the parameters from command prompt. Then call this shell script in any of the market
available schedulers. The 2nd option is schedule these jobs using Data Stage director.
Answer Question Select Best Answer

APR

112012

12:35 AM

4811
Views

1
Ans

How do you handle errors? Exception handlers?


o

upendarkm

DataStage Interview Questions

Jobin Joseph Veliyath


o
May 30th, 2012

The errors can be handled in different way, will list few of the methods which we can
achieve via ETL If it is a critical fact table (Merchandise, location and time (weekly)) then
Non-Critical Data...
Answer Question Select Best Answer

APR

262012

09:38 AM

3818
Views

2
Ans

Constaints
o

upendarkm

DataStage Interview Questions

What are the different types of constraints ?


Difference b/w unique and not null constraints?
SURESH
o

May 6th, 2012

Constraints are used for filter the data as transformer stage and from in sql different type of
constrains having null, isnull, primarykey,foreign key and unique constrains Ans unique
constrain is not allowed duplicate values and Notnull constrain means should be having a
data in that column.
suresh
o

May 4th, 2012

Constrain is nothing but it is an condition, it is might be any type of condition. and second
one is unique and not null means unique is maintain one record do not accept duplicate
record . not null m...
Answer Question Select Best Answer

MAR

132006

03:15 PM

8921
Views

7
Ans

What is data set? and what is file set?

mallikharjuna reddy

DataStage Interview Questions

praveen
o

Apr 28th, 2012

Data set and file set are file/stage ,these to are data extract from db.in the data set does not
have rejected link file set had rejected link
arjun
o

Jan 25th, 2012

The fundamental concept of the Orchestrate framework is the Data Set. Data Sets are the
inputs and outputs of Orchestrate operators. As a concept a Data Set is like a database
table, in so far as it...
Answer Question Select Best Answer

AUG

252005

06:32 AM

2475
Views

8
Ans

What is DS Director used for - did u use it?


o
o

pradeep
o

DataStage Interview Questions

Apr 20th, 2012

Actually Data stage director is the GUI based component and this is the Clint component,
The main use of Director is view logs, we can see view log and status of running job,By
using the multiple instances we can run the job in period of time for testing
propose.schedule the job.
abhijeet

Mar 12th, 2012

DS Administrator is use to see the log of job.Administrator monitor the job.


Answer Question Select Best Answer

APR

112012

12:30 AM

4030
Views

1
Ans

Give me some performance tips on data stage?


o

upendarkm

DataStage Interview Questions

pradeep
o

Apr 20th, 2012

To the best of my knowledge, By using partitioning technique may be used for performance
increases. and to select relevant stage for developing transformation
Answer Question Select Best Answer

242005

MAY

05:52 PM

2010
Views

3
Ans

What versions of DS you worked with?


o
DataStage Interview Questions

DS 7.0.2/6.0/5.2
svr
o

Mar 7th, 2012

7.5 is single execute system 8.01 is multiple execute system


file based repository
database repository
it is used in change capture & change apply stages. Here we r using SCDS like 1,2,3
glaciya

Dec 12th, 2011

Datastage 7.5x2
Answer Question Select Best Answer

APR

262007

04:46 AM

2269
Views

2
Ans

How to recompile the job automatically when the job failed to


compile?
o

sandy123123

DataStage Interview Questions

Seio

Feb 25th, 2012

I'm interested in this answer because for any reason my jobs appear not compile after I
begin the execution and I need an automatic process that auto compile the jobs when it
founds uncompiled.
Thanks.
nikhilanshuman

Mar 19th, 2010

If the job failed to compile, there must be some problem with the job.If the problem is not
fixed then no matter how many times it is recompiled, the compilation will fail each
time.Automatic recompile option does not seem to be there in datastage.
Answer Question Select Best Answer

242005

MAY

05:47 PM

4010
Views

2
Ans

Explain your last project and your role in it.?">

Explain your last project and your role in it.?


o
o

KRISHNA
o

DataStage Interview Questions

Feb 19th, 2012

How to explain insurance project in datastage?


nikhilanshuman

Dec 17th, 2010

Explain the interviewer about your project, its architecture and your role in that project. The
role can be developer, Tester, BA, lead etc.. Also, explain what you did in that project like
coding, preparing UTC, contributing effectively with testers to complete the UAT phase of
the project etc...

Answer Question Select Best Answer

JUN

212010

10:10 PM

3844
Views

2
Ans

Commit and Rollback


o

gantaravindranath

DataStage Interview Questions

How do you perform commit and rollback in loading jobs? What happens if the job fails in between and
What will you do?
Sivaramakrishna
o
Feb 23rd, 2012

There is an option in RDBMS stage as Transaction Isolation tab whether data commit or un
commit.
Gokul21

Jul 12th, 2010

There is an option called "Transsaction Grouping" is available. You can specify the condition
there to commit or rollback if the job fails. This option is available in teradata.
Answer Question Select Best Answer

FEB

172011

12:10 PM

3751
Views

4
Ans

Join Stage Reject Records


o

nag_sree

DataStage Interview Questions

How to get the reject records from join stage?


AvelinCorreia

Feb 23rd, 2012

If you want rejected records also, you need to use an outer join so that all the records are
carried forward and then youll need to use next stage as a transformer. In the transformer,
you can use appropriate condition in the constraints and redirect the unwanted records to a
sequential file.
Bharath
o

Feb 14th, 2012

Outside the join stage, you can put a filter to filter out the records having null values for the
columns coming from right part of the join
Answer Question Select Best Answer

AUG

122008

10:39 AM

3548
Views

3
Ans

User Created Stages


o

rajashekar kuraku

DataStage Interview Questions

Explain Wrapped, Build and Custom stages?


Phuzzball
o

Feb 15th, 2012

Datastage allows you to create your own stages with custom properties. There are three
types of custom stages: Custom: This is Orchestrate based custom stage. Orchestrate
operators are used for def...
MArcus_Datastage

Mar 17th, 2009

Custom: This allows knowledgeable Orchestrate users to specifyan Orchestrate operator as


a DataStage stage. This is then availableto use in DataStage Parallel jobs.Build: This allows
you to design and...
Answer Question Select Best Answer

FEB

082012

05:25 AM

2691
Views

0
Ans

Transformer stage Compilation Error in Ascential Datastage 7.5,


Windows XP
o

MrReddy

DataStage Interview Questions

Hi,
I have installed the datastage 7.5 in XP operating system, also installedvisual studio

.net 2008 for c++

compiler. But when trying to compile the datastage job which contain the transformer stage, I am unable
to compile.
Below warning I am getting:

Warning: the PWD


environment variable is not defined. Remote process start up may be
unsuccessful as...
Answer Question

FEB

262011

06:08 AM

3208
Views

1
Ans

What happens if we run 7.5 vesion job in 8 version


o

praveen.bollu

DataStage Interview Questions

What happens if we run 7.5 vesion job in 8 version? what is the error?
Mohammadsadiq
o
Feb 2nd, 2012

If we compile Datastage 7.5x2 job in compatible mode then it will run....if you wont complied
and run in 8.x....then it will pop up some error
Answer Question Select Best Answer

JUN

012010

12:29 AM

2209
Views

4
Ans

Nodes
o

gantaravindranath

DataStage Interview Questions

Where do you specify nodes for datastage jobs?


srinivas
o

Jan 30th, 2012

Node is nothing but an identified by the number of nodes on which parallel jobs can run r
not
j.padma89

Jul 17th, 2011

In app config file which will be named as default.apt


Answer Question Select Best Answer

OCT

232007

03:44 AM

3653
Views

4
Ans

How to connect from client machine to Unix server in Datastage?


o

amarnreddy09

DataStage Interview Questions

vjviji86

Jan 30th, 2012

I work on Datastage 8.1 ...Most of the Clients framework prefers Connect DirectProtocol
when compared to SFTP?

nikhilanshuman

Mar 17th, 2010

In Datastage there are the client components like Designer, Manager, Director.When you
open these applications, you are asked the user id, password and the Project name to
which to want to connect.The...
Answer Question Select Best Answer

NOV

072005

07:16 AM

2658
Views

1
Ans

What is DataStage Multi-byte, Single-byte file conversions in


Mainframe jobs? what is UTF 8 ? whats use of UTF 8 ?
o

ramireddy

DataStage Interview Questions

anil_k_nayaka

Jan 24th, 2012

UTF 8 is a type of file format to read files in dos format by datastage jobs.
This file may contain characters not only in English but also in foreign language
Answer Question Select Best Answer

FEB

282007

02:36 AM

7465
Views

3
Ans

What is the difference between a Filter and a Switch Stage ?


o

infinity

DataStage Interview Questions

kaps3157
o

Jan 24th, 2012

Both stages functionality and responsibilities is same. But the difference way of execution
like.. In filter stage, we have to give the multiple conditions, on multiple columns. But every
time data ...
nikhilanshuman

Dec 25th, 2009

A switch stage can have maximum 128 output links.A filter stage can have any number of
output links.

Flat File as Source


o

murali.d

DataStage Interview Questions

How to extract your flat files using sequntial file or dataset? From where to get these flat files as source?
manish parashar
o
Jan 10th, 2012

Flat files can be accessed from unix by using sequential file stage. For example if we are
having csv file . we can give the path in the sequential stage for this file and we can access
the file directly.
Answer Question Select Best Answer

NOV

01:58 AM

142011

2957
Views

2
Ans

Cannot find stage variables definition in the trasformer stage


properties
o

Gina Ying

DataStage Interview Questions

I can only see the name of the stage variables but cannot see the definition, where can I find them?
manish_toy

Jan 10th, 2012

If you are not able to view the definition then its has not been given. You have to provide the
definition by double clicking to the left of the variable and give the definition as per your
requirement.
manish_toy

Jan 10th, 2012

If it is not visible in the transformer stage then the definition has not been given. You have to
provide the definition for the stage variable by double clicking to the left of the variable, then
you can give the definition as per your requirement.
Answer Question Select Best Answer

MAR

232006

12:25 PM

1478
Views

1
Ans

All the user defined environmental variables and assignee values


for it. but Encrypted IE password i am not able to read. if it reads
only in the server job it is able to read. not from the sequencer. it
says password error...
o

James

DataStage Interview Questions

glaciya

Dec 16th, 2011

Please check your language, it may be the language of your computer did not match your
servers language.
Answer Question Select Best Answer

AUG

112010

05:05 AM

4229
Views

3
Ans

Stage Variables Initialization


o

greek143

DataStage Interview Questions

What are the initial values of stage variables? How can we set this values?
glaciya

Dec 12th, 2011

In my thought, the inital variables are the variables which can be defined when you just
have installed the software, in another word they are usually called default variable or you
have the first chance to create and config some variables that are used in a stage.
G. Venu

Oct 14th, 2011

In Transformer stage properties we can set the stage variable initial value.
Answer Question Select Best Answer

NOV

142011

01:51 AM

2825
Views

0
Ans

Unknown transformer variable definition


o

Gina Ying

DataStage Interview Questions

I can see the names of newly defined variables in the transformer stage property, but I cannot see the
definition and the logic under it. Is it normal? Where can I find the definition of the stage variables?
Answer Question

OCT

202011

03:49 AM

4983
Views

0
Ans

How to use change capture stage in scds

chandukommuri

DataStage Interview Questions

Could any one please explain me how can we use change capture stage for scd1,2,3 types?
Answer Question

APR

222010

01:28 PM

1918
Views

2
Ans

Lookup Stage
o

sujan544

DataStage Interview Questions

What is meant by Lookup Stage?


shri
o

Oct 18th, 2011

Lookup stage is similar as join stage,what ever the records from source to target we can
use by using lookup, join and in the lookup we have reject link option and join doesn't have
reject link option.
j.padma89

Jul 17th, 2011

Look up stage is used to join dataset which has similar functionality of join stage with some
extras.
the difference between join and look up is here
http://www.geekinterview.com/question_details/82179
Hope it helps !
Answer Question Select Best Answer

JAN

262008

07:22 AM

5475
Views

4
Ans

How do you remove duplicates in dataset


o

pradeep.dwh

DataStage Interview Questions

in how many ways we can delete dataset? If a record is duplicated 3 times then how to get middle
duplicated record? Is it advisable to use basic Tfr in Parallel jobs?
chhavis928

Oct 8th, 2011

We can use filter stage here to get middle record from 3 duplicate records.
srkreddy111
o

Aug 12th, 2011

First you have to open the data set and click on the partitioning and after click hash partition
and next click perform sort after click on unique and after ok and after compile and run the
job.open the target output,the duplicate records are removed..
Answer Question Select Best Answer

AUG

032011

02:18 AM

8583
Views

4
Ans

What is job commit in Datastage?


o

Mohit

DataStage Interview Questions

What is job commit in Datastage?


If I am working on parallel jobs, how can I give this option(in each odbc satge) that commit should perform
after specific number of records transaction.
Please help
RAJESH
o

Sep 11th, 2011

Read Committed and arraysize u option in target ODBC stage.


Venkat Duvvuri
o
Sep 9th, 2011

Hi Im not sure which version of datastage you r using..this is pretty much simple in
Datastage8.5 by using DTS
Thanks n Regards
Venkat Duvvuri
Answer Question Select Best Answer

AUG

272007

03:13 AM

4302
Views

3
Ans

What is the max size of Data set stage?

rangagopi

DataStage Interview Questions

Venkat Duvvuri
o
Sep 9th, 2011

S I agree with Hari's answer..Thae max size of the dataset stage is entirely depends upon
the size of the resource disk space, which we have specified under config file.
Regards,
Venkat Duvvuri
harishsj

Mar 14th, 2008

The Max size of a Dataset is equal to the summation of the space available in the Resource
disk specified in the configuration file.
Answer Question Select Best Answer

AUG

262011

10:37 AM

3272
Views

0
Ans

Rate Routing & Standard routing difference


o

shyam

DataStage Interview Questions

Can anybody tell in detail why we use Rate Routing in Repetitive Mfg. instead Normal Standard Routing?
Answer Question

242005

MAY

05:50 PM

2619
Views

3
Ans

Did you Parameterize the job or hard-coded the values in the jobs?
">

Did you Parameterize the job or hard-coded the values in the


jobs?
o
DataStage Interview Questions

Always parameterized the job. Either the values are coming from Job Properties or from a Parameter
Manager a third part tool. There is no way you will hardcode some parameters in your jobs. The
often Parameterized variables in a job are: DB DSN name, username, password, dates W.R.T for the
data to be looked against at.
venkat
o

Aug 18th, 2011

Never hard codes hardcode parameters in your jobs,Always parameterized the job. Either
the values are coming from Job Properties or parameters set
venkat
09940692102
Bimal Pradhan

May 17th, 2011

Never hard-code , until and unless it is required.Always parametrize the job. If the i/p
parameters are dynamic or not in your control then , it should taken by scheduling jobs.
Answer Question Select Best Answer

MAR

06:18 PM

282005

7746
Views

2
Ans

How do you execute datastage job from command line prompt?


o
DataStage Interview Questions

Using "dsjob" command as follows. dsjob -run -jobstatus projectname jobname


Mishra G
o

Aug 17th, 2011

$dsjob -server 97.253.32.155:31540 -user username -password passwd -run -jobstatus


-mode RESET $project_name $job_name
hanug

May 6th, 2007

U can use dsjob executable command from unix or command line.The previous post was
correct.Hanu.
Answer Question Select Best Answer

FEB

172011

12:09 PM

2498
Views

2
Ans

InvocationId

nag_sree

DataStage Interview Questions

What is invocation Id? How to find InvoicationID? What is the need of Invocation Id?
tapan8984

Aug 12th, 2011

Invocation ID is the unique identifier for distinguishing the instances in multiple instance job.
It is needed for uniquely define the instances. While running the job there is a field for
Invocation ID in the run window (pop-up).
narra satish

Mar 7th, 2011

Invocation id is nothing about instance. Normaly job run in only one instance.By enabling it
will run in multiple instance. As long as running invocation Id is unique.
Answer Question Select Best Answer

FEB

262011

06:11 AM

2232
Views

1
Ans

How to improve the performance while reading data through


sequentilal file?
o

praveen.bollu

DataStage Interview Questions

tapan8984

Aug 12th, 2011

We can use the Number of readers per Node property to enhance parallelism in Sequential
file, hence enhancing the performance..
Answer Question Select Best Answer

OCT

202005

05:48 AM

7435
Views

4
Ans

What is NLS in datastage? how we use NLS in Datastage ? what


advantages in that ? at the time of installation i am not choosen
that NLS option , now i want to use that options what can i do ?
to reinstall that datastage or first uninstall and install once
again ?
o

sreedhar

DataStage Interview Questions

mahesh
o

Aug 4th, 2011

With NLS enabled, datastage can process data in a wide range of languages & accept data
in any character set.
Anwar
o

Nov 1st, 2006

NLS is basically Local language setting(characterset) .Once u install the DS u wil get NLS
present.Just login into Admin and u can set the NLS of your project based on your project
requirement.Just ne...
Answer Question Select Best Answer

232011

MAY

03:17 PM

7014
Views

1
Ans

How to configure ODBC drivers to work with Data stage ?


o

dhora9999

DataStage Interview Questions

SinhaS

Aug 4th, 2011

1) First check that the Driver has been installed and the library env variable has been set
correctly to point to it by the UNIX admin who installed it ($LD_LIBRARY_PATH). 2) Go to
$DSHOME and update ...
Answer Question Select Best Answer

132011

JUL

02:02 AM

3347
Views

1
Ans

What is base partitioning


o

saratgunji

DataStage Interview Questions

What is base partitioning in DataStage how we can implement


Ch S Rajesh
o

Aug 1st, 2011

There is no such a question like base partitioning.Based on the requirement the partitioning
method of the stage changes.Mostly we will follow auto i.e., datastage will take RoundRobin
by default(auto).Partitioning method same will improve the performance characteristics of
datastage job.
Answer Question Select Best Answer

JUN

232011

06:17 AM

4867
Views

1
Ans

Ways of implementing scd2


o

rupam

DataStage Interview Questions

How many number of ways that you can implement SCD2 ? Explain them
RAJESH
o

Aug 1st, 2011

3 ways to construct the scd2 in datastage 8.0.1


1)using SCD stage in processing stage
2)using change capture and change applay stages
3)using source file,lookup,transformers,filters,surrogate key generator or stored
procedures,target tables
Answer Question Select Best Answer

NOV

05:47 AM

142007

2526
Views

4
Ans

Datastage Questions
o

Nagoor

DataStage Interview Questions

what is multiple instances?how u trigger the job in windows?how can u see,delete dataset in unix and
windows?what is the function of shared container?why don't we use odd nodes to run the job?what is the
normal view and materialized view?when do u use separate sort stage and inbuilt sort utility in a stage?
what is the difference?if a job is aborted?can v run the job with out compilation?how do u commit...
j.padma89

Jul 15th, 2011

How to check the no. of nodes while running the job in UNIX? Answer : type in the directory
path where the apt_config_file is located,then u can see the no of nodes and other config
information like ...
j.padma89

Jul 15th, 2011

If a job is aborted?can we run the job with out compilation


Answer:For any condition ("abort " or "reject"), one should run a job only after compilation
(strictly)

FEB

212011

06:59 AM

1575
Views

1
Ans

Job Endtime

blueboys.dsdw

Profile Answers by blueboys.dsdw Questions by blueboys.dsdw


DataStage Interview Questions

How to find job endtime?


j.padma89

Jul 14th, 2011

Job end time can be found in Director > job log


Answer Question Select Best Answer

FEB

262011

06:12 AM

3436
Views

2
Ans

How to run 10 jobs at the same time in DataStage 8.0.1?

praveen.bollu

DataStage Interview Questions

j.padma89

Jul 14th, 2011

To run 10 or any number of jobs one after the other or in any sequence, One can use "job
sequences".
In each sequence one can assign a job, so after the 10 sequences , you can run a
sequence job.
Using sequential jobs developer can even decide what should a job do if it is aborted.
amulas

Jul 13th, 2011

By using multiple job compile from ds director, we can run multiple jobs at a time.....
Answer Question Select Best Answer

132011

JUL

02:36 PM

4425
Views

1
Ans

Incremental loading

amulas

DataStage Interview Questions

How to design a job with incremental loading in datastage?

Divya
o

Jul 14th, 2011

There are ways to perform the incremental load. While performing incremental load Pls take
care of below points: 1. Retrieve changed records from last load run from source by using
appropriate extract...
Answer Question Select Best Answer

APR

092007

09:57 AM

3305
Views

4
Ans

What is the difference between Datastage Server jobs and Datastage


Parallel jobs?

Gnaneshwar

DataStage Interview Questions

saratgunji

Jul 13th, 2011

Both Server jobs and Parallel jobs run on Data Stage server only. where are server jobs
performance is slower than parallel jobs because parallel jobs run on SMP and MPP.
parallel jobs has partition...
nikhilanshuman

Nov 29th, 2009

Datastage parallel jobs can run in parallel on multiple nodes. Server jobs do not run on
multiple node.Parallel jobs support partition parallelism(Round robin,Hash,modulus etc.),
server jobs don...
Answer Question Select Best Answer

JUN

232011

07:50 AM

3536
Views

1
Ans

How do we pass parameters from UNIX?

amulas

DataStage Interview Questions

How do we pass parameters from UNIX? & How do we pass parameters by using UNIX shell scripting?
mcrao1

Jul 11th, 2011

Hi Amulaswe can pass parameter in UNIX command by using '-PARAM'.syntax:


-PARAM parameter_name %parameter value%Example: -PARAM business_date
%20110711% RegardsChalapathirao.MHCL Technologies( Malaysia)

+60182296096.

Answer Question Select Best Answer

AUG

112006

08:08 AM

3587
Views

5
Ans

How can we filter in sequential stage

madycool

DataStage Interview Questions

dhora9999

May 22nd, 2011

Ya you are right, If you do not know the answer please do not cut paste here. Please
answer the questions promptly. do not confuse anybody..Thanks D!!
rajani
o

Feb 21st, 2007

In the Filter option,u can use the UNIX command like this to filter the input records..head -3
input_file.dat
Answer Question Select Best Answer

SEP

06:44 AM

052005

5562
Views

8
Ans

Importance of Surrogate Key in Data warehousing?

o
DataStage Interview Questions

Surrogate Key is a Primary Key for a Dimension table. Most importance of using it is it is independent of
underlying database. i.e Surrogate Key is not affected by the changes going on with a database.
Datastage Etl

May 13th, 2011

Surrogate Key is used to produce the Sequence numbers. So that, based on the
Surrogate Key generated, we can identify the Unique Id in any column. Surrogate
Key is mainly implemented in the...
ASHOK1324

Apr 29th, 2009

Primary Key does not allow data duplications for actual source data. We cannot maintain
historical data of each record using PK.SID acts as a primary key in target WH systems,
that allows data duplications and maintains complete historical data along with current data.
Answer Question Select Best Answer

FEB

06:03 AM

262011

4038
Views

2
Ans

Improve performance using copy stage

praveen.bollu

DataStage Interview Questions

How can we improve performance using copy stage


Datastage Etl

May 13th, 2011

Copy stage is used to send the data to the multiple sources. Here we can change the
column names . Rather than using Transformer Stage where ever necessary , we can use
the copy stage. Like this, c...
narra satish

Mar 21st, 2011

By using copy stage we can increase the performance. In this stage we can do sorting,
Removing unwanted columns
Answer Question Select Best Answer

MAR

282005

06:18 PM

12229
Views

12
Ans

What are Stage Variables, Derivations and Constants?

o
DataStage Interview Questions

Stage Variable - An intermediate processing variable that retains value during read and doesnt pass the
value into target column. Derivation - Expression that specifies value to be passed on to the target
column. Constant - Conditions that are either true or false that specifies flow of data with a link.
narra satish

Mar 25th, 2011

Hi, Stage Variables:This are used in thr transformer For passing a Value in the Input but not
effected on output Data.It is applicable for all the output links.Dervation:This are the values
which pass...
amit101here

Oct 20th, 2010

Read below from IBM pdf :1. Any before-stage subroutine is executed. If ErrorCode is nonzero, the job aborts. 2. A row is obtained from the stream input link. 3. For each reference
input li...
Answer Question Select Best Answer

192005

MAY

11:01 AM

2808
Views

7
Ans

What is the order of execution done internally in the transformer with


the stage editor having input links on the lft hand side and output
links?

o
DataStage Interview Questions

Amit_Mishra

Apr 11th, 2011

Please follow the following details:Stage Variables-> Constraints->


Derivations/Expressions.Link ordering is mainly based to which link you want to process
first.Link Ordering is mainly useful f...
Bublu

Aug 4th, 2009

The following is the order of execution done internally in the transformer:* Stage
variables*Constraints*Derivations or ExpressionsNote: We can't change the order of
execution by using Link Ordering option. Link Ordering is no where related to this.
Answer Question Select Best Answer

242005

MAY

05:58 PM

2415
Views

1
Ans

How did u connect with DB2 in your last project?

o
DataStage Interview Questions

Most of the times the data was sent to us in the form of flat files. The data is dumped and sent to us. In
some cases were we need to connect to DB2 for look-ups as an instance then we used ODBC drivers to
connect to DB2 (or) DB2-UDB depending the situation and availability. Certainly DB2-UDB is better in
terms of performance as you know the native drivers are always better than ODBC drivers. 'iSeries...
Read Best Answer

Editorial / Best Answer


nikhilanshuman

Member Since Nov-2009 | Dec 17th, 2010

By making use of "DB2/UDB Stage".There we basically need to specify connection


parameters like client instance name,default server,default database.
nikhilanshuman

Dec 17th, 2010

By making use of "DB2/UDB Stage".There we basically need to specify connection


parameters like client instance name,default server,default database.
Improve Answer

242005

MAY

05:57 PM

4846
Views

1
Ans

What are other Performance tunings you have done in your last
project to increase the performance of slowly running jobs?

o
DataStage Interview Questions

Staged the data coming from ODBC/OCI/DB2UDB stages or any database on the server using
Hash/Sequential files for optimum performance also for data recovery in case job aborts.Tuned the OCI
stage for 'Array Size' and 'Rows per Transaction' numerical values for faster inserts, updates and
selects.Tuned the 'Project Tunables' in Administrator for better performance.Used sorted data for
Aggregator.Sorted...
Read Best Answer

Editorial / Best Answer


sistlasatish

Member Since Nov-2005 | Nov 14th, 2005

Minimise the usage of Transformer (Instead of this use Copy, modify, Filter, Row Generator)
Use SQL Code while extracting the data
3. Handle the nulls
4. Minimise the warnings
5. Reduce the number of lookups in a job design
6. Use not more than 20stages in a job
7. Use IPC stage between two passive stages Reduces processing time
8. Drop indexes before data loading and recreate after loading data into tables
9. Gen\'ll we cannot avoid no of lookups if our requirements to do lookups compulsory.
10. There is no limit for no of stages like 20 or 30 but we can break the job into small jobs then we use
dataset Stages to store the data.
1.
2.

11. IPC Stage that is provided in Server Jobs not in Parallel Jobs
12. Check the write cache of Hash file. If the same hash file is used for Look up and as well as target,
13.
14.
15.
16.
17.
18.
19.
20.

disable this Option.


If the hash file is used only for lookup then \"enable Preload to memory\". This will improve the
performance. Also, check the order of execution of the routines.
Don\'t use more than 7 lookups in the same transformer; introduce new transformers if it exceeds 7
lookups.
Use Preload to memory option in the hash file output.
Use Write to cache in the hash file input.
Write into the error tables only after all the transformer stages.
Reduce the width of the input record - remove the columns that you would not use.
Cache the hash files you are reading from and writting into. Make sure your cache is big enough to
hold the hash files.
Use ANALYZE.FILE or HASH.HELP to determine the optimal settings for your hash files.
This would also minimize overflow on the hash file.

If possible, break the input into multiple threads and run multiple instances of the job.
Staged the data coming from ODBC/OCI/DB2UDB stages or any database on the server using

Hash/Sequential files for optimum performance also for data recovery in case job aborts.
Tuned the OCI stage for 'Array Size' and 'Rows per Transaction' numerical values for faster inserts,

o
o
o

updates and selects.


Tuned the 'Project Tunables' in Administrator for better performance.
Used sorted data for Aggregator.
Sorted the data as much as possible in DB and reduced the use of DS-Sort for better performance of

jobs
Removed the data not used from the source as early as possible in the job.
Worked with DB-admin to create appropriate Indexes on tables for better performance of DS queries
Converted some of the complex joins/business in DS to Stored Procedures on DS for faster

execution of the jobs.


If an input file has an excessive number of rows and can be split-up then use standard logic to run

jobs in parallel.
Before writing a routine or a transform, make sure that there is not the functionality required in one of

the standard routines supplied in the sdk or ds utilities categories.


Constraints are generally CPU intensive and take a significant amount of time to process. This may
be the case if the constraint calls routines or external macros but if it is inline code then the overhead
will be minimal.
Try to have the constraints in the 'Selection' criteria of the jobs itself. This will eliminate the

o
o

o
o
o
o

unnecessary records even getting in before joins are made.


Tuning should occur on a job-by-job basis.
Use the power of DBMS.
Try not to use a sort stage when you can use an ORDER BY clause in the database.
Using a constraint to filter a record set is much slower than performing a SELECT WHERE.

Make every attempt to use the bulk loader for your particular database. Bulk loaders are generally
faster than using ODBC or OLE.

sistlasatish

Nov 14th, 2005

Minimise the usage of Transformer (Instead of this use Copy, modify, Filter, Row
Generator)Use SQL Code while extracting the dataHandle the nullsMinimise the
warningsReduce the number of lookups in a ...
Improve Answer

242005

MAY

05:55 PM

1972
Views

1
Ans

Where do you use Link-Partitioner and Link-Collector ?

o
DataStage Interview Questions

Link Partitioner - Used for partitioning the data.Link Collector - Used for collecting the partitioned data.
Read Best Answer

Editorial / Best Answer


nikhilanshuman

Member Since Nov-2009 | Dec 17th, 2010

Link Partitioner and collecter are basically used to introduce data parallellism in server
jobs.link partitioner,splits the data on many links.Once the data is processed,link collector
collects the data and passes it to a single link.These are used in server jobs.In datastage
parallel jobs,these things are inbuilt and automatically taken care of.
nikhilanshuman

Dec 17th, 2010

Link Partitioner and collecter are basically used to introduce data parallellism in server
jobs.link partitioner,splits the data on many links.Once the data is processed,link collector
collects the da...
Improve Answer

242005

MAY

05:54 PM

6129
Views

1
Ans

What are OConv () and Iconv () functions and where are they used?

o
o

DataStage Interview Questions

IConv() - Converts a string to an internal storage formatOConv() - Converts an expression to an output


format.
Read Best Answer

Editorial / Best Answer


sekr

Member Since Dec-2005 | Jan 30th, 2006

iconv is used to convert the date into into internal format i.e only datastage can understand
example :- date comming in mm/dd/yyyy format
datasatge will conver this ur date into some number like :- 740
u can use this 740 in derive in ur own format by using oconv.
suppose u want to change mm/dd/yyyy to dd/mm/yyyy
now u will use iconv and oconv.
ocnv(iconv(datecommingfromi/pstring,SOMEXYZ(seein help which is
iconvformat),defineoconvformat))
sekr

Jan 30th, 2006

iconv is used to convert the date into into internal format i.e only datastage can understand
example :- date comming in mm/dd/yyyy format datasatge will conver this ur date into some
number like :- 7...
Improve Answer

242005

MAY

05:54 PM

1406
Views

1
Ans

How did you connect to DB2 in your last project?

o
DataStage Interview Questions

Read Best Answer

Editorial / Best Answer


gagan8877

Member Since May-2007 | May 8th, 2007

The following stages can connect to DB2 Database:


ODBC
DB2 Plug-in Stage
Dynamic Relational Stage
gagan8877

May 8th, 2007

The following stages can connect to DB2 Database:ODBCDB2 Plug-in StageDynamic


Relational Stage
Improve Answer

242005

MAY

05:53 PM

4167
Views

6
Ans

What is Metastage?

o
DataStage Interview Questions

Read Best Answer

Editorial / Best Answer


spartankiya

Member Since Feb-2007 | Feb 19th, 2007

MetaStage is a persistent metadata Directory that uniquely synchronizes metadata across


multiple separate silos, eliminating re keying and the manual establishment of cross-tool
relationships. Based on patented technology, it provides seamless cross-tool integration
throughout the entire Business Intelligence and data integration life cycle and tool sets
jinti001

Nov 13th, 2008

Data Modeling Tool


spartankiya

Feb 19th, 2007

MetaStage is a persistent metadata Directory that uniquely synchronizes metadata across


multiple separate silos, eliminating re keying and the manual establishment of cross-tool
relationships. Based o...
Improve Answer

MAR

06:18 PM

282005

2254
Views

3
Ans

What does a Config File in parallel extender consist of?

o
DataStage Interview Questions

Config file consists of the following. a) Number of Processes or Nodes. b) Actual Disk Storage Location.
Venkat Poonati

Mar 29th, 2011

1. fastname-----usually networkname,It is physicall node name,That stages use to open


...
aladap12

Jun 26th, 2009

Biggest strength of Datastage PX is its configuration file.. Datastage engine learns about
size and shape of the system by reading configuration file.It contains Nodes and resource
available for running the job.
Answer Question Select Best Answer

AUG

092010

09:54 AM

2666
Views

4
Ans

ETL Tools Lookup Stage

nagarjuna900

DataStage Interview Questions

By using the Lookup stage we will combine the data from multiple table. By using join stage also we will
combine the data from multiple tables. Then what is the need of Lookup stage?
Venkat Poonati

Mar 29th, 2011

There are three things to consider here1.memory usage As with joiner fewer rows has to be
in memory at any time.This is not with lookup2.treating of rows with unmatched keys
Joiner w...
narra satish

Mar 21st, 2011

Join and lookup is differ in terms of memory usage. Join is light weight than lookup stage.
Join stage won't use much system resources. While performing lookup will used more
system resources. In case of lookup it fetch the whole data from refernce link into RAM
memory and then it perform lookup.
Answer Question Select Best Answer

OCT

122006

03:48 AM

2400
Views

5
Ans

How can we generate a surrogate key in server/parallel jobs?

praveen

DataStage Interview Questions

narra satish

Mar 28th, 2011

Hi,By using transformer we can generate the servgate key.In stagevariablesand write the
derviation.
neeraj82

Jul 3rd, 2009

You can generate through system variable in transformer which is built-in name is keymgt.
You don't have remember full name and in parallel you can generate through surrogate key
generator.
Answer Question Select Best Answer

FEB

262011

06:05 AM

4656
Views

3
Ans

Reading multiple files through sequential file

praveen.bollu

DataStage Interview Questions

What are the ways to read multiiple files from sequentila file if the both files are differnt
reddiraja

Mar 10th, 2011

Number of nodes = true that time we read multiple files in sequential file stage.
bupesh

Mar 9th, 2011

This can be achieved by selecting the File pattern option and the path of the the files in
the sequential stage.

Change File Attribute


o

subhalaxmipanda

DataStage Interview Questions

How to change the attribute of a .dsx format file to make it read only?
narra satish

Mar 7th, 2011

First import the job in .dsxformat and change the attributes ther 0-Editable1-Read onlyAnd
import again and override it.
Answer Question Select Best Answer

JAN

08:54 PM

292006

7317
Views

3
Ans

Explain about Error Handling in Datastage and best practices.


o

Amar

DataStage Interview Questions

madhavsandireddi

Feb 25th, 2011

In sequence job we have to run a job and if a job is aborted and if we do not have
terminator activity then control goes to the exception handler where it checks for what
reason the job is aborted..ThanksMadhav
nikhilanshuman

Dec 29th, 2009

In datastage sequence,there is "Exception Handler" activity.When you are calling your jobs
from a Datastage Sequence you should do following :Step 1: Go to properties of master
sequence and ...
Answer Question Select Best Answer

NOV

032007

03:39 AM

2866
Views

3
Ans

What is the exact difference b/w LOOKUP stage , JOIN stage and
MERGE stage?
o

vijay

DataStage Interview Questions

madhavsandireddi

Feb 24th, 2011

The major difference between LKP stage, Join stage, merge stage are LKP & JOIN: If the
reference data in the LKP stage is huge when compared to the primary data. there is no
reject requiremen...
Pavan Batchu
o

Nov 12th, 2007

Lookup stage: 1.Can only return multiple matching rows from one reference.2. Can reject
rows based on constraint.3. Can set failure.4. Does not need partitioned and sorted input.
Merge stage: 1. Can...
Answer Question Select Best Answer

JAN

162011

09:53 PM

3522
Views

1
Ans

Passing parameters from one job to another job using command


prompt
o

AlamAlam

DataStage Interview Questions

How can we pass parameters from one job to another job by using command line prompt?plz send ans...
thanks in advance.Alam
cevvavijay

Feb 10th, 2011

We can pass parameter to a job using two ways .. using dsjob- command line or from a
sequencer.Other way would be -You configure single parameter set ( version 8.0 onwards)
and use the same in both th...
Answer Question Select Best Answer

OCT

292010

08:56 AM

3775
Views

3
Ans

How to Create new user in Datastage 7.5


o

na.sreedhar

DataStage Interview Questions

I want to create new user in datastage 7.5 I dont know how to do it?
manoj kumar ganji

Jan 22nd, 2011

User creation in 7.5 is os dependant i.e.., you can make creation of users in OS. For
example when you load DS on Windows XP , then create new user in user accounts in control
panel.
tisha24

Dec 13th, 2010

The user created in and for windows will be a user for datastage by default..You can't crete
a user from datastage environment. If you want to create a user create through Control
panel user.. The...
Answer Question Select Best Answer

AUG

082010

12:47 AM

1725
Views

2
Ans

Job Unlock
o

skyboyfli

DataStage Interview Questions

How to unlock a job in Datastage 8.0?


sivakumar.datastagepx

Jan 19th, 2011

$ps -EF$kill PIDBy using above 2 commands in unix ...then we can kill the jobotherwise
tools-> cleanup resourses
svasu.r

Aug 14th, 2010

If you want unlock the jobDS DIRECTOR----------->JOB TAB-------->Clean up on resourcesIn


admin enable priviliges to director
Answer Question Select Best Answer

OCT

302009

12:18 AM

3055
Views

2
Ans

OCI Stage
o

srikanth.ds

DataStage Interview Questions

Why array list is used in Oracle

OCI stage? Brief about Important Options in OCI Stage.

Mili Jon

Jan 16th, 2011

Array size is a property of OCI stage which help to reduce the context switch between
DataStage and Oracle database. We should keep this in mind that however it help us
reducing the context switch but...
DSQuest

May 7th, 2010

I Suppose it is called "Array Size"....Array Size is mainly used to increase the buffer during
Write operations into a Oracle DB in a Server job using the Oracle OCI stage.This comes of
grea...
Answer Question Select Best Answer

192005

MAY

11:01 AM

1897
Views

4
Ans

What happens if the job fails at night?


o
o

DataStage Interview Questions

Mili Jon

Jan 16th, 2011

Job aborts and logs can be seen on Director client. However a notification can be set using
the notification Stage to send emails during failure.

gagan8877

May 8th, 2007

U can define a job sequence to send an email using SMTP activity if the job fails. Or log the
failure to a log file using DSlogfatal/DSLogEvent from controlling job or using a After Job
Routine. orUse dsJob -log from CLI.
Answer Question Select Best Answer

252008

JUL

02:36 AM

3477
Views

1
Ans

Orchestrate Schema
o

nishu_so_005

DataStage Interview Questions

What is Orchestrate Schema? Distinguish internal data type (Orchestrate schema) vs external data type
nikhilanshuman

Dec 24th, 2010

Orchestrate schema defines the fields and their datatypes.There are two types of
Orchestarte shcemas a) Input Schems b)Output schema.Orchestrate Schema can be
compared with "STRUCT" of C.Example : RollNo - int32
Answer Question Select Best Answer

242005

MAY

06:00 PM

3640
Views

1
Ans

Tell me one situation from your last project, where you had faced
problem and How did u solve it?
o
DataStage Interview Questions

A. The jobs in which data is read directly from OCI stages are running extremely slow. I had to stage the
data before sending to the transformer to make the jobs run faster.B. The job aborts in the middle of
loading some 500,000 rows. Have an option either cleaning/deleting the loaded data and then run the
fixed job or run the job again from the row the job has aborted. To make sure the load...
nikhilanshuman

Dec 18th, 2010

a) We had a big job with around 40 stages.The job was taking too long tocompile and
run.We broke the job into 3 smaller jobs.After this ,we observed that the performance was
slighly improved and maint...
Answer Question Select Best Answer

JUN

062005

05:22 AM

1894
Views

1
Ans

When should we use ODS?


o
o

DataStage Interview Questions

DWH's are typically read only, batch updated on a scheduleODS's are maintained in more real time,
trickle fed constantly
nikhilanshuman

Dec 18th, 2010

An opearation data store contains the data which are constantly updated through the course
of the business operations. ODS is specially designed such that it can quickly perform
relatively simply que...
Answer Question Select Best Answer

112008

JUL

04:42 AM

3023
Views

1
Ans

DSParams file
o

arcteetc

DataStage Interview Questions

Is it like that if we define the values of variables in DSParams file then there is no need to give the values
at job level ar Project level ?& how to configure this file at job level ?so that we need not hardcode the
values.....
nikhilanshuman

Dec 18th, 2010

yes.DSParams file contains all the project level parameters which are set up in Datastage
Administrator.If the values are provided/modified in this file,the changes are automatically
reflected in the ...
Answer Question Select Best Answer

FEB

152008

04:48 AM

2333
Views

1
Ans

How can we load the flat file


o

etl_bhargavi

DataStage Interview Questions

How can i load a flat file into target as fast as i can?Assuming that the source bottleneck is not there,that
is there is no performance issues in the source side.
nikhilanshuman

Dec 18th, 2010

A Flat file can be read using sequential file stage.To make data load faster,try implementing
least stages in you job and use minimal transformations.if you are trying to load the data of
a flat file ...
Answer Question Select Best Answer

DEC

072005

02:41 PM

893
Views

1
Ans

Does the BibhudataStage Oracle plug-in better than OCI plug-in


coming from DataStage? What is the BibhudataStage extra
functions?

MustageemRaees

DataStage Interview Questions

nikhilanshuman

Dec 18th, 2010

There is no such thing like "BibhudataStage Oracle plug-in" in Datastage.The question


seems to be wrong.
Answer Question Select Best Answer

JUN

072006

01:49 PM

992
Views

1
Ans

Hican any one can explain what areDB2 UDB utilitiesub


o

ub

DataStage Interview Questions

nikhilanshuman

Dec 18th, 2010

"DB2 UDB" is a stage in Datastage using which the connectivity could be made to DB2
databases.It could be used to fetch the data from DB2 databases or to perform DML
operations in DB2(e.g. Insert/update) or Bulk load etc..
Answer Question Select Best Answer

AUG

09:18 AM

202007

1862
Views

1
Ans

A batch is running and it is scheduled to run in 5 minutes. But


after 10 days the time changes to 10 minutes. What type of error
is this and how to fix it?
o

subharatanjena

DataStage Interview Questions

nikhilanshuman

Dec 18th, 2010

If the batch is taking more time to execute(10 mins),it may be due to the performance
issues.In such cases the performance optimizstion mesures should be taken.If without
making any changes,time taken...
Answer Question Select Best Answer

FEB

282007

02:45 AM

1803
Views

1
Ans

Does type of partitioning change for SMP and MPP systems?


o

infinity

DataStage Interview Questions

nikhilanshuman

Dec 18th, 2010

Datastage is platform independent.A same job,when designed properly can run on SMP as
well as MPP systems.Partitioning does not need to be changed for SMP/MPP systems.
Answer Question Select Best Answer

NOV

222005

09:45 AM

1777
Views

1
Ans

What is trouble shhoting in server jobs ? what are the diff kinds
of errors encountered while running any job?
o

Ajju2005

DataStage Interview Questions

nikhilanshuman

Dec 18th, 2010

Troubleshooting in datastage server jobs involves monitoring the job log for fatal errors and
taking appropriate actions to resolve them.There can be various errors which could be
encountered while ru...
Answer Question Select Best Answer

242005

MAY

05:58 PM

6286
Views

1
Ans

Read the String functions in DS


o
DataStage Interview Questions

Functions like [] -> sub-string function and ':' -> concatenation operatorSyntax: string [ [ start, ] length ]
string [ delimiter, instance, repeats ]
nikhilanshuman

Dec 17th, 2010

It seems that it is being asked to explain string function.Following are some string functions
used in datastage:Compare,Field,Convert,Padstring,TrimB,TrimF,TrimLeadingTrailingSome
string conversion functions
:StringToDate,StringToDecimal,StringToTime,StaringToUstring,StringToTimestamp etc...
Answer Question Select Best Answer

032006

JUL

01:02 AM

1092
Views

1
Ans

Actually my requirement is like that :Here is the codification


suggested:
SALE_HEADER_XXXXX_YYYYMMDD.PSVSALE_LINE_XXXXX_Y
YYYMMDD.PSVXXXXX = LVM sequence to ensure unicity and
continuity of file exchangesCaution, there will an increment to
implement.YYYYMMDD = LVM date of file creation

COMPRESSION AND DELIVERY TO:


SALE_HEADER_XXXXX_YYYYMMDD.ZIP AND
SALE_LINE_XXXXX_YYYYMMDD.ZIPif we run that job the target
file names are like this sale_header_1_20060206 &
sale_line_1_20060206.if we run next time means the target files
we like this sale_header_2_20060206 & sale_line_2_20060206.if
we run the same in next day means the target files we want like
this sale_header_3_20060306 & sale_line_3_20060306.Ie if
whenever we run the same job the target files automatically
changing in to filename_increment to previous
number(previousnumber + 1)_currentdate;please do needful by
repling this question..
o

sreedhar kancherla

DataStage Interview Questions

nikhilanshuman

Dec 17th, 2010

Here,the basic thing required is to add 1 to the number which was used in last
load.Following is the logic for this:a) Initially take a file and write 0 to it.b) Now in the
sequence,create a user vari...

Anda mungkin juga menyukai