Você está na página 1de 2

key change and cluster key change do the exactly same, only difference cluster k ey change require sorted

or sorted group data. That means sort key mode should not be sort for cluster key change. -------------------------------------------------------------------------------------------------------Why Entire partition is used in LOOKUP stage ? Entire partition has all data across the nodes So while matching(in lookup) the records all data should be present across all nodes. For lookup sorting is not required.so when we are not using entire partition the n reference data splits into all nodes. Then each primary record need check with all nodes for matched reference record.Then we face performance issue.If we use entire in lookup then one primary record needs to look into 1 node is enough.if match found then that record goes to target otherwise it move to reject,drop etc(base d on requirement)no need check in another node.In this case if we are running jo b in 4 nodes then at a time 4 records should process. Note:Please remember we go for lookup only we have small reference data.If we go for big data it is performance issue(I/O work will increase here) and also some times job will abort. ----------------------------------------------------------------------------------------------------How to load data from the record it failed to load I am working with a DS job responsible for loading data into the dimension table . I have noticed that of the 15000 records that have to be loaded only 5000 have been loaded as the job aborted prematurely. Is there a way to start loading data from 5001 record (without the need to drop the initial 5000 records and reload all the 15000 records)? ******** the one way you can do this is to do a lookup to the target table based on necce ssay keys and load only the ones that do not find a match. if you are inserting into the table using the "Upsert" option then you can set t he environment related to the commit frequency to zero (assuming the target is O racle) - in which case your records are only committed once the entire transaction has gone through successfully. Havent tried this yet though. ***** we can also use "Tail stage" to get all the remaining records from 5001 to 15000 and the target is the same SeqF stage with append mode. -------------------------------------------------------------------------------------------------------interview-scenario 1. src flat file - emp, dept tables 2. target i need ename,deptno,dname,sal 3.in output i must have avg sal on deptno pls give me sql query for the same ********************8 select deptno,(select avg(sal) from emp b where a.deptno=b.deptno)from emp a ord er by a.deptno;

****************88 in datastage take source(emp)--->aggrigator(group by deptno,calculation on sal and select avg sal column) {this output is as referance to the---->look-up{primary link file i s (emp)){key column is deptno}---->dataset run this job u will get that output. -----------------------------------------------------------------------------------------------------------

Você também pode gostar