|  |  | Archives of the TeradataForumMessage Posted: Mon, 24 Feb 2003 @ 19:57:02 GMT
 
 
  
| Subj: |  | Re: Couting duplicate rows thru multi load |  |  |  | From: |  | Venkata Panabakam |  
 |  | My experience has been that the writing of that UV table may be performed on a row-by-row basis, not in a block fashion.  So it may
be slow. |  | 
 
 
 |  | Why not multiload the 25 million to a multiset staging table.  Then use SQL and grouping to find the duplicates and non-duplicates.
At least you stay with parallel operations. |  | 
 
 
 
 finding duplicates rows on teradata,  does it slow down teradata? usually data validation will done before loading data into
teradata .... I am assuming loading duplicate rows into a table (with multi set) will slow down the teradata.... that is why we are
separating duplicate rows on flat file itself......... pls..correct me if i am wrong.... 
 
 |  |