|
|
Archives of the TeradataForum
Message Posted: Mon, 24 Feb 2003 @ 19:57:02 GMT
Subj: | | Re: Couting duplicate rows thru multi load |
|
From: | | Venkata Panabakam |
| My experience has been that the writing of that UV table may be performed on a row-by-row basis, not in a block fashion. So it may
be slow. | |
| Why not multiload the 25 million to a multiset staging table. Then use SQL and grouping to find the duplicates and non-duplicates.
At least you stay with parallel operations. | |
finding duplicates rows on teradata, does it slow down teradata? usually data validation will done before loading data into
teradata .... I am assuming loading duplicate rows into a table (with multi set) will slow down the teradata.... that is why we are
separating duplicate rows on flat file itself.........
pls..correct me if i am wrong....
| |