Archives of the TeradataForum
Message Posted: Mon, 24 Feb 2003 @ 19:57:02 GMT
| Subj: || || Re: Couting duplicate rows thru multi load |
| From: || || Venkata Panabakam |
| ||My experience has been that the writing of that UV table may be performed on a row-by-row basis, not in a block fashion. So it may
be slow.|| |
| ||Why not multiload the 25 million to a multiset staging table. Then use SQL and grouping to find the duplicates and non-duplicates.
At least you stay with parallel operations.|| |
finding duplicates rows on teradata, does it slow down teradata? usually data validation will done before loading data into
teradata .... I am assuming loading duplicate rows into a table (with multi set) will slow down the teradata.... that is why we are
separating duplicate rows on flat file itself.........
pls..correct me if i am wrong....