Archives of the TeradataForum
Message Posted: Sat, 08 May 2004 @ 15:18:26 GMT
You mentioned that converting the source data might be difficult, however, using the native data formats would be much faster than making the load utility search byte by byte for a delimiter.
Now that I have that out of my system, I suggest you use a different delimiter, one that does not appear anywhere in your data. I realize this is changing the "source data" but you didn't indicate how the data is created. If an unload utility is being used, it should have an option to specify the delimiter, like Fastload, MultiLoad and TPump, even Queryman.
As far as your second question, the input data should fit into the memory location defined in the .layout. I might suggest making the field the largest that you might ever be sent (sounds like a problem on the creation side of the equation). Then, load your table or a temporal table and then run a query to do whatever you want with rows that contain a data value greater than CHAR or CHARACTER_LENGTH or OCTET_LENGTH bytes long. You might as well scrub your bad data in parallel instead of serially.
Hope these suggestions at least give you a place to start.
|Copyright 2016 - All Rights Reserved|
|Last Modified: 27 Dec 2016|