|
Archives of the TeradataForumMessage Posted: Fri, 18 Jan 2008 @ 12:22:18 GMT
The first thing to remember is that Teradata stored Unicode in UTF16 i.e. double byte in all cases. So, if the character set is Unicode, they are all 2 byte characters! when they get to Teradata- hence it is not really possible. If the character set on the column is Latin, you cannot have a multibyte character. Any latin characters with a hex value of 80 or greater will result in a 2 byte character in a UTF8 file. Characters may be corrupted moving between systems (between Oracle systems, between Teradata systems or between the two) if you are not careful. If the column uses 2 byte characters on Oracle, and you have the NLS_LANG environment variable set to UTF8, any characters over X'80' go out as 2 bytes. If you load this as UTF8 (specify the character set in Fastload), it will go back fine. If you do not specify UTF8 to fastload, you will get 2 bytes loaded to the field, neither of which are the correct characters. Check that you specified the UTF8 character set in Fastload - you need this even if you are loading to Latin (if the output file comes from source with multipart characters. If this is OK, extract the column to a file on both systems and check the contents by opening the file in binary using something like Textpad or Ultraedit. Hope that helps Dave Clough
| ||||||||||||||||||||||||||||||||||||||||||||||||
| ||||||||||||||||||||||||||||||||||||||||||||||||
Copyright 2016 - All Rights Reserved | ||||||||||||||||||||||||||||||||||||||||||||||||
Last Modified: 15 Jun 2023 | ||||||||||||||||||||||||||||||||||||||||||||||||