爱心技术专栏专题

nationofUFT-8andUnicode

摘录:java基础 来源:java基础 加入时间:2006年08月19日
摘要:
nationofUFT-8andUnicode
nation of UFT-8 and Unicode

What is unicode?

  A mapping with characters and a index, we use u+xxxx to represent it.

Confuse with unicode and UTF-8?    Unicode is a standard char set, UTF-8 …

转载:转载请保留本信息,本文来自
http://www.51dibs.com
/html/2006/article/info16/a_ef73c11671589353.htm

nationofUFT-8andUnicode

站点:爱心种子小博士 关键字:nationofUFT-8and

   
nation of UFT-8 and Unicode

What is unicode?

  A mapping with characters and a index, we use u+xxxx to represent it.

Confuse with unicode and UTF-8?    Unicode is a standard char set, UTF-8 is one of implementation, just one of UCS-2, UCS-4 and so forth, but it becomes standard way of encoding. but note one thing, when we are talking about some english characters, those two standard are the same, it means

U-00000000 - U-0000007F:  0xxxxxxx

    sometimes, especially the programmer, since U-00000000 - U-0000007F is enough for their dialy use(26 english and some symbols), so, there is no different between the character set standards(unicode) and implementation standard(UTF-8) for them. when they are talking with you, you may confuse.

Why is UTF-8?    You may ask why not use UCS-4 or UCS-2? do people like 8 more(in cantonese, it means become rich)?        The answer is no. Using UCS-2 (or UCS-4) under Unix would lead to very severe problems. Strings with these encodings can contain as parts of many wide characters bytes like \0 or / which have a special meaning in filenames and other C library function parameters.

(An ASCII or Latin-1 file can be transformed into a UCS-2 file by simply inserting a 0x00 byte in front of every ASCII byte. If we want to have a UCS-4 file, we have to insert three 0x00 bytes instead before every ASCII byte.)

    In addition, the majority of UNIX tools expects ASCII files and cant read 16-bit words as characters without major modifications.(In UTF-8U+0000 to U+007F (ASCII) are encoded simply as bytes 0x00 to 0x7F (ASCII compatibility).

This means that files and strings which contain only 7-bit ASCII characters have the same encoding under both ASCII and UTF-8)

------------prove the ASCII and UTF-8 are the same---------package unicode;public class CharTest {    public static void main(String[] args) throws Exception {        char[] chars = new char[]{\u007F};        String str = new String(chars);        System.out.println("within 0000 - 007F : " + str);       //for the character whose unicode less than u0080, it is no different      with encode by  //ISO-8859-1 or UTF-8. they are compatiable.        System.out.println("   UTF-8 - UTF-8      " + new String(str.getBytes("UTF-8"),

"ISO-8859-1"));         System.out.println("   ISO-8859-1 - UTF-8 " +new String(str.getBytes("ISO-8859-1"),

"UTF-8"));        chars = new char[]{\u00F2};        str = new String(chars); //The above principle can not apply to the character lager than 007F        System.out.println("out of 0000 - 007F : " + str);        System.out.println("   UTF-8 - UTF-8      "  + new String(str.getBytes("UTF-8"),

"ISO-8859-1"));        System.out.println("   ISO-8859-1 - UTF-8 "  + new String(str.getBytes("ISO-8859-

1"), "UTF-8"));    }}---------------------------------------------------------------------------------

How long is the UTF-8 encoding?    Theoretically, it can be 6 bytes, but actually, 3 byte is enough for us since BMP is not longer than 3(The most commonly used characters, including all those found in major older encoding standards, have been placed into the first plane (0x0000 to 0xFFFD), which is called the Basic Multilingual Plane (BMP))   

Important UTF-8 features:  1. UCS characters U+0000 to U+007F (ASCII) are encoded simply as bytes 0x00 to 0x7F (ASCII compatibility). This means that files and strings which contain only 7-bit ASCII characters have the same encoding under both ASCII and UTF-8.   2. All UCS characters >U+007F are encoded as a sequence of several bytes, each of which has the most significant bit set. Therefore, no ASCII byte (0x00-0x7F) can appear as part of any other character.   3. The first byte of a multibyte sequence that represents a non-ASCII character is always in the range 0xC0 to 0xFD and it indicates how many bytes follow for this character. All further bytes in a multibyte sequence are in the range 0x80 to 0xBF. This allows easy resynchronization and makes the encoding stateless and robust against missing bytes. (?? the further investigate is necessary. can explain this currently)  4. All possible 231 UCS codes can be encoded.   5. UTF-8 encoded characters may theoretically be up to six bytes long, however 16-bit BMP

characters are only up to three bytes long.   6. The bytes 0xFE and 0xFF are never used in the UTF-8 encoding. ------------Prove the features(1,2,3)-----------------package unicode;

public class UTF8Features {    public static void main(String[] args) throws Exception {        //Why not write some no-ASCII character in the src?        //Since it will depends on your system rather than        //a UTF-8 as your image        char[] chars = new char[]{\u007F};        String str = new String(chars);        System.out.println("Point 1 : " + str);        System.out.println("   UTF-8 - UTF-8      "                + new String(str.getBytes("UTF-8"), "ISO-8859-1"));        System.out.println("   ISO-8859-1 - UTF-8 "                + new String(str.getBytes("ISO-8859-1"), "UTF-8"));        System.out.println();

        chars = new char[]{\uE840};        str = new String(chars);        System.out.println("Point 2 : " + str);        //just a sample you can use this method to verify more characters        System.out.println("   No less than 7F      " + getHexString(str));

        chars = new char[]{\u2260};        str = new String(chars);        //just a sample you can use this method to verify more characters        System.out.println("Point 3 : " + str);        System.out.println("   Range of 1st Byte      " + getHexString(str));    }

    public static String getHexString(String num) throws Exception {        StringBuffer sb = new StringBuffer();        //You must specify UTF-8 here, else it will use the defaul encoding        //which depends on your enviroment        byte[] bytes = num.getBytes("UTF-8");        for (int i = 0; i < bytes.length; i++) {            sb.append(Integer.toHexString((bytes[i] >= 0 ?                     bytes[i] : 256 + bytes[i])).toUpperCase() + " ");        }        return sb.toString();    }}---------------------------------------------------------------------------------Pinciple of presenting a unicode use UTF-8:

U-00000000 - U-0000007F:  0xxxxxxx  U-00000080 - U-000007FF:  110xxxxx 10xxxxxx  U-00000800 - U-0000FFFF:  1110xxxx 10xxxxxx 10xxxxxx  U-00010000 - U-001FFFFF:  11110xxx 10xxxxxx 10xxxxxx 10xxxxxx  U-00200000 - U-03FFFFFF:  111110xx 10xxxxxx 10xxxxxx 10xxxxxx 10xxxxxx  U-04000000 - U-7FFFFFFF:  1111110x 10xxxxxx 10xxxxxx 10xxxxxx 10xxxxxx 10xxxxxx 

How to use the principle above?

Sample:The Unicode character U+00A9 = 1010 1001 (copyright sign) is encoded in UTF-8 as

    11000010 10101001 = 0xC2 0xA9

Explain :

A:1010

9:1001

principle 2 : 00000080 <  00A9 < 000007FF

from low to high

1. There 6 x in the low bit    we cut last 6 bit from  - 10101001(A9)  which is 101001

2.There 5 x in the high bit. we cut the rest 2 bit of A9 which is 10 and extend it to 5 bit with three 0 which is 00010

complete the low byte with 10. ----> (10) combine (101001) -> 10101001

complete the high byte with 110, ---> (110) combine (00010) -> 11000010

the Result is

11000010 10101001 = 0xC2 0xA9

you can also verify the following unicode with principle 3 use the way above:

U-00000800 - U-0000FFFF:  1110xxxx 10xxxxxx 10xxxxxx 

character U+2260 = 0010 0010 0110 0000 (not equal to) is encoded as:

    11100010 10001001 10100000 = 0xE2 0x89 0xA0



Reference:

http://www.cl.cam.ac.uk/~mgk25/unicode.html#unicode



客户服务中心信箱:[email protected] [email protected] 网站地图

声明

合作伙伴: