bigsplash opened this issue on Jan 03, 2007 · 17 posts
kuroyume0161 posted Sun, 22 April 2007 at 4:36 PM
stewer, what were the encodings for non-ASCII characters in previous versions (allowing for the variety causing potential troubles)? Was there a BOM in use in any of these situations?
Miss Nancy - Wrong! The entire idea of Unicode is to allow encodings of large character sets - especially Sino-Japanese Kana/Kanji, Arabic, Hebrew, European, Korean, Cyrillic, and a great number of others - You can't specify 20,000,000 different glyph characters in 8-bit ASCII (ever). You can specify different ones - individually - this way in a limited scope (European characters can be included in 8-bit UNsigned specifications such as ASCII-Latin 1), but that fails for ideographics like Sino-Japanese and Korean. This is why larger encodings were necessary - and this is exactly what Unicode is for.
http://en.wikipedia.org/wiki/Unicode
http://www.unicode.org
C makes it easy to shoot yourself in the
foot. C++ makes it harder, but when you do, you blow your whole leg
off.
-- Bjarne
Stroustrup
Contact Me | Kuroyume's DevelopmentZone