Unicode normalization problem

Brendan Eich brendan at mozilla.org
Thu Apr 2 22:33:59 UTC 2015

It was the 90s, when 16 bits seemed enough. Wish we could go back. Even 
in 1995 this was obviously going to fail, but the die had been cast 
years earlier in Windows and Java APIs and language/implementation designs.


Claude Pache wrote:
> (So, taking your example, the 💩 character is internally represented as a sequence of two 16-bit-units, not “characters”. And, very confusingly, the String methods that contain “char” in their name have nothing to do with “characters”.)
> —Claude

More information about the es-discuss mailing list