Unicode normalization problem
Brendan Eich
brendan at mozilla.org
Thu Apr 2 22:33:59 UTC 2015
It was the 90s, when 16 bits seemed enough. Wish we could go back. Even
in 1995 this was obviously going to fail, but the die had been cast
years earlier in Windows and Java APIs and language/implementation designs.
/be
Claude Pache wrote:
> (So, taking your example, the 💩 character is internally represented as a sequence of two 16-bit-units, not “characters”. And, very confusingly, the String methods that contain “char” in their name have nothing to do with “characters”.)
>
> —Claude
More information about the es-discuss
mailing list