mc4ndr3 commented: "... Explicit width types have better defined behavior. ""
TazeTSchnitzel story link reference to c & c type sizes.
Type widths/sizes do not impart encoding information, just limit how many bits referencing in a given programming instanc. (use of 32 bits means don't have to chain 4 8bit groups together).
java script still has the same C type size issues, just instead of integer & (void), it's disguised as a visual package called unicode.
Historically, just had to worry about confusing letter/digit
o/0 and l/ (simple to just check the ASCIC, Ebcdic, or other numerical value of the encoded for character used.
Unicode raises that to a whole new level.
What's the collation order for a printed unicode character that's defined by 4 different unicodes?
What's the collation order for printed unicode character that can be defined 6 different ways using one or more different unicode characters?
Does it really matter if it's UTF_16, UTF_32, UTF_64?
mc4ndr3 commented: "... Explicit width types have better defined behavior. ""
TazeTSchnitzel story link reference to c & c type sizes.
Type widths/sizes do not impart encoding information, just limit how many bits referencing in a given programming instanc. (use of 32 bits means don't have to chain 4 8bit groups together).
java script still has the same C type size issues, just instead of integer & (void), it's disguised as a visual package called unicode.
https://seriot.ch/resources/talks_papers/20141106_asfws_unic... )
Simple example:
Historically, just had to worry about confusing letter/digit o/0 and l/ (simple to just check the ASCIC, Ebcdic, or other numerical value of the encoded for character used.
Unicode raises that to a whole new level.
What's the collation order for a printed unicode character that's defined by 4 different unicodes?
What's the collation order for printed unicode character that can be defined 6 different ways using one or more different unicode characters?
Does it really matter if it's UTF_16, UTF_32, UTF_64?