The least useful property of unicode is that a trivial subset of it can appear to be 'simple text'.
I fully agree. It fails hard instead of failing safe.
Perl could mitigate that problem by keeping track of whether a string is decoded or not.
Recognise that unicode isn't a single format, but many formats all lumped together in a confused and confusing mess.
I don't follow. Who thinks UTF-8 and UTF-16le are the same format?
rationalise the formats to a single, fixed-width, self-identifying format.
Not sure "self-identifying" makes sense. length($a) + length($b) == length($a . $b) is a nice property. It's possible to cause hard failures on misuse without self-identification.
In reply to Re^4: JSON::XS (and JSON::PP) appear to generate invalid UTF-8 for character in range 127 to 255
by ikegami
in thread JSON::XS (and JSON::PP) appear to generate invalid UTF-8 for character in range 127 to 255
by Ovid
| For: | Use: | ||
| & | & | ||
| < | < | ||
| > | > | ||
| [ | [ | ||
| ] | ] |