Many—including the OP, apparently—assume it indicates whether the characters[1] of the string are Code Points or bytes. It does not.
It's a bit that indicates the internal storage format of the string.
When 0, the string is stored in the "downgraded" format.
The characters are stored as an array of C char objects.
When 1, the string is stored in the "upgraded" format.
The characters—whatever they may be—are encoded using utf8 (not UTF-8).
Being internal, you have no reason to access it unless debugging an XS module (which must deal with the two formats) or Perl itself. In such cases, you can use aforementioned utf8::is_utf8 or Devel::Peek's Dump. C code has access to the similar SvUTF8 and sv_dump.
In reply to Re^3: How to set the UTF8 flag?
by ikegami
in thread How to set the UTF8 flag?
by dissident
| For: | Use: | ||
| & | & | ||
| < | < | ||
| > | > | ||
| [ | [ | ||
| ] | ] |