UTF-32 (32-bitUnicode Transformation Format), sometimes called UCS-4, is a fixed-lengthencoding used to encode Unicodecode points that uses exactly 32 bits (fourbytes) per code point (but a number of leading bits must be zero as there are far fewer than 232 Unicode code points, needing actually only 21 bits).[1] In contrast, all other Unicode transformation formats are variable-length encodings. Each 32-bit value in UTF-32 represents one Unicode code point and is exactly equal to that code point's numerical value.
The main advantage of UTF-32 is that the Unicode code points are directly indexed. Finding theNth code point in a sequence of code points is aconstant-time operation. In contrast, avariable-length code requireslinear-time to countN code points from the start of the string. This makes UTF-32 a simple replacement in code that usesintegers that are incremented by one to examine each location in astring, as was commonly done forASCII. However, Unicode code points are rarely processed in complete isolation, such ascombining character sequences and for emoji.[2]
The main disadvantage of UTF-32 is that it is space-inefficient, using fourbytes per code point, including 11 bits that are always zero. Characters beyond theBMP are relatively rare in most texts (except, for example, in the case of texts with some popular emojis), and can typically be ignored for sizing estimates. This makes UTF-32 close to twice the size ofUTF-16. It can be up to four times the size ofUTF-8 depending on how many of the characters are in theASCII subset.[2]
The originalISO/IEC 10646 standard defines a 32-bitencoding form calledUCS-4, in which each code point in theUniversal Character Set (UCS) is represented by a 31-bit value from 0 to 0x7FFFFFFF (the sign bit was unused and zero). In November 2003, Unicode was restricted by RFC 3629 to match the constraints of theUTF-16 encoding: explicitly prohibiting code points greater than U+10FFFF (and also the high and low surrogates U+D800 through U+DFFF). This limited subset defines UTF-32.[3][1] Although the ISO standard had (as of 1998 in Unicode 2.1) "reserved for private use" 0xE00000 to 0xFFFFFF, and 0x60000000 to 0x7FFFFFFF[4] these areas were removed in later versions. Because the Principles and Procedures document ofISO/IEC JTC 1/SC 2 Working Group 2 states that all future assignments of code points will be constrained to the Unicode range, UTF-32 will be able to represent all UCS code points and UTF-32 and UCS-4 are identical.[5]
A fixed number of bytes per code point has theoretical advantages, but each of these has problems in reality:
The main use of UTF-32 is in internal APIs where the data is single code points orglyphs, rather than strings of characters. For instance, in modern text rendering, it is common[citation needed] that the last step is to build a list of structures each containingcoordinates (x, y), attributes, and a single UTF-32 code point identifying the glyph to draw. Often non-Unicode information is stored in the "unused" 11 bits of each word.[citation needed]
Use of UTF-32 strings on Windows (wherewchar_t is 16 bits) is almost non-existent. On Unix systems, UTF-32 strings are sometimes, but rarely, used internally by applications, due to the typewchar_t being defined as 32-bit.
UTF-32 is also forbidden as an HTML character encoding.[8][9]
Python versions up to 3.2 can be compiled to use them[clarification needed] instead ofUTF-16; from version 3.3 onward, Unicode strings are stored in UTF-32 if there is at least 1 non-BMP character in the string, but with leading zero bytes optimized away "depending on the [code point] with the largest Unicode ordinal (1, 2, or 4 bytes)" to make all code points that size.[10]
Seed7[11] andLasso[citation needed] programming languages encode all strings with UTF-32, in the belief that direct indexing is important, whereas theJulia programming language moved away from built-in UTF-32 support with its 1.0 release, simplifying the language to having only UTF-8 strings (with all the other encodings considered legacy and moved out of the standard library to package[12]) following the "UTF-8 Everywhere Manifesto".[13]
C++11 has 2 built-in data types that use UTF-32. Thechar32_t
data type stores 1 character in UTF-32. Theu32string
data type stores a string of UTF-32-encoded characters. A UTF-32-encoded character or string literal is marked withU
before the character or string literal.[14][15]
#include<string>char32_tUTF32_character=U'🔟';// also written as U'\U0001F51F'std::u32stringUTF32_string=U"UTF–32-encoded string";// defined as `const char32_t*´
C# has aUTF32Encoding
class which represents Unicode characters as bytes, rather than as a string.[16]
Though technically invalid, the surrogate halves are often encoded and allowed. This allows invalid UTF-16 (such as Windows filenames) to be translated to UTF-32, similar to how theWTF-8 variant of UTF-8 works. Sometimes paired surrogates are encoded instead of non-BMP characters, similar toCESU-8. Due to the large number of unused 32-bit values, it is also possible to preserve invalid UTF-8 by using non-Unicode values to encode UTF-8 errors, though there is no standard for this.
UTF-32 has 2 versions for big-endian and little-endian:UTF-32-BE andUTF-32-LE.
Clause 9.4: "Because surrogate code points are not UCS scalar values, UTF-32 code units in the range 0000 D800-0000 DFFF are ill-formed". Clause 4.57: "[UCS codespace] consisting of the integers from 0 to 10 FFFF (hexadecimal)". Clause 4.58: "[UCS scalar value] any UCS code point except high-surrogate and low-surrogate code points".
It [UCS-4] is now treated simply as a synonym for UTF-32, and is considered the canonical form for representation of characters in 10646.
Folks start implying that code points mean something, and that O(1) indexing or slicing at code point boundaries is a useful operation.