|
|
.. _unicode-howto:
|
|
|
|
|
|
*****************
|
|
|
Unicode HOWTO
|
|
|
*****************
|
|
|
|
|
|
:Release: 1.12
|
|
|
|
|
|
This HOWTO discusses Python's support for the Unicode specification
|
|
|
for representing textual data, and explains various problems that
|
|
|
people commonly encounter when trying to work with Unicode.
|
|
|
|
|
|
|
|
|
Introduction to Unicode
|
|
|
=======================
|
|
|
|
|
|
Definitions
|
|
|
-----------
|
|
|
|
|
|
Today's programs need to be able to handle a wide variety of
|
|
|
characters. Applications are often internationalized to display
|
|
|
messages and output in a variety of user-selectable languages; the
|
|
|
same program might need to output an error message in English, French,
|
|
|
Japanese, Hebrew, or Russian. Web content can be written in any of
|
|
|
these languages and can also include a variety of emoji symbols.
|
|
|
Python's string type uses the Unicode Standard for representing
|
|
|
characters, which lets Python programs work with all these different
|
|
|
possible characters.
|
|
|
|
|
|
Unicode (https://www.unicode.org/) is a specification that aims to
|
|
|
list every character used by human languages and give each character
|
|
|
its own unique code. The Unicode specifications are continually
|
|
|
revised and updated to add new languages and symbols.
|
|
|
|
|
|
A **character** is the smallest possible component of a text. 'A', 'B', 'C',
|
|
|
etc., are all different characters. So are 'È' and 'Í'. Characters vary
|
|
|
depending on the language or context you're talking
|
|
|
about. For example, there's a character for "Roman Numeral One", 'Ⅰ', that's
|
|
|
separate from the uppercase letter 'I'. They'll usually look the same,
|
|
|
but these are two different characters that have different meanings.
|
|
|
|
|
|
The Unicode standard describes how characters are represented by
|
|
|
**code points**. A code point value is an integer in the range 0 to
|
|
|
0x10FFFF (about 1.1 million values, the
|
|
|
`actual number assigned <https://www.unicode.org/versions/latest/#Summary>`_
|
|
|
is less than that). In the standard and in this document, a code point is written
|
|
|
using the notation ``U+265E`` to mean the character with value
|
|
|
``0x265e`` (9,822 in decimal).
|
|
|
|
|
|
The Unicode standard contains a lot of tables listing characters and
|
|
|
their corresponding code points:
|
|
|
|
|
|
.. code-block:: none
|
|
|
|
|
|
0061 'a'; LATIN SMALL LETTER A
|
|
|
0062 'b'; LATIN SMALL LETTER B
|
|
|
0063 'c'; LATIN SMALL LETTER C
|
|
|
...
|
|
|
007B '{'; LEFT CURLY BRACKET
|
|
|
...
|
|
|
2167 'Ⅷ'; ROMAN NUMERAL EIGHT
|
|
|
2168 'Ⅸ'; ROMAN NUMERAL NINE
|
|
|
...
|
|
|
265E '♞'; BLACK CHESS KNIGHT
|
|
|
265F '♟'; BLACK CHESS PAWN
|
|
|
...
|
|
|
1F600 '😀'; GRINNING FACE
|
|
|
1F609 '😉'; WINKING FACE
|
|
|
...
|
|
|
|
|
|
Strictly, these definitions imply that it's meaningless to say 'this is
|
|
|
character ``U+265E``'. ``U+265E`` is a code point, which represents some particular
|
|
|
character; in this case, it represents the character 'BLACK CHESS KNIGHT',
|
|
|
'♞'. In
|
|
|
informal contexts, this distinction between code points and characters will
|
|
|
sometimes be forgotten.
|
|
|
|
|
|
A character is represented on a screen or on paper by a set of graphical
|
|
|
elements that's called a **glyph**. The glyph for an uppercase A, for example,
|
|
|
is two diagonal strokes and a horizontal stroke, though the exact details will
|
|
|
depend on the font being used. Most Python code doesn't need to worry about
|
|
|
glyphs; figuring out the correct glyph to display is generally the job of a GUI
|
|
|
toolkit or a terminal's font renderer.
|
|
|
|
|
|
|
|
|
Encodings
|
|
|
---------
|
|
|
|
|
|
To summarize the previous section: a Unicode string is a sequence of
|
|
|
code points, which are numbers from 0 through ``0x10FFFF`` (1,114,111
|
|
|
decimal). This sequence of code points needs to be represented in
|
|
|
memory as a set of **code units**, and **code units** are then mapped
|
|
|
to 8-bit bytes. The rules for translating a Unicode string into a
|
|
|
sequence of bytes are called a **character encoding**, or just
|
|
|
an **encoding**.
|
|
|
|
|
|
The first encoding you might think of is using 32-bit integers as the
|
|
|
code unit, and then using the CPU's representation of 32-bit integers.
|
|
|
In this representation, the string "Python" might look like this:
|
|
|
|
|
|
.. code-block:: none
|
|
|
|
|
|
P y t h o n
|
|
|
0x50 00 00 00 79 00 00 00 74 00 00 00 68 00 00 00 6f 00 00 00 6e 00 00 00
|
|
|
0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23
|
|
|
|
|
|
This representation is straightforward but using it presents a number of
|
|
|
problems.
|
|
|
|
|
|
1. It's not portable; different processors order the bytes differently.
|
|
|
|
|
|
2. It's very wasteful of space. In most texts, the majority of the code points
|
|
|
are less than 127, or less than 255, so a lot of space is occupied by ``0x00``
|
|
|
bytes. The above string takes 24 bytes compared to the 6 bytes needed for an
|
|
|
ASCII representation. Increased RAM usage doesn't matter too much (desktop
|
|
|
computers have gigabytes of RAM, and strings aren't usually that large), but
|
|
|
expanding our usage of disk and network bandwidth by a factor of 4 is
|
|
|
intolerable.
|
|
|
|
|
|
3. It's not compatible with existing C functions such as ``strlen()``, so a new
|
|
|
family of wide string functions would need to be used.
|
|
|
|
|
|
Therefore this encoding isn't used very much, and people instead choose other
|
|
|
encodings that are more efficient and convenient, such as UTF-8.
|
|
|
|
|
|
UTF-8 is one of the most commonly used encodings, and Python often
|
|
|
defaults to using it. UTF stands for "Unicode Transformation Format",
|
|
|
and the '8' means that 8-bit values are used in the encoding. (There
|
|
|
are also UTF-16 and UTF-32 encodings, but they are less frequently
|
|
|
used than UTF-8.) UTF-8 uses the following rules:
|
|
|
|
|
|
1. If the code point is < 128, it's represented by the corresponding byte value.
|
|
|
2. If the code point is >= 128, it's turned into a sequence of two, three, or
|
|
|
four bytes, where each byte of the sequence is between 128 and 255.
|
|
|
|
|
|
UTF-8 has several convenient properties:
|
|
|
|
|
|
1. It can handle any Unicode code point.
|
|
|
2. A Unicode string is turned into a sequence of bytes that contains embedded
|
|
|
zero bytes only where they represent the null character (U+0000). This means
|
|
|
that UTF-8 strings can be processed by C functions such as ``strcpy()`` and sent
|
|
|
through protocols that can't handle zero bytes for anything other than
|
|
|
end-of-string markers.
|
|
|
3. A string of ASCII text is also valid UTF-8 text.
|
|
|
4. UTF-8 is fairly compact; the majority of commonly used characters can be
|
|
|
represented with one or two bytes.
|
|
|
5. If bytes are corrupted or lost, it's possible to determine the start of the
|
|
|
next UTF-8-encoded code point and resynchronize. It's also unlikely that
|
|
|
random 8-bit data will look like valid UTF-8.
|
|
|
6. UTF-8 is a byte oriented encoding. The encoding specifies that each
|
|
|
character is represented by a specific sequence of one or more bytes. This
|
|
|
avoids the byte-ordering issues that can occur with integer and word oriented
|
|
|
encodings, like UTF-16 and UTF-32, where the sequence of bytes varies depending
|
|
|
on the hardware on which the string was encoded.
|
|
|
|
|
|
|
|
|
References
|
|
|
----------
|
|
|
|
|
|
The `Unicode Consortium site <https://www.unicode.org>`_ has character charts, a
|
|
|
glossary, and PDF versions of the Unicode specification. Be prepared for some
|
|
|
difficult reading. `A chronology <https://www.unicode.org/history/>`_ of the
|
|
|
origin and development of Unicode is also available on the site.
|
|
|
|
|
|
On the Computerphile Youtube channel, Tom Scott briefly
|
|
|
`discusses the history of Unicode and UTF-8 <https://www.youtube.com/watch?v=MijmeoH9LT4>`_
|
|
|
(9 minutes 36 seconds).
|
|
|
|
|
|
To help understand the standard, Jukka Korpela has written `an introductory
|
|
|
guide <http://jkorpela.fi/unicode/guide.html>`_ to reading the
|
|
|
Unicode character tables.
|
|
|
|
|
|
Another `good introductory article <https://www.joelonsoftware.com/2003/10/08/the-absolute-minimum-every-software-developer-absolutely-positively-must-know-about-unicode-and-character-sets-no-excuses/>`_
|
|
|
was written by Joel Spolsky.
|
|
|
If this introduction didn't make things clear to you, you should try
|
|
|
reading this alternate article before continuing.
|
|
|
|
|
|
Wikipedia entries are often helpful; see the entries for "`character encoding
|
|
|
<https://en.wikipedia.org/wiki/Character_encoding>`_" and `UTF-8
|
|
|
<https://en.wikipedia.org/wiki/UTF-8>`_, for example.
|
|
|
|
|
|
|
|
|
Python's Unicode Support
|
|
|
========================
|
|
|
|
|
|
Now that you've learned the rudiments of Unicode, we can look at Python's
|
|
|
Unicode features.
|
|
|
|
|
|
The String Type
|
|
|
---------------
|
|
|
|
|
|
Since Python 3.0, the language's :class:`str` type contains Unicode
|
|
|
characters, meaning any string created using ``"unicode rocks!"``, ``'unicode
|
|
|
rocks!'``, or the triple-quoted string syntax is stored as Unicode.
|
|
|
|
|
|
The default encoding for Python source code is UTF-8, so you can simply
|
|
|
include a Unicode character in a string literal::
|
|
|
|
|
|
try:
|
|
|
with open('/tmp/input.txt', 'r') as f:
|
|
|
...
|
|
|
except OSError:
|
|
|
# 'File not found' error message.
|
|
|
print("Fichier non trouvé")
|
|
|
|
|
|
Side note: Python 3 also supports using Unicode characters in identifiers::
|
|
|
|
|
|
répertoire = "/tmp/records.log"
|
|
|
with open(répertoire, "w") as f:
|
|
|
f.write("test\n")
|
|
|
|
|
|
If you can't enter a particular character in your editor or want to
|
|
|
keep the source code ASCII-only for some reason, you can also use
|
|
|
escape sequences in string literals. (Depending on your system,
|
|
|
you may see the actual capital-delta glyph instead of a \u escape.) ::
|
|
|
|
|
|
>>> "\N{GREEK CAPITAL LETTER DELTA}" # Using the character name
|
|
|
'\u0394'
|
|
|
>>> "\u0394" # Using a 16-bit hex value
|
|
|
'\u0394'
|
|
|
>>> "\U00000394" # Using a 32-bit hex value
|
|
|
'\u0394'
|
|
|
|
|
|
In addition, one can create a string using the :func:`~bytes.decode` method of
|
|
|
:class:`bytes`. This method takes an *encoding* argument, such as ``UTF-8``,
|
|
|
and optionally an *errors* argument.
|
|
|
|
|
|
The *errors* argument specifies the response when the input string can't be
|
|
|
converted according to the encoding's rules. Legal values for this argument are
|
|
|
``'strict'`` (raise a :exc:`UnicodeDecodeError` exception), ``'replace'`` (use
|
|
|
``U+FFFD``, ``REPLACEMENT CHARACTER``), ``'ignore'`` (just leave the
|
|
|
character out of the Unicode result), or ``'backslashreplace'`` (inserts a
|
|
|
``\xNN`` escape sequence).
|
|
|
The following examples show the differences::
|
|
|
|
|
|
>>> b'\x80abc'.decode("utf-8", "strict") #doctest: +NORMALIZE_WHITESPACE
|
|
|
Traceback (most recent call last):
|
|
|
...
|
|
|
UnicodeDecodeError: 'utf-8' codec can't decode byte 0x80 in position 0:
|
|
|
invalid start byte
|
|
|
>>> b'\x80abc'.decode("utf-8", "replace")
|
|
|
'\ufffdabc'
|
|
|
>>> b'\x80abc'.decode("utf-8", "backslashreplace")
|
|
|
'\\x80abc'
|
|
|
>>> b'\x80abc'.decode("utf-8", "ignore")
|
|
|
'abc'
|
|
|
|
|
|
Encodings are specified as strings containing the encoding's name. Python
|
|
|
comes with roughly 100 different encodings; see the Python Library Reference at
|
|
|
:ref:`standard-encodings` for a list. Some encodings have multiple names; for
|
|
|
example, ``'latin-1'``, ``'iso_8859_1'`` and ``'8859``' are all synonyms for
|
|
|
the same encoding.
|
|
|
|
|
|
One-character Unicode strings can also be created with the :func:`chr`
|
|
|
built-in function, which takes integers and returns a Unicode string of length 1
|
|
|
that contains the corresponding code point. The reverse operation is the
|
|
|
built-in :func:`ord` function that takes a one-character Unicode string and
|
|
|
returns the code point value::
|
|
|
|
|
|
>>> chr(57344)
|
|
|
'\ue000'
|
|
|
>>> ord('\ue000')
|
|
|
57344
|
|
|
|
|
|
Converting to Bytes
|
|
|
-------------------
|
|
|
|
|
|
The opposite method of :meth:`bytes.decode` is :meth:`str.encode`,
|
|
|
which returns a :class:`bytes` representation of the Unicode string, encoded in the
|
|
|
requested *encoding*.
|
|
|
|
|
|
The *errors* parameter is the same as the parameter of the
|
|
|
:meth:`~bytes.decode` method but supports a few more possible handlers. As well as
|
|
|
``'strict'``, ``'ignore'``, and ``'replace'`` (which in this case
|
|
|
inserts a question mark instead of the unencodable character), there is
|
|
|
also ``'xmlcharrefreplace'`` (inserts an XML character reference),
|
|
|
``backslashreplace`` (inserts a ``\uNNNN`` escape sequence) and
|
|
|
``namereplace`` (inserts a ``\N{...}`` escape sequence).
|
|
|
|
|
|
The following example shows the different results::
|
|
|
|
|
|
>>> u = chr(40960) + 'abcd' + chr(1972)
|
|
|
>>> u.encode('utf-8')
|
|
|
b'\xea\x80\x80abcd\xde\xb4'
|
|
|
>>> u.encode('ascii') #doctest: +NORMALIZE_WHITESPACE
|
|
|
Traceback (most recent call last):
|
|
|
...
|
|
|
UnicodeEncodeError: 'ascii' codec can't encode character '\ua000' in
|
|
|
position 0: ordinal not in range(128)
|
|
|
>>> u.encode('ascii', 'ignore')
|
|
|
b'abcd'
|
|
|
>>> u.encode('ascii', 'replace')
|
|
|
b'?abcd?'
|
|
|
>>> u.encode('ascii', 'xmlcharrefreplace')
|
|
|
b'ꀀabcd޴'
|
|
|
>>> u.encode('ascii', 'backslashreplace')
|
|
|
b'\\ua000abcd\\u07b4'
|
|
|
>>> u.encode('ascii', 'namereplace')
|
|
|
b'\\N{YI SYLLABLE IT}abcd\\u07b4'
|
|
|
|
|
|
The low-level routines for registering and accessing the available
|
|
|
encodings are found in the :mod:`codecs` module. Implementing new
|
|
|
encodings also requires understanding the :mod:`codecs` module.
|
|
|
However, the encoding and decoding functions returned by this module
|
|
|
are usually more low-level than is comfortable, and writing new encodings
|
|
|
is a specialized task, so the module won't be covered in this HOWTO.
|
|
|
|
|
|
|
|
|
Unicode Literals in Python Source Code
|
|
|
--------------------------------------
|
|
|
|
|
|
In Python source code, specific Unicode code points can be written using the
|
|
|
``\u`` escape sequence, which is followed by four hex digits giving the code
|
|
|
point. The ``\U`` escape sequence is similar, but expects eight hex digits,
|
|
|
not four::
|
|
|
|
|
|
>>> s = "a\xac\u1234\u20ac\U00008000"
|
|
|
... # ^^^^ two-digit hex escape
|
|
|
... # ^^^^^^ four-digit Unicode escape
|
|
|
... # ^^^^^^^^^^ eight-digit Unicode escape
|
|
|
>>> [ord(c) for c in s]
|
|
|
[97, 172, 4660, 8364, 32768]
|
|
|
|
|
|
Using escape sequences for code points greater than 127 is fine in small doses,
|
|
|
but becomes an annoyance if you're using many accented characters, as you would
|
|
|
in a program with messages in French or some other accent-using language. You
|
|
|
can also assemble strings using the :func:`chr` built-in function, but this is
|
|
|
even more tedious.
|
|
|
|
|
|
Ideally, you'd want to be able to write literals in your language's natural
|
|
|
encoding. You could then edit Python source code with your favorite editor
|
|
|
which would display the accented characters naturally, and have the right
|
|
|
characters used at runtime.
|
|
|
|
|
|
Python supports writing source code in UTF-8 by default, but you can use almost
|
|
|
any encoding if you declare the encoding being used. This is done by including
|
|
|
a special comment as either the first or second line of the source file::
|
|
|
|
|
|
#!/usr/bin/env python
|
|
|
# -*- coding: latin-1 -*-
|
|
|
|
|
|
u = 'abcdé'
|
|
|
print(ord(u[-1]))
|
|
|
|
|
|
The syntax is inspired by Emacs's notation for specifying variables local to a
|
|
|
file. Emacs supports many different variables, but Python only supports
|
|
|
'coding'. The ``-*-`` symbols indicate to Emacs that the comment is special;
|
|
|
they have no significance to Python but are a convention. Python looks for
|
|
|
``coding: name`` or ``coding=name`` in the comment.
|
|
|
|
|
|
If you don't include such a comment, the default encoding used will be UTF-8 as
|
|
|
already mentioned. See also :pep:`263` for more information.
|
|
|
|
|
|
|
|
|
Unicode Properties
|
|
|
------------------
|
|
|
|
|
|
The Unicode specification includes a database of information about
|
|
|
code points. For each defined code point, the information includes
|
|
|
the character's name, its category, the numeric value if applicable
|
|
|
(for characters representing numeric concepts such as the Roman
|
|
|
numerals, fractions such as one-third and four-fifths, etc.). There
|
|
|
are also display-related properties, such as how to use the code point
|
|
|
in bidirectional text.
|
|
|
|
|
|
The following program displays some information about several characters, and
|
|
|
prints the numeric value of one particular character::
|
|
|
|
|
|
import unicodedata
|
|
|
|
|
|
u = chr(233) + chr(0x0bf2) + chr(3972) + chr(6000) + chr(13231)
|
|
|
|
|
|
for i, c in enumerate(u):
|
|
|
print(i, '%04x' % ord(c), unicodedata.category(c), end=" ")
|
|
|
print(unicodedata.name(c))
|
|
|
|
|
|
# Get numeric value of second character
|
|
|
print(unicodedata.numeric(u[1]))
|
|
|
|
|
|
When run, this prints:
|
|
|
|
|
|
.. code-block:: none
|
|
|
|
|
|
0 00e9 Ll LATIN SMALL LETTER E WITH ACUTE
|
|
|
1 0bf2 No TAMIL NUMBER ONE THOUSAND
|
|
|
2 0f84 Mn TIBETAN MARK HALANTA
|
|
|
3 1770 Lo TAGBANWA LETTER SA
|
|
|
4 33af So SQUARE RAD OVER S SQUARED
|
|
|
1000.0
|
|
|
|
|
|
The category codes are abbreviations describing the nature of the character.
|
|
|
These are grouped into categories such as "Letter", "Number", "Punctuation", or
|
|
|
"Symbol", which in turn are broken up into subcategories. To take the codes
|
|
|
from the above output, ``'Ll'`` means 'Letter, lowercase', ``'No'`` means
|
|
|
"Number, other", ``'Mn'`` is "Mark, nonspacing", and ``'So'`` is "Symbol,
|
|
|
other". See
|
|
|
`the General Category Values section of the Unicode Character Database documentation <https://www.unicode.org/reports/tr44/#General_Category_Values>`_ for a
|
|
|
list of category codes.
|
|
|
|
|
|
|
|
|
Comparing Strings
|
|
|
-----------------
|
|
|
|
|
|
Unicode adds some complication to comparing strings, because the same
|
|
|
set of characters can be represented by different sequences of code
|
|
|
points. For example, a letter like 'ê' can be represented as a single
|
|
|
code point U+00EA, or as U+0065 U+0302, which is the code point for
|
|
|
'e' followed by a code point for 'COMBINING CIRCUMFLEX ACCENT'. These
|
|
|
will produce the same output when printed, but one is a string of
|
|
|
length 1 and the other is of length 2.
|
|
|
|
|
|
One tool for a case-insensitive comparison is the
|
|
|
:meth:`~str.casefold` string method that converts a string to a
|
|
|
case-insensitive form following an algorithm described by the Unicode
|
|
|
Standard. This algorithm has special handling for characters such as
|
|
|
the German letter 'ß' (code point U+00DF), which becomes the pair of
|
|
|
lowercase letters 'ss'.
|
|
|
|
|
|
::
|
|
|
|
|
|
>>> street = 'Gürzenichstraße'
|
|
|
>>> street.casefold()
|
|
|
'gürzenichstrasse'
|
|
|
|
|
|
A second tool is the :mod:`unicodedata` module's
|
|
|
:func:`~unicodedata.normalize` function that converts strings to one
|
|
|
of several normal forms, where letters followed by a combining
|
|
|
character are replaced with single characters. :func:`normalize` can
|
|
|
be used to perform string comparisons that won't falsely report
|
|
|
inequality if two strings use combining characters differently:
|
|
|
|
|
|
::
|
|
|
|
|
|
import unicodedata
|
|
|
|
|
|
def compare_strs(s1, s2):
|
|
|
def NFD(s):
|
|
|
return unicodedata.normalize('NFD', s)
|
|
|
|
|
|
return NFD(s1) == NFD(s2)
|
|
|
|
|
|
single_char = 'ê'
|
|
|
multiple_chars = '\N{LATIN SMALL LETTER E}\N{COMBINING CIRCUMFLEX ACCENT}'
|
|
|
print('length of first string=', len(single_char))
|
|
|
print('length of second string=', len(multiple_chars))
|
|
|
print(compare_strs(single_char, multiple_chars))
|
|
|
|
|
|
When run, this outputs:
|
|
|
|
|
|
.. code-block:: shell-session
|
|
|
|
|
|
$ python3 compare-strs.py
|
|
|
length of first string= 1
|
|
|
length of second string= 2
|
|
|
True
|
|
|
|
|
|
The first argument to the :func:`~unicodedata.normalize` function is a
|
|
|
string giving the desired normalization form, which can be one of
|
|
|
'NFC', 'NFKC', 'NFD', and 'NFKD'.
|
|
|
|
|
|
The Unicode Standard also specifies how to do caseless comparisons::
|
|
|
|
|
|
import unicodedata
|
|
|
|
|
|
def compare_caseless(s1, s2):
|
|
|
def NFD(s):
|
|
|
return unicodedata.normalize('NFD', s)
|
|
|
|
|
|
return NFD(NFD(s1).casefold()) == NFD(NFD(s2).casefold())
|
|
|
|
|
|
# Example usage
|
|
|
single_char = 'ê'
|
|
|
multiple_chars = '\N{LATIN CAPITAL LETTER E}\N{COMBINING CIRCUMFLEX ACCENT}'
|
|
|
|
|
|
print(compare_caseless(single_char, multiple_chars))
|
|
|
|
|
|
This will print ``True``. (Why is :func:`NFD` invoked twice? Because
|
|
|
there are a few characters that make :meth:`casefold` return a
|
|
|
non-normalized string, so the result needs to be normalized again. See
|
|
|
section 3.13 of the Unicode Standard for a discussion and an example.)
|
|
|
|
|
|
|
|
|
Unicode Regular Expressions
|
|
|
---------------------------
|
|
|
|
|
|
The regular expressions supported by the :mod:`re` module can be provided
|
|
|
either as bytes or strings. Some of the special character sequences such as
|
|
|
``\d`` and ``\w`` have different meanings depending on whether
|
|
|
the pattern is supplied as bytes or a string. For example,
|
|
|
``\d`` will match the characters ``[0-9]`` in bytes but
|
|
|
in strings will match any character that's in the ``'Nd'`` category.
|
|
|
|
|
|
The string in this example has the number 57 written in both Thai and
|
|
|
Arabic numerals::
|
|
|
|
|
|
import re
|
|
|
p = re.compile(r'\d+')
|
|
|
|
|
|
s = "Over \u0e55\u0e57 57 flavours"
|
|
|
m = p.search(s)
|
|
|
print(repr(m.group()))
|
|
|
|
|
|
When executed, ``\d+`` will match the Thai numerals and print them
|
|
|
out. If you supply the :const:`re.ASCII` flag to
|
|
|
:func:`~re.compile`, ``\d+`` will match the substring "57" instead.
|
|
|
|
|
|
Similarly, ``\w`` matches a wide variety of Unicode characters but
|
|
|
only ``[a-zA-Z0-9_]`` in bytes or if :const:`re.ASCII` is supplied,
|
|
|
and ``\s`` will match either Unicode whitespace characters or
|
|
|
``[ \t\n\r\f\v]``.
|
|
|
|
|
|
|
|
|
References
|
|
|
----------
|
|
|
|
|
|
.. comment should these be mentioned earlier, e.g. at the start of the "introduction to Unicode" first section?
|
|
|
|
|
|
Some good alternative discussions of Python's Unicode support are:
|
|
|
|
|
|
* `Processing Text Files in Python 3 <http://python-notes.curiousefficiency.org/en/latest/python3/text_file_processing.html>`_, by Nick Coghlan.
|
|
|
* `Pragmatic Unicode <https://nedbatchelder.com/text/unipain.html>`_, a PyCon 2012 presentation by Ned Batchelder.
|
|
|
|
|
|
The :class:`str` type is described in the Python library reference at
|
|
|
:ref:`textseq`.
|
|
|
|
|
|
The documentation for the :mod:`unicodedata` module.
|
|
|
|
|
|
The documentation for the :mod:`codecs` module.
|
|
|
|
|
|
Marc-André Lemburg gave `a presentation titled "Python and Unicode" (PDF slides)
|
|
|
<https://downloads.egenix.com/python/Unicode-EPC2002-Talk.pdf>`_ at
|
|
|
EuroPython 2002. The slides are an excellent overview of the design of Python
|
|
|
2's Unicode features (where the Unicode string type is called ``unicode`` and
|
|
|
literals start with ``u``).
|
|
|
|
|
|
|
|
|
Reading and Writing Unicode Data
|
|
|
================================
|
|
|
|
|
|
Once you've written some code that works with Unicode data, the next problem is
|
|
|
input/output. How do you get Unicode strings into your program, and how do you
|
|
|
convert Unicode into a form suitable for storage or transmission?
|
|
|
|
|
|
It's possible that you may not need to do anything depending on your input
|
|
|
sources and output destinations; you should check whether the libraries used in
|
|
|
your application support Unicode natively. XML parsers often return Unicode
|
|
|
data, for example. Many relational databases also support Unicode-valued
|
|
|
columns and can return Unicode values from an SQL query.
|
|
|
|
|
|
Unicode data is usually converted to a particular encoding before it gets
|
|
|
written to disk or sent over a socket. It's possible to do all the work
|
|
|
yourself: open a file, read an 8-bit bytes object from it, and convert the bytes
|
|
|
with ``bytes.decode(encoding)``. However, the manual approach is not recommended.
|
|
|
|
|
|
One problem is the multi-byte nature of encodings; one Unicode character can be
|
|
|
represented by several bytes. If you want to read the file in arbitrary-sized
|
|
|
chunks (say, 1024 or 4096 bytes), you need to write error-handling code to catch the case
|
|
|
where only part of the bytes encoding a single Unicode character are read at the
|
|
|
end of a chunk. One solution would be to read the entire file into memory and
|
|
|
then perform the decoding, but that prevents you from working with files that
|
|
|
are extremely large; if you need to read a 2 GiB file, you need 2 GiB of RAM.
|
|
|
(More, really, since for at least a moment you'd need to have both the encoded
|
|
|
string and its Unicode version in memory.)
|
|
|
|
|
|
The solution would be to use the low-level decoding interface to catch the case
|
|
|
of partial coding sequences. The work of implementing this has already been
|
|
|
done for you: the built-in :func:`open` function can return a file-like object
|
|
|
that assumes the file's contents are in a specified encoding and accepts Unicode
|
|
|
parameters for methods such as :meth:`~io.TextIOBase.read` and
|
|
|
:meth:`~io.TextIOBase.write`. This works through :func:`open`\'s *encoding* and
|
|
|
*errors* parameters which are interpreted just like those in :meth:`str.encode`
|
|
|
and :meth:`bytes.decode`.
|
|
|
|
|
|
Reading Unicode from a file is therefore simple::
|
|
|
|
|
|
with open('unicode.txt', encoding='utf-8') as f:
|
|
|
for line in f:
|
|
|
print(repr(line))
|
|
|
|
|
|
It's also possible to open files in update mode, allowing both reading and
|
|
|
writing::
|
|
|
|
|
|
with open('test', encoding='utf-8', mode='w+') as f:
|
|
|
f.write('\u4500 blah blah blah\n')
|
|
|
f.seek(0)
|
|
|
print(repr(f.readline()[:1]))
|
|
|
|
|
|
The Unicode character ``U+FEFF`` is used as a byte-order mark (BOM), and is often
|
|
|
written as the first character of a file in order to assist with autodetection
|
|
|
of the file's byte ordering. Some encodings, such as UTF-16, expect a BOM to be
|
|
|
present at the start of a file; when such an encoding is used, the BOM will be
|
|
|
automatically written as the first character and will be silently dropped when
|
|
|
the file is read. There are variants of these encodings, such as 'utf-16-le'
|
|
|
and 'utf-16-be' for little-endian and big-endian encodings, that specify one
|
|
|
particular byte ordering and don't skip the BOM.
|
|
|
|
|
|
In some areas, it is also convention to use a "BOM" at the start of UTF-8
|
|
|
encoded files; the name is misleading since UTF-8 is not byte-order dependent.
|
|
|
The mark simply announces that the file is encoded in UTF-8. For reading such
|
|
|
files, use the 'utf-8-sig' codec to automatically skip the mark if present.
|
|
|
|
|
|
|
|
|
Unicode filenames
|
|
|
-----------------
|
|
|
|
|
|
Most of the operating systems in common use today support filenames
|
|
|
that contain arbitrary Unicode characters. Usually this is
|
|
|
implemented by converting the Unicode string into some encoding that
|
|
|
varies depending on the system. Today Python is converging on using
|
|
|
UTF-8: Python on MacOS has used UTF-8 for several versions, and Python
|
|
|
3.6 switched to using UTF-8 on Windows as well. On Unix systems,
|
|
|
there will only be a filesystem encoding if you've set the ``LANG`` or
|
|
|
``LC_CTYPE`` environment variables; if you haven't, the default
|
|
|
encoding is again UTF-8.
|
|
|
|
|
|
The :func:`sys.getfilesystemencoding` function returns the encoding to use on
|
|
|
your current system, in case you want to do the encoding manually, but there's
|
|
|
not much reason to bother. When opening a file for reading or writing, you can
|
|
|
usually just provide the Unicode string as the filename, and it will be
|
|
|
automatically converted to the right encoding for you::
|
|
|
|
|
|
filename = 'filename\u4500abc'
|
|
|
with open(filename, 'w') as f:
|
|
|
f.write('blah\n')
|
|
|
|
|
|
Functions in the :mod:`os` module such as :func:`os.stat` will also accept Unicode
|
|
|
filenames.
|
|
|
|
|
|
The :func:`os.listdir` function returns filenames, which raises an issue: should it return
|
|
|
the Unicode version of filenames, or should it return bytes containing
|
|
|
the encoded versions? :func:`os.listdir` can do both, depending on whether you
|
|
|
provided the directory path as bytes or a Unicode string. If you pass a
|
|
|
Unicode string as the path, filenames will be decoded using the filesystem's
|
|
|
encoding and a list of Unicode strings will be returned, while passing a byte
|
|
|
path will return the filenames as bytes. For example,
|
|
|
assuming the default filesystem encoding is UTF-8, running the following
|
|
|
program::
|
|
|
|
|
|
fn = 'filename\u4500abc'
|
|
|
f = open(fn, 'w')
|
|
|
f.close()
|
|
|
|
|
|
import os
|
|
|
print(os.listdir(b'.'))
|
|
|
print(os.listdir('.'))
|
|
|
|
|
|
will produce the following output:
|
|
|
|
|
|
.. code-block:: shell-session
|
|
|
|
|
|
$ python listdir-test.py
|
|
|
[b'filename\xe4\x94\x80abc', ...]
|
|
|
['filename\u4500abc', ...]
|
|
|
|
|
|
The first list contains UTF-8-encoded filenames, and the second list contains
|
|
|
the Unicode versions.
|
|
|
|
|
|
Note that on most occasions, you should can just stick with using
|
|
|
Unicode with these APIs. The bytes APIs should only be used on
|
|
|
systems where undecodable file names can be present; that's
|
|
|
pretty much only Unix systems now.
|
|
|
|
|
|
|
|
|
Tips for Writing Unicode-aware Programs
|
|
|
---------------------------------------
|
|
|
|
|
|
This section provides some suggestions on writing software that deals with
|
|
|
Unicode.
|
|
|
|
|
|
The most important tip is:
|
|
|
|
|
|
Software should only work with Unicode strings internally, decoding the input
|
|
|
data as soon as possible and encoding the output only at the end.
|
|
|
|
|
|
If you attempt to write processing functions that accept both Unicode and byte
|
|
|
strings, you will find your program vulnerable to bugs wherever you combine the
|
|
|
two different kinds of strings. There is no automatic encoding or decoding: if
|
|
|
you do e.g. ``str + bytes``, a :exc:`TypeError` will be raised.
|
|
|
|
|
|
When using data coming from a web browser or some other untrusted source, a
|
|
|
common technique is to check for illegal characters in a string before using the
|
|
|
string in a generated command line or storing it in a database. If you're doing
|
|
|
this, be careful to check the decoded string, not the encoded bytes data;
|
|
|
some encodings may have interesting properties, such as not being bijective
|
|
|
or not being fully ASCII-compatible. This is especially true if the input
|
|
|
data also specifies the encoding, since the attacker can then choose a
|
|
|
clever way to hide malicious text in the encoded bytestream.
|
|
|
|
|
|
|
|
|
Converting Between File Encodings
|
|
|
'''''''''''''''''''''''''''''''''
|
|
|
|
|
|
The :class:`~codecs.StreamRecoder` class can transparently convert between
|
|
|
encodings, taking a stream that returns data in encoding #1
|
|
|
and behaving like a stream returning data in encoding #2.
|
|
|
|
|
|
For example, if you have an input file *f* that's in Latin-1, you
|
|
|
can wrap it with a :class:`~codecs.StreamRecoder` to return bytes encoded in
|
|
|
UTF-8::
|
|
|
|
|
|
new_f = codecs.StreamRecoder(f,
|
|
|
# en/decoder: used by read() to encode its results and
|
|
|
# by write() to decode its input.
|
|
|
codecs.getencoder('utf-8'), codecs.getdecoder('utf-8'),
|
|
|
|
|
|
# reader/writer: used to read and write to the stream.
|
|
|
codecs.getreader('latin-1'), codecs.getwriter('latin-1') )
|
|
|
|
|
|
|
|
|
Files in an Unknown Encoding
|
|
|
''''''''''''''''''''''''''''
|
|
|
|
|
|
What can you do if you need to make a change to a file, but don't know
|
|
|
the file's encoding? If you know the encoding is ASCII-compatible and
|
|
|
only want to examine or modify the ASCII parts, you can open the file
|
|
|
with the ``surrogateescape`` error handler::
|
|
|
|
|
|
with open(fname, 'r', encoding="ascii", errors="surrogateescape") as f:
|
|
|
data = f.read()
|
|
|
|
|
|
# make changes to the string 'data'
|
|
|
|
|
|
with open(fname + '.new', 'w',
|
|
|
encoding="ascii", errors="surrogateescape") as f:
|
|
|
f.write(data)
|
|
|
|
|
|
The ``surrogateescape`` error handler will decode any non-ASCII bytes
|
|
|
as code points in a special range running from U+DC80 to
|
|
|
U+DCFF. These code points will then turn back into the
|
|
|
same bytes when the ``surrogateescape`` error handler is used to
|
|
|
encode the data and write it back out.
|
|
|
|
|
|
|
|
|
References
|
|
|
----------
|
|
|
|
|
|
One section of `Mastering Python 3 Input/Output
|
|
|
<http://pyvideo.org/video/289/pycon-2010--mastering-python-3-i-o>`_,
|
|
|
a PyCon 2010 talk by David Beazley, discusses text processing and binary data handling.
|
|
|
|
|
|
The `PDF slides for Marc-André Lemburg's presentation "Writing Unicode-aware
|
|
|
Applications in Python"
|
|
|
<https://downloads.egenix.com/python/LSM2005-Developing-Unicode-aware-applications-in-Python.pdf>`_
|
|
|
discuss questions of character encodings as well as how to internationalize
|
|
|
and localize an application. These slides cover Python 2.x only.
|
|
|
|
|
|
`The Guts of Unicode in Python
|
|
|
<http://pyvideo.org/video/1768/the-guts-of-unicode-in-python>`_
|
|
|
is a PyCon 2013 talk by Benjamin Peterson that discusses the internal Unicode
|
|
|
representation in Python 3.3.
|
|
|
|
|
|
|
|
|
Acknowledgements
|
|
|
================
|
|
|
|
|
|
The initial draft of this document was written by Andrew Kuchling.
|
|
|
It has since been revised further by Alexander Belopolsky, Georg Brandl,
|
|
|
Andrew Kuchling, and Ezio Melotti.
|
|
|
|
|
|
Thanks to the following people who have noted errors or offered
|
|
|
suggestions on this article: Éric Araujo, Nicholas Bastin, Nick
|
|
|
Coghlan, Marius Gedminas, Kent Johnson, Ken Krugler, Marc-André
|
|
|
Lemburg, Martin von Löwis, Terry J. Reedy, Serhiy Storchaka,
|
|
|
Eryk Sun, Chad Whitacre, Graham Wideman.
|