ne
supports UTF-8. It can use UTF-8 for its
input/output, and it can also interpret one or more documents as containing
UTF-8 encoded text, acting accordingly. Note that the document content is
actual UTF-8 text—ne
does not use wide characters. As a
positive side-effect, ne
can support fully the ISO-10646
standard, but nonetheless non-UTF-8 texts occupy exactly one byte per
character.
More precisely, any piece of text in ne
is classified as
US-ASCII, 8-bit or UTF-8. A US-ASCII text contains only US-ASCII
characters. An 8-bit text sports a one-to-one correspondence between
characters and bytes, whereas an UTF-8 text is interpreted in UTF-8. Of
course, this raises a difficult question: when should a document be
classified as UTF-8?
Character encodings are a mess. There is nothing we can do to change this fact, as character encodings are metadata that modify data semantics. The same file may represent different texts of different lengths when interpreted with different encodings. Thus, there is no safe way of guessing the encoding of a file.
ne
stays on the safe side: it will never try to convert a file
from an encoding to another one. It can, however, interpret data
contained in a document depending on an encoding: in other words,
encodings are truly treated as metadata. You can switch off UTF-8
at any time, and see the same document as a standard 8-bit file.
Moreover, ne
uses a lazy approach to the problem: first of
all, unless the UTF-8 automatic detection flag is set
(see UTF8Auto), no attempt is ever made to consider a file as UTF-8
encoded. Every file, clip, command line, etc., is firstly scanned for
non-US-ASCII characters: if it is entirely made of US-ASCII characters,
it is classified as US-ASCII. An US-ASCII piece of text is compatible
with anything else—it may be pasted in any document, or, if it is a
document, it may accept any form of text. Documents classified as US-ASCII
are distinguished by an ‘A’ on the status bar.
As soon as a user action forces a choice of encoding (e.g., an accented
character is typed, or an UTF-8-encoded clip is pasted), ne
fixes
the mode to 8-bit or UTF-8 (when there is a choice, this depends on
the value of the UTF8Auto flag). Of course, in some cases this may
be impossible, and in that case an error will be reported.
All this happens behind the scenes, and it is designed so that in 99% of
the cases there is no need to think of encodings. In any case, should
ne
’s behaviour not match your needs, you can always change at run
time the level of UTF-8 support.