ASCII
ASCII
The Internet Assigned Numbers Authority (IANA) prefers the name US-ASCII for this
character encoding.
UNICODE
The Unicode Standard consists of a set of code charts for visual reference, an encoding
method and set of standard character encodings, a set of reference data files, and a
number of related items, such as character properties, rules for normalization,
decomposition, collation, rendering, and bidirectional display order (for the correct
display of text containing both right-to-left scripts, such as Arabic and Hebrew, and left-
to-right scripts)
ASCII and Unicode are two character encodings. Basically, they are standards on how
to represent difference characters in binary so that they can be written, stored,
transmitted, and read in digital media. The main difference between the two is in the
way they encode the character and the number of bits that they use for each. ASCII
originally used seven bits to encode each character. This was later increased to eight
with Extended ASCII to address the apparent inadequacy of the original. In contrast,
Unicode uses a variable bit encoding program where you can choose between 32, 16,
and 8-bit encodings. Using more bits lets you use more characters at the expense of
larger files while fewer bits give you a limited choice but you save a lot of space. Using
fewer bits (i.e. UTF-8 or ASCII) would probably be best if you are encoding a large
document in English.
One of the main reasons why Unicode was the problem arose from the many non-
standard extended ASCII programs. Unless you are using the prevalent page, which is
used by Microsoft and most other software companies, then you are likely to encounter
problems with your characters appearing as boxes. Unicode virtually eliminates this
problem as all the character code points were standardized.