ReferenceMarch 15, 20263 min read

Binary, ASCII, and Text Encoding: How Computers Store Letters

Every letter on your screen is stored as a number. Understand how binary, ASCII, and Unicode work together to represent text in computers.

Everything Is a Number

When you type the letter A on your keyboard, your computer does not store an "A" shape somewhere in memory. It stores the number 65. When your screen needs to display that character, it looks up number 65 in a font table and draws the corresponding glyph.

This mapping between characters and numbers is called character encoding. Understanding it explains why text sometimes looks garbled when transferred between systems, why file sizes vary for the same content in different encodings, and why emoji work on some platforms and not others.

ASCII: The Original Character Map

ASCII (American Standard Code for Information Interchange) was defined in 1963. It assigns numbers 0 through 127 to 128 characters: uppercase and lowercase English letters, digits 0–9, punctuation marks, and control characters (like newline and tab).

Key ASCII values:

  • A–Z: 65–90
  • a–z: 97–122
  • 0–9: 48–57
  • Space: 32
  • Newline: 10

With 128 values, ASCII fits in 7 bits. In practice, it uses one byte (8 bits) per character, with the highest bit unused. This means ASCII text is compact: a 1,000-character document takes exactly 1,000 bytes.

The limitation is obvious — ASCII only covers English. No accented characters, no Chinese, no Arabic, no emoji.

Unicode: Every Character Ever

Unicode is the universal character encoding standard that aims to assign a unique number (called a code point) to every character in every writing system. As of 2026, Unicode defines over 154,000 characters covering 168 scripts.

The most common Unicode encoding is UTF-8, which uses 1 to 4 bytes per character. ASCII characters (0–127) use exactly 1 byte in UTF-8, making it backward-compatible with ASCII. Characters from other scripts use 2, 3, or 4 bytes.

This variable-length encoding is why a file containing only English text and a file containing Chinese text with the same number of characters will have different file sizes. Each Chinese character requires 3 bytes in UTF-8 versus 1 byte for an ASCII character.

Binary Representation

At the hardware level, everything is binary — zeros and ones. The number 65 (the letter A) in binary is 01000001. Each digit is a bit. Eight bits make a byte.

Binary-to-text conversion is the process of taking a sequence of binary digits and interpreting them as text using a character encoding. The binary string 01001000 01100101 01101100 01101100 01101111 decodes to "Hello" in ASCII — each group of 8 bits maps to one character.

This works in reverse too. Any text can be represented as a binary string by converting each character to its numeric code and then to binary.

Why This Matters in Practice

File encoding issues. When you open a file and see garbled characters like é instead of é, the file was saved in one encoding but opened in another. The bytes are the same — the interpretation is different.

Data transmission. Network protocols and APIs often work with binary data. Converting between binary and text is essential for debugging, logging, and data inspection.

Programming. String manipulation, character comparison, and sorting all depend on the numeric values behind characters. Knowing that uppercase A is 65 and lowercase a is 97 explains why case-insensitive comparison requires special handling.

How to Use the Toobits Binary to Text Converter

Type or paste binary digits and see the text equivalent instantly. Switch direction to convert text to binary. The tool handles ASCII and UTF-8 encoding, showing you exactly how each character maps to its binary representation. Everything runs in your browser — no data is sent anywhere.

Try These Tools

Related Articles