you are viewing a single comment's thread.

view the rest of the comments →

[–][deleted] 5 insightful - 1 fun5 insightful - 0 fun6 insightful - 1 fun -  (0 children)

No, English is still the master race for text encoding. The default 1 byte representation of a character is ASCII in major programming languages. Since other languages can't fit all of their symbols in one byte they have to use multiple bytes for a character. Before the proliferation of the internet, other countries had competing character encoding standards for representing text. This basically made text files from other countries unreadable because you would need to know ahead of time what country the text file came from and which character encoding standard it used. To solve this problem of reading and displaying text they formed a committee and worked with other inferior languages to figure out what range of values would represent the characters of their written language. There is still debate on which unicode standard should be standard for the internet because Utf-8 is anglo supremacist because it is backwards compatible with ASCII. Asians like utf-16 because it is easier to parse text if you assume most of your letters are going to be 2 bytes. A turbo sperg was annoyed that sometimes letters would be a variable length so they created utf-32 which is a fixed length character encoding of 4 bytes. Since it would make the length of text files up to four times as large for Anglos nobody used it.