• houseofleft@slrpnk.net
    link
    fedilink
    English
    arrow-up
    3
    ·
    3 days ago

    Ascii needs seven bits, but is almost always encoded as bytes, so every ascii letter has a throwaway bit.

      • anton@lemmy.blahaj.zone
        link
        fedilink
        arrow-up
        1
        ·
        17 hours ago

        That boolean can indicate if it’s a fancy character, that way all ASCII characters are themselves but if the boolean is set it’s something else. We could take the other symbol from a page of codes to fit the users language.
        Or we could let true mean that the character is larger, allowing us to transform all of unicode to a format consisting of 8 bits parts.

    • FuckBigTech347@lemmygrad.ml
      link
      fedilink
      arrow-up
      1
      ·
      3 days ago

      Some old software does use 8-Bit ASCII for special/locale specific characters. Also there is this Unicode hack where the last bit is used to determine if the byte is part of a multi-byte sequence.