• @[email protected]
    link
    fedilink
    199 months ago

    Today we have 64-bit computers (e.g. amd64), which descended from 32-bit computers (i386), which descended from 16-bit computers (Intel 8086), which descended from 8-bit computers (Intel 8008). Bit widths in our world naturally follow powers of two.

    However, some 1960s computers used word sizes that weren’t powers of two. Both IBM and DEC, among others, made 18- and 36-bit systems. Suppose that computing had continued to follow a multiples-of-nine pattern instead of the powers-of-two pattern?


    For one thing, hexadecimal is less common. If you’re writing 9-, 18-, 36-bit values, you typically write in octal, not hex. (In our world, Unix permissions modes are written in octal; Unix originated on the PDP-7, an 18-bit system.)

    IPv4 addresses are 36 bits wide instead of 32, and you write them in octal instead of decimal. localhost is 700.0.0.1, and a typical LAN subnet mask is 777.777.777.0.

    No hexadecimal means no 0xDEADBEEF or 0xCAFEBABE jokes. However, memory or files that get overwritten with junk are said to be “525’d”, because binary 101010101... is octal 525252....


    char would be nine bits wide instead of eight. This affects the development of character sets.

    In our world, ASCII was originally a 6-bit encoding, expanded to 7-bits to support lowercase. IBM then extended it to 8-bits with code pages for different European languages, creating 8-bit PC extended ASCII. However, no single code page supports all European languages, to say nothing of non-European ones. This led to the invention of multibyte character encodings and ultimately Unicode.

    In 9-bit world, multibyte characters are adopted earlier, using the high bit to indicate an extended character. Code pages don’t get invented; mojibake never happens.


    With 36-bit time_t, the Year 2038 problem doesn’t happen; the time_t’s don’t wrap around until the year 3058!


    A 3½" high-density floppy disk stores one megabyte of 9-bit bytes.

  • Izzy
    link
    fedilink
    69 months ago

    Lengths such as inches, feet and miles they are all unrelated to one another so here is my proposal.

    Instead of the kilobyte range we would have a bet. This is the size of the extended ascii table off 255 ascii characters that are 7 bits each or 1785 bytes. The bet comes from alphabet.

    Instead of megabytes we have the img which is based off the average size of a photo at 1.44mb which is only coincidentally the same size of a floppy disk.

    Instead of a gigabyte we have the bloat which is derived from total install size of Windows XP at around 2.4gb.

    • raubarno
      link
      fedilink
      5
      edit-2
      9 months ago

      Windows 11 installation requirements in Crown English:

      • 27 bloats of free space in the hard disk drive;
      • 1⅔ bloats of RAM;
      • Processor with at least of 60 milliard beats per minute clock rate;
    • @[email protected]OP
      link
      fedilink
      59 months ago

      In addition to that we may add a Jude because a the average MP3 file of Hey Jude by the Beatles is around 8 MB.

  • 10_0
    link
    fedilink
    69 months ago

    1 byte = 827.76787ths of a French fry

  • @[email protected]
    cake
    link
    fedilink
    4
    edit-2
    9 months ago

    Storage could be measured in what is needed for various files. They would have to be of various sizes, but not linearly increasing much like inch, foot, yard.

    Launch Codes, Pledge of Allegiances, Constitutions, God Bless the U.S.As, Average individual’s Patriot Act file (Pafs for short), etc.

  • Bobby Turkalino
    link
    English
    19 months ago

    Freedom binary digits are 7 and 4 🇺🇸

  • @[email protected]
    link
    fedilink
    19 months ago

    We currently are, but not really. A gigabyte is 1024 megabytes, and a megabyte is 1024 kilobytes, etc. However macOS and hardware manufacturers use 1000 instead of 1024 to calculate storage space. So you could say Apple uses the metric version of storage and Windows uses the imperial version.

    • @[email protected]
      link
      fedilink
      59 months ago

      Technically a gigabyte is 1000 megabyte. Megabytes are 1000 kilobytes and kilobytes are 1000 bytes. Which are all proper metric units but sadly don’t make any sense. So datasystem manufactures and computer generally calculate with their proper counterparts that you mentions gibibyte mebibyte and kibibyte which sre actualy 1024 of their previous ones. Small but crucial difference.