• @SigHunter
      link
      English
      7
      edit-2
      6 months ago

      Back in the day, the rule was mbit (megabit) for data in transfer (network speed) and MB (megabyte) for data at rest, like on HDDs

        • @bitwaba@lemmy.world
          link
          fedilink
          English
          16 months ago

          The real answer?

          Data is transmitted in packets. Each packet has a packet header, and a packet payload. The total data transmitted is the header + payload.

          If you’re transmitting smaller packet sizes, it means your header is a larger percentage of the total packet size.

          Measuring in megabits is the ISP telling you “look, your connection is good for X amount of data. How you choose to use that data is up to you. If you want more of it going to your packet headers instead of your payload, fine.” A bit is a bit is a bit to your ISP.

        • rhys the great
          link
          fedilink
          16 months ago

          @Moneo @SigHunter Networking came to be when there were lots of different implementations of a ‘byte’. The PDP-10 was prevalent at the time the internet was being developed for example, which supported variable byte lengths of up to 36-bits per byte.

          Network protocols had to support every device regardless of its byte size, so protocol specifications settled on bits as the lowest common unit size, while referring to 8-bit fields as ‘octets’ before 8-bit became the de facto standard byte length.

    • @lud@lemm.ee
      link
      fedilink
      English
      36 months ago

      The best format imo is MB/s and Mbit/s

      It avoids all confusion.