• SigHunter@lemmy.kde.social
    link
    fedilink
    English
    arrow-up
    7
    ·
    edit-2
    6 months ago

    Back in the day, the rule was mbit (megabit) for data in transfer (network speed) and MB (megabyte) for data at rest, like on HDDs

      • rhys the great@mastodon.rhys.wtf
        link
        fedilink
        arrow-up
        1
        ·
        6 months ago

        @Moneo @SigHunter Networking came to be when there were lots of different implementations of a ‘byte’. The PDP-10 was prevalent at the time the internet was being developed for example, which supported variable byte lengths of up to 36-bits per byte.

        Network protocols had to support every device regardless of its byte size, so protocol specifications settled on bits as the lowest common unit size, while referring to 8-bit fields as ‘octets’ before 8-bit became the de facto standard byte length.

      • bitwaba@lemmy.world
        link
        fedilink
        English
        arrow-up
        1
        ·
        6 months ago

        The real answer?

        Data is transmitted in packets. Each packet has a packet header, and a packet payload. The total data transmitted is the header + payload.

        If you’re transmitting smaller packet sizes, it means your header is a larger percentage of the total packet size.

        Measuring in megabits is the ISP telling you “look, your connection is good for X amount of data. How you choose to use that data is up to you. If you want more of it going to your packet headers instead of your payload, fine.” A bit is a bit is a bit to your ISP.