Submitted by ArchariosMaster t3_127czbb in explainlikeimfive
Comments
phiwong t1_jedorky wrote
Partly tradition, partly the hardware and partly the software.
For computing devices data is moved around, operated and stored in bytes consisting of 8 bits. (although modern ones operate just as well in double bytes of 16 bits) So the byte became the default means of designating the fundamental data element. Therefore the size of "stuff" is conveniently expressed in the number of bytes.
For transmission, however, most modern communications and networks operate on a single line transmission. There are parallel communication methods but they tend to be local. But things like USB, wi-fi, internet have hardware that send data in a stream of bits. Therefore the speed of transmission is conveniently measured in bits per second
Jarix t1_jedp0fc wrote
Also there in the early days everything was new and changed a lot so people didnt care that much which one anyone used anyways
yalloc t1_jedpw6q wrote
Marketing for the most part. Bigger number sounds better for speeds so things are advertised with the bigger number. They don’t expect grandma to know the difference between bit and byte and grandma of course is gonna take the speed that’s 8 times bigger.
spikecurtis t1_jeds1fs wrote
The difference is largely historical.
Data storage devices are byte-addressable, in that the smallest unit of data you can read one write is one byte. We long ago standardized on 8-bits per byte, but in early computing there were different byte sizes.
Early data transmission devices (what we now call network hardware) measured their speed in bits per second. The byte wasn’t standardized as 8 bits yet, and different ends of the transmission might have different byte sizes.
Modern networking gear basically always sends data in some whole number of 8-bit bytes (in the context of networking they are sometimes called octets to be absolutely clear they are 8-bit units), but the bits per second terminology persists.
You wouldn’t want to be the manufacturer who unilaterally shifts to bytes per second when you peers are still marketing bits per second, for fear of people thinking your stuff is slower than it is!
satans_toast t1_jedw2yf wrote
Computer networks send data one bit at a time. It’s a state change, either electrical or with light pulses. It happens faster and faster as technology improves, but it’s still just one bit at a time. Hence, bits per second.
It doesn’t matter whether it’s a file or a digitized real-time data stream or a bunch of random characters, it’s sending one bit at a time.
Also, when you’re transmitting a file over a network, it’s not just the file that’s being sent. Overhead bits are added to the stream, whether it’s the source & destination addresses, fragmentation flags, error checking bits, and a slew of other bits used for controlling the traffic somehow.
[deleted] t1_jeed6bo wrote
[removed]
spikecurtis t1_jefl68l wrote
Some modulation schemes send multiple bits at a time. For example in phase shift keying, the phase of an oscillating wave encodes the information. If you do this with 4 phases, 90 degrees apart, then you send 2 bits at a time, 8 phases 3 bits, etc.
[deleted] t1_jedog4y wrote
[removed]