In computing, bandwidth is the maximum rate of data transfer across a given path. Bandwidth may be characterized as network bandwidth,[1] data bandwidth,[2] or digital bandwidth.[3] [4]
This definition of bandwidth is in contrast to the field of signal processing, wireless communications, modem data transmission, digital communications, and electronics, in which bandwidth is used to refer to analog signal bandwidth measured in hertz, meaning the frequency range between lowest and highest attainable frequency while meeting a well-defined impairment level in signal power. The actual bit rate that can be achieved depends not only on the signal bandwidth but also on the noise on the channel.
The term bandwidth sometimes defines the net bit rate peak bit rate, information rate, or physical layer useful bit rate, channel capacity, or the maximum throughput of a logical or physical communication path in a digital communication system. For example, bandwidth tests measure the maximum throughput of a computer network. The maximum rate that can be sustained on a link is limited by the Shannon–Hartley channel capacity for these communication systems, which is dependent on the bandwidth in hertz and the noise on the channel.
The consumed bandwidth in bit/s, corresponds to achieved throughput or goodput, i.e., the average rate of successful data transfer through a communication path. The consumed bandwidth can be affected by technologies such as bandwidth shaping, bandwidth management, bandwidth throttling, bandwidth cap, bandwidth allocation (for example bandwidth allocation protocol and dynamic bandwidth allocation), etc. A bit stream's bandwidth is proportional to the average consumed signal bandwidth in hertz (the average spectral bandwidth of the analog signal representing the bit stream) during a studied time interval.
Channel bandwidth may be confused with useful data throughput (or goodput). For example, a channel with x bit/s may not necessarily transmit data at x rate, since protocols, encryption, and other factors can add appreciable overhead. For instance, much internet traffic uses the transmission control protocol (TCP), which requires a three-way handshake for each transaction. Although in many modern implementations the protocol is efficient, it does add significant overhead compared to simpler protocols. Also, data packets may be lost, which further reduces the useful data throughput. In general, for any effective digital communication, a framing protocol is needed; overhead and effective throughput depends on implementation. Useful throughput is less than or equal to the actual channel capacity minus implementation overhead.
The asymptotic bandwidth (formally asymptotic throughput) for a network is the measure of maximum throughput for a greedy source, for example when the message size (the number of packets per second from a source) approaches close to the maximum amount.[5]
Asymptotic bandwidths are usually estimated by sending a number of very large messages through the network, measuring the end-to-end throughput. As with other bandwidths, the asymptotic bandwidth is measured in multiples of bits per seconds. Since bandwidth spikes can skew the measurement, carriers often use the 95th percentile method. This method continuously measures bandwidth usage and then removes the top 5 percent.[6]
Digital bandwidth may also refer to: multimedia bit rate or average bitrate after multimedia data compression (source coding), defined as the total amount of data divided by the playback time.
Due to the impractically high bandwidth requirements of uncompressed digital media, the required multimedia bandwidth can be significantly reduced with data compression.[7] The most widely used data compression technique for media bandwidth reduction is the discrete cosine transform (DCT), which was first proposed by Nasir Ahmed in the early 1970s.[8] DCT compression significantly reduces the amount of memory and bandwidth required for digital signals, capable of achieving a data compression ratio of up to 100:1 compared to uncompressed media.[9]
In Web hosting service, the term bandwidth is often incorrectly used to describe the amount of data transferred to or from the website or server within a prescribed period of time, for example bandwidth consumption accumulated over a month measured in gigabytes per month.[10] The more accurate phrase used for this meaning of a maximum amount of data transfer each month or given period is monthly data transfer.
A similar situation can occur for end-user Internet service providers as well, especially where network capacity is limited (for example in areas with underdeveloped internet connectivity and on wireless networks).
This table shows the maximum bandwidth (the physical layer net bit rate) of common Internet access technologies. For more detailed lists see
Bit rate | Connection type | |
---|---|---|
56 kbit/s | Dialup | |
1.5 Mbit/s | ADSL Lite | |
1.544 Mbit/s | T1/DS1 | |
2.048 Mbit/s | E1 / E-carrier | |
4 Mbit/s | ADSL1 | |
10 Mbit/s | Ethernet | |
11 Mbit/s | Wireless 802.11b | |
24 Mbit/s | ADSL2+ | |
44.736 Mbit/s | ||
54 Mbit/s | Wireless 802.11g | |
100 Mbit/s | Fast Ethernet | |
155 Mbit/s | ||
600 Mbit/s | Wireless 802.11n | |
622 Mbit/s | OC12 | |
1 Gbit/s | Gigabit Ethernet | |
1.3 Gbit/s | Wireless 802.11ac | |
2.5 Gbit/s | OC48 | |
5 Gbit/s | SuperSpeed USB | |
7 Gbit/s | Wireless 802.11ad | |
9.6 Gbit/s | ||
10 Gbit/s | 10 Gigabit Ethernet, SuperSpeed USB 10 Gbit/s | |
20 Gbit/s | SuperSpeed USB 20 Gbit/s | |
40 Gbit/s | Thunderbolt 3 | |
100 Gbit/s | 100 Gigabit Ethernet |
See main article: Edholm's law.
Edholm's law, proposed by and named after Phil Edholm in 2004,[11] holds that the bandwidth of telecommunication networks double every 18 months, which has proven to be true since the 1970s.[11] [12] The trend is evident in the cases of Internet,[11] cellular (mobile), wireless LAN and wireless personal area networks.
The MOSFET (metal–oxide–semiconductor field-effect transistor) is the most important factor enabling the rapid increase in bandwidth.[13] The MOSFET (MOS transistor) was invented by Mohamed M. Atalla and Dawon Kahng at Bell Labs in 1959,[14] [15] [16] and went on to become the basic building block of modern telecommunications technology.[17] [18] Continuous MOSFET scaling, along with various advances in MOS technology, has enabled both Moore's law (transistor counts in integrated circuit chips doubling every two years) and Edholm's law (communication bandwidth doubling every 18 months).[13]