Network Latency
From Computing and Software Wiki
m |
|||
Line 18: | Line 18: | ||
==Measuring Latency== | ==Measuring Latency== | ||
- | + | Network conditions constantly fluctuate, and so measuring the latency in transmitting a packet between the same two endpoints multiple times can have differing results. Because of this, the latency of any single packet may not be a meaningful measurement. Another issue is the fact that any latency measurements exchanged between hosts will themselves be subject to delay on the network. | |
- | A simple solution to both of these problems is calculating latency using average round-trip time. | + | A simple solution to both of these problems is calculating latency using average round-trip time. Round-trip communication times can be measured from a single host, and taking the average latency over several packets provides a reasonable estimate of the delay needed for future packets to arrive. Extra steps may need to be taken, as dropped packets and temporary disconnections can skew the average latency measurement much higher. Also, the time a host takes to process packets is typically not included when calculating a round-trip time, but users and applications waiting for data will experience this delay just the same. |
- | + | Using [[Ping]] is a simple way to find your latency relative to some host address. | |
- | ==Issues | + | |
+ | |||
+ | ==Issues== | ||
===Quantum Computing=== | ===Quantum Computing=== | ||
===Real-Time Gaming=== | ===Real-Time Gaming=== |
Revision as of 23:59, 12 April 2009
As an Engineering term, latency refers to the span of time taken from when some action is initiated to when it actually takes effect.
In the context of packet-switching networks, latency can refer to any of the following:
- The time from when a packet is sent to when that packet reaches its destination
- The round-trip time of a packet
- The perceived delay in communication between hosts
In online multiplayer games, the round-trip time of a packet is commonly known as ping.
Contents |
Causes
Traffic Congestion
Any packets which are prevented from reaching their destination for any period of time will result in an increase in latency. Heavy network traffic can therefore increase latency, as bandwidth limitations and routing issues contribute to the time that a message spends in transit.
Application performance
Since every packet must at some point be created and sent by an application, any time taken in processing the information necessary to create or read a packet will cause additional latency. The perception of latency is also created when communication is delayed due to packets being dropped (from events such as packet collisions), because the user will only see the time from when the request was sent to when the message was successfully received.
Distance
Communication is naturally limited by the speed of light. Therefore the round-trip time of packets is unavoidably linked to the distance over which the packets are being sent, subject to the laws of Relativity. This is particularly an issue in the field of Space Exploration, where the round-trip time of communication is commonly measured in minutes or hours. Because of this, rovers must be programmed with some level of artificial intelligence so that moment-to-moment decisions can be made autonomously.
Measuring Latency
Network conditions constantly fluctuate, and so measuring the latency in transmitting a packet between the same two endpoints multiple times can have differing results. Because of this, the latency of any single packet may not be a meaningful measurement. Another issue is the fact that any latency measurements exchanged between hosts will themselves be subject to delay on the network.
A simple solution to both of these problems is calculating latency using average round-trip time. Round-trip communication times can be measured from a single host, and taking the average latency over several packets provides a reasonable estimate of the delay needed for future packets to arrive. Extra steps may need to be taken, as dropped packets and temporary disconnections can skew the average latency measurement much higher. Also, the time a host takes to process packets is typically not included when calculating a round-trip time, but users and applications waiting for data will experience this delay just the same.
Using Ping is a simple way to find your latency relative to some host address.
Issues
Quantum Computing
Real-Time Gaming
Real-time online multiplayer games suffer from network latency, causing a delay between the players' input and the game's response. Modern online games are typically designed to use a client-server networking model, though peer-to-peer implementations are also possible. Latency means that the players' instructions to the game do not reach the server instantaneously, and the server's description of the current gamestate is slightly outdated by the time it reaches the players. During the infancy of online gaming the only workaround was for players to manually compensate for the delay by performing actions earlier than actually desired.
Modern game engines such as Valve Software's Source Engine [x] implement a number of lag compensation techniques. To eliminate the perceived delay on a client, character animations and other responses which do not affect gamestate are played immediately. Actions which do affect the gamestate (such as a player walking forward) have their outcomes predicted on the client's end as if they have already happened, causing the player to immediately see themselves walking forward. The client is therefore seeing a real-time approximation of the server's current gamestate at any given moment. If any discrepancy is found once the server finally responds to the client, the client adjusts itself to match the server's new state in order to avoid desynchronization (this however can cause a sudden "jerking" effect known as "rubber-banding"). This method is known as client-side prediction.
In order to mitigate the discrepancies between the server and clients, latency can be factored into critical gamestate calculations. For example, suppose the server wants to check if player A shoots player B. Since the clients' predictions are not perfectly accurate, clients can never be sure of the server's exact gamestate, and a naive collision check would make it impossible to hit player B consistently. Instead, the server looks back in time (according to player A's round-trip latency) and checks to see if player A's hit would be successful according to that state. Though some mathematical imprecision may result, this much more closely aligns the events of the game with what each player is seeing (shown on the right).
Compensating for latency
When a noticeable latency is unavoidable, some method of hiding the latency from the user needs to be employed. One strategy for this is prefetching
Suppose that a piece of real-time software needs to communicate with a server acting as an authority on the application's state, and that the network has a round-trip latency of 100 milliseconds.
Prefetching
Some web browsers (such as Mozilla Firefox [x])
Prediction
Interpolation
Programs which
<ref>Template:Cite journal</ref>
References
Template loop detected: Template:Reflist http://en.wikipedia.org/wiki/Link_prefetching http://developer.valvesoftware.com/wiki/Source_Multiplayer_Networking
See Also
http://www.michaelnygard.com/blog/2007/11/architecting_for_latency.html