The 4 factors which slowdown data trasfer and how to identify them
Every IT Operations team faces complaints indicating that the network is slowing down or poorly delivering applications. Their very first focus is on checking the network performance (latency, packet loss, etc.).
Nevertheless, the network is not the sole driver of data transfer speed and of the end user experience.
Many other factors directly impact how fast application queries and responses will flow through the network. If one wants to troubleshoot performance degradation, this check list of the factors which can badly impact the transfer speed, will come in very handy.
Let’s start with what may be network-led:
1. NETWORK LATENCY
Network latency refers to the time needed to send a packet from the source to the destination. This time varies depending:
- on the physical
- the number of
network devices which have to be crossed (also referred to as
number of hops)
- and to a lesser extent,
to the performance of each of the devices.
The relationship between latency and transfer rates depends on the protocol which carries the data. To keep the focus on the most common ones: for a UDP flow, latency may not have an impact. As for TCP applications, generally the most commonly used protocol, it will have a drastic impact.
2. NETWORK CONGESTION
Network congestion refers to the saturation of a path used by packets to flow between the source and the destination. The element on the path can be either an active device (e.g. Router or Switch) or a physical link (e.g. cable).
When the maximum capacity of the element is reached, the packet cannot be transferred in a timely manner as it is either put in a queue (e.g. in a router) or dropped if a no queue system is available to retain them. It may even become impossible to setup new sessions.
The consequence will then vary, depending on the level of delay generated by the congestion:
- Packets are delayed for a
short period of time.
- The latency will
- Some re transmissions will occur (for TCP flows) as the acknowledgment packets are not received
fast enough by the sender.
- Duplicate acknowledgment
packets will also be received.
- Packets are lost or
dropped (packet loss).
- The re transmission increases significantly: as packets are not acknowledged, they will then
be massively resent.
- Disconnections: sessions
are dropped as too many packets are lost:
- You might see TTL
exceeded, session time-outs.
- TCP sessions not being
3. INFRASTRUCTURE PARAMETER (QOS, FILTERING, ROUTING)
Although the overall network path is free of any congestion (lack of bandwidth or system resources), some devices apply policies:
- Prioritization: some traffic is either more strategic (critical
applications) or more performance sensitive (real time applications, VoIP,
video conferencing) and gets allocated a higher priority than the rest of
the applications using a given network path. In case the maximum capacity
on the network path is reached, lower priority flows will start
experiencing re transmission, packet loss or disconnection depending on how
long and important the congestion is.
- Filtering / encryption: there may be many kinds of filtering’s in place to
scan viruses, to prevent users to reach non-recommended sites, to prevent
threats on web servers, etc. Filtering has an impact on data transfer:
depending on how much processing time it requires. This might have an
impact on the latency between the client and the server. If the processing
time becomes excessive, it can generate re transmission and packet loss.
- Routing / load balancing: some devices distribute the load across a group of
servers / devices or route the traffic to the most adequate path from a
performance and / or an economic standpoint. The devices may also be
overloaded or misconfigured which could lead to re transmission, packet
loss or disconnection issues.
While troubleshooting slow transfer rates, it is important to list the devices on the path between clients and servers. You can then identify at which point in time and for which flow: re transmissions, duplicate acknowledgments, packet loss, TTL expired and session time-out or incomplete TCP start can be observed.
4. CLIENT OR SERVER HEALTH
It is probably the last item you will consider if you are focused on network performance. But these systems also have limited resources which can lead to a congestion situation and slowdown the data transfer rates.
If a server lacks hardware resources, such as RAM, CPU, I/O, it will process user queries slower.
At a given moment, a client or a server reaching a congestion point will slow down the transfer using standard TCP mechanisms.
Here is how you can identify that situation:
- 0 window events: one of
the parties is asking to reduce the throughput. You can interpret this
indicator as a sign of lack of resources and investigates on the host to
identify which resource is not sufficiently available.
- RST – Resets events: one
of the parties disconnects the session abruptly. Keep in mind that some
applications may use RST as a standard way to terminate a session, even if
it is not a best practice!
It is easy to gather this information from your network traffic and to quickly pinpoint where your data transfer slowdowns are coming from.