Pages

Search This Blog

Thursday, November 17, 2016

THE 4 FACTORS WHICH SLOWDOWN DATA TRASFER AND HOW TO IDENTIFY THEM

The 4 factors which slowdown data trasfer and how to identify them
Every IT Operations team faces complaints indicating that the network is slowing down or poorly delivering applications. Their very first focus is on checking the network performance (latency, packet loss, etc.). 
Nevertheless, the network is not the sole driver of data transfer speed and of the end user experience.
Many other factors directly impact how fast application queries and responses will flow through the network. If one wants to troubleshoot performance degradation, this check list of the factors which can badly impact the transfer speed, will come in very handy. 
Let’s start with what may be network-led:
1.    NETWORK LATENCY
Network latency refers to the time needed to send a packet from the source to the destination. This time varies depending:
  • on the physical distance,
  • the number of network devices which have to be crossed (also referred to as number of hops)
  • and to a lesser extent, to the performance of each of the devices.
The relationship between latency and transfer rates depends on the protocol which carries the data. To keep the focus on the most common ones: for a UDP flow, latency may not have an impact. As for TCP applications, generally the most commonly used protocol, it will have a drastic impact.
2.    NETWORK CONGESTION
Network congestion refers to the saturation of a path used by packets to flow between the source and the destination. The element on the path can be either an active device (e.g. Router or Switch) or a physical link (e.g. cable).
When the maximum capacity of the element is reached, the packet cannot be transferred in a timely manner as it is either put in a queue (e.g. in a router) or dropped if a no queue system is available to retain them. It may even become impossible to setup new sessions.
The consequence will then vary, depending on the level of delay generated by the congestion:
  • Packets are delayed for a short period of time.
    • The latency will increase.
    • Some re transmissions will occur (for TCP flows) as the acknowledgment packets are not received fast enough by the sender.
    • Duplicate acknowledgment packets will also be received.
  • Packets are lost or dropped (packet loss).
    • The re transmission increases significantly: as packets are not acknowledged, they will then be massively resent.

  • Disconnections: sessions are dropped as too many packets are lost:
    • You might see TTL exceeded, session time-outs.
    • TCP sessions not being terminated properly. 
3.    INFRASTRUCTURE PARAMETER (QOS, FILTERING, ROUTING)
Although the overall network path is free of any congestion (lack of bandwidth or system resources), some devices apply policies:
  • Prioritization: some traffic is either more strategic (critical applications) or more performance sensitive (real time applications, VoIP, video conferencing) and gets allocated a higher priority than the rest of the applications using a given network path. In case the maximum capacity on the network path is reached, lower priority flows will start experiencing re transmission, packet loss or disconnection depending on how long and important the congestion is.
  • Filtering / encryption: there may be many kinds of filtering’s in place to scan viruses, to prevent users to reach non-recommended sites, to prevent threats on web servers, etc. Filtering has an impact on data transfer: depending on how much processing time it requires. This might have an impact on the latency between the client and the server. If the processing time becomes excessive, it can generate re transmission and packet loss.
  • Routing / load balancing: some devices distribute the load across a group of servers / devices or route the traffic to the most adequate path from a performance and / or an economic standpoint. The devices may also be overloaded or misconfigured which could lead to re transmission, packet loss or disconnection issues.
While troubleshooting slow transfer rates, it is important to list the devices on the path between clients and servers. You can then identify at which point in time and for which flow: re transmissions, duplicate acknowledgments, packet loss, TTL expired and session time-out or incomplete TCP start can be observed.
4.    CLIENT OR SERVER HEALTH
It is probably the last item you will consider if you are focused on network performance. But these systems also have limited resources which can lead to a congestion situation and slowdown the data transfer rates.
If a server lacks hardware resources, such as RAM, CPU, I/O, it will process user queries slower.
At a given moment, a client or a server reaching a congestion point will slow down the transfer using standard TCP mechanisms.
Here is how you can identify that situation:
  • 0 window events: one of the parties is asking to reduce the throughput. You can interpret this indicator as a sign of lack of resources and investigates on the host to identify which resource is not sufficiently available.
  • RST – Resets events: one of the parties disconnects the session abruptly. Keep in mind that some applications may use RST as a standard way to terminate a session, even if it is not a best practice! 
 It is easy to gather this information from your network traffic and to quickly pinpoint where your data transfer slowdowns are coming from. 





No comments:

Post a Comment