Discharging the Network from Its Flow Control Headaches

Packet Drops and HOL Blocking

Journal Article (2016)
Authors

Nikolaos Chrysos (Zurich Lab, Foundation for Research and Technology - Hellas (FORTH))

Y. Chen (Zurich Lab)

Christoforos Kachris (Athens Information Technology-AIT)

Manolis Katevenis (Zurich Lab, University of Crete)

Affiliation
External organisation
To reference this document use:
https://doi.org/10.1109/TNET.2014.2378012
More Info
expand_more
Publication Year
2016
Language
English
Affiliation
External organisation
Issue number
1
Volume number
24
Pages (from-to)
15-28
DOI:
https://doi.org/10.1109/TNET.2014.2378012

Abstract

Congestion control becomes indispensable in highly utilized consolidated networks running demanding applications. In this paper, proactive congestion management schemes for Clos networks are described and evaluated. The key idea is to move the congestion avoidance burden from the data fabric to a scheduling network, which isolates flows using per-flow request counters. The scheduling network comprises per-output arbiters that grant data packets after reserving space for them in the buffer memories in front of fabric outputs. Computer simulations show that this strategy eliminates head-of-line (HOL) blocking and its adversarial effects throughout the fabric, without having to drop packets. In particular, a simplified model describes this result as a synergy between proactive buffer reservations and fine-grained multipath routing. Two alternative designs are presented. The first one places all arbiters in a central control unit, is simpler, and has superior performance. The second is more scalable by distributing the arbiters over the switching elements of the Clos network and by routing the control messages to and from endpoint adapters via multiple paths. Computer simulations of the complete system demonstrate high throughput and low latency under any number of congested outputs. Weighted max-min fair allocation of fabric-output link bandwidth is also demonstrated. Furthermore, delay breakdowns show that the time that packets wait in fabric and resequencing buffers is minimized as a result of the reduced (and equalized across all fabric paths) in-fabric contention. Finally, the high throughput capability of the system is corroborated by a Markov chain analysis of output buffer credits.

No files available

Metadata only record. There are no files for this record.