data center tcp (dctcp). introduction in recent years, data centers have trans- formed computing...

71
Data Center TCP (DCTCP)

Upload: kory-spencer

Post on 21-Dec-2015

225 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Data Center TCP (DCTCP). Introduction In recent years, data centers have trans- formed computing With large scale consolidation of enterprise IT into

Data Center TCP (DCTCP)

Page 2: Data Center TCP (DCTCP). Introduction In recent years, data centers have trans- formed computing With large scale consolidation of enterprise IT into

Introduction

• In recent years, data centers have trans-formed computing

• With large scale consolidation of enterprise IT into data center hubs

• With the emergence of cloud computing service providers like Amazon, Microsoft and Google

Page 3: Data Center TCP (DCTCP). Introduction In recent years, data centers have trans- formed computing With large scale consolidation of enterprise IT into

Introduction

• Build highly available, highly performant computing and storage infrastructure using low cost, commodity components.

• Authors focus on soft real-time applications, supporting web search, retail, advertising, and recommendation systems that have driven much data center construction.

• These applications generate a diverse mix of short and long flows

Page 4: Data Center TCP (DCTCP). Introduction In recent years, data centers have trans- formed computing With large scale consolidation of enterprise IT into

Introduction

• This traffic pattern requires three things from the data center network: low latency for short flows, high burst tolerance, and high utilization for long flows.

• The first two requirements stem from the Partition/Aggregate workflow pattern that many of these applications use

• The third requirement stems from the need to continuously update internal data structures of these applications

Page 5: Data Center TCP (DCTCP). Introduction In recent years, data centers have trans- formed computing With large scale consolidation of enterprise IT into

Introduction

• In this paper, authors make two major contributions. – measure and analyze production traffic, extracting

application patterns and needs, and identify impairments that hurt performance

– propose Data Center TCP (DCTCP), which addresses these impairments to meet the needs of applications

Page 6: Data Center TCP (DCTCP). Introduction In recent years, data centers have trans- formed computing With large scale consolidation of enterprise IT into

Introduction

• The measurements reveal that 99.91% of traffic in the data center is TCP traffic.

• The traffic consists of– query traffic (2KB to 20KB in size) – delay sensitive short messages (100KB to 1MB) – throughput sensitive long flows (1MB to 100MB)

• Query and delay-sensitive short messages experience long latencies due to long flows consuming some or all of the available buffer in the switches.

Page 7: Data Center TCP (DCTCP). Introduction In recent years, data centers have trans- formed computing With large scale consolidation of enterprise IT into

Introduction

• DCTCP combines Explicit Congestion Notification (ECN) with a novel control scheme at the sources.

• DCTCP operate with very low buffer occupancies while still achieving high throughput

Page 8: Data Center TCP (DCTCP). Introduction In recent years, data centers have trans- formed computing With large scale consolidation of enterprise IT into

Communication in Data Center

• Authors first describe a common application structure in order to understand the challenges facing data center transport protocols

• Authors measure the synchronized and bursty traffic patterns that result from these appli-cation structure

• Identify three performance impairments these patterns cause.

Page 9: Data Center TCP (DCTCP). Introduction In recent years, data centers have trans- formed computing With large scale consolidation of enterprise IT into

Partition/Aggregate

Page 10: Data Center TCP (DCTCP). Introduction In recent years, data centers have trans- formed computing With large scale consolidation of enterprise IT into

Partition/Aggregate

• Requests from higher layers of the application are broken into pieces and farmed out to workers in lower layers.

• The responses of these workers are aggregated to produce a result.

• For soft-real-time applications, latency is the key metric.• After subtracting typical Internet and rendering delays,

the “backend” part of the application is typically allocated between 230-300ms. – This limit is called an all-up SLA.

Page 11: Data Center TCP (DCTCP). Introduction In recent years, data centers have trans- formed computing With large scale consolidation of enterprise IT into

Partition/Aggregate

• Many applications have a multi-layer partition/aggregate pattern workflow, with lags at one layer delaying the initiation of others.

• Answering a request may require iteratively invoking the pattern, with an aggregator making serial requests to the workers below it to prepare a response

Page 12: Data Center TCP (DCTCP). Introduction In recent years, data centers have trans- formed computing With large scale consolidation of enterprise IT into

Partition/Aggregate

• To prevent the all-up SLA from being violated– worker nodes are assigned tight deadlines, usually

on the order of 10-100ms. – When a node misses its deadline, the computation

continues without that response, lowering the quality of the result.

• With such tight deadlines, network delays within the data center play a significant role in application design.

Page 13: Data Center TCP (DCTCP). Introduction In recent years, data centers have trans- formed computing With large scale consolidation of enterprise IT into

Workload Characterization

• Authors instrumented a total of over 6000 servers in over 150 racks.

• The three clusters support soft real-time query traffic, integrated with – urgent short message traffic that coordinates the

activities in the cluster – continuous background traffic that ingests and or-

ganizes the massive data needed to sustain the quality of the query responses.

Page 14: Data Center TCP (DCTCP). Introduction In recent years, data centers have trans- formed computing With large scale consolidation of enterprise IT into

Query Traffic

• The query traffic consists of very short, latency-critical flows – A high-level aggregator (HLA) partitions queries to

a large number of mid-level aggregators (MLAs) that in turn partition each query over the 43 other servers in the same rack as the MLA.

– Servers act as both MLAs and workers, so each server will be acting as an aggregator for some queries at the same time it is acting as a worker for other queries.

Page 15: Data Center TCP (DCTCP). Introduction In recent years, data centers have trans- formed computing With large scale consolidation of enterprise IT into

Query Traffic

• Figure shows the CDF of time between arrivals of queries at mid-level aggregators.

Page 16: Data Center TCP (DCTCP). Introduction In recent years, data centers have trans- formed computing With large scale consolidation of enterprise IT into

Background Traffic

• Background traffic consists of both large and small flows – most background flows are small, but most of the

bytes in background traffic are part of large flows

Page 17: Data Center TCP (DCTCP). Introduction In recent years, data centers have trans- formed computing With large scale consolidation of enterprise IT into

Background Traffic

• Figure shows the time between arrival of new background flows

Page 18: Data Center TCP (DCTCP). Introduction In recent years, data centers have trans- formed computing With large scale consolidation of enterprise IT into

Background Traffic

• The variance in interarrival time is very high, with a very heavy tail.

• Embedded spikes occur, for example the 0ms inter-arrivals that explain the CDF hugging the y-axis up to the 50th percentile.

• Relatively large numbers of outgoing flows occur periodically, resulting from workers periodically polling a number of peers looking for updated files.

Page 19: Data Center TCP (DCTCP). Introduction In recent years, data centers have trans- formed computing With large scale consolidation of enterprise IT into

Flow Concurrency and Size

• When all flows are considered, the median number of concurrent flows is 36– results from the breadth of the Partition/Aggregate traffic pattern in

which each server talks to 43 other servers • The 99.99th percentile is over 1,600

– one server with a median of 1,200 connections.

Page 20: Data Center TCP (DCTCP). Introduction In recent years, data centers have trans- formed computing With large scale consolidation of enterprise IT into

Flow Concurrency and Size

• Only large flows (> 1MB) are considered, the degree of statistical multiplexing is very low – the median number of con- current large flows is 1, and the

75th percentile is 2 – Yet they last several RTTs, and can consume significant buffer

space by causing queue buildup

Page 21: Data Center TCP (DCTCP). Introduction In recent years, data centers have trans- formed computing With large scale consolidation of enterprise IT into

Understanding Performance Impairments – Switches

• Shared memory switches that aim to exploit statistical multiplexing gain through use of logically common packet buffers available to all switch ports.

• If the interface has hit its maximum memory allocation or the shared pool itself is depleted, then the packet is dropped.

• Building large multi-ported memories is very expensive, so most cheap switches are shallow buffered, with packet buffer being the scarcest resource.

Page 22: Data Center TCP (DCTCP). Introduction In recent years, data centers have trans- formed computing With large scale consolidation of enterprise IT into

Understanding Performance Impairments – Incast

• If many flows converge on the same interface of a switch over a short period of time, the packets may exhaust either the switch memory or the maximum permitted buffer for that interface, resulting in packet losses

• This traffic pattern arises naturally from use of the Partition/Aggregate design pattern, as the request for data synchronizes the workers’ responses and creates incast at the queue of the switch port connected to the aggregator.

Page 23: Data Center TCP (DCTCP). Introduction In recent years, data centers have trans- formed computing With large scale consolidation of enterprise IT into

Understanding Performance Impairments – Incast

• Incast degrades both performance and, more importantly, user experience.

• A response that incurs incast will almost certainly miss the aggregator deadline and be left out of the final results.

Page 24: Data Center TCP (DCTCP). Introduction In recent years, data centers have trans- formed computing With large scale consolidation of enterprise IT into

Understanding Performance Impairments – Incast

• Developers have made two major changes to the application code to avoid timeouts on worker responses. – deliberately limited the size of the response to

2KB to improve the odds that all the responses will fit in the memory of the switch

– developers added application-level jittering to desynchronize the responses by deliberating delaying them by a random amount of time (typically a mean value of 10ms)

Page 25: Data Center TCP (DCTCP). Introduction In recent years, data centers have trans- formed computing With large scale consolidation of enterprise IT into

Understanding Performance Impairments – Incast

Page 26: Data Center TCP (DCTCP). Introduction In recent years, data centers have trans- formed computing With large scale consolidation of enterprise IT into

Understanding Performance Impairments – Queue buildup

• Long-lived, greedy TCP flows will cause the length of the bottleneck queue to grow until packets are dropped

• When long and short flows traverse the same queue, two impairments occur – packet loss on the short flows can cause incast problems,

described before – there is a queue buildup impairment: even when no packets

are lost, the short flows experience in- creased latency as they are in queue behind packets from the large flows

• Since the latency is caused by queueing, the only solution is to reduce the size of the queues.

Page 27: Data Center TCP (DCTCP). Introduction In recent years, data centers have trans- formed computing With large scale consolidation of enterprise IT into

Understanding Performance Impairments – Buffer pressure

• Given the mix of long and short flows in our data center, it is very common for short flows on one port to be impacted by activity on any of the many other ports

• The loss rate of short flows in this traffic pattern depends on the number of long flows traversing other ports– different ports is coupled by the shared memory pool

• Buffer space is a shared resource, the queue build up reduces the amount of buffer space available to absorb bursts of traffic from Partition/Aggregate traffic – result is packet loss and timeouts, as in incast, but without

requiring synchronized flows.

Page 28: Data Center TCP (DCTCP). Introduction In recent years, data centers have trans- formed computing With large scale consolidation of enterprise IT into

DCTCP Algorithm

• The goal of DCTCP:achieve high burst tolerance, low latency, and high throughput, with commodity shallow buffered switches.

• DCTCP is designed to operate with small queue occupancies, without loss of throughput.

• DCTCP uses a simple marking scheme at switches – sets the Congestion Experienced (CE) codepoint of packets as

soon as the buffer occupancy exceeds a fixed small threshold. – The DCTCP source reacts by reducing the window by a factor

that depends on the fraction of marked packets: the larger the fraction, the bigger the decrease factor.

Page 29: Data Center TCP (DCTCP). Introduction In recent years, data centers have trans- formed computing With large scale consolidation of enterprise IT into

DCTCP Algorithm – Simple Marking at the Switch

• There is only a single parameter, the marking threshold, K.

• An arriving packet is marked with the CE codepoint if the queue occupancy is greater than K upon it’s arrival.

• Otherwise, it is not marked. • This scheme ensures that sources are quickly

notified of the queue overshoot.

Page 30: Data Center TCP (DCTCP). Introduction In recent years, data centers have trans- formed computing With large scale consolidation of enterprise IT into

DCTCP Algorithm – ECN-Echo at the Receiver

• The DCTCP receiver uses the trivial two state state-machine shown in Figure 10 to determine whether to set ECN- Echo bit. – The states correspond to whether the last received packet was

marked with the CE codepoint or not.

Page 31: Data Center TCP (DCTCP). Introduction In recent years, data centers have trans- formed computing With large scale consolidation of enterprise IT into

DCTCP Algorithm – Controller at the Sender

• a <- (1-g)×a+g×F– F is the fraction of packets that were marked in the

last window of data – 0 < g < 1 is the weight given to new samples against

the past in the estimation of a– a estimates the probability that the queue size is

greater than K – a close to 0 indicates low, and a close to 1 indicates

high levels of congestion • cwnd <- cwnd × (1-a/2)

Page 32: Data Center TCP (DCTCP). Introduction In recent years, data centers have trans- formed computing With large scale consolidation of enterprise IT into

DCTCP Algorithm - Benefit

• Queue buildup – DCTCP senders start reacting as soon as queue length on an

interface exceeds K – Reduces queueing delays on congested switch ports, which

minimizes the impact of long flows on the completion time of small flows.

• Buffer pressure – A congested port’s queue length does not grow exceedingly large.

• Incast – It is often bursts in subsequent RTTs that lead to drops – DCTCP sources receive enough marks during the first one or two

RTTs to tame the size of follow up bursts.

Page 33: Data Center TCP (DCTCP). Introduction In recent years, data centers have trans- formed computing With large scale consolidation of enterprise IT into

DCTCP Algorithm - Analysis

• Consider N infinitely long-lived flows with identical round-trip times RTT , sharing a single bottleneck link of capacity C.

• N flows are synchronized • The queue size at time t is given by – Q(t) = NW(t) – C×RTT – W (t) is the window size of a single source

Page 34: Data Center TCP (DCTCP). Introduction In recent years, data centers have trans- formed computing With large scale consolidation of enterprise IT into

DCTCP Algorithm - Analysis

• Interested in computing the following quantities which completely specify the saw- tooth – the maximum queue size (Qmax) – the amplitude of queue oscillations (A) – the period of oscillations (TC )

Page 35: Data Center TCP (DCTCP). Introduction In recent years, data centers have trans- formed computing With large scale consolidation of enterprise IT into

DCTCP Algorithm - Analysis

• Compute the fraction of marked packets, a, by simply dividing the number of packets sent during the last RTT of the period by the total number of packets sent during a full period of the sawtooth, TC

• Let W*=(C×RTT+K)/N. This is the critical window size at which the queue size reaches K

Page 36: Data Center TCP (DCTCP). Introduction In recent years, data centers have trans- formed computing With large scale consolidation of enterprise IT into

DCTCP Algorithm - Analysis

• S(W1,W2) denote the number of packets sent by the sender, while its window size increases from W1 to W2 > W1.

Page 37: Data Center TCP (DCTCP). Introduction In recent years, data centers have trans- formed computing With large scale consolidation of enterprise IT into

DCTCP Algorithm - Analysis

• D is the amplitude of oscillation in window size of a single flow

Page 38: Data Center TCP (DCTCP). Introduction In recent years, data centers have trans- formed computing With large scale consolidation of enterprise IT into

DCTCP Algorithm - Analysis

• Authors have evaluated the accuracy of the above results using NS-2 simulations in a variety of scenarios – For large N, as in the N = 40 case, de-synchronization of the

flows leads to smaller queue variations than predicted by the analysis

Page 39: Data Center TCP (DCTCP). Introduction In recent years, data centers have trans- formed computing With large scale consolidation of enterprise IT into

Guidelines for choosing parameters

• Marking Threshold: The minimum value of the queue occupancy sawtooth is given by:

• To find a lower bound on K, we minimize it over N, and choose K so that this minimum is larger than zero

Page 40: Data Center TCP (DCTCP). Introduction In recent years, data centers have trans- formed computing With large scale consolidation of enterprise IT into

Guidelines for choosing parameters

• Estimation Gain: g must be chosen small enough to ensure the exponential moving average “spans” at least one congestion event

• Plugging in expression for Tc with the worst case value N = 1

Page 41: Data Center TCP (DCTCP). Introduction In recent years, data centers have trans- formed computing With large scale consolidation of enterprise IT into

Discussion: AQM is not enough • Active Queue Management (AQM): RED and PI

- Do not modify TCP’s congestion control mechanism.- Do not work well when there is low statistical multiplexing and

traffic is bursty.- Have to make a tradeoff between throughput and delay.

• PI suffers from queue underflows and a loss of utilization with few flows (< 5), while with many flows (20), queue oscillations get worse, causing latency of time-critical flows.

• RED suffers from high variability of the queue length, thus query traffic experiences timeouts.

• DCTCP achieves both high through-put and low delay.

Page 42: Data Center TCP (DCTCP). Introduction In recent years, data centers have trans- formed computing With large scale consolidation of enterprise IT into

Discussion: Convergence and Synchronization

• DCTCP sacrifices convergence time: the time required for a new flow to grab its share of the bandwidth from an existing flow with a large window size.

• Because DCTCP source must make incremental adjustments to its window size based on the accumulated multi-bit feedback in :

Page 43: Data Center TCP (DCTCP). Introduction In recent years, data centers have trans- formed computing With large scale consolidation of enterprise IT into

Discussion: Convergence and Synchronization

Why sacrifices convergence time?• DC RTT: 0.1ms, not substantial compared to RTT in

Internet. Convergence times for DCTCP is just 2 or 3 times more than TCP.

• DC mostly are microbursts: too small to converge, i.e., last too short, no need to converge.

• Big flows:can tolerate a small convergence delay over their long lifetimes.

Page 44: Data Center TCP (DCTCP). Introduction In recent years, data centers have trans- formed computing With large scale consolidation of enterprise IT into

Discussion: Convergence and Synchronization

Does the “on-off” style marking cause synchronization between flows?

• DCTCP’s reaction to congestion is not severe, so it is less critical to avoid synchronization.

Page 45: Data Center TCP (DCTCP). Introduction In recent years, data centers have trans- formed computing With large scale consolidation of enterprise IT into

Discussion: Practical considerations

In reality, hosts tend to send more bursts (30-40 packets): must make allowances for such bursts when selecting the value of K.

• K = 20 for 1Gbps ( K > (C×RTT)/7 ). • K = 65 for 10Gbps to avoid loss of throughput.

Page 46: Data Center TCP (DCTCP). Introduction In recent years, data centers have trans- formed computing With large scale consolidation of enterprise IT into

Results: DCTCP Performance

Throughput and queue length:• Question: Whether DCTCP achieves the same

throughput as TCP on long-lived flow?• Settings: – One receiver and some senders are connected to the

Triumph switch with 1Gbps link.– Senders establish long-lived connections and send data as

fast as they can.– K=20.– Sample the instantaneous queue length at the receiver’s

switch port every 125ms.

Page 47: Data Center TCP (DCTCP). Introduction In recent years, data centers have trans- formed computing With large scale consolidation of enterprise IT into

DCTCP Performance

Throughput and queue length: • Similarity: both TCP and DCTCP achieve the

maximum throughput of 0.95Gbps, and link utilization is nearly 100%.

• Difference: queue length at the receiver interface.

Page 48: Data Center TCP (DCTCP). Introduction In recent years, data centers have trans- formed computing With large scale consolidation of enterprise IT into

DCTCP Performance• DCTCP queue length is stable

around 20 packets, while the TCP queue length is 10X larger and varies widely.

Page 49: Data Center TCP (DCTCP). Introduction In recent years, data centers have trans- formed computing With large scale consolidation of enterprise IT into

DCTCP Performance

Throughput: • With 1Gbps link, performance is insensitive to value

of K.• With 10Gbps link, K>65, DCTCP gets the same

throughput as TCP.

Page 50: Data Center TCP (DCTCP). Introduction In recent years, data centers have trans- formed computing With large scale consolidation of enterprise IT into

DCTCP PerformanceQueue length: • Comparison: RED shows wide queue length

oscillations, requiring twice as much buffer to achieve the same throughput as DCTCP.

Page 51: Data Center TCP (DCTCP). Introduction In recent years, data centers have trans- formed computing With large scale consolidation of enterprise IT into

DCTCP Performance

Fairness and convergence:• Flows converge to their fair share?• Settings: – 1 receiver 5 senders connected via 1Gbps links to

the Triumph switch.– K=20. – One sender starts a single long-lived flow, and

then sequentially start and then stop the other senders, spaced by 30 seconds.

Page 52: Data Center TCP (DCTCP). Introduction In recent years, data centers have trans- formed computing With large scale consolidation of enterprise IT into

DCTCP Performance Fairness and convergence:• DCTCP converges quickly, and all flows achieve their

fair share.• TCP throughput is fair on average, but has much

higher variation.

Page 53: Data Center TCP (DCTCP). Introduction In recent years, data centers have trans- formed computing With large scale consolidation of enterprise IT into

DCTCP Performance Multi-hop, multi-bottleneck environment:• Settings: S1 and S3 send to R1; S2 send to R2 (two bottlenecks in

red circles).• Result: Senders get fair-share (deviation<10%). DCTCP can cope

with the multiple bottlenecks and differing RTT.• TCP: queue length fluctuations at the two bottlenecks cause

timeouts.

Page 54: Data Center TCP (DCTCP). Introduction In recent years, data centers have trans- formed computing With large scale consolidation of enterprise IT into

Impairment microbenchmarks Incast:• Settings:

– 1 client, 40 servers connected with 1Gbps links. – Static buffer size of 100 packets each port in switch.– Client repeatedly requests 1MB/n bytes from n different

servers (n=1,5,10...40).• Measured query completion time and timeouts fraction

with 300ms and 10ms timeout values.

Page 55: Data Center TCP (DCTCP). Introduction In recent years, data centers have trans- formed computing With large scale consolidation of enterprise IT into

Impairment microbenchmarks Incast:• DCTCP performs better than TCP and then converges at 35

senders.• TCP begins to suffer timeouts when the number of servers

exceeds 10, while DCTCP only suffers timeouts once the number of senders is large enough so that the static buffer run out.

Page 56: Data Center TCP (DCTCP). Introduction In recent years, data centers have trans- formed computing With large scale consolidation of enterprise IT into

Impairment microbenchmarks Incast:• With dynamic buffering:– DCTCP no longer suffers incast timeouts.– TCP continues to suffer, even with 700KB buffer size.

Page 57: Data Center TCP (DCTCP). Introduction In recent years, data centers have trans- formed computing With large scale consolidation of enterprise IT into

Impairment microbenchmarks Incast:• All-to-all incast: • Settings: Each machine requests 25KB of data from the

remaining 40 machines (total 1MB).• Results:

– DCTCP has less response time than TCP.– DCTCP suffers no timeouts, while for TCP 55% of the queries suffer

from timeout.

Page 58: Data Center TCP (DCTCP). Introduction In recent years, data centers have trans- formed computing With large scale consolidation of enterprise IT into

Impairment microbenchmarks

Incast:• Summary:

DCTCP handles incast without timeouts until there are so many senders that the traffic sent in the first RTT overflows the buffers.

Page 59: Data Center TCP (DCTCP). Introduction In recent years, data centers have trans- formed computing With large scale consolidation of enterprise IT into

Impairment microbenchmarks Queue buildup:• Settings:

– 1 receiver and 3 senders connected with 1Gbps links. – 2 senders send big flows, 1 sender responds to receiver’s repeated

request (20KB).

• DCTCP’s completion time (median delay < 1 ms) is much lower than TCP (median delay 19ms).

• DCTCP improves latency caused by the big flows (queue build-up).

Page 60: Data Center TCP (DCTCP). Introduction In recent years, data centers have trans- formed computing With large scale consolidation of enterprise IT into

Impairment microbenchmarks Buffer pressure:• TCP: long flows use up the shared buffer space, so

there’s less headroom to absorb incast bursts.• DCTCP should alleviate this problem by limiting the

queue build up on the interfaces used by long flows.• Settings: – Connected 44 hosts to a Triumph switch with 1Gbps links. – 1 client repeatedly requests a total of 1MB data from 10

servers (100KB from each). – 33 hosts start “background” traffic of long-lived flows to

consume the shared buffer.

Page 61: Data Center TCP (DCTCP). Introduction In recent years, data centers have trans- formed computing With large scale consolidation of enterprise IT into

Impairment microbenchmarks Buffer pressure:• Results: With TCP, query completion time is

substantially worse in the presence of long-lived flows, while DCTCP’s performance is unchanged.

• Conclusion: DCTCP improves the performance isolation between flows by reducing the buffer pressure that would otherwise couple them.

Page 62: Data Center TCP (DCTCP). Introduction In recent years, data centers have trans- formed computing With large scale consolidation of enterprise IT into

Benchmark Traffic How would DCTCP perform under the traffic patterns

found in real production cluster?• Settings:

– Generate all three types of traffic found in the cluster: query, short-message, and background.

– RTO_min = 10ms in both TCP and DCTCP.– For DCTCP, K=20 for 1Gbps, K=65 for 10Gbps.– Using dynamic buffer allocation.

Page 63: Data Center TCP (DCTCP). Introduction In recent years, data centers have trans- formed computing With large scale consolidation of enterprise IT into

Benchmark Traffic For query and short-message flows: completion time.• Results (query and short-messages):

– DCTCP performs better than TCP because of low queuing delay.– Due to DCTCP’s amelioration of queue buildup, short-messages has lower

latency.– TCP: 1.15% queries suffer timeouts. – DCTCP: No queries suffer from timeouts.

Page 64: Data Center TCP (DCTCP). Introduction In recent years, data centers have trans- formed computing With large scale consolidation of enterprise IT into

Benchmark TrafficResults for background flows:– No timeout (both TCP and DCTCP).– Flows from 100KB to 1MB benefit most from DCTCP.

Page 65: Data Center TCP (DCTCP). Introduction In recent years, data centers have trans- formed computing With large scale consolidation of enterprise IT into

Benchmark Traffic Scaling the traffic in the benchmark:• Developers restrict the size of query responses and update

frequency because of SLA bounds on response time.• Then how would DCTCP perform if query response sizes are

larger? Or how would DCTCP perform if background traffic characteristics are different?

• Settings: – Increase the size of update flows larger than 1MB by a factor of

10.– Increase total size of query response to 1MB (44 servers, each

25KB).• Try other solutions: (1) deep-buffered switch, (2) RED with

ECN.

Page 66: Data Center TCP (DCTCP). Introduction In recent years, data centers have trans- formed computing With large scale consolidation of enterprise IT into

Benchmark Traffic Results for scaled traffic:• DCTCP performs significantly better than TCP for

both update and query traffic.• TCP, over 92% of the queries suffer from timeouts,

while only 0.3% suffer from timeouts with DCTCP.

Page 67: Data Center TCP (DCTCP). Introduction In recent years, data centers have trans- formed computing With large scale consolidation of enterprise IT into

Benchmark Traffic Deep buffered switches:• Solves incast problem, query completion time is

comparable to DCTCP due to less timeout.• Completion time for short-message traffic is even larger

than TCP, because deep buffers cause queue buildup.

Page 68: Data Center TCP (DCTCP). Introduction In recent years, data centers have trans- formed computing With large scale consolidation of enterprise IT into

Benchmark Traffic RED:• Improves performance of short transfers by keeping average queue

length low, but the high variability of the queue length leads to poor performance for the query traffic (95% of queries experience a timeout).

• Another possible factor is that RED marks packets based on average queue length, and is thus slower to react to bursts.

Page 69: Data Center TCP (DCTCP). Introduction In recent years, data centers have trans- formed computing With large scale consolidation of enterprise IT into

Benchmark TrafficSummary:• First, if data center uses DCTCP it could handle 10X

larger query responses and 10X larger background flows while performing better than it does with TCP today.

• Second, while using deep buffered switches improves performance of query traffic, it makes performance of short transfers worse, due to queue build up.

• Third, while RED improves performance of short transfers, it does not improve the performance of query traffic, due to queue length variability.

Page 70: Data Center TCP (DCTCP). Introduction In recent years, data centers have trans- formed computing With large scale consolidation of enterprise IT into

Conclusion

• Several performance impairments are due to TCP’s congestion control.

• To meet the needs of the observed diverse mix of short and long flows, switch buffer occupancies need to be persistently low, while maintaining high throughput for the long flows.

• DCTCP relies on Explicit Congestion Notification (ECN) and succeeds through early reaction using the multi-bit feedback derived from the series of ECN marks.

Page 71: Data Center TCP (DCTCP). Introduction In recent years, data centers have trans- formed computing With large scale consolidation of enterprise IT into

Questions?

Thank you!