Networking The Cloud

38
1 Albert Greenberg, ICDCS 2009 keynote Networking The Cloud Albert Greenberg Principal Researcher [email protected] (work with James Hamilton, Srikanth Kandula, Dave Maltz, Parveen Patel, Sudipta Sengupta, Changhoon Kim)

description

Networking The Cloud. Albert Greenberg Principal Researcher [email protected] (work with James Hamilton, Srikanth Kandula, Dave Maltz, Parveen Patel, Sudipta Sengupta, Changhoon Kim). Agenda. Data Center Costs Importance of Agility Today’s Data Center Network A Better Way?. - PowerPoint PPT Presentation

Transcript of Networking The Cloud

Page 1: Networking The Cloud

Networking The Cloud

Albert GreenbergPrincipal [email protected]

(work with James Hamilton, Srikanth Kandula, Dave Maltz, Parveen Patel, Sudipta Sengupta, Changhoon Kim)

Page 2: Networking The Cloud

2Albert Greenberg, ICDCS 2009 keynoteAlbert Greenberg, ICDCS 2009 keynote

Agenda

• Data Center Costs– Importance of Agility

• Today’s Data Center Network• A Better Way?

Page 3: Networking The Cloud

3Albert Greenberg, ICDCS 2009 keynote

Data Center Costs

• Total cost varies– Upwards of $1/4 B for mega data center– Server costs dominate– Network costs significant

Amortized Cost* Component Sub-Components~45% Servers CPU, memory, disk

~25% Power infrastructure UPS, cooling, power distribution

~15% Power draw Electrical utility costs

~15% Network Switches, links, transit

*3 yr amortization for servers, 15 yr for infrastructure; 5% cost of money

The Cost of a Cloud: Research Problems in Data Center Networks. Sigcomm CCR 2009. Greenberg, Hamilton, Maltz, Patel.

Page 4: Networking The Cloud

4

Server Costs

Ugly secret: 30% utilization considered “good” in data centersCauses include:• Uneven application fit:

– Each server has CPU, memory, disk: most applications exhaust one resource, stranding the others

• Long provisioning timescales:– New servers purchased quarterly at best

• Uncertainty in demand:– Demand for a new service can spike quickly

• Risk management:– Not having spare servers to meet demand brings failure just when

success is at hand• Session state and storage constraints

– If the world were stateless servers, life would be good

Page 5: Networking The Cloud

5

Goal: Agility – Any service, Any Server

• Turn the servers into a single large fungible pool– Let services “breathe” : dynamically expand and contract their

footprint as needed• Benefits

– Increase service developer productivity– Lower cost– Achieve high performance and reliability

The 3 motivators of most infrastructure projects

Page 6: Networking The Cloud

6

Achieving Agility

• Workload management– Means for rapidly installing a service’s code on a server– Virtual machines, disk images

• Storage Management– Means for a server to access persistent data– Distributed filesystems (e.g., blob stores)

• Network– Means for communicating with other servers, regardless of where

they are in the data center

Page 7: Networking The Cloud

7

Network Objectives

Developers want a mental model where all their servers, and only their servers, are plugged into an Ethernet switch1. Uniform high capacity

– Capacity between servers limited only by their NICs– No need to consider topology when adding servers

2. Performance isolation– Traffic of one service should be unaffected by others

3. Layer-2 semantics– Flat addressing, so any server can have any IP address– Server configuration is the same as in a LAN– Legacy applications depending on broadcast must work

Page 8: Networking The Cloud

8

Agenda

• Data Center Costs– Importance of Agility

• Today’s Data Center Network• A Better Way?

Page 9: Networking The Cloud

9Albert Greenberg, ICDCS 2009 keynoteAlbert Greenberg, ICDCS 2009 keynote

The Network of a Modern Data Center

• Hierarchical network; 1+1 redundancy• Equipment higher in the hierarchy handles more traffic, more

expensive, more efforts made at availability scale-up design• Servers connect via 1 Gbps UTP to Top of Rack switches• Other links are mix of 1G, 10G; fiber, copper

Ref: Data Center: Load Balancing Data Center Services, Cisco 2004

InternetCR CR

AR AR AR AR…

SSLB LB

Data CenterLayer 3

Internet

SS

A AA …

SS

A AA …

Layer 2

Key:• CR = L3 Core Router• AR = L3 Access Router• S = L2 Switch• LB = Load Balancer• A = Rack of 20 servers with Top of Rack switch

~ 4,000 servers/pod

Page 10: Networking The Cloud

10Albert Greenberg, ICDCS 2009 keynoteAlbert Greenberg, ICDCS 2009 keynote

Internal Fragmentation Prevents Applications from Dynamically Growing/Shrinking

• VLANs used to isolate properties from each other• IP addresses topologically determined by ARs• Reconfiguration of IPs and VLAN trunks painful, error-

prone, slow, often manual

InternetCR CR

…AR AR

SSLB LB

SS

A AA …

SS

A AA …

AR AR

SSLB LB

SS

A AA …

SS

A AA …A

Page 11: Networking The Cloud

11Albert Greenberg, ICDCS 2009 keynoteAlbert Greenberg, ICDCS 2009 keynote

No Performance Isolation

• VLANs typically provide reachability isolation only• One service sending/receiving too much traffic hurts all

services sharing its subtree

InternetCR CR

…AR AR

SSLB LB

SS

A AA …

SS

A AA …

AR AR

SSLB LB

SS

A AA …

SS

A AA …A

Collateral damage

Page 12: Networking The Cloud

12Albert Greenberg, ICDCS 2009 keynoteAlbert Greenberg, ICDCS 2009 keynote

Network has Limited Server-to-Server Capacity, and Requires Traffic Engineering to Use What It Has

• Data centers run two kinds of applications:– Outward facing (serving web pages to users)– Internal computation (computing search index – think HPC)

InternetCR CR

…AR AR

SSLB LB

SS

A AA …

SS

A AA …

AR AR

SSLB LB

SS

A AA …

SS

A AA …

10:1 over-subscription or worse (80:1, 240:1)

Page 13: Networking The Cloud

13Albert Greenberg, ICDCS 2009 keynoteAlbert Greenberg, ICDCS 2009 keynote

Network Needs Greater Bisection BW, and Requires Traffic Engineering to Use What It Has

• Data centers run two kinds of applications:– Outward facing (serving web pages to users)– Internal computation (computing search index – think HPC)

InternetCR CR

…AR AR

SSLB LB

SS

A AA …

SS

A AA …

AR AR

SSLB LB

SS

A AA …

SS

A AA …

Dynamic reassignment of servers and Map/Reduce-style computations mean

traffic matrix is constantly changing

Explicit traffic engineering is a nightmare

Page 14: Networking The Cloud

14Albert Greenberg, ICDCS 2009 keynoteAlbert Greenberg, ICDCS 2009 keynote

What Do Data Center Faults Look Like?

•Need very high reliability near top of the tree

– Very hard to achieve Example: failure of a

temporarily unpaired core switch affected ten million users for four hours

– 0.3% of failure events knocked out all members of a network redundancy group

Ref: Data Center: Load Balancing Data Center Services, Cisco 2004

CR CR

AR AR AR AR…

SSLB LB

SS

A AA …

SS

A AA …

VL2: A Flexible and Scalable Data Center Network. Sigcomm 2009. Greenberg, Jain, Kandula, Kim, Lahiri, Maltz, Patel, Sengupta.

Page 15: Networking The Cloud

15Albert Greenberg, ICDCS 2009 keynoteAlbert Greenberg, ICDCS 2009 keynote

Agenda

• Data Center Costs– Importance of Agility

• Today’s Data Center Network• A Better Way?

Page 16: Networking The Cloud

16Albert Greenberg, ICDCS 2009 keynoteAlbert Greenberg, ICDCS 2009 keynote

Agenda

• Data Center Costs– Importance of Agility

• Today’s Data Center Network• A Better Way?

– Building Blocks– Traffic– New Design

Page 17: Networking The Cloud

17Albert Greenberg, ICDCS 2009 keynoteAlbert Greenberg, ICDCS 2009 keynote

Switch on Chip ASICsGeneral purpose

CPUfor control plane

Switch-on-a-chip ASIC

ASIC floorplan

X-c

eive

r(S

erD

es)

Packet bufferMemory

Forwarding tables

Forwarding pipeline

• Current design points– 24 port 1G, 4 10G 16K IPv4 fwd entries, 2 MB buff– 24 port 10G Eth, 16K IPv4 fwd entries, 2 MB buff

• Future– 48 port 10G, 16K fwd entries, 4 MB buff– Trends towards more ports, faster port speed

Page 18: Networking The Cloud

18Albert Greenberg, ICDCS 2009 keynote

Packaging

• Switch– Combine ASICs

Silicon fab costs drive ASIC price

Market size drives packaged switch price

– Economize links: On chip < on PCB < on

chassis < between chasses

– Example: 144 port 10G switch,

built from 24 port switch ASICs in single chassis.

• Link technologies– SFP+ 10G port– $100, MM fiber– 300m reach

• QSFP (Quad SFP)– 40G port avail today– 4 10G bound together

• Fiber “ribbon cables”– Up to 72 fibers per cable, to

a single MT connector

Page 19: Networking The Cloud

19Albert Greenberg, ICDCS 2009 keynoteAlbert Greenberg, ICDCS 2009 keynote

Latency

• Propagation delay in the data center is essentially 0– Light goes a foot in a nanosecond; 1000’ = 1 usec

• End to end latency comes from– Switching latency

10G to 10G:~ 2.5 usec (store&fwd); 2 usec (cut-thru)– Queueing latency

Depends on size of queues and network load• Typical times across a quiet data center: 10-20usec• Worst-case measurement (from our testbed, not real DC, with all2all

traffic pounding and link util > 86%): 2-8 ms• Comparison:

– Time across a typical host network stack is 10 usec• Application developer SLAs > 1 ms granularity

Page 20: Networking The Cloud

20Albert Greenberg, ICDCS 2009 keynoteAlbert Greenberg, ICDCS 2009 keynote

Agenda

• Data Center Costs– Importance of Agility

• Today’s Data Center Network• A Better Way?

– Building Blocks– Traffic– New Design

Page 21: Networking The Cloud

21Albert Greenberg, ICDCS 2009 keynoteAlbert Greenberg, ICDCS 2009 keynote

Measuring Traffic in Today’s Data Centers

• 80% of the packets stay inside the data center– Data mining, index computations, back end to front end– Trend is towards even more internal communication

• Detailed measurement study of data mining cluster– 1,500 servers, 79 Top of Rack (ToR) switches– Logged: 5-tuple and size of all socket-level R/W ops– Aggregated in flows – all activity separated by < 60 s– Aggregated into traffic matrices every 100 s

Src, Dst, Bytes of data exchange

Page 22: Networking The Cloud

22Albert Greenberg, ICDCS 2009 keynoteAlbert Greenberg, ICDCS 2009 keynote

Flow Characteristics

Median of 10 concurrent flows per server

Most of the flows: various mice

Most of the bytes: within 100MB flows

DC traffic != Internet traffic

Page 23: Networking The Cloud

23Albert Greenberg, ICDCS 2009 keynoteAlbert Greenberg, ICDCS 2009 keynote

Traffic Matrix Volatility

- Traffic pattern changes nearly constantly

- Run length is 100s to 80% percentile; 99th is 800s

- Collapse similar traffic matrices (over 100sec) into “clusters”

- Need 50-60 clusters to cover a day’s traffic

Page 24: Networking The Cloud

24Albert Greenberg, ICDCS 2009 keynoteAlbert Greenberg, ICDCS 2009 keynote

Today, Computation Constrained by Network*

*Kandula, Sengupta, Greenberg,Patel

Figure: ln(Bytes/10sec) between servers in operational cluster • Great efforts required to place communicating servers under the same ToR

Most traffic lies on the diagonal (w/o log scale all you see is the diagonal)• Stripes show there is need for inter-ToR communication

Page 25: Networking The Cloud

25Albert Greenberg, ICDCS 2009 keynoteAlbert Greenberg, ICDCS 2009 keynote

Congestion: Hits Hard When it Hits*

*Kandula, Sengupta, Greenberg, Patel

Page 26: Networking The Cloud

26Albert Greenberg, ICDCS 2009 keynoteAlbert Greenberg, ICDCS 2009 keynote

Agenda

• Data Center Costs– Importance of Agility

• Today’s Data Center Network• A Better Way?

– Building Blocks– Traffic– New Design

• VL2: A Flexible and Scalable Data Center Network. Sigcomm 2009. Greenberg, Hamilton, Jain, Kandula, Kim, Lahiri, Maltz, Patel, Sengupta.

• Towards a Next Generation Data Center Architecture: Scalability and Commoditization.  Presto 2009. Greenberg, Maltz, Patel, Sengupta, Lahiri.

• PortLand: A Scalable Fault-Tolerant Layer 2 Data Center Network Fabric. Mysore, Pamboris, Farrington, Huang, Miri, Radhakrishnan, Subramanya, Vahdat

• BCube: A High Performance, Server-centric Network Architecture for Modular Data Centers.  Guo, Lu, Li, Wu, Zhang, Shi, Tian, Zhang, Lu

Page 27: Networking The Cloud

27Albert Greenberg, ICDCS 2009 keynoteAlbert Greenberg, ICDCS 2009 keynote

VL2: Distinguishing Design Principles• Randomizing to Cope with Volatility

– Tremendous variability in traffic matrices• Separating Names from Locations

– Any server, any service• Embracing End Systems

– Leverage the programmability & resources of servers– Avoid changes to switches

• Building on Proven Networking Technology– We can build with parts shipping today– Leverage low cost, powerful merchant silicon ASICs,

though do not rely on any one vendor

Page 28: Networking The Cloud

28Albert Greenberg, ICDCS 2009 keynote

What Enables a New Solution Now?

• Programmable switches with high port density– Fast: ASIC switches on a chip (Broadcom, Fulcrum, …)– Cheap: Small buffers, small forwarding tables– Flexible: Programmable control planes

• Centralized coordination– Scale-out data centers are

not like enterprise networks– Centralized services already

control/monitor health and role of each server (Autopilot)

– Centralized directory and control plane acceptable (4D) 20 port 10GE switch. List price: $10K

Page 29: Networking The Cloud

29Albert Greenberg, ICDCS 2009 keynoteAlbert Greenberg, ICDCS 2009 keynote

An Example VL2 Topology: Clos Network

10GD/2 ports

D/2 ports

Aggregationswitches

. . .

. . .D switches

D/2 switches

Intermediate node switches in VLBD ports

Top Of Rack switch

[D2/4] * 20 Servers

20 ports

Node degree (D) of available switches & # servers supported

D # Servers in pool4 80

24 2,88048 11,520

144 103,680

• A scale-out design with broad layers• Same bisection capacity at each layer no oversubscription• Extensive path diversity Graceful degradation under failure

Page 30: Networking The Cloud

30Albert Greenberg, ICDCS 2009 keynoteAlbert Greenberg, ICDCS 2009 keynote

Use Randomization to Cope with Volatility

• Valiant Load Balancing– Every flow “bounced” off a random intermediate switch– Provably hotspot free for any admissible traffic matrix– Servers could randomize flow-lets if needed

Node degree (D) of available switches & # servers supported

D # Servers in pool4 80

24 2,88048 11,520

144 103,68010G

D/2 ports

D/2 ports

. . .

. . .D switches

D/2 switches

Intermediate node switches in VLBD ports

Top Of Rack switch

[D2/4] * 20 Servers

20 ports

Aggregationswitches

Page 31: Networking The Cloud

31Albert Greenberg, ICDCS 2009 keynoteAlbert Greenberg, ICDCS 2009 keynote

Separating Names from Locations:How Smart Servers Use Dumb Switches

• Encapsulation used to transfer complexity to servers– Commodity switches have simple forwarding primitives– Complexity moved to computing the headers

• Many types of encapsulation available– IEEE 802.1ah defines MAC-in-MAC encapsulation; VLANs; etc.

Source (S)

ToR (TS)Dest: N Src: SDest: TD Src: SDest: D Src: SPayload

Intermediate Node (N)

Dest (D)

ToR (TD)

1

2 3

4

Dest: TD Src: SDest: D Src: SPayload…

Payload…Dest: D Src: S

Dest: N Src: SDest: TD Src: SDest: D Src: SPayload…

Headers

Page 32: Networking The Cloud

32Albert Greenberg, ICDCS 2009 keynoteAlbert Greenberg, ICDCS 2009 keynote

Embracing End Systems

• Data center OSes already heavily modified for VMs, storage clouds, etc.– A thin shim for network support is no big deal

• No change to applications or clients outside DC

TCP

IP

NIC

ARP

EncapsulatorMAC

Resolution Cache

VL2 AgentUserKernel

ResolveremoteIP

DirectorySystem

ServerRole

ServerHealth

NetworkHealth

Server machine

Page 33: Networking The Cloud

33Albert Greenberg, ICDCS 2009 keynoteAlbert Greenberg, ICDCS 2009 keynote

VL2 Prototype

• 4 ToR switches, 3 aggregation switches, 3 intermediate switches• Experiments conducted with both 40 and 80 servers

Page 34: Networking The Cloud

34Albert Greenberg, ICDCS 2009 keynoteAlbert Greenberg, ICDCS 2009 keynote

VL2 Achieves Uniform High Throughput

• Experiment: all-to-all shuffle of 500 MB among 75 servers – 2.7 TB• Excellent metric of overall efficiency and performance• All2All shuffle is superset of other traffic patterns

• Results:• Ave goodput: 58.6 Gbps; Fairness index: .995; Ave link util: 86%

• Perfect system-wide efficiency would yield aggregate goodput of 75G– VL2 efficiency is 78% of perfect– 10% inefficiency due to duplexing issues; 7% header overhead– VL2 efficiency is 94% of optimal

Page 35: Networking The Cloud

35Albert Greenberg, ICDCS 2009 keynoteAlbert Greenberg, ICDCS 2009 keynote

VL2 Provides Performance Isolation

- Service 1 unaffected by service 2’s activity

Page 36: Networking The Cloud

36Albert Greenberg, ICDCS 2009 keynoteAlbert Greenberg, ICDCS 2009 keynote

VL2 is resilient to link failures

- Performance degrades and recovers gracefully as links are failed and restored

Page 37: Networking The Cloud

37Albert Greenberg, ICDCS 2009 keynoteAlbert Greenberg, ICDCS 2009 keynote

SummaryAmortized Cost

Component Sub-Components

~45% Servers CPU, memory, disk

~25% Infrastructure UPS, cooling, power distribution

~15% Power draw Electrical utility costs

~15% Network Switches, links, transit

10G

D/2 ports

D/2 ports

. . .

. . .

D switches

D/2 switches

Intermediate node switches in VLB

D ports

Top Of Rack switch

[D 2/4] * 20 Servers

20 port

s

Aggregationswitches

• It’s about agility– Increase data center capacity

Any service on any server, anywhere in the data center• VL2 enables agility and ends oversubscription

– Results have near perfect scaling– Increases service developer productivity

A simpler abstraction -- all servers plugged into one huge Ethernet switch

– Lowers cost High scale server pooling, with tenant perf isolation Commodity networking components

– Achieves high performance and reliability Gives us some confidence that design will scale-out

– Prototype behaves as theory predicts

Page 38: Networking The Cloud

38Albert Greenberg, ICDCS 2009 keynote

Thank You