Lecture 4 Part 1 Cloud Infrastructure Data Center

  • Slides: 14
Download presentation
Lecture 4 (Part 1): Cloud Infrastructure Data Center Network Architecture

Lecture 4 (Part 1): Cloud Infrastructure Data Center Network Architecture

Socrative • www. socrative. com • Room: CMU 14848 • https: //api. socrative. com/rc/Nfu

Socrative • www. socrative. com • Room: CMU 14848 • https: //api. socrative. com/rc/Nfu 6 Lp

(K/2)2 core routers Recall Fat Trees Note the replacement of aggregation layer switches with

(K/2)2 core routers Recall Fat Trees Note the replacement of aggregation layer switches with 2 layers of K/2 K-port switches (K/2)2 servers per pod K-port switches support K 3/4 servers

Using Multiple Paths • Must pick different paths (“path diversity”) or will have a

Using Multiple Paths • Must pick different paths (“path diversity”) or will have a hotspot • Unless sessions use the same path, reordering will be a problem and need to be resolved with buffering higher up • Static paths may not respond to actual, dynamic workloads • Can be done at different levels. • Higher levels, e. g. transport, are more flexible, but likely more effort and slower • Lower levels are likely less adaptive, but simpler and faster. • Ability to weight or remove paths can aid fault tolerance

Portland Solution • Use commodity switches and off-load services into software on commodity server

Portland Solution • Use commodity switches and off-load services into software on commodity server • Start With Fat Tree for a topology without hot spots • Use layer-2 to avoid routing, forwarding, and related complexity • Separate host identifier from host location • IP addresses identify host, but not location, just and ID • Use “Pseudo MAC Address” to identify location at Level-2

Port. Land Addresses • Normally MAC addresses are arbitrary – no clue about location

Port. Land Addresses • Normally MAC addresses are arbitrary – no clue about location • IP normally is hierarchical, but here we are using it only as a host identifier • If MAC addresses are not tied to location, switch tables grow linearly with growth of network, i. e. O(n) • Port. Land uses hierarchical MAC addresses, called “Pseudo MAC” or PMAC addresses to provide for switch location • <pod: port: position: vmid> • <16, 8, 8, 16> bits

0 2 Positio n Port. Land PMAC Addresses 1 3 0 1 0 PMAC:

0 2 Positio n Port. Land PMAC Addresses 1 3 0 1 0 PMAC: <pod. position. port. vmid> 1 0 1 48 bits: <16 -bits. 8 -bits. 16 -bits>

0 1 2 3 Portland PMAC Addresses <pod: port: position: vmid> <16, 8, 8,

0 1 2 3 Portland PMAC Addresses <pod: port: position: vmid> <16, 8, 8, 16> bits

Name Resolution: MAC PMAC IP • End hosts continue to use Actual MAC (AMAC)

Name Resolution: MAC PMAC IP • End hosts continue to use Actual MAC (AMAC) addresses • Switches convert PMAC<->AMAC for the host • Edge switch responsible for creating PMAC: AMAC mapping and telling Fabric Manager • Software on commodity server, can be replicated, etc. Simplicity is a virtue. • Mappings timed out of Fabric Manager’s cache, if not used. • ARPs are for PMACs • First ask fabric manager which keeps cache. Then, if needed, broadcast.

VM Migration • Flat address space. • IP address unchanged after migration, higher level

VM Migration • Flat address space. • IP address unchanged after migration, higher level doesn’t see state change • • After migration IP<->PMAC changes, as PMAC is location dependent VM sends gratuitous ARP with new mapping. Fabric Manager receives ARP and sends invalidation to old switch Old switch sets flow table to software, causing ARP to be sent to any stray packets • Forwarding the packet is optional, as retransmit (if reliable) will fix delivery

Location Discovery: Configuring Switch IDs • Humans = Not right Answer • Discovery =

Location Discovery: Configuring Switch IDs • Humans = Not right Answer • Discovery = Right Answer • Send messages to neighbors – Get Tree Level • • • Hosts don’t reply, so edge only hears back from above Aggregate hears back from both levels Core hears back only from aggregate • Contact Fabric Manager with tree level to get ID • • • Fabric Manager is service running on commodity host Assigns ID Soft state

No loops, No Spanning Trees • Forwarding can only go up the tree. •

No loops, No Spanning Trees • Forwarding can only go up the tree. • Cycles not possible.

Failure • • Keep-alives like the link discovery messages Miss a keep alive? Tattle

Failure • • Keep-alives like the link discovery messages Miss a keep alive? Tattle to the Fabric Manager Fabric manager tells effected switches, which adjust own tables. O(N) vs O(N 2) for traditional routing algorithms (Fabric Manager tells every switch vs every switch tells every switch)

Looking Back • Connectivity – Hosts can talk! No possibility of loops • Efficiency

Looking Back • Connectivity – Hosts can talk! No possibility of loops • Efficiency – Much less memory needed in switches, O(N) fault handlingh • Self configuring – Discovery protocol + ARP • Robust – Failure handling coordinated by FM • VMs and Migration – Each has own IP address, each has own MAC address • Commodity hardware – Nothing magic.