4-4 1-4 delivers great performance guarantees in traditional (non-virtualized) setting, due to location based static IP address allocation to all network elements.
Download this ppt first and then open in powerpoint to view without merged figures and with animations.
3. INTRODUCTION
• Data center network
• Traditional architecture
• Agility
• Virtualization
• 4-4 1-4 Data center network
Large clusters of servers interconnected
by network switches, concurrently provide
large number of different services for
different client-organizations.
Design Goals:
• Availability and Fault tolerance
• Scalability
• Throughput
• Economies of scale
• Load balancing
• Low Opex
A number of virtual servers are
consolidated onto a single physical server.
Advantages:
• Each customer gets his own VM
• Virtualization provides Agility
Fig 1. Traditional Data Center
• In case of hardware failure VM can be
cloned & migrated to diff server.
• Synchronized Fig 3: 4-replicated 4 1-4 Data Center
VM images
instead of redundant servers
• Easier to test, upgrade and move
virtual servers across locations
• Virtual devices in a DCN
• Reduced Capex and Opex.
4. 4-4 1-4 ARCHITECTURE
• 4-4 1-4 is a location based forwarding
architecture for DCN which utilizes IP-hierarchy.
• Forwarding of packets is done by masking
the destination IP address bits.
• No routing or forwarding table maintained
at switches
• No convergence overhead.
• Uses statically assigned, location based
IP addresses for all network nodes.
A 3-Level 4-4 1-4 Data Center Network
6. MOTIVATION FOR THIS WORK
4-4 1-4 delivers great performance guarantees in traditional (non-virtualized)
setting, due to location based static IP address allocation to all
network elements.
Agility is essential in current data-centers, run by cloud service providers, to
reduces cost by increasing infrastructure utilization.
Server Virtualization provides the required agility.
Whether the 4-4 1-4 network delivers performance guarantees in a
virtualized setting, suitable to modern Data Centers, is the major
motivation for this work.
7. PROBLEM STATEMENT
How to virtualize the 4-4 1-4 data center network with the following constraints:
• Use static IP allocation along with dynamic VMs.
• No modification of network elements or end hosts.
Design Goals: To design a virtualized data center using 4-4 1-4 topology, that is
• Agile
• Scalable and Robust
• Minimize overhead incurred due to virtualization.
• Minimum end-to-end Latency and maximum Throughput
• Suitable for all kinds of data center usage scenarios:
• Compute Intensive: HPC
• Data Intensive: Video and File Streaming
• Balanced: Geographic Information System
8. PROPOSED SOLUTION
• Separation of Location-IP and VM-IP
• Tunneling at source
• Directory structure
• Query process
• Directory Update mechanism
Packet tunneled through physical network using location-IP header
Packet sending at a server running a type-1 hypervisor
9. PROPOSED SOLUTION
• Separation of Location-IP and VM-IP
• Tunneling at source
• Directory structure
• Query process
• Directory Update mechanism
Directory structure
Physical Machines = 2^16.
Virtual Machines = 2^17 (2 VMs/PM).
Virtual Machines = 2^20 (16 VMs/PM).
Directory Servers = 64.
Number of Update Servers = 16.
Hence, one DS per 1024 PMs and
one US per (4 * 1024) PMs.
This implies 64 DSs, for minimum 131072 VMs
10. PROPOSED SOLUTION
• Separation of Location-IP and VM-IP
• Tunneling at source
• Directory structure
• Query process
• Directory Update mechanism
Data Structure of Directory
11. EXPERIMENTAL SET UP
Simulation environment: Extension of NS2/Greencloud:
• Packet-level simulation of communications in DCN, unlike CloudSim,
MDCSim, etc.
• DCN Entities are modeled as C++ and OTCL objects.
DCN Workloads: Categories of Experiments setup
• Computation Intensive Workload (CIW): The servers are considerably
loaded, but negligible inter server communication.
• Data Intensive Workload (DIW): Huge inter server data transfers, but
negligible load at the computing servers
• Balanced Workload (BW): Communication links and computing servers are
proportionally loaded.
12. EXPERIMENTAL SET UP
• In CIW and BW, tasks are scheduled in a Round Robin fashion by the Data
Center object, onto VMs on servers fulfilling task resource requirement.
• A task is sent to allocated VM by DCobject through core switches. Output is
returned to the same core switch, which then forwards it to the DCobject.
• In DIW and BW, intra-DCN comm or data-transfer is modelled by 1:1:1 TCP
flows between servers.
S: Source and Destination within same Level-0
D: Source and Destination are in different Level-0 but same Level-1
R: Random selection of Source and Destination pairs inside Level-
1.
15. PERFORMANCE METRICS
• Average packet delay
• Network Throughput
• End to End Aggregate/data Throughput
• Average hop count
• Packet drop rate
• Normalized Routing overhead
17. RESULTS: COMPUTE INTENSIVE WORKLOAD
• DVR vs LocR in 16 Servers: 50% less Delay and more throughput
• 16 vs 64 Servers: Almost same.
• Routing Overhead in DVR increases with more number of servers.
18. RESULTS: DATA INTENSIVE WORKLOAD
• Average Packet Delay:
• DVR vs LocR: Less Delay in
LocR
• 16 vs 64: Delay reduces by 54%
• DVR vs LocR: More in LocR
• 16 vs 64: Increases by 54%
• Network throughput:
• End-to-end aggregate Throughput:
• DVR vs LocR: More in LocR
• 16 vs 64: Increases by 53%
19. RESULTS: BALANCED WORKLOAD
• Average Packet Delay:
• DVR vs LocR: Less Delay in
LocR
• 16 vs 64: Delay reduces by 42%
• DVR vs LocR: More in LocR
• 16 vs 64: Increases by 42%
• Network throughput:
• End-to-end aggregate Throughput:
• DVR vs LocR: More in LocR
• 16 vs 64: Increases by 41%
20. CONCLUSION
Creation of a packet level simulation prototype in NS2/Greencloud for 4-4 1-
4 DCN.
Modelling of compute-intensive, data-intensive and balanced workloads
We conclude that our framework for virtualization in 4-4 1-4 DCN in has the
following significance:
• Routing over-head: No convergence overhead in location-based routing
• Networking loops: network is free from networking loops
• Faster hop-by-hop forwarding: as per-packet-per-hop mask operation is
faster than table lookup and update operation.
• Efficiency: Location- IP based routing delivers two to ten times more
throughput than DVR with same traffic and same topology.
• Scalable: In DIW and BW the performance increases by 50% when number
of servers is increased by four times.
21. LIMITATION
4-4 1-4 is highly scalable in Data Intensive and Balanced workload data
centers but moderately for heavy-computing data centers .
In computation intensive workloads, the performance of 4-4 1-4 DCN with
location based routing, either remains the same or increases marginally.
22. FUTURE WORK
Simulation
test-bed is
ready
Trace-driven
workload
Dynamic VM
migration
Optimum
task
Scheduling
for 4-4 1-4
Energy
consump
tion
23. REFERENCES
1. A. Kumar, S. V. Rao, and D. Goswami, “4-4, 1-4: Architecture for Data Center
Network Based on IP Address Hierarchy for Efficient Routing," in Parallel and
Distributed Computing (ISPDC), 2012 11th International Symposium on,
2012, pp. 235-242.
2. D. Chisnall, The defitive guide to the xen hypervisor, 1st ed. Upper Saddle
River, NJ, USA: Prentice Hall Press, 2007.
3. D. Kliazovich, P. Bouvry, and S. Khan, “Greencloud: a packet-level simulator
of energy-aware cloud computing data centers," The Journal of
Supercomputing, pp.1{21, 2010, 10.1007/s11227-010-0504-1. Available:
http://dx.doi.org/10.1007/s11227-010-0504-1
4. “The Network Simulator NS-2," http://www.isi.edu/nsnam/ns/.
25. There are mysteries in the
universe,
We were never meant to solve,
But who we are, and why we are
here,
Are not one of them.
Those answers we carry inside.
26. RESULTS: DATA INTENSIVE WORKLOAD
• Average Packet Delay:
• DVR vs LocR: Less Delay in
LocR
• 16 vs 64: Delay reduces by 54%
• DVR vs LocR: More in LocR
• 16 vs 64: Increases by 54%
• Network throughput:
• End-to-end aggregate Throughput:
• DVR vs LocR: More in LocR
• 16 vs 64: Increases by 53%
• Routing overhead using dynamic routing:
27. RESULTS: BALANCED WORKLOAD
• Average Packet Delay:
• DVR vs LocR: Less Delay in
LocR
• 16 vs 64: Delay reduces by 54%
• DVR vs LocR: More in LocR
• 16 vs 64: Increases by 54%
• Network throughput:
• End-to-end aggregate Throughput:
• DVR vs LocR: More in LocR
• 16 vs 64: Increases by 53%
• DVR Routing overhead:
Notas do Editor
In virtualization a system pretends to be two or more of the same system
CloudSim, MDCSim are event based simulators. Fire events.
whenever a data message has to be transmitted between simulator entities a packet
structure with its protocol headers is allocated in the memory and all the associated
protocol processing is performed.
On the contrary, CloudSim and MDCSim are event-based simulators.
They avoid building and processing small simulation objects
(like packets) individually. Instead, the effect of object interaction is captured. Such
a method reduces simulation time considerably, improves scalability, but lacks in the
simulation accuracy.
Average packet delay: Delay of a packet is the actual time it took to reach its destination
node, from its source node. Average packet delay is computed as:
P
p2P pdp
N , where P is
the set of all packets, pdp is the delay of pth packet and N is the total number of packets.
Network Throughput: It is dened as the the end-to-end bytes transmitted in the network
per unit of time. It is computed as:
P
p2P
psp
pdp
, where psp is the size of the pth packet, in
bits.
End to End Aggregate/data Throughput: It is dened as the useful data bytes transmitted
in the network per unit of time. It is computed as:
P
p2P
psp
pdp
, where psp is the size of
data payload in pth packet and pdp is the delay of pth packet.
Packet drop rate: Ratio of the number of packets dropped to the toatal number of packets
sent.
Average hop count: Average number of hops taken by packets to reach their destination,
from their source.
42
Normalized Routing overhead: Number of routing packets per data packet. It is computed
as the ratio of number of routing packets, to the number of data packets.
Theoretical network limit
rough estimation: rate < (MSS/RTT)*(C/sqrt(Loss)) [ C=1 ] (based on the Mathis et.al. formula)network limit (MSS 1540 byte, RTT: 0.4 ms, Loss: 10-08 (10-06%)) : 305970.51 Mbit/sec.
Bandwidth-delay Product and buffer size
BDP (1000 Mbit/sec, 0.4 ms) = 0.05 MByte required tcp buffer to reach 1000 Mbps with RTT of 0.4 ms >= 49.2 KByte maximum throughput with a TCP window of 32 KByte and RTT of 0.4 ms <= 666.67 Mbit/sec.
We have created the test-bed for simulations on virtualized 4-4 1-4 data center network in NS2/Greencloud.
The next step would be to run tests for dynamic virtual machine migration
and server provisioning in this network. We have modelled different kinds of workloads possible
for any data center network but trace-driven workload generation on this network would present
the most relevant and practical metrics. Performance of Multi-casting groups in this network
is also an unexplored domain. VM Scheduling policy plays a very important role in the power
consumption and performance of a network. A scheduling algorithm optimum for 4-4 1-4
network is needed for better computation Centric workload performance.