41
A Distributed Energy Saving Approach for Ethernet Switches in Data Centers Weisheng Si 1 , Javid Taheri 2 , Albert Zomaya 2 1 School of Computing, Engineering, and Mathematics, University of Western Sydney 2 School of IT, University of Sydney 1

A Distributed Energy Saving Approach for Ethernet Switches in Data Centers

Embed Size (px)

DESCRIPTION

A Distributed Energy Saving Approach for Ethernet Switches in Data Centers. Weisheng Si 1 , Javid Taheri 2 , Albert Zomaya 2 1 School of Computing, Engineering, and Mathematics, University of Western Sydney 2 School of IT, University of Sydney. Outline. Background Related work - PowerPoint PPT Presentation

Citation preview

Page 1: A Distributed Energy Saving Approach for Ethernet Switches in Data Centers

A Distributed Energy Saving Approach for Ethernet Switches in Data Centers

Weisheng Si 1, Javid Taheri 2, Albert Zomaya 2

1 School of Computing, Engineering, and Mathematics,

University of Western Sydney2 School of IT, University of Sydney

1

Page 2: A Distributed Energy Saving Approach for Ethernet Switches in Data Centers

Outline

Background Related work Main idea of our approach Details of our approach Evaluation Conclusions

2

Page 3: A Distributed Energy Saving Approach for Ethernet Switches in Data Centers

Energy consumption of data centers Data centers (DCs) contribute a significant

portion of global electricity consumption. DCs mainly consist of

Servers: consuming about 60% of the total energy Ethernet switches (shortened as switches

hereafter): consuming about 10%-20% Infrastructure(distribution, cooling, etc.): 20%-30%

This paper considers energy saving for switches

3

Page 4: A Distributed Energy Saving Approach for Ethernet Switches in Data Centers

Importance of saving energy in switches The percentage of energy consumed by switches

will grow considerably in the near future. To accommodate peak traffic load, the network

topologies of DCs have to employ a rich connectivity of switches. For instance, the fat tree topology

The peak-traffic time in DCs is much shorter than the non-peak time.

4

Page 5: A Distributed Energy Saving Approach for Ethernet Switches in Data Centers

The Fat Tree Topology Three levels of switches, bipartite graph in each

pod

Pod 0 Pod 1 Pod 2 Pod 3

switches

servers

5

Page 6: A Distributed Energy Saving Approach for Ethernet Switches in Data Centers

Outline

Background Related work Main idea of our approach Description of our approach Evaluation Conclusions

6

Page 7: A Distributed Energy Saving Approach for Ethernet Switches in Data Centers

Hardware features for energy saving There are altogether three hardware features

exploited by the existing approaches to save energy: Idle an entire switch

Idle a port

Adaptive Link Rate (ALR)

7

Page 8: A Distributed Energy Saving Approach for Ethernet Switches in Data Centers

Centralized approaches Most of the existing approaches are centralized

approaches.

They assume a central controller who knows global information including The entire network topology

The traffic rates among the servers in DCs

They obtain an NP-hard problem and propose heuristics to solve the problem.

They show that their heuristic algorithms can achieve close results to the optimal solutions.

8

Page 9: A Distributed Energy Saving Approach for Ethernet Switches in Data Centers

Limitations of centralized approaches Because a central controller needs to collect the global information and then distributes the calculated results, the centralized approaches suffer from the following limitations: They cannot quickly respond to the traffic change in

DCs

They are not scalable.

9

Page 10: A Distributed Energy Saving Approach for Ethernet Switches in Data Centers

Outline

Background Related work Main idea of our approach Description of our approach Evaluation Conclusions

10

Page 11: A Distributed Energy Saving Approach for Ethernet Switches in Data Centers

Power model IEEE approved the Energy Efficient Ethernet

as a standard (IEEE Std 802.3az) in 2010. In this standard, a new state for ports called

Low Power Idle (LPI) is introduced and the mechanisms for entering and exiting LPI are defined.

Therefore, we assume a port has two states: idle and active.

Also assume a switch has two states: idle and active.

11

Page 12: A Distributed Energy Saving Approach for Ethernet Switches in Data Centers

Power model (cont’d) The formula for calculating the power consumed

by a switch:

12

0

0

_

___

activeswidle

activeptactiveactiveptidleidleswbase

sw

nifP

nifPnPnP

P

There is a penalty for the port state transition between idle and active: this transition takes a short period of time, during which no packets can be transmitted and the same amount of power is consumed as the active state.

Page 13: A Distributed Energy Saving Approach for Ethernet Switches in Data Centers

We proposed a distributed approach We refer to our approach as eAware hereafter. The main idea of eAware is as follows:

Each switch monitors the queue lengths and utilizations at its ports;

If the queue length at a port exceeds certain threshold, certain other ports will be activated;

If a port has a zero queue length and its utilization is below certain threshold, this port will be made idle;

If all ports on a switch are idle, the entire switch will be made idle.

13

Page 14: A Distributed Energy Saving Approach for Ethernet Switches in Data Centers

Why examining queue length? Energy saving should not considerably impair the

network performance of DCs. The most important metrics on network

performance are: Packet end-to-end delay

Packet loss ratio

Both of them are closely related to queue length: Long queues lead to long end-to-end delays

If queue length is close to buffer capacity, packet drops are prone to happen.

14

Page 15: A Distributed Energy Saving Approach for Ethernet Switches in Data Centers

Why examining utilization when idling a port? A port can have a zero queue length but a high

utilization at the same time E.g., when a port transmits a network flow which has a

constant bit rate slightly less than the port capacity.

Measuring utilization needs a period of time, which can prevent ports from oscillating between idle and active states too frequently.

15

Page 16: A Distributed Energy Saving Approach for Ethernet Switches in Data Centers

Outline Background Related work Main idea of our approach Details of our approach

Some concepts The increasing port algorithm The decreasing port algorithm

Evaluation Conclusions

16

Page 17: A Distributed Energy Saving Approach for Ethernet Switches in Data Centers

Some concepts Port order: we stipulate an order on the ports of a

switch as their order of appearance from left to right in a network topology. We require the network topology satisfies that if port i

appears left to port j in a switch, port i will connect to a switch appearing left to the switch that port j connects to.

The idea here is to consolidate traffic to the ports and switches in the left, such that the ports and switches in the right can idle to save energy.

Port ID: to indicate the above order, if a port p appears left to a port q on a switch, port p has a smaller ID than port q.

17

Page 18: A Distributed Energy Saving Approach for Ethernet Switches in Data Centers

Some concepts (cont’d) port utilization: the traffic transmission rate at a port

during certain period divided by the capacity of this port. Its value ranges between 0 and 1.

utilization period: the period for measuring the port utilization.

high_length: the threshold of the queue length to trigger the increase of active ports.

low_util: the threshold of the port utilization to trigger the decrease of active ports when the queue length is zero.

18

Page 19: A Distributed Energy Saving Approach for Ethernet Switches in Data Centers

Some concepts (cont’d)

port state constraints: (1) if port u becomes active, port v must also become active; (2) port u can idle only when port v can also idle. The signaling between ports u and v to maintain these

two constraints is supported by the link layer mechanisms in IEEE Std 802.3az.

Separation of routing algorithm and our algorithms

The routing algorithm will be unaware of the active/idle states of the ports.

19

U V

Page 20: A Distributed Energy Saving Approach for Ethernet Switches in Data Centers

Separation of routing algorithm and our algorithms

The routing algorithm maintains the routing table

20

Switchport i port j port k

output buffers

for portsi, j, k

A packet arrives, and ports i, j, k are returned from routing table

Our algorithms will maintain the active/idle states of ports and decide to which port to forward a packet

Page 21: A Distributed Energy Saving Approach for Ethernet Switches in Data Centers

The increasing port algorithm This algorithm is invoked after the routing table lookup for every packet, dealing with the increase of active ports.

21

Reasons for invocation for every packet: Only after the routing table is looked up for a packet, we

know which ports can be used to forward this packet and thus which ports should be made active to alleviate long queues if present.

The queuing delay of a packet should be reduced whenever possible.

Page 22: A Distributed Energy Saving Approach for Ethernet Switches in Data Centers

The increasing port algorithm (cont’d)

22

Input: (1) a packet and (2) the n ports returned for this packet from the routing table.

Output: (1) the port for forwarding this packet and (2) a port to activate if there should be one.

Page 23: A Distributed Energy Saving Approach for Ethernet Switches in Data Centers

The increasing port algorithm (cont’d)

1. Assume m of these n (m≤n) ports are currently in active state. One of these m active ports is selected as the port (denoted by port f hereafter) for forwarding this packet.

2. After sending this packet to the buffer of port f, if queue length of port f > high_length, go to step 3. Otherwise, the algorithm ends here.

3. If all n ports are currently active, the algorithm ends. Otherwise, determine the leftmost idle port (denoted by port u hereafter) among those n ports.

4. Set port u to active and also signal port v to become active.

23

Page 24: A Distributed Energy Saving Approach for Ethernet Switches in Data Centers

The decreasing port algorithm This algorithm is invoked every utilization period, dealing with the decrease of active ports.

This algorithm is triggered by a periodical timer instead of packet arrivals, because a port should be able to become idle even when no packet arrives.

24

Page 25: A Distributed Energy Saving Approach for Ethernet Switches in Data Centers

The decreasing port algorithm (cont’d)

25

Input: the queue length and utilization of each port on this switch. Output: the list of ports to idle.

1. For each port u in this switch, execute the following steps.

2. Calculate the utilization of port u.

3. If (port u is active ) and (queue length of port u = 0) and (utilization of port u < low_util):

a) Signal port v to see whether port v can idle.

b) If (queue length of port v = 0) and (utilization of port v < low_util), both ports u and v transit to idle state. Otherwise, both ports stay in active state.

Page 26: A Distributed Energy Saving Approach for Ethernet Switches in Data Centers

Outline

Background Related work Main idea of our approach Details of our approach Evaluation Conclusions

26

Page 27: A Distributed Energy Saving Approach for Ethernet Switches in Data Centers

Experiment Setup We implemented eAware using the network simulator ns-2,

and made the source codes available at my homepage.

We compare eAware with the approach described in the fat tree paper which considers no energy saving. Hereafter, we label this approach eOblivious.

Without loss of generality, the fat tree topology is used as the network topologies.

Pareto traffic generator shipped in ns-2 is used to generate the network traffic.

27

Page 28: A Distributed Energy Saving Approach for Ethernet Switches in Data Centers

A fat tree topology formed in ns-2

28

Page 29: A Distributed Energy Saving Approach for Ethernet Switches in Data Centers

Default experiment parameters

29

Page 30: A Distributed Energy Saving Approach for Ethernet Switches in Data Centers

Parameter variations number of pods: we use four values of 4, 6, 8, 10. Note

that a larger number of pods implies a greater redundancy of switches in a network.

traffic_rate: we use six values of 50M, 100M, 150M, 200M, 250M, 300M.

(high_length, low_util): we use three pairs of threshold values: (125 pkts, 0.25), (25 pkts, 0.05), and (5 pkts, 0.01). Note that the first pair of values favors the most on energy efficiency and the least on network performance, and the last pair, vice versa.

30

Page 31: A Distributed Energy Saving Approach for Ethernet Switches in Data Centers

Evaluation Metrics Percentage of energy saving

31

%100eObliviousbyenergy

eAwarebyenergyeObliviousbyenergyPES

Percentage of packet loss increase

Percentage of packet end-to-end delay increase

%100eObliviousdelayavg

eObliviousdelayavgeAwaredelayavgPDI

%100eObliviousratioloss

eObliviousratiolosseAwareratiolossPLI

Page 32: A Distributed Energy Saving Approach for Ethernet Switches in Data Centers

Percentage of energy saving (fixing thresholds)

32

Page 33: A Distributed Energy Saving Approach for Ethernet Switches in Data Centers

Percentage of energy saving (fixing k)

33

Page 34: A Distributed Energy Saving Approach for Ethernet Switches in Data Centers

Percentage of delay increase (fixing thresholds)

34

Page 35: A Distributed Energy Saving Approach for Ethernet Switches in Data Centers

Percentage of delay increase (fixing k)

35

Page 36: A Distributed Energy Saving Approach for Ethernet Switches in Data Centers

Percentage of loss increase (fixing thresholds)

36

Page 37: A Distributed Energy Saving Approach for Ethernet Switches in Data Centers

Percentage of loss increase (fixing k)

37

Page 38: A Distributed Energy Saving Approach for Ethernet Switches in Data Centers

Conclusions This paper proposed eAware, a distributed

approach based on queue length and utilization to idle/activate ports and switches to save energy for DCs.

Due to its distributed and localized nature, eAware can quickly adapt to the traffic changes in DCs and is also scalable.

38

Page 39: A Distributed Energy Saving Approach for Ethernet Switches in Data Centers

Conclusions (on experiments) EAware can save 30%-50% energy compared with the

energy oblivious approach. EAware increases the average packet end-to-end delay by

3%-20%, and the packet loss ratio by 0%-0.9%, thus only impairing the network performance slightly.

If the two thresholds high_length and low_util are set to very low values, eAware can achieve similar network performance to the energy oblivious approach but can still save energy in those long periods of low traffic load.

39

Page 40: A Distributed Energy Saving Approach for Ethernet Switches in Data Centers

Conclusions (a discussion) EAware is independent of the multipath routing protocols

adopted by DCs, which brings the advantages of: Simplicity Portability

A further reason that we do not combine eAware with the multipath routing protocol (e.g., making queue length and utilization part of the routing messages) is: The queue length and utilization at a switch port (especially the

former) can change quickly in a DC environment, while the period of sending routing messages (usually 30s) is too long to propagate these two metrics.

Need to experiment to find out.

Theoretically model the approach

40

Page 41: A Distributed Energy Saving Approach for Ethernet Switches in Data Centers

Thank you!

Questions and Suggestions?

41