16
Ahmed M. Bedewy Joint work with Yin Sun, Ness B. Shroff Departments of ECE, The Ohio State University Nov. 7 th 2016 Optimizing Data Freshness, Throughput, and Delay in Multi-Server Information-Update Systems IEEE ISIT 2016

Optimizing Data Freshness, Throughput, and Delay in Multi

  • Upload
    others

  • View
    5

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Optimizing Data Freshness, Throughput, and Delay in Multi

Ahmed M. Bedewy

Joint work with Yin Sun, Ness B. Shroff Departments of ECE,

The Ohio State University

Nov. 7th 2016

Optimizing Data Freshness, Throughput, and Delay in Multi-Server Information-Update Systems

IEEE ISIT 2016

Page 2: Optimizing Data Freshness, Throughput, and Delay in Multi

Update is generated at time

What is the Age of Information?

1(14)

time

0

age of info/data staleness

A stream of messages generated at an information source

To be sent to a destination via multiple communication channels/servers

Dii

Di Di+1

45o

�(t)

t

and delivered at time

commun. servers

information source

user screen �(0)

Information-update system

Definition: at any time , the age-of-information is the “age” of the freshest message available at the destination

t �(t)

si

�(t) = t�max{si : Di t}

si si+1

Queue

Page 3: Optimizing Data Freshness, Throughput, and Delay in Multi

Difference between Delay & Age

2(14)

Age is different from delay. [Kaul, Yates, Gruteser, Infocom 2012]

• High arrival rate: long waiting time, packets becomes stale • Low arrival rate: short queue, but infrequent updates High age

(

• For delay, Little’s Law says: • Delay grows linearly wrt queue length

L = �W

High arrival rate Low arrival rate

Age is not monotonic wrt queue length

Page 4: Optimizing Data Freshness, Throughput, and Delay in Multi

Open Questions

3(14)

info arrivals server

Enqueue-and-forward model

Age Characterization and Age Reduction: • M/M/1, M/D/1, D/M/1 [Kaul, Yates, Gruteser 2012] • M/M/2 [Kam, Kompella, Ephremides 2014] • Multi-sources [Yates, Kaul ISIT 2012] [Huang, Modiano 2015] • Packet management, LCFS [Kam, Kompella, Ephremides

2013, 2014] • Channel state info [Costa, Valentin, Ephremides, 2015] • LCFS (single server) with & without preemption [Kaul, Yates,

Gruteser 2012]

Question 1: Which policy is age-optimal?

Page 5: Optimizing Data Freshness, Throughput, and Delay in Multi

Open Questions (cont.)

4(14)

Question 2: Is there a policy which simultaneously optimizes age, throughput, and delay?

Information Updates • News, emails, notifications,

stock quotes, …

Users may be interested in not just the latest updates, but also past news.

• This presentation will answer these two questions

Page 6: Optimizing Data Freshness, Throughput, and Delay in Multi

System Model

5(14)

!"#$%&'$%(&)*"*"+,*--".+

/".0".+1

/".0".+2

/".0".+3

4&5(1%&6*78'$"#

An information-update system with • m i.i.d. servers • Queue with buffer size

Packet is generated at time ,and arrives at time ( ) • Arbitrary arrival process (including non-stationary) • Update packets can arrive out of order (e.g., but )

The set of all causal policies is denoted by

i si ai

ai > ai+1

B

si < si+1

s1 s2 . . .

Page 7: Optimizing Data Freshness, Throughput, and Delay in Multi

• The preempted packet will be stored back into the queue

• A server can switch to send any packet Definition. Service Preemption: At any time

Definitions

6(14)

• To be sent at a later time when the servers are available again.

Definition. Stochastic Ordering: Let and be two random variables. Then, is said to be stochastically smaller than (denoted as ), if

X YX Y X st Y

In other words: iff there exist two random variable and , defined on the same probability space, such that

and

X st Y X Y

Y =st Y X =st X&

(here denotes equality in law)=st

P{X > x} P{Y > x}, 8x 2 R.

P{X Y } = 1

Page 8: Optimizing Data Freshness, Throughput, and Delay in Multi

Definitions

7(14)

Definition. Stochastic Ordering of Stochastic Processes: Two random processes and satisfies

iff there exist two random processes and , defined on the same probability space, such that

and

{X(t), t 2 [0,1)} {Y (t), t 2 [0,1)}

{X(t), t 2 [0,1)} st {Y (t), t 2 [0,1)}

Definition. Age Optimality: A policy is said to be age-optimal, if for all ⇡ 2 ⇧

{X(t), t 2 [0,1)} {Y (t), t 2 [0,1)}

{Y (t), t 2 [0,1)} =st {Y (t), t 2 [0,1)} {X(t), t 2 [0,1)} =st {X(t), t 2 [0,1)}&

� 2 ⇧

{��(t), t 2 [0,1)} st {�⇡(t), t 2 [0,1)}.

P[X(t) Y (t), t 2 [0,1)] = 1

Page 9: Optimizing Data Freshness, Throughput, and Delay in Multi

Policy P Algorithm

8(14)

Arrival• Stale packet • Fresh packet

s2

s3

s1 s1

s2

s3

Departure

s2

s3

s1

Policy P is serving the freshest packets among all arrived packets. Preempted packet is stored back in the queue to preserve the throughput.

Policy P: Preemptive Last Generated First Served (LGFS)

Page 10: Optimizing Data Freshness, Throughput, and Delay in Multi

• is the largest time stamp of the delivered packets

• is the i-th smallest time stamp of the packets being transmitted

Theorem 1: For

The preemptive LGFS policy is age-optimal.

Main Theorem

9(14)

The system state of policy is , where⇡ V⇡(t) = (U⇡(t),↵1,⇡(t), . . . ,↵m,⇡(t))

U⇡(t)

↵i,⇡(t)

1. i.i.d. exponential service time distribution 2. any packet generation and arrival times 3. any buffer sizes B

3-Servers system

Destination

U⇡

↵1,⇡

↵2,⇡

↵3,⇡

s3

s4

s2

s1

Page 11: Optimizing Data Freshness, Throughput, and Delay in Multi

Proof sketch

10(14)

Step 1: Construction• •

Step 2: Coupling

Step 3: Forward induction in time

↵1,P

↵i,P

↵m,P

Policy P Policy ⇡

↵m,⇡

↵i,⇡

↵1,⇡

{VP (t), t 2 [0,1)}�st {V⇡(t), t 2 [0,1)}. 8⇡ 2 ⇧

Def.

{�P (t), t 2 [0,1)}st{�⇡(t), t 2 [0,1)}, 8⇡ 2 ⇧

• Prove

V⇡(t) = (U⇡(t),↵1,⇡(t), . . . ,↵m,⇡(t))

Exponential service time distribution is memory-less

{VP (t), t 2 [0,1)} =st {VP (t), t 2 [0,1)}.{V⇡(t), t 2 [0,1)} =st {V⇡(t), t 2 [0,1)}.

{VP (t), t 2 [0,1)} =st {VP (t), t 2 [0,1)}.

{V⇡(t), t 2 [0,1)} =st {V⇡(t), t 2 [0,1)}.

P[VP (t) � V⇡(t), t 2 [0,1)] = 1

Policy vs policy P ⇡

Page 12: Optimizing Data Freshness, Throughput, and Delay in Multi

Corollary: For the same system setting as Theorem 1, the preemptive LGFS policy minimizes:

1. The time-average age 2. Average peak age 3. Time-average age penalty

Corollary: The age performance of the preemptive LGFS policy remains the same for any queue size

Corollaries

11(14)

B � 0

• Stale packets are stored back in the queue.

• These age metrics are increasing function of the age process.

Page 13: Optimizing Data Freshness, Throughput, and Delay in Multi

Theorem

12(14)

Theorem 2: For

The preemptive LGFS policy is throughput-optimal and mean-delay-optimal among all policies in .

B = 1

1. i.i.d. exponential service time distribution 2. Any packet generation and arrival times 3. Infinite buffer size

Page 14: Optimizing Data Freshness, Throughput, and Delay in Multi

Simulation Result

13(14)

servers, buffer size B m = 5inter-generation times: i.i.d. Erlang-2 distribution is modeled to be either 1 or 100 with equal probability(ai � si)

0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1ρ

1.5

2

2.5

3

3.5

4

4.5

5

5.5

6

6.5

Aver

age

Age

FCFS, B=∞FCFS, B=10FCFS, B=1Non-preemptive LGFS, B=1Preemptive LGFS, any B≥ 0

Observations: 1. Preemptive LGFS outperforms all other policies. 2. The age performance of the preemptive LGFS is invariant for any B 3. FCFS: The age gets worse as B increases.

Invariant with B

Page 15: Optimizing Data Freshness, Throughput, and Delay in Multi

Summary & Future Work

14(14)

Exponential service time. The preemptive LGFS optimizes age, throughput and delay among all causally feasible policies for • Arbitrary packet generation times • Arbitrary arrival process (could be non-stationary, non-

ergodic, out of order arrivals) • Any number of servers

Other service time distributions?

Service time dist. ExponentialNWU

Hyperexponential distribution

NBUErlang

distribution

Preemptive LGFS Age-optimal (any arrival orders)

Not age-optimal in the same policy

space

Non-preemptive LGFS Near age-optimal Near age-optimal

s1, s2, . . .a1, a2, . . .

Page 16: Optimizing Data Freshness, Throughput, and Delay in Multi