• Aucun résultat trouvé

Scheduling Games in the Dark Nguyen Kim Thang (joint work with Christoph Durr)

N/A
N/A
Protected

Academic year: 2022

Partager "Scheduling Games in the Dark Nguyen Kim Thang (joint work with Christoph Durr)"

Copied!
77
0
0

Texte intégral

(1)

Scheduling Games in the Dark

Nguyen Kim Thang

(joint work with Christoph Durr)

Aarhus

March 9th, 09

Tuesday, November 16, 2010

(2)

Outline

Scheduling Games

Definition & Motivation Summary of results

Existence of pure Nash equilibrium

Inefficiency of equilibria

Potential argument

Price of anarchy

(3)

Scheduling Games

Tuesday, November 16, 2010

(4)

Scheduling Games

jobs (players) and machines: a job chooses a machine to execute. The processing time of job on machine is

n m

i j pij

(5)

Scheduling Games

jobs (players) and machines: a job chooses a machine to execute. The processing time of job on machine is

n m

i j pij

Such a job-machine assignment is a strategy profile. σ The load of machine is j !j = !

i:σ(i)=j

pij

Tuesday, November 16, 2010

(6)

Scheduling Games

Each machine specifies a policy how jobs assigned to the machine are to be scheduled (e.g., SPT, LPT, ...).

jobs (players) and machines: a job chooses a machine to execute. The processing time of job on machine is

n m

i j pij

The cost of a job is its completion time. ci i

Such a job-machine assignment is a strategy profile. σ The load of machine is j !j = !

i:σ(i)=j

pij

(7)

Coordination Mechanism

Tuesday, November 16, 2010

(8)

Coordination Mechanism

Coordination mechanism aim to design policies such that the self-interested behaviors lead to equilibria with small PoA.

(9)

Coordination Mechanism

Coordination mechanism aim to design policies such that the self-interested behaviors lead to equilibria with small PoA.

Communication is hard or costly (large-scale autonomous systems in the Internet).

Tuesday, November 16, 2010

(10)

Coordination Mechanism

Policies are designed based on local information.

Coordination mechanism aim to design policies such that the self-interested behaviors lead to equilibria with small PoA.

Communication is hard or costly (large-scale autonomous systems in the Internet).

(11)

Coordination Mechanism

Policies are designed based on local information.

Coordination mechanism aim to design policies such that the self-interested behaviors lead to equilibria with small PoA.

Communication is hard or costly (large-scale autonomous systems in the Internet).

Strongly local policy: a machine looks only at proc. time of jobs assigned to the machine.

σ(i) = j pij

Tuesday, November 16, 2010

(12)

Coordination Mechanism

Policies are designed based on local information.

Coordination mechanism aim to design policies such that the self-interested behaviors lead to equilibria with small PoA.

Communication is hard or costly (large-scale autonomous systems in the Internet).

Strongly local policy: a machine looks only at proc. time of jobs assigned to the machine.

σ(i) = j pij

Local policy: depends only on the parameters

(13)

Policies

Strongly local policies:

Tuesday, November 16, 2010

(14)

Policies

Strongly local policies:

Shortest Processing Time First (SPT)

(15)

Policies

Strongly local policies:

Shortest Processing Time First (SPT) machine 1

machine 2 machine 3

Tuesday, November 16, 2010

(16)

Policies

Strongly local policies:

Shortest Processing Time First (SPT) machine 1

machine 2 machine 3

(17)

Policies

Strongly local policies:

Shortest Processing Time First (SPT) machine 1

machine 2 machine 3

Longest Processing Time First (LPT)

Tuesday, November 16, 2010

(18)

Policies

Strongly local policies:

Shortest Processing Time First (SPT) machine 1

machine 2 machine 3

Longest Processing Time First (LPT) MAKESPAN: if

=

σ(i) = j

(19)

Policies

Tuesday, November 16, 2010

(20)

Policies

Local policies:

Inefficiency-based policy: greedily schedule jobs in increasing order of where ρi = pij/qi

qi = min

j! pij!

(21)

Policies

Local policies:

Inefficiency-based policy: greedily schedule jobs in increasing order of where ρi = pij/qi

qi = min

j! pij!

Typically, a policy depends on the processing time of jobs assigned to the machine.

Tuesday, November 16, 2010

(22)

Non-clairvoyant policies

(23)

Non-clairvoyant policies

What about policies that do not require this knowledge?

Incomplete information games Private information of jobs

Jobs can influence on their completion time by misreporting their processing time

Tuesday, November 16, 2010

(24)

Non-clairvoyant policies

What about policies that do not require this knowledge?

Incomplete information games Private information of jobs

Jobs can influence on their completion time by misreporting their processing time

Non-clairvoyant policies existence

Nash equilibrium small PoA

(25)

Natural policies

RANDOM: schedules jobs in a random order.

Tuesday, November 16, 2010

(26)

Natural policies

RANDOM: schedules jobs in a random order.

ci = pij + 1 2

!

i!:σ(i!)=j,i!!=i

pi!j

σ i j

In the strategy profile , is assigned to

= 1

2 (pij + !j)

(27)

Natural policies

RANDOM: schedules jobs in a random order.

A job on machine has an incentive to move to machine iff: i j! j

pij + !j > 2pij! + !j! ci = pij + 1

2

!

i!:σ(i!)=j,i!!=i

pi!j

σ i j

In the strategy profile , is assigned to

= 1

2 (pij + !j)

Tuesday, November 16, 2010

(28)

Natural policies

EQUI: schedules jobs in parallel, assigning each job an equal fraction of the processor.

(29)

Natural policies

EQUI: schedules jobs in parallel, assigning each job an equal fraction of the processor.

A B C

C D

pA = 1 pB = 1 pC = 2

pD = 3

0 4 6 7

Tuesday, November 16, 2010

(30)

Natural policies

EQUI: schedules jobs in parallel, assigning each job an equal fraction of the processor.

ci = p1j + . . . + pi1,j + (k i + 1)pij

If there are jobs on machine s.t: k j p1j . . . pkj

A B C

C D

pA = 1 pB = 1 pC = 2

pD = 3

0 4 6 7

(31)

Natural policies

EQUI: schedules jobs in parallel, assigning each job an equal fraction of the processor.

ci = p1j + . . . + pi1,j + (k i + 1)pij

If there are jobs on machine s.t: k j p1j . . . pkj

A B C

C D

pA = 1 pB = 1 pC = 2

pD = 3

0 4 6 7

p1j . . . pij . . . pkj

sum number

Tuesday, November 16, 2010

(32)

Models

Def of models:

Def: A job is balanced if i max pij/ min pij 2

Identical machines: for some length Uniform machines: for some speed Unrelated machines: arbitrary

pij = pi ∀j pij = pi/sj

pij

sj pi

(33)

Existence of equilibrium

The game under EQUI policy is a potential game.

Theorem:

The game under RANDOM policy is a potential game for 2 unrelated machines with balanced jobs but it is not for more than 3 machines. For uniform machines,

balanced jobs, there always exists equilibrium.

Tuesday, November 16, 2010

(34)

Inefficiency

Theorem: For unrelated machines, the PoA of policy EQUI is at most 2m – interestingly, that matches the best clairvoyant policy.

PoA is not increased when processing times are unknown to the

machines.

the worst NE

PoA

OPT

(35)

Existence of equilibrium

Tuesday, November 16, 2010

(36)

Standard definitions

Def: A job is unhappy if it can decrease its cost by

changing the strategy (other players’ strategies are fixed) Def: a best response (best move) of a job is the

strategy which minimizes the cost of the job (while other players’ strategies are fixed)

Def: Best-response dynamic is a process that let an arbitrary unhappy job make a best response.

(37)

RANDOM, uniform machines

Theorem:

The game under RANDOM policy is a potential game for uniform machines with balanced jobs

(balanced speeds).

Tuesday, November 16, 2010

(38)

RANDOM, uniform machines

(39)

RANDOM, uniform machines

Jobs have length

Machines have speed

p1 p2 . . . pn

s1 s2 . . . sm

p

ij

= p

i

/s

j

Tuesday, November 16, 2010

(40)

RANDOM, uniform machines

Jobs have length

Machines have speed

p1 p2 . . . pn

s1 s2 . . . sm

p

ij

= p

i

/s

j

Lemma: Consider a job making a best move from to . If there is a new unhappy job with index greater than , then sa > sb

i i

a b

(41)

RANDOM, uniform machines

Jobs have length

Machines have speed

p1 p2 . . . pn

s1 s2 . . . sm

p

ij

= p

i

/s

j

Lemma: Consider a job making a best move from to . If there is a new unhappy job with index greater than , then sa > sb

i i

a b

Proof: let be a new unhappy job. i!

Tuesday, November 16, 2010

(42)

RANDOM, uniform machines

Jobs have length

Machines have speed

p1 p2 . . . pn

s1 s2 . . . sm

p

ij

= p

i

/s

j

Lemma: Consider a job making a best move from to . If there is a new unhappy job with index greater than , then sa > sb

i i

a b

Proof: let be a new unhappy job. i!

was happy on machine and now has an incentive to move to machine i

! c i!

a

(43)

RANDOM, uniform machines

Jobs have length

Machines have speed

p1 p2 . . . pn

s1 s2 . . . sm

p

ij

= p

i

/s

j

Lemma: Consider a job making a best move from to . If there is a new unhappy job with index greater than , then sa > sb

i i

a b

Proof: let be a new unhappy job. i!

was happy on machine and now has an incentive to move to machine i

! i!

c b

was happy on machine and now has an incentive to move to machine i

! c i!

a

Tuesday, November 16, 2010

(44)

RANDOM, uniform machines

Proof: let be a new unhappy job. i!

was happy on machine and now has an incentive to move to machine i

! c i!

a

!c + pi!/sc !b + 2pi!/sb

!c + pi!/sc > (!a pi/sa) + 2pi!/sa

!a + pi/sa > !b + 2pi/sb

(45)

RANDOM, uniform machines

Proof: let be a new unhappy job. i!

was happy on machine and now has an incentive to move to machine i

! c i!

a

!c + pi!/sc !b + 2pi!/sb

!c + pi!/sc > (!a pi/sa) + 2pi!/sa

!a + pi/sa > !b + 2pi/sb

(sa sb)(pi! pi) > 0 Hence,

Tuesday, November 16, 2010

(46)

RANDOM, uniform machines

Proof: let be a new unhappy job. i!

was happy on machine and now has an incentive to move to machine i

! i!

c b

(pi pi!)(2sb sc) > 0

(47)

RANDOM, uniform machines

Proof: let be a new unhappy job. i!

was happy on machine and now has an incentive to move to machine i

! i!

c b

(pi pi!)(2sb sc) > 0 The lemma follows.

Tuesday, November 16, 2010

(48)

Potential function

(49)

Potential function

Dynamic: among all unhappy jobs, let the one with the greatest index make a best move.

Tuesday, November 16, 2010

(50)

Potential function

Dynamic: among all unhappy jobs, let the one with the greatest index make a best move.

For any strategy profile , let be the unhappy job with

greatest index. σ t

fσ(i) =

!1 if 1 i t, 0 otherwise.

1 t 0

(51)

Potential function

Dynamic: among all unhappy jobs, let the one with the greatest index make a best move.

For any strategy profile , let be the unhappy job with

greatest index. σ t

fσ(i) =

!1 if 1 i t, 0 otherwise.

1 t 0

lex. decreasesΦ(σ) = (fσ(1), sσ(1), . . . , fσ(n), sσ(n))

Tuesday, November 16, 2010

(52)

Potential function

lex. decreasesΦ(σ) = (fσ(1), sσ(1), . . . , fσ(n), sσ(n))

(53)

Potential function

lex. decreasesΦ(σ) = (fσ(1), sσ(1), . . . , fσ(n), sσ(n)) is the unhappy player with greatest index in t! σ!

Tuesday, November 16, 2010

(54)

Potential function

lex. decreasesΦ(σ) = (fσ(1), sσ(1), . . . , fσ(n), sσ(n)) is the unhappy player with greatest index in t! σ!

t! < t

Φ(σ) = (1, sσ(1), . . . , 1, sσ(t!), 1, sσ(t!+1), . . . , 1, sσ(t), . . .) Φ(σ!) = (1, sσ(1), . . . , 1, sσ(t!), 0, sσ(t!+1), . . . , 0, sσ!(t), . . .)

If

0 0 1

1 t!

t

(55)

Potential function

lex. decreasesΦ(σ) = (fσ(1), sσ(1), . . . , fσ(n), sσ(n)) is the unhappy player with greatest index in t! σ!

t! < t

Φ(σ) = (1, sσ(1), . . . , 1, sσ(t!), 1, sσ(t!+1), . . . , 1, sσ(t), . . .) Φ(σ!) = (1, sσ(1), . . . , 1, sσ(t!), 0, sσ(t!+1), . . . , 0, sσ!(t), . . .)

If

0 0 1

1 t!

t

t! > t

Φ(σ) = (1, sσ(1), . . . , 1, sσ(t), . . . , 0, sσ(t!), . . .) Φ(σ!) = (1, sσ(1), . . . , 1, sσ!(t), . . . , 0, sσ(t!), . . .)

If

0 0 1

1 t!

t

Tuesday, November 16, 2010

(56)

Potential function

lex. decreasesΦ(σ) = (fσ(1), sσ(1), . . . , fσ(n), sσ(n)) is the unhappy player with greatest index in t! σ!

t! < t

Φ(σ) = (1, sσ(1), . . . , 1, sσ(t!), 1, sσ(t!+1), . . . , 1, sσ(t), . . .) Φ(σ!) = (1, sσ(1), . . . , 1, sσ(t!), 0, sσ(t!+1), . . . , 0, sσ!(t), . . .)

If

0 0 1

1 t!

t

t! > t

Φ(σ) = (1, sσ(1), . . . , 1, sσ(t), . . . , 0, sσ(t!), . . .) Φ(σ!) = (1, s , . . . , 1, s ! , . . . , 0, s ! , . . .)

If

0 0 1

1 t!

t

(57)

RANDOM, unrel. machines

Tuesday, November 16, 2010

(58)

RANDOM, unrel. machines

Theorem:

The game under RANDOM policy is a potential

game for 2 unrelated machines with balanced jobs but it is not for more than 3 machines.

(59)

RANDOM, unrel. machines

Theorem:

The game under RANDOM policy is a potential

game for 2 unrelated machines with balanced jobs but it is not for more than 3 machines.

Proof:

Let be the current profile.σ : {1, . . . , n} →{ 1, 2}

The following potential decreases strictly Φ = |!1 !2| + 3 !

i

max{piσ(i) piσ(i), 0}

Tuesday, November 16, 2010

(60)

EQUI

(61)

EQUI

Theorem:

The game under EQUI policy is a strong potential game.

Tuesday, November 16, 2010

(62)

EQUI

Theorem:

The game under EQUI policy is a strong potential game.

Proof:

Let be the current profile.

The following exact potential decreases strictly σ

Φ = 1 2

!(ci + piσ(i))

(63)

Inefficiency of equilibria

Tuesday, November 16, 2010

(64)

Inefficiency

Theorem: For unrelated machines, the PoA of policy EQUI is at most 2m.

The knowledge about jobs’ characteristics is not necessarily needed.

the worst NE

PoA

OPT

(65)

Proof (sketch)

Tuesday, November 16, 2010

(66)

Proof (sketch)

ci = p1j + . . . + pi1,j + (k i + 1)pij

If there are jobs on machine s.t: k j p1j . . . pkj

(67)

Proof (sketch)

Proof: qi := min

j pij

Q(i) := arg min

j pij

!n

i=1

qi m · OP T

Rename jobs such that q1 q2 . . . qn ci = p1j + . . . + pi1,j + (k i + 1)pij

If there are jobs on machine s.t: k j p1j . . . pkj

Tuesday, November 16, 2010

(68)

Proof (sketch)

Proof: qi := min

j pij

Q(i) := arg min

j pij

!n

i=1

qi m · OP T

Rename jobs such that q1 q2 . . . qn Lemma: In any NE,

ci 2q1 + . . . + 2qi1 + (n i + 1)qi ci = p1j + . . . + pi1,j + (k i + 1)pij

If there are jobs on machine s.t: k j p1j . . . pkj

(69)

Proof (sketch)

Proof: qi := min

j pij

Q(i) := arg min

j pij

!n

i=1

qi m · OP T

Rename jobs such that q1 q2 . . . qn Lemma: In any NE,

ci 2q1 + . . . + 2qi1 + (n i + 1)qi ci = p1j + . . . + pi1,j + (k i + 1)pij

If there are jobs on machine s.t: k j p1j . . . pkj

Proof: c1 nq1

c2 q1 + (n 1)q2

= the worst cost on Q(1)

= the worst cost on Q(2)

Tuesday, November 16, 2010

(70)

Proof (sketch)

By monotonicity of (qi)ni=1 makespan = max

i ci

max

i (2q1 + . . . 2qi + (n i + 1)qi)

2 !

i

qi

2m · OP T P oA 2m

(71)

Lower bound

Theorem: The strong PoA of EQUI is at least (m+1)/4.

Tuesday, November 16, 2010

(72)

Lower bound

Theorem: The strong PoA of EQUI is at least (m+1)/4.

Proof (idea):

groups of jobs m J1, . . . , Jm

Jobs in can be scheduled only on machines Ji i, i + 1

(73)

Lower bound

Theorem: The strong PoA of EQUI is at least (m+1)/4.

Proof (idea):

groups of jobs m J1, . . . , Jm

Jobs in can be scheduled only on machines Ji i, i + 1 OPT

J1 J2

Tuesday, November 16, 2010

(74)

Lower bound

Theorem: The strong PoA of EQUI is at least (m+1)/4.

Proof (idea):

groups of jobs m J1, . . . , Jm

Jobs in can be scheduled only on machines Ji i, i + 1 NE

J2

OPT

(75)

Conclusion

Knowledge of jobs’ characteristics is not necessarily needed while restricting to strongly local policies.

Tuesday, November 16, 2010

(76)

Conclusion

Knowledge of jobs’ characteristics is not necessarily needed while restricting to strongly local policies.

Study the existence of equilibrium for RANDOM in two unrelated machines and in uniform machines.

(77)

Conclusion

Knowledge of jobs’ characteristics is not necessarily needed while restricting to strongly local policies.

Designing local policy with PoA = o(log m)

Study the existence of equilibrium for RANDOM in two unrelated machines and in uniform machines.

Tuesday, November 16, 2010

Références

Documents relatifs

The present paper presents an extension of this framework by introducing a differentiation within intermediate goods sector. Indeed, we assume that there are two upstream sectors:

(2016) Learning motor coordination under resistive viscous force fields at the joint level with an upper-limb robotic exoskeleton.. Converging Clinical and Engineering Research

Under those assumptions, we can distinguish two main phases the eective game will go through: an initial phase which can be described by the Gaussian ansatz introduced section 4.1

In section III we then consider the corresponding ergodic state which role in the long optimization time limit is fundamental both because it describes a signicant part of the

L’archive ouverte pluridisciplinaire HAL, est destinée au dépôt et à la diffusion de documents scientifiques de niveau recherche, publiés ou non, émanant des

the role of inequality is subtle. Allowing subjects to choose an option that provides equal payoffs to both players clearly increases the mechanism effectiveness, presumably because

The overall objective of these projects is to lead tasks as closely as possible to the sector in order to propose concrete solutions (methods and tools software) to

Monday, August 27 – Symposia presentations 2 Monday, August 27 – Oral presentations 9 Tuesday, August 28 – Poster presentations 14 Wednesday, August 29 – Symposia presentations