• Keine Ergebnisse gefunden

2 Recap: No-Regret Learning

N/A
N/A
Protected

Academic year: 2022

Aktie "2 Recap: No-Regret Learning"

Copied!
5
0
0

Wird geladen.... (Jetzt Volltext ansehen)

Volltext

(1)

Max-Flow via Experts

Instructor: Thomas Kesselheim

Today, we will get to know another very surprising application of the experts framework.

We will use it to solve the Maximum-Flow Problem. Our algorithm will be slow but it follows the same pattern that the fastest known algorithms for this problem use.

1 Max-Flow Problem

We are given a graph G = (V, E) with edge capacities (ce)e∈E and a dedicated source node s ∈ V and sink node t ∈ V. Let P be the set of all paths from s to t. Our goal is to assign flow values (xP)P∈P to the s-t-paths such thatxP ≥0 for allP, no edge has more flow than its capacity, i.e., P

P:e∈PxP ≤ce for all e∈E, and P

P∈PxP is maximized.

This problem can also be stated as a linear program as follows.

maximize X

P∈P

xP subject to X

P:e∈P

xP ≤ce for all e∈E

xP ≥0 for all P ∈ P

2 Recap: No-Regret Learning

Let us quickly recap the framework of no-regret learning. We rephrase it slightly to better fit our needs for today. There are m actions (experts) we can choose from in every step. There is a sequence of initially unknown gain vectors g(1), . . . , g(T). Choosing action iin step tgives gain g(t)i ∈[0,1]. In step t, the algorithm first chooses a probability vector y(t), then it incurs gainPm

i=1y(t)i gi(t) and gets to know the entire vectorg(t). The regret of the algorithm is defined as

Regret(T) =G(Tmax)

T

X

t=1 m

X

i=1

yi(t)g(t)i ,

whereG(Tmax) = maxiPT t=1gi(t).

The Multiplicative Weights algorithm guarantees

T

X

t=1 m

X

i=1

yi(t)g(t)i ≥(1−η)G(Tmax) −lnm η . So, Regret(T)≤ηG(Tmax) +lnηm.

3 Algorithm Intuition

We design an algorithm based on the experts framework. It is, indeed, more or less the same algorithm that was proposed by Garg and K¨onemann, although they actually do not talk about regret. The algorithm actually works, just as it is, for multi-commodity flow.

(2)

The idea behind the algorithm is simple but maybe not intuitive. Like many other flow algorithms, we choose shortest paths from s tot and route as much flow along these edges as possible. The Edmonds-Karp algorithm chooses a path that minimizes the number of edges and then changes the network to a residual network. Our algorithm is different: It changes the lengths of the edges. At this point, the experts algorithm comes into play: We let it define the edge lengths. This is done by considering each edge as an expert and the probability that it puts on an expert as the respective edge length.

4 Flows and Edge Lengths

There is an important connection between edge lengths and flows, which we state in the following necessary condition for the existence of a flow.

Lemma 27.1. There is a flow of value F only if for all choices of edge lengths (ye)e∈E with P

e∈Eye= 1 there is a path P such that P

e∈P ye

ceF1. Proof. For any feasible LP solution x we have

X

P:e∈P

1

cexP ≤1 for alle∈E . This also implies

X

e∈E

ye

X

P:e∈P

1 ce

xP ≤X

e∈E

ye= 1 . We can also reorder the left-hand side to

X

e∈E

ye X

P:e∈P

1

cexP = X

P∈P

X

e∈P

ye ce

! xP .

IfP

e∈P ye

ce > F1 for all paths P, then this immediately implies that also X

P∈P

xP < F X

P∈P

X

e∈P

ye ce

!

xP ≤F .

The point is that this lemma is not only necessary but also sufficient. More on this later.

5 Algorithm

We now formally define the algorithm. We use an arbitrary experts algorithm. To avoid any confusion with the paths, we call the probability vectory(t) today.

• Fort= 1, . . . , T

– Get probability distribution y(t) from the experts algorithm.

– Compute P(t) as the shortest path with edge lengths yc(t)e

e

– Letc(t)= mine∈P(t)ce

– Let (x(t)P )P∈P be a vector such thatxP(t) =c(t) and xP = 0 for P 6=P(t).

(3)

– Return g(t) back to the experts algorithm, where ge(t)=

(c(t)

ce ife∈P(t) 0 otherwise

• Compute ¯x=PT

t=1x(t),G(Tmax) = maxe∈EPT t=1g(t)e

• Return x= 1

G(T)max

¯ x

Interestingly, using any no-regret algorithm, this algorithm always computes a 1−-approximate flow if the number of iterations, T, is chosen large enough.

Lemma 27.2. The algorithm computes a feasible flow x.

Proof. Note that

G(Tmax) = max

e∈E T

X

t=1

g(t)e = max

e∈E

X

P:e∈P

¯ xP .

So G(Tmax) is exactly the maximum factor by which ¯x exceeds an edge capacity. Therefore, it is clear that the flow x is feasible.

Lemma 27.3. The flow x has value at least F(1− 1

G(T)max

Regret(T)), where F is the value of an optimal flow.

Proof. By the regret definition

T

X

t=1

X

e∈E

y(t)e ge(t) = max

e∈E T

X

t=1

ge(t)−Regret(T)=G(Tmax) −Regret(T) . Furthermore, for all t

X

e∈E

ye(t)ge(t)= X

e∈P(t)

ye(t)c(t)

ce =c(t) X

e∈P(t)

ye(t)

ce .

Recall thatP(t)was is a shortest path with respect to edge lengths

y(t)e

ce

e∈E

. So, by Lemma 27.1,

X

e∈P(t)

y(t)e

ce ≤ 1 F . In combination, this gives us

1 F

T

X

t=1

c(t)≥G(Tmax) −Regret(T) . Note that

T

X

t=1

c(t)= X

P∈P

¯ xP and so

X

P∈P

xP = 1 G(Tmax)

T

X

t=1

c(t)≥F 1−Regret(T) G(Tmax)

! .

(4)

Note that this bound only is meaningful if G(Tmax) is large. Fortunately, this is true in our case.

Lemma 27.4. The gain vectorsg(1), . . . , g(T) generated by the algorithm fulfill G(Tmax) ≥ T

m .

Proof. Observe that in each step t there is an edgeesuch thatge(t)= 1, therefore G(Tmax) = max

e∈E T

X

t=1

g(t)e ≥ 1 m

X

e∈E T

X

t=1

ge(t) ≥ T m .

If we combine these lemmas, then as long as we use a no-regret algorithm, that is, Regret(T)= o(T), then the flow value approaches F asymptotically for larger and larger T.

6 Guarantee with Multiplicative Weights

Let us now derive a quantitative bound if we use Multiplicative Weights. It actually pays off to be a little careful and to not just use the O(√

Tlogm) regret guarantee. Recall that the regret guarantee in case ofm experts is

Regret(T)≤ηG(Tmax) +lnm η , so the above guarantee becomes

X

P∈P

xP ≥F 1−η− 1 G(Tmax)

lnm η

!

≥F

1−η−m T

lnm η

.

If we choose η= 2 and T = 42mlnm, thenP

P∈PxP ≥F(1−).

Theorem 27.5. With Multiplicative Weights, the algorithm computes a (1−)-approximate flow using 21mlnm shortest-path computations. It overall running time isO(1m2lnm).

7 What is really happening?

One may wonder: Why does this work? As often, the answer is simple and complicated at the same time: It is because of strong LP duality. The dual of the flow LP (in the path formulation above is)

minimize X

e∈E

ceze subject to X

e∈P

ze≥1 for all P ∈ P

ze≥0 for all e∈E

The experts algorithm tries to find a solution to the primal and the dual LP. It iteratively adapts the primal and dual solution in a way similar to the algorithm for online set cover that we saw earlier.

(5)

References

• Naveen Garg, Jochen K¨onemann: Faster and Simpler Algorithms for Multicommodity Flow and Other Fractional Packing Problems. FOCS 1998

• Sanjeev Arora, Elad Hazan, Satyen Kale: The Multiplicative Weights Update Method: a Meta-Algorithm and Applications. Theory of Computing 8(1): 121-164 (2012): Survey on Multiplicative Weights Technique including this algorithm and others

Referenzen

ÄHNLICHE DOKUMENTE

We mention theoretical work on prediction properties of Pad´e approximants and related algorithms like the epsilon algorithm, and the iterated Aitken and Theta algorithms

For not only are they able to give free reign to their imagination and creativity to create the kind of algorithm visualization they want, their construction

In this context, Poland should push forward a euro- accession roadmap by the end of the year and ensure dialogue with both France and Germany to influence a

E. 2 we apply local improvement on solutions generated by the crossover operator. The MA al- ternates between two neighborhood structures until no further improvements can be

The challenge lies in mapping the different virtual networks with all the resources they require into the available physical network, which is the core of the Virtual Network

We then analysed publicly available human breast cancer data for TFAP2A expression and also studied the expression (by mRNA sequencing), activity (by monitoring the expression of

According to our proposed approach, during DDoS attack scenarios, most packets are generated by attackers, so the packet score value is higher than the packet score value during

broad suite of approaches with both observations and models to broad suite of approaches with both observations and models to examine both ocean carbon uptake and ocean carbon