• Keine Ergebnisse gefunden

Planning and Optimization D2. Abstractions: Additive Abstractions Gabriele R¨oger and Thomas Keller

N/A
N/A
Protected

Academic year: 2022

Aktie "Planning and Optimization D2. Abstractions: Additive Abstractions Gabriele R¨oger and Thomas Keller"

Copied!
31
0
0

Wird geladen.... (Jetzt Volltext ansehen)

Volltext

(1)

D2. Abstractions: Additive Abstractions

Gabriele R¨oger and Thomas Keller

Universit¨at Basel

October 29, 2018

(2)

Content of this Course

Planning

Classical

Tasks Progression/

Regression Complexity Heuristics

Probabilistic

MDPs Uninformed Search

Heuristic Search Monte-Carlo

Methods

(3)

Content of this Course: Heuristics

Heuristics

Delete Relaxation

Abstraction

Abstractions in General

Pattern Databases

Merge &

Shrink Landmarks

Potential Heuristics Cost Partitioning

(4)

Multiple Abstractions

(5)

Multiple Abstractions

One important practical question is how to come up with a suitable abstraction mapping α.

Indeed, there is usually ahuge number of possibilities, and it is important to pick good abstractions

(i.e., ones that lead to informative heuristics).

However, it is generallynot necessary to commit to a single abstraction.

(6)

Combining Multiple Abstractions

Maximizingseveral abstractions:

Each abstraction mapping gives rise to an admissible heuristic.

By computing the maximumof several admissible heuristics, we obtain another admissible heuristic which dominates the component heuristics.

Thus, we can always compute several abstractions and maximize over the individual abstract goal distances.

Addingseveral abstractions:

In some cases, we can even compute thesum of individual estimates and still stay admissible.

Summation often leads to much higher estimates than maximization, so it is important to understand

under which conditions summation of heuristics is admissible.

(7)

Adding Several Abstractions: Example (1)

LRR LLL LLR

LRL

ALR

ALL

BLL

BRL

ARL

ARR

BRR

BLR

RRR RRL

RLR

RLL

h(LRR) = 4

(8)

Adding Several Abstractions: Example (2)

LRR LLR

LLL

LRL LLR

LRL LLL

ALR ARL

ALL ARR

BLL

BRL BRR

BLR ALR ARL

ARR ALL

BLL BRR

BLR BRL

RRR RRL

RLR RLLRLL RRL

RLR RRR

hα1(LRR) = 3

LRR LLR

LLL

LRL ALR

ALL

BLL

BRL LLR

LLL

LRL ALR

ALL

BLL

BRL ARL

ARR

BLR BRR

RRR RRL

RLR RLL ARL

ARR

BLR BRR

RRR RRL

RLR RLL

hα2(LRR) = 2

(9)

Adding Several Abstractions: Example (3)

LRR LLR

LLL

LRL LLR

LRL LLL

ALR ARL

ALL ARR

BLL

BRL BRR

BLR ALR ARL

ARR ALL

BLL BRR

BLR BRL

RRR RRL

RLR RLLRLL RRL

RLR RRR

hα1(LRR) = 3

LRR LLR

LLL

LRL ALR

ALL

BLL

BRL LRR

LLR

LLL

LRL ALR

ALL

BLL

BRL ARL

ARR

BLR BRR

RRR RRL

RLR RLL ARL

ARR

BLR BRR

RRR RRL

RLR RLL

hα2(LRR) = 1

(10)

Additivity

(11)

Orthogonality of Abstractions

Definition (Orthogonal)

Letα1 andα2 be abstractions of transition system T.

We say thatα1 andα2 are orthogonalif for all transitions s −→` t ofT, we haveαi(s) =αi(t) for at least one i ∈ {1,2}.

(12)

Affecting Transition Labels

Definition (Affecting Transition Labels)

LetT be a transition system, and let` be one of its labels.

We say that`affectsT ifT has a transitions −→` t with s 6=t.

Theorem (Affecting Labels vs. Orthogonality)

Letα1 andα2 be abstractions of transition system T. If no label ofT affects both Tα1 andTα2,

thenα1 andα2 are orthogonal.

(Easy proof omitted.)

(13)

Orthogonality and Additivity

Theorem (Additivity for Orthogonal Abstractions)

Let hα1, . . . ,hαn be abstraction heuristics of the same transition system such thatαi andαj are orthogonal for all i 6=j .

ThenPn

i=1hαi is a safe, goal-aware, admissible and consistent heuristic forΠ.

(14)

Orthogonality and Additivity: Example (1)

LRR LLR

LLL

LRL LLR

LRL LLL

ALR ARL

ALL ARR

BLL

BRL BRR

BLR ALR ARL

ARR ALL

BLL BRR

BLR BRL

RRR RRL

RLR RLLRLL RRL

RLR RRR

hα1(LRR) = 3

LRR LLR

LLL

LRL ALR

ALL

BLL

BRL LLR

LLL

LRL ALR

ALL

BLL

BRL ARL

ARR

BLR BRR

RRR RRL

RLR RLL ARL

ARR

BLR BRR

RRR RRL

RLR RLL

hα2(LRR) = 2

(15)

Orthogonality and Additivity: Example (2)

LRR LLR

LLL

LRL LLR

LRL LLL

ALR ARL

ALL ARR

BLL

BRL BRR

BLR ALR ARL

ARR ALL

BLL BRR

BLR BRL

RRR RRL

RLR RLLRLL RRL

RLR RRR

hα1(LRR) = 3

LRR LLR

LLL

LRL ALR

ALL

BLL

BRL LRR

LLR

LLL

LRL ALR

ALL

BLL

BRL ARL

ARR

BLR BRR

RRR RRL

RLR RLL ARL

ARR

BLR BRR

RRR RRL

RLR RLL

hα2(LRR) = 1

(16)

Orthogonality and Additivity: Proof (1)

Proof.

We prove goal-awareness and consistency;

the other properties follow from these two.

LetT =hS,L,c,T,s0,S?ibe the concrete transition system.

Leth=Pn i=1hαi.

Goal-awareness: For goal statess ∈S?, h(s) =Pn

i=1hαi(s) =Pn

i=10 = 0 because all individual

abstraction heuristics are goal-aware. . . .

(17)

Orthogonality and Additivity: Proof (1)

Proof.

We prove goal-awareness and consistency;

the other properties follow from these two.

LetT =hS,L,c,T,s0,S?ibe the concrete transition system.

Leth=Pn i=1hαi.

Goal-awareness: For goal states s ∈S?, h(s) =Pn

i=1hαi(s) =Pn

i=10 = 0 because all individual

abstraction heuristics are goal-aware. . . .

(18)

Orthogonality and Additivity: Proof (2)

Proof (continued).

Consistency: Let s −→o t ∈T. We must prove h(s)≤c(o) +h(t).

Because the abstractions are orthogonal,αi(s)6=αi(t) for at most onei ∈ {1, . . . ,n}.

Case 1: αi(s) =αi(t) for all i ∈ {1, . . . ,n}.

Thenh(s) =Pn

i=1hαi(s)

=Pn

i=1hTαii(s))

=Pn

i=1hTαii(t))

=Pn

i=1hαi(t)

=h(t)≤c(o) +h(t).

. . .

(19)

Orthogonality and Additivity: Proof (2)

Proof (continued).

Consistency: Let s −→o t ∈T. We must prove h(s)≤c(o) +h(t).

Because the abstractions are orthogonal,αi(s)6=αi(t) forat most one i ∈ {1, . . . ,n}.

Case 1: αi(s) =αi(t) for all i ∈ {1, . . . ,n}.

Thenh(s) =Pn

i=1hαi(s)

=Pn

i=1hTαii(s))

=Pn

i=1hTαii(t))

=Pn

i=1hαi(t)

=h(t)≤c(o) +h(t).

. . .

(20)

Orthogonality and Additivity: Proof (2)

Proof (continued).

Consistency: Let s −→o t ∈T. We must prove h(s)≤c(o) +h(t).

Because the abstractions are orthogonal,αi(s)6=αi(t) forat most one i ∈ {1, . . . ,n}.

Case 1: αi(s) =αi(t) for all i ∈ {1, . . . ,n}.

Thenh(s) =Pn

i=1hαi(s)

=Pn

i=1hTαii(s))

=Pn

i=1hTαii(t))

=Pn

i=1hαi(t)

=h(t)≤c(o) +h(t).

. . .

(21)

Orthogonality and Additivity: Proof (2)

Proof (continued).

Consistency: Let s −→o t ∈T. We must prove h(s)≤c(o) +h(t).

Because the abstractions are orthogonal,αi(s)6=αi(t) forat most one i ∈ {1, . . . ,n}.

Case 1: αi(s) =αi(t) for all i ∈ {1, . . . ,n}.

Thenh(s) =Pn

i=1hαi(s)

=Pn

i=1hTαii(s))

=Pn

i=1hTαii(t))

=Pn

i=1hαi(t)

=h(t)≤c(o) +h(t).

. . .

(22)

Orthogonality and Additivity: Proof (3)

Proof (continued).

Case 2: αi(s)6=αi(t) for exactly one i ∈ {1, . . . ,n}.

Letk ∈ {1, . . . ,n}such that αk(s)6=αk(t).

Thenh(s) =Pn

i=1hαi(s)

=P

i∈{1,...,n}\{k}hTαii(s)) +hαk(s)

≤P

i∈{1,...,n}\{k}hTαii(t)) +c(o) +hαk(t)

=c(o) +Pn

i=1hαi(t)

=c(o) +h(t),

where the inequality holds becauseαi(s) =αi(t) for all i 6=k andhαk is consistent.

(23)

Orthogonality and Additivity: Proof (3)

Proof (continued).

Case 2: αi(s)6=αi(t) for exactly one i ∈ {1, . . . ,n}.

Letk ∈ {1, . . . ,n}such that αk(s)6=αk(t).

Thenh(s) =Pn

i=1hαi(s)

=P

i∈{1,...,n}\{k}hTαii(s)) +hαk(s)

≤P

i∈{1,...,n}\{k}hTαii(t)) +c(o) +hαk(t)

=c(o) +Pn

i=1hαi(t)

=c(o) +h(t),

where the inequality holds becauseαi(s) =αi(t) for all i 6=k andhαk is consistent.

(24)

Outlook

(25)

Using Abstraction Heuristics in Practice

In practice, there are conflicting goals for abstractions:

we want to obtain an informative heuristic, but want to keep its representation small.

Abstractions have small representations if there are few abstract statesand there is a succinct encoding forα.

(26)

Counterexample: One-State Abstraction

LRR

LLR

LLL

LRL

ALR

ALL

BLL

BRL

ARL

ARR

BRR

BLR

RRR RRL

RLR LRR RLL

LLR

LLL

LRL

ALR

ALL

BLL

BRL

ARL

ARR

BRR

BLR

RRR RRL

RLR

RLL

One-state abstraction: α(s) := const.

+ very few abstract states andsuccinct encoding forα

− completely uninformative heuristic

(27)

Counterexample: Identity Abstraction

LRR LLL LLR

LRL

ALR

ALL

BLL

BRL

ARL

ARR

BRR

BLR

RRR RRL

RLR

RLL

Identity abstraction: α(s) :=s.

+ perfect heuristicand succinct encoding forα

− too many abstract states

(28)

Counterexample: Perfect Abstraction

LRR

LLR

LLL

LRL LLR

LRL LLL

ALR

ALL

BLL

BRL ALR

BRL ALL

BLL

ARL

ARR

BRR

BLR ARL

BLR ARR

BRR

RRR RRL

RLR

RLLRLL RRL

RLR RRR

Perfect abstraction: α(s) :=h(s).

+ perfect heuristicand usually few abstract states

− usually no succinct encoding forα

(29)

Automatically Deriving Good Abstraction Heuristics

Abstraction Heuristics for Planning: Main Research Problem Automatically derive effective abstraction heuristics

for planning tasks.

we will study two state-of-the-art approaches in Chapters D3–D8

(30)

Summary

(31)

Summary

Often, multiple abstractions are used.

They can always be maximizedadmissibly.

Adding abstraction heuristics is not always admissible.

When it is, it leads to a stronger heuristic than maximizing.

Abstraction heuristics fromorthogonal abstractions can be addedwithout losing admissibility or consistency.

One sufficient condition for orthogonality is that all abstractions areaffectedbydisjoint sets of labels.

Practically useful abstractions are those which give informative heuristics, yet have a small representation.

Coming up with good abstractions automatically is the main research challenge when applying abstraction heuristics in planning.

Referenzen

ÄHNLICHE DOKUMENTE

Abstractions: Formal Definition and Heuristics Homomorphisms and Isomorphisms. D1.2 Homomorphisms

Abstractions α induce abstraction heuristics h α : h α (s ) is the goal distance of α(s ) in the abstract transition system. Abstraction heuristics are safe, goal-aware, admissible

I Coming up with good abstractions automatically is the main research challenge when applying abstraction heuristics

D3.1 Projections and Pattern Database Heuristics D3.2 Implementing PDBs: Precomputation.. D3.3 Implementing PDBs: Lookup

Formally, pattern database heuristics are abstraction heuristics induced by a particular class of abstractions called projections.

I Hence we now look at the transition systems for atomic projections of our example task, including transition labels. I We abbreviate operator names as in

Hence we now look at the transition systems for atomic projections of our example task, including transition labels. We abbreviate operator names as in

The objective value of an integer program that minimizes this cost subject to the flow constraints is a lower bound on the plan cost (i.e., an admissible heuristic estimate)..