• Keine Ergebnisse gefunden

Efficient Epistemic Updates in Rank-based Belief Networks

N/A
N/A
Protected

Academic year: 2022

Aktie "Efficient Epistemic Updates in Rank-based Belief Networks"

Copied!
255
0
0

Wird geladen.... (Jetzt Volltext ansehen)

Volltext

(1)

E E U

R ank -B ased B elief N etworks

D issertation zur E rlangung des akademischen G rades eines D oktors der P hilosophie (D r . phil .)

an der

Geisteswissenschaftliche Sektion Fachbereich Philosophie

vorgelegt von Stefan Alexander Hohenadel

Tag der mündlichen Prüfung: 4 . September 2012

1 . Referent: Prof. Dr. Wolfgang Spohn

2 . Referent: PD Dr. Sven Kosub

(2)
(3)

E E U

R ank -B ased B elief N etworks

Stefan Alexander Hohenadel November 9 , 2013

A Dissertation Submitted in Partial Fulfillment of the Requirements for the Degree of Doctor in Philosophy.

Department of Philosophy

Faculty of Humanities

University of Konstanz

(4)
(5)

belief networks. The update is performed on two input values: the current doxastic state, rep- resented by the network, and, second, a doxastic evidence that is represented as a change on a subset of the variables in the network. From these inputs, a Lauritzen-Spiegelhalter-styled up- date strategy can compute the updated posterior doxastic state of the network. The posterior state reflects the combination of the evidence and the prior state. This strategy is well-known for Bayesian networks. The thesis transfers the strategy to those networks whose semantics is specified by epistemic ranking functions instead of probability measures. As a foundation, the construction of rank-based belief networks is discussed, which are graphical models for rank- ing functions. It is shown that global, local and pairwise Markov properties are equivalent in rank-based belief networks and, furthermore, that the Hammersley-Clifford-Theorem holds for such ranking networks. This means that from the equivalence of the Markov properties it follows that a potential representation of the actual ranking function can be derived from the network structure. It is shown how by this property the update strategy of the Lauritzen- Spiegelhalter-algorithm can be transferred to ranking networks. For this purpose, the solution of the two main problems is demonstrated: first, the triangulation of the moralized input network and the decompositon of this triangulation to a clique tree. Then, second, message passingcan be performed on this clique tree to incorporate the evidence into the clique tree.

The entire approach is in fact a technical description of belief revision.

Zusammenfassung– Diese Dissertation stellt einen Ansatz für die effiziente Aktualisierung von Rangfunktion-basierten doxastischen Netzwerken vor. Die Aktualisierung eines doxasti- schen Netzwerks erfolgt auf der Basis von zwei Eingaben: dem aktuellen doxastischen Zu- stand, repräsentiert durch das Netzwerk, sowie einer doxastischen Evidenz, die als Änderung von Aktualwerten einer Untermenge von Variablen des Netzwerks formalisiert wird. Aus diesen Eingaben kann mithilfe der Strategie des Algorithmus von Lauritzen & Spiegelhalter der aktualisierte Folgezustand des Netzwerks berechnet werden, in welchem die Evidenz re- flektiert ist. Dieses Vorgehen ist für Bayessche Netze bereits seit langem bekannt und wird hier auf Netze angewandt, deren Semantik statt durch Wahrscheinlichkeitsmaße durch Rang- funktionen spezifiziert ist. Als Grundlegung wird auch die Bildung grafischer Modelle für Rangfunktionen, also Rangfunktion-basierter doxastischer Netzwerke ausführlich diskutiert.

Unter anderem wird dabei gezeigt, dass globale, lokale und paarweise Markov-Eigenschaften in Rangfunktion-basierten Netzwerken äquivalent sind und dass weiterhin für solche Rang- netzwerke das Hammersley-Clifford-Theorem gilt. Dies bedeutet, dass aus der Äquivalenz der genannten Markov-Eigenschaften folgt, dass stets eine cliquen-basierte Potentialdarstel- lung der jeweiligen Rangfunktion aus der Netzwerkstruktur abgeleitet werden kann. Es wird gezeigt, wie durch diese Eigenschaft die Aktualisierungsstrategie des Lauritzen-Spiegelhalter- Algorithmus auf Rangfunktion-basierte Netzwerke übertragen werden kann. Hierzu wird die Lösung der beiden Hauptaufgaben gezeigt: Zunächst ist die Triangulierung des moralisierten Netzwerks sowie die Dekomposition des triangulierten Netzwerks zu einem Cliquenbaum auszuführen, danach kann auf dem so gewonnenen Cliquenbaummessage passingausgeführt werden, um die doxastische Evidenz in den Cliquenbaum zu inkorporieren. Dies entspricht einer technischen Darstellung des Revisionsvorgangs von Überzeugungszuständen.

(6)
(7)

C

Preface . . . 9 I Belief, BeliefStates,andBeliefChange . . . 15

1.1Introduction·151.1.1 Remark on Sources and Citation·16 – 1.2A Normative Per- spective on Epistemology·161.2.1 Descriptive and Normative Perspective·161.2.2 The Philosopher and the Engineer·171.2.3 Belief Revision and the Problem of Induction·23 – 1.3 Elements of a Belief Theory · 26 – 1.4 Propositions as Epistemic Units · 281.4.1The Concept of Proposition·281.4.2Propositions Form Algebras·291.4.3 Atoms and Atomic Algebras·321.4.4 Beliefs, Contents, and Concepts·33 – 1.5Epistemic States and Rationality·371.5.1 Rationality Postulates·371.5.2 Rational Belief Sets·401.5.3 Belief Cores·41 – 1.6 Transitions Between Epistemic States ·421.6.1 Descrip- tion of Epistemic Updates · 421.6.2 Transition by Consistent Evidence· 431.6.3 The Inconsistent Case·441.6.4 The Transition Function·45

II RankingFunctions andRank-basedConditionalIndependence. . . 47 2.1 Introduction · 47 – 2.2 Ranking Functions · 482.2.1 Ranking Functions on Pos- sibilities ·482.2.2 Negative Ranking Functions·492.2.3 Minimitivity, Completeness, and Naturalness ·522.2.4 Two-sided Ranking Functions ·562.2.5 Conditional Nega- tive Ranks ·572.2.6 Conditional Two-sided Ranks·622.2.7 A Digression on Positive Ranking Functions ·65 – 2.3 Conditionalization and Revision of Ranking Functions ·682.3.1 Plain Conditionalization·682.3.2 Spohn-conditionalization·702.3.3 Shenoy- conditionalization·73 – 2.4 Rank-based Conditional Independence·74

III GraphicalModels forRankingFunctions . . . 81 3.1Introduction·813.1.1Content of this Chapter·813.1.2Historical Remarks·823.1.3Measurability and Variables·873.1.4Algebras Over Variable Sets·903.1.5Graph- theoretic Preliminaries· 91 – 3.2 Graphoids and Conditional Independence Among Vari- ables ·1003.2.1 Conditional Independence Among Variables·1003.2.2 RCI is a Gra- phoid·1013.2.3Agenda·103 – 3.3Ranking Functions and Their Markov Graphs·1043.3.1 D-Maps, I-Maps, and Markov Properties·1043.3.2 Markov Blankets and Markov Boundaries·106 – 3.4 Ranking Functions and Given Markov Graphs·1093.4.1 Po- tential Representation of Negative Ranking Functions·1093.4.2 Representations of Negative Ranking Functions by Markov Graphs·111 – 3.5 RCI and Undirected Graphs·118 – 3.6 Ranking Networks · 1193.6.1 DAGs as Graphical Models· 1193.6.2 Strict Linear

(8)

Orderings on Variables· 1213.6.3 Separation in Directed Graphs· 1223.6.4 Directed Markov Properties · 1233.6.5 Factorization in Directed Graphs · 1283.6.6 Potential Representation of Ranking Networks· 130 – 3.7 Perfect Maps of Ranking Functions · 1313.7.1 Ranking Functions and DAGs·1313.7.2 Characterization of CLs that have Perfect Maps·1333.7.3 Outlook: Can CLs Be Made DAG-isomorphic?·136

IV BeliefPropagation inRankingNetworks. . . 137

4.1 Introduction · 137 – 4.2 Hunter’s Algorithm for Polytrees · 140 – 4.3 The LS- Strategy on Ranking Networks: An Outline·142 – 4.4 Phase1– Triangulation and De- composition of the Network·1464.4.1Methods for Obtaining the Clique Tree from the Initial Network·1464.4.2 Triangulating Graphs: A Brief Survey ·1494.4.3 Desired Criteria for Triangulations of Graphs·1534.4.4 Generating the Elimination Ordering·1554.4.5 TheMCS-MAlgorithm·1584.4.6 Determining the Set of Cliques of the Fill-In-Graph·1594.4.7 Inline Recognition of Cliques· 1614.4.8 Inline Tree Construction·1664.4.9 An Algorithm for Decomposing a Moralized Ranking Network·1714.4.10 A Digression on Triangulatedness and the Epistemic Domain· 174 – 4.5 Phase2 – Message Passing on the Clique Tree·1764.5.1 Local Computation on Cliques·1764.5.2 Locally Available In- formation·1824.5.3 Pre-Initializing the Permanent Belief Base·1834.5.4 Bottom-Up Propagation: Conditional Ranks of the Cliques·1844.5.5Top-Down Propagation: Joint Ranks of the Separators·1864.5.6 Processing Update Information·1884.5.7 Queries on the Clique Tree·188 – 4.6 Conclusion·1894.6.1Achievements·1894.6.2 Remarks on Aspects Not Discussed·191 – 4.7 Outlook: Learning Ranking Networks and Induction to the Unknown·191 A A ComputedExample forDecomposition . . . 193

B A ComputedExample forUpdating . . . 199

2.1The Ranking Network·199 – 2.2Initialization Phase·2002.2.1Going Bottom-Up: Computing the Conditional Ranks·2032.2.2Going Top-Down: Joint Ranks of the Cliques·205 – 2.3 Update By New Evidence·206 Acknowledgements . . . 209

Index ofDefinitions . . . 211

Index ofSymbols . . . 215

Index ofAlgorithms. . . 219

Definitions andTheorems from“TheLaws ofBelief”. . . 221

Literature. . . 225

Index . . . 251

(9)

P

This thesis introduces an efficient algorithm for iterated belief change. It is based on the concept of belief modelled by the devices of ranking theory as developed by Wolfgang Spohn since1983in a series of papers and recently presented comprehensively in his (2012).

The algorithm will introduce a concrete description of how a prior epistemic state is changed by an available evidence into a posterior epistemic state, where epistemic states are imple- mented by Spohnian ranking functions among a set of variables. Since ranking functions are the foundation of the formal modeling of epistemic states, we will speak of “rank-based”

belief states and an algorithm for “epistemic updates”.

The current research for this thesis was inspired by the philosophical discussion of belief revision. However, the specific argumentation of this thesis is characterized by a high level of technical concreteness and uses ranking theory not as a philosophical position but as a mathematical framework for constructing data structures. The thesis is interdisciplinary in the sense that it enriches the work on epistemological problems utilizing devices common in computer science. Since the author does not wish to assign his inquiry specifically to either philosophy or computer science, nor recognizes any requirement to do so, he will just present his arguments without much theorizing on meta-levels.

The main goal of this thesis is to develop a comprehensive algorithmical treatment of rank- based epistemic updates illustrated with a concrete proposal. The first step will be to show how belief states and evidences are represented by ranking functions. The second step will be to construct a graph-based data structure that represents the belief state adequately as a rank-based belief network, and in a third step the algorithm for updating the belief network is introduced. The algorithm requires a prior belief state and a new evidence as input and generates a posterior belief state reflecting the evidence as an output.

While the first step is mostly reproductive in that it presents many argumentative material brought in previously by other authors, the second and the third step prevailingly introduce arguments developed originally by the author.

An important inspiration for the argumentation of this thesis is the well-known concept of a Bayesian network. Bayesian networks are considered as a tool for representing belief states and furthermore a foundation for the mechanism of transition from a prior to a posterior belief state. This transition is considered to be triggered by the accessibility of new evidence.

Ongoing research over the last25years led to a differentiated understanding of the strengths and weaknesses of Bayesian networks considering the representation of belief states and the capabilities of epistemic updates in relation to different fields of application.

(10)

We will represent epistemic states by ranking functions instead of probability measures, but nevertheless we will utilize many known facts about Bayesian networks that serve as an important blueprint for modeling the update mechanism.

During the work on this thesis, the author was frequently confronted with the requirement of re-introducing rank-based versions of concepts already common in probability theory. This requirement always contains the danger of drifting into “re-inventing the wheel”, yet it was very insightful to start formal theorizing about ranks on the very basic foundations. The author decided to not only provide his arguments but also a general introduction to the topic. Therefore, we will explicitly discuss the minimal set of indispensable algebra basics and, later on, also address some specific graph-theoretic problems, such as finding a perfect vertex ordering or complete recognition of cliques in a triangulated graph. Those topics are neither part of ranking theory in particular nor yet of epistemology in general, but they rather play an important role in the argumentation, which means that they had to be respected for presentation in an adequate manner. It is therefore characteristic for this thesis that it utilizes formal tools of measure theory, graph-theory and algorithmics for a contribution to formal epistemology.

Instead of just compiling relevant research results about Bayesian networks for the use on an epistemological topic, the author transfers already available knowledge to the case of ranking theory with the aim to make a progress in the research topic of ranking theory towards a general mechanism of efficient updating.

What this thesis doesnotdo is to introduce a philosophical position of the author. Instead, the author concentrates on showing that ranking theory, taking it as a philosophical position on its own, is well-suited to provide a foundation of concrete applications in computer science.

This of course has implications for epistemology since it shows on a very concrete formal level how iterated belief change works. The contribution therefore is the transfer of research results from probability theory to the context of ranking theory as well as extending ranking theory with new concepts and arguments.

To the best knowledge of the author, no other publication has described an update algorithm for multiply connected rank-based networks so far. Nonetheless, as we will see later, highly relevant considerations of updates for singly connected networks have already been discussed insightfully in the works of Daniel Hunter.

ChapterI of this thesis presents a sketch of the philosophical field of questions to which this thesis intends to contribute, and introduces the relevant concepts formally, but only on a very high level of abstraction.

Additionally, chapterIsketches rather shortly the basic problems of belief revision with only a few pointers to the connected philosophical discussions1. It also introduces the algebraic foundation of belief theory in the form of propositional algebras. It is argued why propositions are supposed to be the objects of belief during this inquiry and how propositions are formally defined. The chapter further defines epistemic states and epistemic updates from the most coarse-grained abstraction level.

1Those discussions will be mostly known to readers with a more philosophical background. Readers more common with the parts of the work related to computer science may feel that the historical and theoretical background of the philosophical problems may not seem to be highly relevant for the concrete arguments.

(11)

The concepts and many of the arguments in chapterIare not invented originally by the au- thor. They are mostly reproduced from chapters1,2and4of (Spohn,2012) but with a strong focus on the formal and technical aspects, leaving out most of the references to the philosoph- ical discussion. The author added a brief reflection about the use of “engineering-like” formal tools in the philosophical domain (section1.2.2) and a short digression about belief content and the related discussion originating from the arguments of Kripke and Putnam (section 1.4.4). The entire other argumentative substance in the chapter is owed to Wolfgang Spohn.

Chapter II presents the most important aspects of the current research status in ranking theory. It formally introduces ranking functions among propositions and discusses the most basic formal extensions and variations that were introduced into discussion so far. Namely there are two-sided ranking functions proposed in (Spohn, 2012, p.76)2as well as the vary- ing concepts of conditional ranks. Additionally, two different methods of conditionalization (Spohn-conditionalization and Shenoy-conditionalization) are discussed that directly corre- spond to Jeffrey- and Field-conditionalization in probability theory. Other important contents described in this chapter are for instance a rank-based version of Bayes’ theorem and a “chain rule” for ranks. For the formal concepts that are introduced, additionally some notable prop- erties are shown.

The chapter compiles the relevant material that was widely published before, also dis- cussing distinct variations in terminology and also properties of the formal concepts. Taken on its own, chapterIIshould be adequate as a general introduction to the mathematical foun- dation of ranking theory. Like chapterI, it is mostly reproductive and contributes compiled introduction of formal material. The arguments and concepts presented have already been described by other authors, mainly Wolfgang Spohn.

It seems very reasonable to the author to present a compiled version of this material for different reasons:

1. A clean and detailed introduction of all formal concepts, adjusted in terminology and formal language to a uniform manner, is very important for a concise presentation of the arguments of this thesis. Since the formal concepts stem from different discussions and multiple disciplines, the author considered it necessary to introduce all relevant concepts to ensure uniformity. For this reason, chapter II introduces ranking theory while chapterIIIintroduces a considerable set of concepts from graph theory.

2. Before the advent of (Spohn, 2012), the foundation of ranking theory was scattered about a variety of texts; some of them were separated from each other by many years during which the concepts underwent some variation concerning the conventions of naming, notation, and also mathematical properties. (For instance, Wolfgang Spohn at one time decided to use natural numbers instead of ordinal numbers as codomain for ranking functions. Additionally, a variety of concepts representing conditional ranks were proposed, mostly with different properties.) Presenting own introductions of the formal concepts ensures maximal clarity about what concepts are to be discussed. Of course, this is cum grano salissince the reproduction obviously suffers from a lack of

2Confer definition5.12.

(12)

philosophical depth, which is a consequence of considering ranking functions as mere material for algorithmic work. Whenever Wolfgang Spohn added insightful comments on why he introduces things in a particular manner and not in the style of argumentx, the author of this thesis mostly just joined his view whenever it did not entail difficulties in the algorithmic part. One could also state that the thesis is written from an engineer- ing point of view – nevertheless, this fact is philosophically reflected in the beginning of chapterI.

3. The reader not familiar with ranking theory may use chapter IIas both, a general in- troduction to this topic and a reference. Therefore, chapterIIand partly chapterIIIare written in the style of a textbook and should be easy to read for readers used to formal concepts. In general, the proof level of chapter II is extraordinary low. The level of some proofs is surely almost trivial for readers with more than the most basic mathe- matical knowledge and indeed, along some passages in this chapter, the argumentation progresses quite slow. Although, it may demand some patience from the reader with appropriate background, this is perfectly congruent with the intention of the author since he wants to show from which basic mathematical structures the rank-based con- cepts originate, without leaving behind the reader less experienced in formal topics.

Since all proofs are consequently marked, it is easy to just skip a particular proof, if the reader is not interested in it. In chaptersIII and IV, the proof level returns to a more conventional level.

4. Only a complete introduction of all concepts will give a “vertical” insight to the matter introduced. To clarify this: the most coarse-grained vertical perspective may state that the author considers propositions as an algebraic structure on which ranking functions can be defined. Ranking functions can form graphical data structures that can be shown to satisfy the Markov properties and can hence be updated by a Lauritzen-Spiegelhalter- style update algorithm. It was the aim of the author to make this vertical view as complete as possible.

The main contribution of chapter II is therefore to present the current research status in ranking theory as well as to introduce the topic to readers not already familiar with it.

ChapterIIIuses the formal material introduced in the previous chapters to develop ranking networks as graphical models of ranking-functions. A ranking network is a data structure that represents the ranking function among a set of variables and is therefore the implementation of an epistemic state at a certain discrete time. In a formal respect, ranking networks are directed acyclic graphs with their vertices being conditional matrices of variables and their edges being subjectively valid relationships of causal influence between the variables. The network has a negative ranking function assigned that defines the conditional information for each vertex.

The concept of a measurable variable is introduced along with many basic concepts from graph theory (which may not be common to all readers). Since it has already been shown that rank-based conditional independence is a graphoid3, the focus lies on what is consequently

3Confer (Hunter,1991a) and recently (Spohn,2012, p.132) with theorem7.10.

(13)

the subsequent step: showing that the Markov properties hold for ranking networks. This is separately discussed for undirected as well as for directed graphs and for three different Markov properties. Those Markov properties are the formal precondition to make the up- date algorithm work since they enable the represented ranking function to be expressed by a potential representation.

Some of the relevant proofs are either very similar to their counterparts in probability the- ory, or they are solvable on the mere graph-theory level at all, not requiring any arguments from ranking theory. Others are more sophisticated and quite non-trivial. For the same argu- mentation as pointed out for chapterII, this material is introduced in uniform terminology.

The presumably most important contribution of chapterIII is a rank-based version of the Hammersley-Clifford-theorem as it is introduced by theorem 3.67on page 114. To the best knowledge of the author, the validity of this theorem in ranking theory was never shown before.

ChapterIVshows how a transition from a prior to a posterior belief state can be formally represented and efficiently executed. The prior as well as the posterior belief state are thereby represented by ranking networks as they are defined in chapterIII.

The update process is separated in two phases: in the first phase, a “compilation” of the ranking network is required in which the permanent belief base is created (or in fact re- created) from the prior ranking network. The permanent belief base is a data structure that represents the prior belief state in a technically feasible manner. It is ensured to be a tree (in the graph-theoretic sense of the word), which ensures that well-known techniques for tree updates can be applied to incorporate the evidence. The compilation phase is only required if the evidence has modified the structure of the network by the insertion or deletion of edges or vertices. If the evidential input does not change the network structure, recompiling can be omitted, although a compilation definitively has to be performed at least once when the iterated update process starts.

Technically, the compilation phase decomposes the ranking network to a clique tree and computes a potential representation of the ranking function that characterizes the prior belief state. This chapter presents adaption of graph algorithms like clique recognition, and the establishing of a perfect elimination ordering to derive a triangulation of the input network.

The second phase is the “update phase”: it incorporates the evidential input in the perma- nent belief base. After the update phase is completed, the belief base reflects the evidential input. In technical terms, this can be performed by Pearl’s message passing algorithm, but the factual arithmetics has to be replaced by operations adequate for the ranking semantics of the vertices.

The update algorithm works on ranking networks in general, which means, on each kind of directed acyclic graph without loops and multi-edges. Especially the network is not supposed to be singly connected. As the reader may already have noticed while reading the former paragraphs, the algorithm is heavily inspired by the technique of Lauritzen and Spiegelhalter for updating multiply connected Bayesian networks, and the author drew many benefits from the former works of Judea Pearl and Richard Neapolitan while developing it. Nonetheless, a number of more recent research results are also considered in this thesis. Improving Neapoli- tan’s approach, the author shows that the compilation phase can be completed by the run of

(14)

one single procedure.

ChapterIVends with an outlook over further tasks that could be reformulated in specific rank-based terms, such as structure learning of ranking networks and implementation of further types of inference.

The argumentation in the thesis leads from basic considerations of propositional belief for- mation towards a directly implementable system of permanent epistemic activity that keeps a belief base up-to-date with the available evidence by passively incorporating the new evi- dence.

(15)

I

B elief , B elief S tates , and B elief C hange

1 . 1 I ntroduction

This chapter elaborates on the relationship between normative epistemology and engineering.

It further describes the contribution this thesis provides to this field of research, and addition- ally introduces the basic notions of belief theory as they will be understood throughout the text.

As already stated on page11, the argumentative material in this chapter is not originally invented by the author but just tries to “distill” the most relevant parts of formal belief rep- resentation for a brief but sufficient introduction. As a blueprint, the author used mainly chapters1,2and4of (Spohn,2012).

Section 1.2 contains the discussion of the perspective of normative epistemology about which the thesis is written. Some of the commonalities with general engineering are em- phasized. The section is not intended to comprehensively discuss the conditions of normative epistemology but rather it seeks to underline why normative epistemology is attractive for working on philosophical topics by the devices provided by mathematics and computer sci- ence. Since this thought perhaps seems unintuitive at first, an introductory remark about this aspect may be reasonable.

Section 1.3 describes the belief revision system the thesis will develop from an intuitive view, emphasizing the mechanistic perspective that will successively become more dominant in the later chapters and eventually supersede the philosophical considerations.

Section1.4, as already stated, introduces a framework for representing propositional belief that is suitable to technical implementation. It reproduces main ideas introduced by Wolfgang Spohn in chapters2and4of his (2012).

Sections 1.5 and 1.6 reproduce a subset of the argumentative material Spohn gives about the rationality of belief sets and the transition between rational epistemic states. We will concentrate on the more technical parts and leave out most of the references to the history of analytical philosophy. The sections may be seen as a “foreshadowing” on the implementation described in chaptersIIIandIV.

(16)

1.1.1 Remark onSources andCitation

In this chapter, we will introduce an algebraic framework to represent belief states and be- lief change. This framework was developed by Wolfang Spohn in chapter 2 of his (2012).

Although, Spohn makes also use of some ideas that are quite common in the particular dis- cussion about belief revision and the adquate representation of belief, the author of this thesis used Spohn’s work as main orientation for his own writing in this chapter.

It is a foremost goal of this thesis to keep compatibility with Spohn’s concepts. In the remainder of this chapter, we will therefore mainly introduce those of Spohn’s formal con- cepts important for our aims, but without reproducing each of his arguments considering the epistemological aspects. Instead, the author discusses the concepts in his own words.

Section 1.2 presents a genuine argumentation of the author of this thesis with the excep- tion of subsection 1.2.3, whose arguments are mainly owed to section 1.1 of (Spohn, 2012).

Especially the essential “two questions” on page 25 were heavily inspired from equivalent questions found in (Spohn,2012, p.6).

Section1.3introduces nearly the same base elements as Spohn does and should therefore be read as just stating the starting point from already well-known concepts. The idea to use Hintikka’s (minimal) requirements of rationality was taken from (Spohn, 2012, p. 48).

Subsection1.4.4presents a genuine argumentation of the author of this thesis.

The reader should explicitly acknowledge that the turns made in section1.4were completely based on the concepts introduced in (Spohn, 2012, chapter 2) and that sections 1.5 and 1.6 entirely consist of material compiled from the sections4.1and4.2in (Spohn,2012, chapter4) but completely presented in own words of the author.

As already stated above, the author of this thesis tried to distill the mere formal parts of Spohn’s arguments whenever possible while commenting them in his own words. Therefore, explicit citations do not occur frequently throughout sections 1.4 – 1.6. To ensure lucidity about the sources, all theorems and definitions stemming from (Spohn,2012) are presented in an index starting on page221, where the source of each item used is made explicit.

The quite generous reproduction of Spohn’s formal concepts seems sensible to the author because in the later chapters, he will connect them to concepts from different research fields and he wishes to keep maximal formal consistency throughout the entire thesis.

1 . 2 A N ormative P erspective on E pistemology

1.2.1 Descriptive andNormativePerspective

Before we take a concrete, detailed, and well-structured look on the questions that form the agenda of this thesis, it has to be pointed out what implications our perspective on the topic will have.

A normative perspective focusses on what rational beingsshouldbelieve, given certain cir- cumstances, and what they should not. This is different from the perspectives of most special sciences like sociology and, mostly, psychology, which just describe the factual belief states of the subject. We will therefore call the perspective of these disciplines the descriptive perspec- tive.

(17)

While a descriptive perspective is interested in understanding and explaining the way how beliefs actuallyareformed and modified in terms of a particular special science, the normative perspective tries to point out how these processesshouldfunction to be acceptable as rational.

A descriptive perspective considers the scientific facts and its interest concentrates on the empirical conditions related to beliefs and their dynamics in respect to the concepts of some special science. The many fascinating questions concerning the empirical aspects of beliefs are scattered widely about the subjects of special sciences, such as the neurosciences, cognitive psychology, linguistics and sociology, to state only the most obvious ones. The perspective that focusses on the concrete empirical questions is by convention not a genuine philosophical perspective, although, its observations take influence on philosophical questions.

The normative perspective tries to find out what beliefs we should acquire under certain circumstances and consequently presupposes the existence of rules that enable us to sketch idealizations about the dynamics of beliefs. The prescriptive aspect in the word “should” im- plies the presence of such an idealization because otherwise no prescription would be justified.

The normative perspective is not interested in practical modifications of cognitive mechanisms but just in gaining a lucid conception of rational belief change. It is a notable fact that engi- neering disciplines like computer science with its topics of machine learning and data mining, to state only two, have a strong family resemblance with normative epistemology concerning the normative perspective.

Idealization about the dynamics of belief enables us to draw a distinction between belief states that are sufficiently justified, correct, coherent, consistent, and which are thus reasonable or at least rational, and other belief states. Because all these adjectives are mortgaged with non-neutral theory contents, in other words, a normative perspective on belief theory claims to be able to make a structured distinction between belief revision processes that would need correction to some extent to fit ideal conditions and belief revision processes that would not need any such correction at all.

This sounds rather vague, because nothing is said so far about the criteria for this distinction.

The terms “correction” and “ideal conditions” appear to be mere placeholders. Furthermore, it has been argued very lucidly by different authors like for instance Jaegwon Kim and Wil- frid Sellars, what normativity could mean and that the term “normative” is not atomic but describes many different ways in which a statement can be normative. We will not engage in this discussions at this point.

Regardless of any reasonable differentiation one could establish, it is sufficient for under- standing the arguments presented in this chapter that the characteristic aspect of the normative perspective sums up to the question what weshouldbelieve under which circumstances. The normative inquiry of belief tries to sketch a lucid picture of the rules or laws that hold for the unbiased acquisition of beliefs and in the course of the unfolding of my investigation, the vague character of the above stated sentence will disappear.

1.2.2 ThePhilosopher and theEngineer

At a first glance, it may seem that figuring out the details of the normative perspective falls naturally and primarily in the responsibility of the philosopher. But this conjecture previously

(18)

has been (and currently remains ) the subject of dispute.

The arguments most interesting for our particular topic were raised in context of a discus- sion about a naturalized version of normative epistemology.

Naturalists usually argue that epistemology is a technical discipline. A common objection against this conjecture is that epistemology has to contain normative parts and that it would be not capable for this discipline to be normative in the sense it need to be. One of the responses of the naturalists was that epistemology could be normative in the sense engineering disciplines are normative. The locus classicusis Quine’s short (1986), where he argues that

“normative epistemology” is “a branch of engineering”, a mere “technology of truth-seeking”:

“Naturalization of epistemology does not jettison the normative and settle for the indiscriminate description of ongoing procedures. For me normative epistemol- ogy is a branch of engineering. It is the technology of truth-seeking, or, in a more cautiously epistemological term, prediction. Like any technology, it makes free use of whatever scientific findings may suit its purpose. It draws upon mathe- matics in computing standard deviation and probable error and in scouting the gambler’s fallacy. It draws upon experimental psychology in exposing perceptual illusions, and upon cognitive psychology in scouting wishful thinking. It draws upon neurology and physics, in a general way, in discounting testimony from occult or parapsychological sources. There is no question here of ultimate value, as in morals; it is a matter of efficacy for an ulterior end, truth or prediction.

The normative here, as elsewhere in engineering, becomes descriptive when the terminal parameter is expressed. We could say the same of morality if we could view it as aimed at reward in heaven.” (Quine,1986, p.664f)

Instead of reading Quine as if he would agree to factually exclude normative epistemology from the core interests of philosophy, the discussion recognized the proposal, that normative epistemology could be “naturalized” by emulating engineering disciplines. For example, an engineer who builds a bridge knows by which properties a “good” bridge is characterized and what he has to do to build a “good” bridge. In the same way, an epistemologist knows what is a “good” revision and how “good” revision can be made even “better”.

Quine’s “engineering reply” to the objection of naturalized epistemology was itself a subject of discussion, see for instance (Wrenn,2006), where also a lucid catalogue of the different types of normativity in question is provided. We will not join this discussion since it is not directly relevant for what we are about to do.

On the other hand, this is in brief exactly what we will do in the remainder of this in- quiry: doing epistemology on an engineering level, using the formal devices provided by engineering-like disciplines, namely computer science and mathematics.

It is therefore nonetheless reasonable to comment on the conditions of such an approach.

Quine is undoubtedly right in stating that many “branches of engineering”, especially but not only in computer science, focus on what could be justifiably entitled to be called “truth- seeking” on a level that could be called “technological”.

Examples for such applications are neither seldom nor special. Think of every day require- ments like a database system that has to preserve data consistency during update procedures,

(19)

the Bayesian filters in most email clients that can learn to make good decisions about which of the messages you receive are undesired, or the fuzzy logic in a digital camera that tries to gen- erate exceptionally good looking pictures of what seems to be reality. The internal processing logic of the camera implements a calculus to decide which visual effects are undesired and which are acceptable. Any knowledge representation system regardless of its purpose has to implement basic requirements of rationality, at least consistency.

It is not very surprising that most sciences rely on their own techniques of truth seeking if we remember that the occidental conception of science is directly and inseparably connected to truth. While the normative part of “truth-oriented” philosophyanalyzesconceptions of truth, most sciences develop truth conceptions implicitly. They are to beimplementedin algorithms, heuristics and strategies within the conceptual scope of the particular discipline or subdis- cipline. One may see this as a “technological” view on truth. But the analysis of the truth conception of a specific discipline is usually not part of the discipline itself with philosophy being the remarkable exception. We will return to this thought a little later.

The differences in perspective and aim of analysis between philosophy and engineering seem to induce a quite lucid distinction between the different spheres of competence: the philosopher’s task seems to be analyzing what truth “is” and the technologist’s task seems to be to develop techniques to “generate” true assertions from knowledge. This seems to be clear and fair on first sight.

But one should not conclude from this observation that the philosopher is supposed to cede a part of the authority concerning the investigation of truth to the protagonists of other sci- ences. Unless one rejects the idea that normative reasoning about truth is a philosophical task, the philosopher is also addressed when it comes to the question which beliefs can – read:

should – be legally generated from beliefs that are already accepted as a part of a subjective knowledge base. Our contemporary experience in this fields shows that mathematics, statis- tics, and especially computer science, provide very useful devices for working out the details of this question.

When philosophy develops a formal conception of truth, it is a legitime question to the philosophers if this conception is applicable in a practical sense. This, in particular, includes the question if it can be implemented technologically. One may conclude that the imple- mentation is not a specific philosophical task, but once the legitimateness of the question of implementability is accepted, it becomes impossible to draw a sharp demarcation line between the philosopher’s part and the engineer’s part in the subject of normative theory of reasoning.

This, of course, sounds like a feuilleton argument but it can be stated more precise by consid- ering the relationship to rationality that is maintained in philosophy on the one hand and in engineering on the other hand.

The best example for engineering in this context is computer science because, among other topics, computer science tries to implement conceptions of intelligence, to understand the formal processes of decision making and inductive inference and to provide systems with the ability to apply these techniques to unknown situations. This seems to directly imply an understanding of at least some currently unexplained capabilities of the human mind.

It may seem that either this claim, formerly the legal territory of philosophy, was usurped by a another discipline or – in an even more pessimistic way – that this shows that at last

(20)

philosophy turned out to be of no more importance for finding out how thinkingworks.

Computer science is as closely connected to logic and linguistics as philosophy. One may refuse to consider computer science as a mere engineering discipline, thinking of subjects like complexity theory and formal languages, which seem to reside in the competence sphere of computer science without being very “engineering-like”. However, it is undebatable that com- puter science is strongly influenced by engineering paradigms. But in contrast to philosophy, computer science does not model rationality from a reflexive point of view, it just uses – as every engineering discipline – an implicit normative perspective which is motivated by the search for techniques to gain precise and good results within the scope of its domain. What

“good” results are is in most cases beyond discussions: it is simply provided as a definition.

For the philosopher a “good” solution to an issue may be a one that does not contain inconsistencies, fits acceptably well with already existing attractive explanations of related issues, does not rise heavy contradictions with existing attractive models of explanation, and is, however, acceptably simple. One may find other criteria and may also argue that they are not uniformly accepted in the entire community but at least, there is some minimal consensus about the properties listed above.

This is also true for computer science, but here, the rules are much stricter: Whoever enters the community claiming to have a “good” solution to a problem has to face three types of questions.

1. Is the proposed solution less resource-consuming than any existing concurrent solution?

(Which means: is it asymptotically faster – or at least empirically faster, for typical situations? Does it need asymptotically less memory or bandwidth?)

2. Does the proposed solution improve the quality of results over any existing concurrent solution? (Which is of special importance on solution types based on heuristics or domains with different conventions of modeling.)

3. Is the proposed solution simpler as any existing concurrent solution? (Which means: is it easier to understand or to implement?)

Each of these questions targets an aspect of what computer science considers as relevant aspects of scientific progress in its scope.

If not at least one of these questions can be answered positively, ideally supported by ap- propriate theoretical proofs or strong arguments and empirical tests, it will typically not be accepted as a valid improvement. The reason for this assessment obviously is that in case of three times “no” the proposal does typically neither promise a contribution to scientific progress nor to practical use.

Hence it seems, computer science does not try to reflect on rationality but merely tries to teach machines to make rational decisions, supposing that rationality is already defined. But this is far too brief: in fact, there are more sciences of “rationality” than only philosophy.

No engineer can implement conceptions, neither in software nor in hardware, that are not fully understood and formally clear. This does not require an engineer to understand some mysteries of mind but it does require her to understand the particular conception of rationality

(21)

she uses to solve her concrete problem. For her claim, it must be completely defined “what it is like to be in a rational state”.

This requirement is in a way weaker as well as stronger than the requirement of the philoso- pher. The philosopher is by tradition more interested in the questions but in exact answers.

Therefore, the requirement of the engineer is stronger, because her conception of rationality criteria must be sufficiently concrete to implement it. The engineer has to understand the in- herent structure of the capability she intends to implement regardless if she uses fuzzy logic, probability theory, neural networks, machine learning algorithms, or other tools.

The philosopher is interested in understanding, what rationality, well, “means”. This ad- dresses any rationality concept. In this regard, the claim of the philosopher is stronger, since her discipline always contains the critical reflection of the notions it uses. But it is obvious that the engineer’s claim comes down to exactly the same question as the philosopher’s: what is the inherent structure of the concept of rationality in consideration?

This equivalence may consequently be interpreted as a kind of concurrence and it may there- fore seem that computer science and the cognition sciences contribute much more substance than contemporary philosophy does.

It appear to be a side effect of scientific progress, that increasingly more of the questions for- merly seeming to be inherently philosophic turned out to be answerable using the conceptions of special sciences.

Of course, this is a statement from the point of view the special sciences keep. Philosophy could answer that most of the questions currently discussed in the neurosciences were first developed in philosophy. Nonetheless, the rise of the neurosciences and computer science does show that not all philosophical questions are eternal.

Some aspects indeed came down to be answerable by concrete approaches of special sci- ences that became feasible by sufficient scientific progress. And, currently it is an ongoing task for philosophy to incorporate the important scientific knowledge brought in by special sciences. Acknowledgement, analysis, and incorporation of the progress made by the special sciences is, and indeed must be, a permanent stimulus for contemporary philosophy.

So far, it is lucid that philosophy and engineering share some questions and insofar it is correct to state that “truth-seeking” has many technological aspects. But how about the converse question: is any approach on rationality that uses formal devices an “engineering- like” approach? Does epistemology turn into an engineering discipline by using “engineering- like” devices?

Of course, one may be seduced to join the naturalist view, accept normative epistemology as a mere technical discipline, and feel committed to the view that reflecting rationality and motivating epistemology is no more part of epistemology itself.

Nonetheless, there are not only commonalities but also extremely important differences between epistemology and computer science.

The mere fact that there exist engineering tasks that yield concrete technical implementa- tions of truth-seeking strategies does not show that the theoretical conception, the rules of acquiring true beliefs, is as a whole a part of engineering.

The philosopher’s attitude becomes suspicious if thoughts can be made concrete to a level such that an engineer can implement them. At a first glance, this might seem to be an argu-

(22)

ment that they lack the abstraction level philosophy prefers and requires being able to reflect its questions.

But in fact each science has its own device at hand to reach progress in knowledge in its special perspective. The rules sufficient in engineering are in many aspects not sufficient in a philosophical perspective.

The learning algorithms of the computer scientist and her solution seeking techniques show some property that disqualifies them as sufficient answers to the questions of the philosopher:

they always serve a certain purposeand are only applicable under specialized, strict, and in some respect uncommon conditions, and to a particular purpose. The criteria of truth in engineering are criteria of optimality to certain purposes. They are specialized and context dependent. They – and this is the strongest difference to philosophy, as already stated above – do not contain their own reflection nor their own motivation, where the latter is purely instru- mental throughout the entire discipline. Building bridges is part of engineering but analyzing reasons why bridges should be build is not, nor why bridge building itself is “good”. This is completely different for epistemology or philosophy in general.

Again, this does not show that truth seeking belongs more to engineering than to philos- ophy but it shows that the questions philosophy tries to answer are highly relevant to other subjects, that the devices philosophy uses are related to the devices of other sciences, and that their devices can be inspiring for philosophical methods andvice versa.

The argument that computer science uses conceptions of truth that are in some way spe- cialized is nonetheless debatable, because it depends on the abstraction level on which the conception is analyzed. The different conceptions of course have the common minimal core of consistency and deductive closure, regardless whether a relational database, a Bayesian filter or an engine for automatic planning is considered. It is therefore not a strong argument that each task adds its own extensions to the conceptions.

It is not the task of the philosopher to develop solutions to the aim of optimality. It is her task to sketch new sensible pictures of old problems and discuss new thoughts about known subjects, bringing aspects to light that were hidden before, but this is a purpose on its own and it does not take place to serve some special purpose or to meet special requirements.

Therefore, the philosopher enjoys more liberty than the engineer (who has to answer the three questions). It seems just impossible to define criteria of optimality for truth-seeking on the level on which a philosopher analyzes truth-seeking. This surely seems to be a strong indication for the engineering sciences that their perspective of truth-seeking cannot catch up with the one of the philosopher, and that reasoning about truth-seeking algorithms may not be a philosophical task. But on the other hand, from the mere fact that the philosopher is not the only researcher who is interested in a certain perspective, it does not follow that she is not entitled to share this perspective or has to give up her task.

Another valid argument against the conjecture that epistemology is “pure” engineering is directly connected to this aspect: the task of the engineer is always the solution of a distinct and special task, and the concepts of truth and consistency are always instruments for her and not object of her scientific reflection of what she does. Reflection of what she does is, as said above, not part of her domain. This is surely different for the case of the philosopher.

An implementation of some heuristic learning algorithm may try to use techniques of mak-

(23)

ing good decisions that are similar to those used by human beings, but the decision made by the algorithm will always be oriented toward some special and narrowed principles. It is not made forreflectionbut for bareuse. Nevertheless, use is a fundamentally different destination than reflection. This is the most important distinction between the normative perspective of the philosopher and the normative perspective of the engineer. The normative perspective of the philosopher is free to producecategorical“goodness”. This is not possible for the engineer, whose categories of goodness are never independent from the underlying unreflected purpose that is not part of engineering itself.

In short, a third aspect is that normative epistemology is the contemporary approach to the problem of induction. This problem definitely lies in the responsibility of the philosopher. We will see in the next section that the two central questions of this thesis sum up to a formal approach to the problem of induction. This problem cannot be analyzed without introducing idealization into epistemology and therefore, philosophy has to consider this topic.

However, the fact that different disciplines are interested in the same conception for differ- ent goals does not in any way show that inquiries in the conception are a definitive task of just one of them. Where some authors may see a strong demarcation line, there is just an interdis- ciplinary discussion about the shared conceptions of truth and rationality. Philosophy in its analysis may utilize formal devices from mathematics, statistics, and even computer science without becoming engineering. Philosophy is not only allowed to do so. There is furthermore a strong indication to use whatever powerful devices are available. On the other hand, phi- losophy without being able to access results from the special sciences, would be cut off from scientific progress. These two aspects form the structure of the interaction between philoso- phy and engineering disciplines. Considering these facts, the question whether epistemology should be engineering or not seems ill-stated and of no relevance for practical scientific work.

1.2.3 BeliefRevision and theProblem ofInduction

The goal of any theory of belief is to describe how a set of beliefs is built up and maintained.

The first is described by belief formation, the second by belief revision. In the course of the inquiry, the author will sometimes use the shortened term “belief theory” that includes belief formation as well as belief revision.

Both aspects provide a question to the beginning of the inquiry. To make clear what the starting point of our reflection is, consider the following analogy.

We want to imagine the computational aspect of belief theory as a kind of update algorithm on a given set of beliefs. Progressing from this thought, the following picture emerges.

There is a cognitive system – perhaps the human mind, but it could also be another, lower- level information processing system – that consists of two functional components. First, there is a set of beliefs that are “held” in the system. This set represents the doxastic state of the system, or, one could say, its “knowledge”.

The second component is a kind of update “cinetics”, a mechanism that operates on this set of beliefs. This component represents the concrete strategy for integrating new information into the knowledge base of the system. This integration process implies the transition from its prior doxastic state to a new state.

(24)

With these two components, the system is capable to perform a kind of administration job on its belief base: whenever it encounters a new piece of information, it has to check whether it can accept the information. If this is the case, it adds the information to its belief base such that the system has at each discrete time an updated belief base. On this basis, we can name it a knowledge management system. This sounds nearly trivial, but it clearly is not.

In the beginning, the system is in some initial state, in which it is characterized by a set of initial or basic beliefs, in which the algorithm is started into a continuous loop of distinct, consecutive modification operations, that are triggered by new information becoming available to the system. Whenever a new information is made present to the system, the algorithm processes the information by adjusting the already present set of beliefs such that the posterior state of the system reflects the new evidence. Immediately we can identify two different cases:

either the new information raises a conflict with the prior state or it does not.

In the second case, where the presence of the new belief does not affect the other beliefs, it can directly be added to the belief base of the system. Because the system is clever, it tries to also obtain knowledge by using this new evidence to draw inferences from. Hence, the integration of the new belief results in a kind of rule-based “interaction” with the belief base.

It may also be the case that the new information represents a very strong evidence that causes the system to erase one or more beliefs that are in conflict with the new information.

The system has then to decide whether the new evidence should be rejected because the sum of conflict-free beliefs it already keeps has more weight than the single new evidence. If it decides to add the new belief to the belief base it has to find a way to resolve the conflicts the new evidence introduces. In the course of this resolving process, some manipulation may be necessary to complete the transition to a conflict-free posterior doxastic state.

In short, the algorithm performs anintegrationof the new information into the set of beliefs that is held in the system in some way that will be the subject of further investigation. The re- sult of such an update operation will be a posterior doxastic state, reflecting the new evidence on the basis of the prior doxastic state. (We also consider the case where the new evidence is rejected as a case of integration.) The crucial point is: knowing this algorithm would provide a solution for the problem of induction since it would provide an objective formal technique to generate new beliefs from current beliefs. This process always involves inductive reasoning.

This picture is very coarse-grained. But it performs two adequate functions in a certain respect: first, it shows us the elements of the theory that will be part of further investigation.

The second and more beneficial effect is, that it leads our attention to the positions that are explanatory empty. Obviously there are unexplained aspects at two crucial points in the picture4.

Although the process of continuous epistemic update seems intuitively very comprehensi- ble, we lack any intuition about the state in which the system starts with its update activity.

How is the initial state of the belief base characterized before considering the first empirical evidence? Which beliefs do we have initially when our mind “starts” to acquire beliefs? In other words:

4The author follows quite direct the argumentation in section1.1of (Spohn,2012). Spohn also narrows down his exposition of the topic to the two questions introduced in this section (cf. (Spohn,2012, p.6)) but he chooses another route.

(25)

How is the initial doxastic state to be characterized?

This question involved continental rationalism and anglo-saxon empirism in a discussion about whether the mind is atabula rasain the beginning of life or whether a human has basic ideas in her mind when she enters the world.

The question for the initial state implies the question for a priori beliefs which immediately involves complex considerations of apriority.

We cannot engage in further investigations in this point, however, the precise nature of the initial belief base will be subject of further inquiries.

We have identified one crucial question, but there is a second important point. Considering the initial belief state as given, the question arises in which way new evidences affect the mod- ification of the belief base. When the presence of new evidence triggers an epistemic update, how is the transition of a prior doxastic state into a posterior state structured? Which beliefs can we derive from already acquired beliefs in the light of new evidence? Which beliefs can we legally derive from beliefs we already have? Stated more briefly: what should we inferen- tially believe? Hence, put in a more formal way, the second question is:

Which rules hold for the transition of one doxastic state to another?

It can easily be seen that these two questions form a condensed version of the problem of induction. The first question is about the belief base and the second about inferential beliefs, thus if we find an approach that answers these questions by giving reasonable theories, a complete inductive account is implemented.

The concepts of derivation and inference used in this chapter and also the following chapters should not be understood as purely deductive because it would be a reduced conjecture not adequate to the problem.

One reason for this is that deduction does not have the capability to “lead us from percep- tions to beliefs about laws, the future or the unobserved” as (Spohn,2012, p.3) says.

But we know that we have the ability to draw such inferences. Hence, deduction is not the only device of inference that we practically use.

The strength of deductive logic is possibly not sufficient to understand the inference pro- cesses involved by the transition from one doxastic state to another. Another demonstration for this comes from the fact that we draw concrete consequences from attitudes that are vague or uncertain.

For example, we are acquainted with some vague attitudes, drawn by strong intuitions5 without having evidences and without fulfilling the rules of deduction in our inference pro- cess. However, we would not say that one does not act rational by following her intuitions to some extent. Those intuitions influence our beliefs as well as our actions. As a result, we have beliefs that are uncertain. To be precise, there are strong indications that none of our beliefs is of absolute certainty. When we speak of inference or derivation, these notions have to be wide

5The notion “intuition” always means intuitions in the colloquial sense, not Kantian intuitions.

(26)

enough to cover uncertainty.

Thus, what on first sight seems quite easy and mechanistic enough to be implemented im- mediately in some algorithms turns out to be all pretty complex and intricate. (The engineer may respond that mere complexity in the details of computation does not introduce a differ- ence in principle and the philosopher will answer that she is not interested in coping with the complexity of implementation but with understanding how the transition works in principle.) The normative perspective of belief revision does of course not formulate the psychological aspects of beliefs, but tries to find out how a perfectly rational mind would acquire beliefs on the basis of uncertain information. That means, it formulates the rules that a transition from one doxastic state to another must fulfill to be a valid epistemic update given the new evidence. The idealization lies in the assumption that, given a new evidence, some updates on the belief base lead to a more “preferable” doxastic state than others. What it precisely means to be more “preferable” will be investigated in section1.5.1.

We clearly see that a formal normative theory of doxastic states and their rules of change will yield an approach to the problem of induction because it would provide a method to infer new beliefs that are not contained in what is already represented in our beliefs.

1 . 3 E lements of a B elief T heory

When we introduced the picture of the mechanistic knowledge management system above, we remarked that one of its benefits would be to clarify the elements that are important in the theorizing about belief theory.

To conceptually introduce the elements of the theory, we substitute the purely mechanistic concept of the knowledge management system by the concept of a subject. This does not make the former example invalid because the knowledge management system can also be a subject and has to act like a subject. Following this thought, the picture contains the following elements.

There are first the objects of belief that take the role of epistemic units. The epistemic units are those objects to which a subject is related by the belief relation. Usually, they are called

“propositions”, but there is a great variety in the philosophical discussion about how propo- sitions should be defined.

Traditional AGM belief revision theory identifies propositions with sentences and that is precisely why AGM theory is considered primarily a logical theory and not an epistemic theory. As a consequence, it is not open to different kinds of interpretations or applications.

Another widely investigated definition of propositions is taking a proposition to describe a set of possible worlds. The thesis will not follow these restrictions but will rather follow Spohn’s definition of propositions, he gives in (Spohn, 2012, p. 17). This definition is more open to interpretation. We will come for propositions in detail in section1.4.

To epistemic units, a subject can have different epistemic attitudes. Gärdenfors speaks of three epistemic attitudes in (Gärdenfors,1988): a subject can accept a proposition, reject it or be indeterminate about it. There is a separate discussion on what it means to accept a belief, or take it to be true. The questions related to this discussion are not in the focus of this thesis.

The epistemic attitude of acceptance of a certain belief can be imagined as keeping this belief

(27)

as a part of the current epistemic statethe subject is in. Rejecting a belief can be defined as accepting its negation. Being indeterminate about a belief means that neither the belief nor its negation are accepted in the current epistemic state.

To make this clearer, epistemic states have to be explained. Epistemic states, or, as they will also be called doxastic states, are the central concept of belief theory. A doxastic state is imagined as the set of all beliefs the subject accepts. Formally an epistemic state is rep- resented by an aggregation of epistemic units. Consequently, an epistemic state is a set of propositions. But this is already a theoretic predetermination, because epistemic states are of different structure in different theories. They can also be expressed as a probability measure in Bayesian models or as a set of possible worlds, to mention only two of the most prominent conceptions.

So far, these are the static aspects of belief. The dynamics of belief come into account by the following argument.

An epistemic state can be altered by the confrontation with newevidences, or, as Gärdenfors calls them, “epistemic inputs”. This name suggests that there is some new entity introduced but this is in fact not the case. A minimal example for an epistemic input can for instance be a proposition becoming in some way “present” to the subject, combined with a particular posterior certainty degree6. If the subject accepts the new epistemic input, she alters her epistemic state by integrating the new proposition into it. This means that she performs a transition from a prior epistemic state to a posterior epistemic state.

This kind of altering is calledepistemic update. It means that the subject changes its prior epistemic state by integrating the new evidence into it. This requires changes on epistemic attitudes to some – possibly many – epistemic units. The result of performing the epistemic update is a posterior epistemic state.

Theorizing about this transition from a prior to a posterior state implies a concept of the re- quirements of rationality because otherwise no assertion can be made about the requirements necessary to ensure that the transition mechanism leads to a rational state, regardless which kind of evidence it processes. The mechanism has to reflect – or, to be more provocative:define – rationality on epistemic states.

Thus on the meta-level, a theory about belief revision makes assumptions on rationality.

The most prominent rationality postulates were brought into discussion in (Hintikka,1962):

consistency and deductive closure. The axioms for modeling the update function have to ensure that the function meets these postulates. A later section will explain them in more detail.

Having completed these considerations, the basic picture is sketched. The remainder of this chapter will propose formal definitions for all the elements described above. In the remainder of this chapter we will formally introduce propositions as epistemic units and show that they can be interpreted to form algebras. We will also analyze which rules should hold for the transition from a prior to a posterior state. The other epistemic entities – as there are attitudes, states, inputs and updates – will be defined in the terms of ranking theory in chapter II.

6Note that also in this most simplified presentation, the evidence is notjustthe proposition but the proposition togetherwith its posterior certainty degree. Leaving out the certainty degree in fact means to assign some default value. This may be interpreted as evidence that comes with maximal certainty. We will discuss the details later.

Referenzen

ÄHNLICHE DOKUMENTE

Now, from the point of view of Howson and Urbach's version of DBA, Dutch Book and Czech Book are on a par: whether the agent would surely lose money or surely win money if

The Laws of Belief—Ranking Theory and its Philosophical Applications. Price H/b £83.00, P/b £30.00.) Reaching print at a time when the Laplacian conception of probability is so

Penalties are imposed in the Criminal Code of the Kyrgyz Republic for crimes related to freedom of religion 15 : direct or indirect violation or restriction of rights and

Diese oder eine ähnliche Frage muß man sich wohl als Studierender immer mal stellen. Wenn man die Zeichen der Zeit bzw. der demo- kratisch legitimierten Regierung zu

Let us go back to the basic idea of confirmational holism. According to confirmational holism, the impact of an experience on the degrees of belief should depend on

agent's prior epistemic state by a set of propositions together with a subjective probability measure, and to use that measure to update the set of propositions

In chapter 2, I introduce agm theory and ranking theory, chapter 3 defines the logics used, in chapter 4, I provide some results on comparing theories of belief revision using

The theory of subjective probabilities is not an adequate description of people's epistemic states (Kahneman et al. It is a normative theory that tells uS how an