Preference Handling for Belief

Preference Handling for Belief-Based Rational Decisions
Samy Sá∗, João Alcântara∗
1
Universidade Federal do Ceará (UFC)
MDCC, Campus do Pici, Bl 910 – Fortaleza, Brazil
[email protected],[email protected]
Abstract. We present an approach to preferences suitable for agents that base
decisions on their beliefs. In our work, agents’ preferences are perceived as a
consequence of their beliefs, but at the same time are used to feed the knowledge base with beliefs about preferences. As a result, agents can reason with
preferences to hypothesize, explain decisions, and review preferences in face of
new information. Finally, we integrate utility-based to reasoning-based criteria
of decision making.
1. Introduction
Autonomous agents are frequently required to make decisions and expected to do so according to their beliefs, goals, and preferences, however, beliefs are rarely connected to
the preferences of the agent. Dietrich and List argue in [Dietrich and List 2013] that logical reasoning and the economic concept of rationality are almost entirely disconnected
in the literature: while logical accounts of reasoning rarely gets to deal with rational decisions in the economic sense, social choice is never worried about the origin of agents’
preferences. But if preferences are disconnected from beliefs, how can an agent explain
their decisions? How could we model the influence of new information in an agent’s
preferences?
In order to reason about options available in a decision, an agent needs a complimentary theory of what the best options are. Since rationality involves trying to maximize
gains (utility), the successful integration of rationality and beliefs requires quantifying the
utility of some of those beliefs. By doing so, agents’ preferences are a reflex of beliefs
about the possible decisions and the agent is capable of different reasoning tasks involving
their preferences such as (i) building arguments to explain decisions; (ii) considering alternative scenarios with different outcomes for the decision; (iii) automatically updating
preferences in face of new information; and (iv) hypothesizing with preferences.
Further, reasoning with preferences under uncertainty is a feature rarely considered in the literature (as mentioned in [Domshlak et al. 2011]), even though it plays a key
role in rational choice and game theory [Osborne and Rubinstein 1994]. By accounting
for knowledge bases as answer set programs [Gelfond and Lifschitz 1991], our approach
can model multiple plausible scenarios of a decision, therefore accounting for uncertainty
while making it easy to perform abduction and belief revision as means to evaluate or update an agent’s preferences. In this paper, we show how comparing available options by
∗
This paper is an extended and revised version of [Sá and Alcântara 2013]. Our work is supported by
CNPq (Universal 2012 - Proc. 473110/2012-1), CAPES (PROCAD 2009), and CNPq/CAPES (Casadinho/PROCAD 2011).
weighing their relevant features can successfully connect rationality and qualitative preferences in a way to relate rational decision criteria [Osborne and Rubinstein 1994] and an
inherently qualitative criteria for reasoning-based decisions [Dubois et al. 2008].
The paper is organized as follows. In Section 2, we discuss some concepts of
the formalism used in the paper. In Section 3, we present our preferences model. In
Section 4, we show how beliefs about preferences can be integrated into a knowledge
base to perform different reasoning tasks revolving preferences. Section 5 shows how
agents make decisions and that utility-based decisions are related to beliefs in our model.
Related work is in Section 7, then our conclusions and references complete the paper.
2. Knowledge Bases
In this paper, we consider agents with knowledge bases (or belief sets) as answer set
programs [Gelfond and Lifschitz 1991]. Such programs can present several models called
answer sets, to which we sometimes refer as possible or plausible scenarios conceived by
the knowledge base. The possible presence of multiple models is an interesting feature to
account for uncertainty and we will later show that they are suitable to express uncertainty
over preferences and arguing about them. The notions of preferences here introduced can
be generalized to other kinds of logic languages with unary predicates.
An Extended Logic Program (ELP) or Answer Set Program
[Gelfond and Lifschitz 1991] is defined over a Herbrand Base HB, the set of all
ground atoms the program might resort to. An ELP consists of a set of rules of the form
r : L0 ← L1 , . . . , Lm , not Lm+1 , . . . , not Ln
such that each Li is a literal, i.e., it is either an atom (A) or its negation (¬A). The symbol
not denotes negation as failure (NAF) and not L is a NAF-literal if L is a literal. We
may speak of literals to generalize literals and NAF-literals. In a rule r as above, we
refer to L0 as the head of r and write head(r) to denote the set {L0 }. We refer to the
conjunction L1 , . . . , Lm , not Lm+1 , . . . , not Ln as the body of r, and body(r) denotes the
set {L1 , . . . , Lm , not Lm+1 , . . . , not Ln }. The literals of the positive and negative parts
of the body i.e. the sets {L1 , . . . , Lm } and {Lm+1 , . . . , Ln }, are denoted by body + (r) and
body − (r) respectively. We also indicate the set of NAF-Literals {not Lm+1 , . . . , not Ln }
as not body − (r). A rule may be written as head(r) ← body + (r), not body − (r) as well.
A rule is an integrity constraint if head(r) = ∅ and it is a fact if body(r) = ∅.
We say that a program, rule or literal without variables is ground. A rule with
variables is seen as a succinct manner to represent all of it’s ground instances, which
can be computed by applying every possible substitution θ = {x1 /t1 , . . . xn /tn } from
variables to terms in HB where x1 , . . . , xn are all distinct variables and all ti is a distinct
term from xi . We often refer to ground predicated formulas as propositions.
The semantics of an ELP is given by the Answer Sets Semantics
[Gelfond and Lifschitz 1991]. A program may have one, zero or multiple answer sets.
An answer set S for KB is consistent if S does not simultaneously contain A and ¬A, for
no atom in the language. The program itself will be said consistent if it has a consistent
answer set. Otherwise, the program is inconsistent. Throughout the paper we will assume
only consistent programs. A goal is a conjunction of literals and NAF-literals. We say
that KB credulously (resp. skeptically) satisfies a goal G if some (resp. all) of its answer
sets satisfy G, in which case we write KB |=c G (resp. KB |=s G).
3. Preferences as Utility + Beliefs
In this section, we introduce our notions of preference profiles and quality thresholds.
Agent preferences are drawn on top of beliefs by attributing weights to unary predicates
expressing relevant features (alike with what is done in weighted propositional formulas
[Lafage and Lang 2000]), which results in an unary utility function. Quality thresholds
are employed to classify options as good, poor or neutral in the eyes of an agent.
Definition 1 (preference profile) Let P red be the set of all unary predicates P (x) used
to express possible features of options (e.g. if a car is expensive) in the language of an
agent. A preference profile is a triple P r = hU t, U p, Lwi, involving a utility function
U t : P red → R, and upper and lower utility thresholds U p, Lw ∈ R, U p ≥ Lw.
An agent can have as many preference profiles as desired for each kind of decision
the agent may get involved (as proposed in [Dietrich and List 2013]). Intuitively, each
preference profile specifies a view about (i) which features (predicates) are relevant to
quality available options (they have U t(P (x)) 6= 0); and (ii) what it takes for an option to
be qualified as good, poor or indifferent. The profiles (both utilities and thresholds) can
be designed together with the knowledge base or obtained from other methods, such as
modeling user preferences. In each case, a preference profile gives us two perspectives of
an agents preferences: quantitative and qualitative.
Definition 2 (available options) We write O = {o1 , . . . , on }, n ≥ 2 to refer to the set of
options (outcomes) available in a decision.
The available options are constants in the language of the agent.
Given a program KB and a preference profile P r = hU t, U p, Lwi, to rank the
available outcomes is straightforward, as each option oi ∈ O, has utility
P
U tS (oi ) =
U t(P (x))
P (oi )∈S
in the answer set S. The utility function takes answer sets as a parameter because
the multiplicity of answer sets may suggest uncertainty about the features satisfied by each
option. In a particular answer set S, oi is perceived as a good option if U tS (oi ) ≥ U p, a
poor option if U tS (oi ) < Lw, and as neutral (neither good or poor) otherwise. Clearly, if
U p = Lw, there are only good and poor options.
The concept of preference profile in Definition 1 induces, for each answer set S
of the agent theory, a total preorder1 over the set of available options
<S = {(oi , oj ) | U tS (oi ) ≥ U tS (oj )}.
The proposition oi <S oj is usually read as “oi is weakly preferred to oj
[Fishburn 1999] in S”. If the agent is indifferent about two options oi , oj , i.e., if both
oi <S oj and oj <S oi hold, we write oi ∼S oj .
From now on, we will account for agents as a pair Ag = (KB, P ref s) involving
a program KB and a set P ref s of preference profiles.
1
A total preorder is a relation that is transitive, reflexive and in which any two elements are related.
Total preorders are also called weak orders.
Example 1 Suppose an agent whose knowledge base includes beliefs about Australia (a),
Brazil (b), and Canada (c), possible destinations (D(x)) of a future vacation trip. Interesting features of a destination to the agent include the presence of famous beaches (Bc(x)),
the availability of cheap flights (Ch(x)) and whether there is a spoken language in x that
is unknown to the agent (U L(x)). A destination x will present unknown languages if there
is at least one language y that is spoken there (SL(y, x)) that is not amongst those spoken
by the agent (S(y)). The agent speaks two languages, namely English (en) and German
(ge). Finally the agent knows their partner has a favorite destination (P F (x)), but does
not know which one is their favorite.
We use the semi-colon to separate rules and a period to mark the last one.
KB : D(a) ← ; D(b) ← ; D(c) ← ;
Bc(a) ← ; Bc(b) ← ; ¬Bc(c) ← ;
Ch(c) ← ;
U L(x) ← D(x), SL(y, x), not S(y) ;
S(en) ← ; S(ge) ← ;
SL(en, a) ← ; SL(pt, b) ← ; SL(en, c) ← ; SL(f r, c) ← ;
P F (a) ← not P F (b), not P F (c);
P F (b) ← not P F (c), not P F (a);
P F (c) ← not P F (a), not P F (b).
The agent has a few preference profiles for this decision:
• P r1 = hU t1 = {(Bc(x), 3), (U L(x), −1), (Ch(x), 4), (D(x), 0), (S(x), 0),
(P F (x), 0)}, U p1 = 3, Lw1 = 2i;
• P r2 = hU t2 = {(Bc(x), 3), (U L(x), 2), (Ch(x), 4), (D(x), 0), (S(x), 0),
(P F (x), 0)}, U p2 = 5, Lw2 = 4i;
• P r3 = hU t3 = {(Bc(x), 3), (U L(x), −1), (Ch(x), 4), (D(x), 0), (S(x), 0),
(P F (x), 3)}, U p3 = 5, Lw3 = 4i.
The program has three answer sets S1 = S ∪ {P F (a)}, S2 = S ∪ {P F (b)} and S3 =
S ∪ {P F (c)}, where S = {Bc(a), Bc(b), U L(b), Ch(c)}. For preference profiles P r1
and P r2 , the utilities are the same in all answer sets. If the agent selects the preference
profile P r1 , it would find utilities U tS1 1 (a) = 3, U tS1 1 (b) = 2, and U tS1 1 (c) = 4. In that
case, option c (going to Canada) is the favorite, but going to Australia is also a good
option and the agent is neutral about going to Brazil. If the agent selects the preference
profile P r2 , however, an option will only be good if it holds two or more relevant features,
since no single trait has utility higher than U p2 on its own. In that case, going to Brazil
would be the only good option, while going to Canada would be still considered neutral.
The last profile P r3 suggests uncertainty, as in S1 Australia is the only good option while
going to Brazil is a poor option and going Canada is neither a good nor a poor option
(it is neutral), but the different answer sets S2 and S3 suggests differently. This models a
state of mind where the preference of the agent’s partner is definitive of their own: Each
answer set favors a different destination according to the partner’s favorite. The utilities
and status of the three options according to each profile are summarized in Table 1.
3.1. On Multiple Preference Profiles
In [Dietrich and List 2013], the authors propose that the preferences of an agent are conditioned by a set of reasons M and each subset of M induces a different preference ordering. In our approach, different motivational states are represented as alternative preference
Australia
Brazil
Canada
P r1 in Si
3, good
2, neutral
4, good
P r2 in Si
3, poor
5, good
4, neutral
P r3 in S1
6, good
2, poor
4, neutral
P r3 in S2
3, poor
5, good
4, neutral
P r3 in S3
3, poor
2, poor
7, good
Table 1. The options according to each profile in each answer set Si , i ∈ {1, 2, 3}.
profiles about the same decision. Special propositions in the theory of the agent can be
used to determine which profile is to be assumed. Dietrich and List also introduce two
axioms which should govern the relationship between an agent’s set of beliefs and their
preferences across variations of their motivational state:
Axiom 1 ([Dietrich and List 2013]) For any two options oi , oj ∈ O and any M ⊆ P red,
if {P (x) ∈ M | P (oi ) holds} = {P (x) ∈ M | P (oj ) holds}, then x ∼ y.
Axiom 2 ([Dietrich and List 2013]) For any oi , oj ∈ O and any M, M 0 ⊆ P red such
that M 0 ⊇ M , if {P (x) ∈ M 0 \ M | P (oi ) or P (oj ) hold} = ∅, then x <M y ⇔ x <M 0 y.
Roughly speaking, Axiom 1 states that two options with the exact same characteristics should be equally preferred. Axiom 2 treats the case where new attributes become
relevant to the decision. In that case, the preferences over any two options that do not
satisfy the extra criteria should remain unchanged.
Theorem 1 Each preference relation <S (Sec. 3) satisfies Axioms 1 and 2.
Proof 1 Consider any consistent answer set S of the agent’s knowledge base KB. If two
options satisfy the exact same predicates in S, the utility attributed to both options will
be the same, so Axiom 1 is satisfied. Now suppose a preference profile P r is updated in
a way the predicate P (x) becomes relevant, i.e., in the updated profile P r0 , U t0 (P (x)) 6=
0, while originally, U t(P (x)) = 0. If neither options oi , oj satisfy P (x) in S, i..e, if
P (oi ), P (oj ) 6∈ S, then their utility remains unchanged. Therefore, Axiom 2 is satisfied.
As a consequence, our approach to model preferences of an agent will adequately
relate their belief sets to their preferences. Variations of the agent’s motivational state
are modeled as different profiles for the same decisions. We will see in the next section
how the preferences are integrated in the knowledge base and briefly comment on how
the different profiles can be considered as required.
4. Rules for the Evaluation of Options and Beliefs About Preferences
In this section, we will show how to build a general theory of preferences so agents are
able to reason about the quality of available options. We use the utility function and
the qualitative thresholds of a preference profile (Sec. 3) to devise a set of special rules
that, when appended to the original knowledge base (as a module), promotes conclusions
about the quality of available options. Therefore, rules encoding preferences should not
interfere much with the models of the original program: It is tolerated to add conclusions
about the quality of options and nothing else.
For that matter, we introduce predicates G(x), N (x), standing respectively for
good (G(x)) and neutral (N (x)) options, while options known not to be good are poor
(¬G(x)). We assume G(x), N (x) are reserved, so they do not appear in knowledge bases.
Given an agent Ag = (KB, P ref s) and a selected preference profile P r ∈
P ref s, P r = hU t, U p, Lwi, the set KBP r of preference rules is built as follows:
Let R∗ ⊆ P red be the set of predicates relevant to a decision, i.e., any P (x) with
U t(P (x)) 6= 0, compute the utility of each subset of features from R,
P
R
Ut =
U t(P (x)),
P (x)∈R
for each R ⊆ R∗ . Then, KBP r will be the set of rules r of the form
r : head(r) ← body + (r), not body − (r) where:
• body + (r) = {P (x)|P (x) ∈ R};
• and not body − (r) = {not P (x)|P (x) ∈ R∗ \ R};
• head(r) = G(x) if R U t ≥ U p, head(r) = ¬G(x) if R U t < Lw, or head(r) =
N (x) otherwise;
Example 2 (Continuation of Example 1) For the agent in Example 1, we would have a
complimentary knowledge base KBP r1 , based on profile P r1 , as follows:
KBP r1 : G(x) ← Bc(x), U L(x), Ch(x);
N (x) ← Bc(x), U L(x), not Ch(x);
G(x) ← Bc(x), not U L(x), Ch(x);
G(x) ← Bc(x), not U L(x), not Ch(x);
G(x) ← not Bc(x), U L(x), Ch(x);
¬G(x) ← not Bc(x), U L(x), not Ch(x);
G(x) ← not Bc(x), not U L(x), Ch(x);
¬G(x) ← not Bc(x), not U L(x), not Ch(x);
The resulting agent knowledge base will be KB 1 = KB ∪ KBP r1 .
The inclusion of rules about preferences provides a description with a global perspective of what good/poor/neutral options are, enriching the original knowledge base
with conclusions about the quality of available options. While employing rules as above,
we have the integration of preferences and beliefs in two ways. First, the beliefs promote
the utility of each option and, therefore, the preferences are based on them. Second, the
calculated utilities of each option are used to complement the knowledge base with predicated formulas regarding beliefs about them. Observe the composed program has the
same answer sets as the original with only extra conclusions about the quality of each option, so proposed set of rules also allow different answer sets to have different conclusions
on the quality of available outcomes (such as proposed in Table 1 based on Example 1).
This models uncertainty about the utility and quality of available options.
It can be observed that the number of rules is exponential in the number |R∗ | of
relevant predicates to a profile. However, this calculation needs to be done only once and
the number of predicates is not likely to be high. Also, this reflects the complexity of
making decision by comparing options according to a particular set of relevant properties.
Regarding the multiplicity of profiles, a complimentary agent theory on its motivational states could be devised to conclude which profile to access in each situation. This
can be attained by linking the rules from preference profiles with new formulas reserved
for that purpose and an extra set of rules with such reserved formulas for conclusions.
5. Making Decisions
Now we will show that our approach to preferences successfully connects beliefs and
preferences as we prove that employing utility-based criteria to decision making satisfies
criteria from reasoning-based decision making.
In Dubois et al. [Dubois et al. 2008], attributes are classified as negative (−),
neutral (0) or positive (+). The positive (resp. negative) attributes are then perceived
as arguments in favor (resp. against) the available outcomes and the decision is based
on comparing the sets of arguments against and in favor, which are here introduced as
features (unary predicates). As the utility of features in our approach ranges in the real
numbers, our own model of preferences can be seen as bipolar. Thus, our proposal induces a bipolar decision process and should satisfy the principle of bivariate monotony
[Dubois et al. 2008] which we adapted to deal with uncertainty and multiple scenarios.
In the following sections, consider oi and oj are two options, KB denotes a
S
knowledge base, and the preference profile P r = hU t, U p, Lwi is selected. Let o+
i
denote the set of positive features satisfied in S, i.e., those P (x) such that P (oi ) ∈ S and
S
U t(P (x)) > 0, and o−
denote the set of negative features satisfied in S, i.e., those P (x)
i
such that P (oi ) ∈ S and U t(P (x)) < 0.
5.1. Single Agent Decision
Definition 3 (bivariate monotonicity) A decision criteria satisfies bivariate monotonicity
−S
S
S
S
if, for each answer set S of KB, whenever o+
⊇ o+
⊆ o−
i
j and oi
j available, it prefers
oi at least as much as oj .
Bivariate monotonicity states that whenever an option has all advantages of another (possibly more), but no disadvantages the second does not (possibly less), the first
option is possibly better. Of course, if the set of pros and cons is exactly the same, one
option is just as good as the other.
Theorem 2 When there is no uncertainty, i.e., there is a single plausible scenario, maximizing utility satisfies bivariate monotonicity.
Proof 2 Remember that positive features have positive utility while negative features have
+
−
−
negative utility. Then, in any specific scenario S, if o+
i ⊇ oj while oi ⊆ oj we will have
U tS (oi ) ≥ U tS (oj ). We conclude that maximizing utility satisfies bivariate monotonicity.
5.2. Single Agent Under Uncertainty
Our notion of preference profile relates to game theory, since multiple plausible scenarios
induces the construction of a payoff matrix. See, for instance, that in Example 1, utility
varies amongst answer sets when the preference profile P r3 is assumed. Table 2 below
depicts the induced payoff matrix with Australia, Brazil and Canada as available moves
and the possible states of nature are S1 , S2 , S3 , the answer sets of KB.
For this reason, it is natural to associate decision making under uncertainty to decision criteria from game theory [Osborne and Rubinstein 1994]. We apply the maximin
[Osborne and Rubinstein 1994] criteria to compare options, which consists of trying to
maximize the minimal gains: For each option oi ∈ O, we calculate M IN ({U tS (oi ) | S ∈
AS}), the minimal utility attributed to each oi amongst scenarios, then we compare these
results to induce a partial preorder <M AXIM IN over the set of alternatives.
S1
S2
S3
Australia 6, good
3, poor
3, poor
Brazil
2, poor
5, good 2, poor
Canada 4, neutral 4, neutral 7, good
Table 2. A payoff matrix with possible scenarios S1 , S2 , andS3 .
Definition 4 (maximin criteria) In a decision problem with uncertainty, let AS denote
the set of possible scenarios (answer sets) compatible with the beliefs of an agent. An
option oi is at least as preferred as the option oj by maximin, i.e. oi <M AXIM IN oj iff
M IN ({U tS (oi ) | S ∈ AS}) ≥ M IN ({U tS (oj ) | S ∈ AS}).
The best options are the maximal elements of <M AXIM IN , i.e., those oi such that
oi <M AXIM IN oj for all oj , j 6= i.
Example 3 (Uncertainty) Consider an agent with preferences as depicted in Table 2. The
agent can choose from Australia, Brasil or Canada, and there are three possible scenarios
where the decision may take effect. The agent has no control about which scenario will
be the case, so making a rational choice involves considering all three. The maximin
criteria proposes the agent to maximize the minimal gain, i.e., to get the best guaranteed
satisfaction. Since the minimal payoffs for Australia, Brazil and Canada are, respectively,
3, 2, and 4, by the maximin criteria, the agent should choose to go to Canada.
Theorem 3 The maximin criteria for decision making under uncertainty satisfies bivariate monotonicity (Definition 3) in the case of multiple scenarios.
S
S
S
S
Proof 3 Let oi , oj be any two available options, and suppose o+
⊇ o+
and o−
⊆ o−
i
j
i
j
hold in every S ∈ AS (if any different, the options cannot be directly compared). Let
the quality of any options with the least quality be M IN ({U tS (oj ) | S ∈ AS}) = m.
Let M ∈ AS be any answer set such that U tM (oj ) = m. By Theorem 2, U tM (oi ) ≥
U tM (oj ), so the least utility attributed to oi is M IN ({U tS (oi ) | S ∈ AS}) ≥ m and
oi <M AXIM IN oj . We conclude oi <M AXIM IN oj , i.e., oi is at least as preferred as oj .
5.3. Quality Instead of Utility
Alternatively, the agent can just try to maximize the quality of it’s choice. In that setting, if the agent consider several different options all as good, picking any good option would be rational, even if the choice does not maximize utility. This kind of decision is usually the case in many settings, as argued, for instance in Approval Voting
[Brams and Fishburn 1978] where agents attribute votes to each of those options they
consider to be good enough. As the next theorem states, attempts to maximize quality as
the criteria for rationality also satisfy bivariate monotonicity.
In the following, QS (oi ) stands for the quality of option oi in scenario S. We
compare the quality of different options by considering good > neutral > poor.
Theorem 4 Maximizing quality satisfies bivariate monotonicity (Definition 3).
+
−
−
Proof 4 In a specific scenario S, if o+
i ⊇ oj while oi ⊆ oj , we have, from Theorem 2,
that U tS (oi ) ≥ U tS (oj ). As quality is obtained from utility, QS (oi ) ≥ QS (oj ) follows.
This is also the case with multiple scenarios (uncertainty) and the criteria of choosing the option with maximal minimal quality.
Theorem 5 The maximin criteria for decision making under uncertainty, when interpreted as maximizing the minimal quality of a choice, satisfies bivariate monotonicity
(Definition 3) in the case of multiple scenarios.
S
S
S
S
Proof 5 Let oi , oj be any two available options, and suppose o+
⊇ o+
and o−
⊆ o−
i
j
i
j
hold in every S ∈ AS (if any different, the options cannot be directly compared). Let
the quality of any options with the least quality be M IN ({QS (oj ) | S ∈ AS}) = q and
M ∈ AS be an answer set such that QM (oj ) = q. By Theorem 4, QM (oi ) ≥ QM (oj ), so
the least quality attributed to oi is M IN ({QS (oi ) | S ∈ AS}) ≥ q. As a consequence,
oi <M AXIM IN oj , i.e., oi is at least as preferred as oj .
6. On Reasoning About Preferences
Besides connecting beliefs and preferences, our approach is flexible enough to deal with
different aspects of reasoning, even simultaneously. One key aspect of our model is the
ability to have different perspectives of preferences integrated in a single formalism. We
argue this opens many doors towards reasoning about preferences, such as building arguments to explain decisions and having preferences easily updated in case the agent learns
new information. We briefly explain some possible features.
6.1. Arguments About The Quality of Options
As a desirable capacity of agents in a society, and especially for those that take part
in collective decisions, being able to explain their decisions or opinions in a collective
decision is very important. Intuitively, an argument consists of a conclusion and some
justification to it (a proof) [Sá and Alcântara 2012]. As we introduce rules on the quality
of options (Sec. 4), the agent becomes able to build arguments to explain their decisions
or argue about options. The opinion of an agent about a particular option can also be used
to determine which roles (proponent, opponent) the agent will play in a group decision
dialogue. For instance, if the agent believes an option oi cannot be good, i.e., KB 6|=c
G(oi ), the agent could oppose (argue against) that the group chooses oi .
6.2. Revising Preferences by Revising Beliefs
Adding or removing formulas in an agent theory KB can change its models and, therefore, the conclusions of an agent about how good each available option is. Therefore, performing belief revision [Alchourrón et al. 1985, Krümpelmann and Kern-Isberner 2012]
reflects in automatic updates to the utilities attributed to available options, while any utility
functions originally used are not affected. As a consequence, the rules about preferences
(from Sec. 4) are kept the same and no extra computation is required for these rules, but
the agent is always prepared to re-evaluate options in face of new information.
6.3. Hypothesizing About Preferences
An agent capable of abductive reasoning2 can hypothesize over options by conceiving
alternative scenarios where the quality of available options would be different. This
2
For a survey on abduction in logic programming, please refer to [Denecker and Kakas 2002].
is particularly useful in a number of interactions with other agents such as to to formulate questions for inquiry, build special arguments to seek consensus in deliberation dialogues [Sá and Alcântara 2012] or to to elaborate new proposals in negotiation
[Sakama and Inoue 2007].
6.4. Dealing with New Options
In a decision process, the agent could learn about other options previously unknown.
If equipped with abduction and belief revision, the agent is able to conceive alternative
situations (abductively) in which such options would be good (or neutral, or poor), in
order to inquiry and learn facts (belief revision) to attribute them utility. As shown in
[Kakas and Mancarella 1990] and [Sakama and Inoue 2003], abductive reasoning can be
used to guide knowledge updates, since explanations indicate consistent ways to update
a program in order to satisfy goal formulas. Based on inquiry or dialogues, the agent is
therefore able to learn about the new options and have proper opinions on their quality.
7. Related Work
Closely related to our work, Dietrich and List [Dietrich and List 2013] observe that logical reasoning and the economic concept of rationality are almost entirely disconnected in
the literature. In their work, Dietrich and List propose preference orderings based on alternative logical contexts as being different psychological states of the agent. We connect
logics and utility-based decisions in this paper by attaching utilities to predicates, so the
best outcomes satisfy the most interesting combinations of predicates amongst available
candidates. In particular, we connect rational behavior from game theory to bipolar preferences due to Dubois et al. [Dubois et al. 2008]. On the other hand, Dietrich and List are
not concerned with reasoning about preferences in [Dietrich and List 2013], but restrict
their analysis to how beliefs can influence decisions.
Lafage and Lang propose in [Lafage and Lang 2000] an approach to group decisions based on weighted logics in which formulas represent constraints and weights are
attributed to each formula to quantify importance. This is related to the way we build utility functions and preferences, however they allow formulas with connectives, while we
restrict utility evaluation to ground formulas with unary predicates. Brafman proposes in
[Brafman 2011] a relational language of preferences where rules may include utility values in the conclusions. The author argues the approach is flexible, as their value functions
can handle a dynamic universe of objects. We consider our approach is also very flexible, as an agent can quickly evaluate a new option, since all the weight is in the features
relevant to the agent.
Several notions of preferences in logic programs have been studied in the
literature, but mainly applied to rank the different models a program may have
[Delgrande et al. 2004]. As it can be observed in Delgrande et al. [Delgrande et al. 2004],
most approaches involve introducing priorities to rules to induce an ordering over the answer sets. An example is the work of Brewka and Eiter [Brewka and Eiter 1999]. In
a different approach Sakama and Inoue introduced priorities over propositions (atoms,
ground predicated literals) from the program language in [Sakama and Inoue 2000]. Our
proposal, on the other hand, takes advantage of multiple answer sets to reason over uncertainty as decision are utility-based and built over unary predicates. This kind of approach
is surprisingly not common, as also stated by Doyle in [Doyle 2004]. A notable exception
is [Labreuche 2011], where Labreuche deals with the issue of explaining a decision made
in a multi-attribute decision model where attributes are weighted in an attempt to circumvent the difficulties of properly explaining utility-based decisions. Anyway, our work has
an alike motivation: arguments are better related to logic-based formalisms, so we attach
weights to attributes (predicates) to calculate preferences and translate them back into a
logical language. For an account of approaches and applications of preferences in Artificial Intelligence, see the more recent survey from Domshlak et al. [Domshlak et al. 2011].
8. Conclusions and Future Work
We have introduced an approach to integrate beliefs and rationality in a way utility-based
criteria for decisions is shown to satisfy criteria from reasoning-based decision making.
This is achieved by attributing utility to unary predicates used to encode relevant qualities
of a decision, together with the notion of qualitative thresholds, both parts of our preference profiles. By doing so, agent preferences are simultaneously perceived in different
perspectives: a utility-based cardinal order, a regular ordinal order and a classification in
good/poor/neutral options. All three perspectives are integrated as we devise rules that
encode the preferences of the agent and append them to its knowledge base for reasoning purposes. The result is a formalism capable of modeling reasoning about preferences
where beliefs are the base for rational decisions. Next, we intend to look deeper in the relations between logical consequences and utility-based choices induced by our proposal
to preferences. We also plan to explore how this model can be employed in group decisions by deliberation and voting. Finally, motivated by work in bipolar decisions as
formal argumentation by Amgoud and Vesic [Amgoud and Vesic 2012], we will analyze
how formalisms of decision making by argumentation relate to rationality criteria of decision making in groups if our approach to preferences is used.
References
Alchourrón, C. E., Gärdenfors, P., and Makinson, D. (1985). On the logic of theory
change: Partial meet contraction and revision functions. J. Symb. Log., 50(2):510–530.
Amgoud, L. and Vesic, S. (2012). On the use of argumentation for multiple criteria decision making. In Greco, S., Bouchon-Meunier, B., Coletti, G., Fedrizzi, M., Matarazzo,
B., and Yager, R. R., editors, IPMU (4), volume 300 of Communications in Computer
and Information Science, pages 480–489. Springer.
Brafman, R. I. (2011). Relational preference rules for control. Artif. Intell., 175(78):1180–1193.
Brams, S. J. and Fishburn, P. C. (1978). Approval voting. The American Political Science
Review, 72(3):pp. 831–847.
Brewka, G. and Eiter, T. (1999). Preferred answer sets for extended logic programs. Artif.
Intell., 109(1-2):297–356.
Delgrande, J. P., Schaub, T., Tompits, H., and Wang, K. (2004). A classification and
survey of preference handling approaches in nonmonotonic reasoning. Computational
Intelligence, 20(2):308–334.
Denecker, M. and Kakas, A. C. (2002). Abduction in logic programming. In Kakas, A. C.
and Sadri, F., editors, Computational Logic: Logic Programming and Beyond, volume
2407 of Lecture Notes in Computer Science, pages 402–436. Springer.
Dietrich, F. and List, C. (2013). A reason-based theory of rational choice*. Noûs,
47(1):104–134.
Domshlak, C., Hüllermeier, E., Kaci, S., and Prade, H. (2011). Preferences in ai: An
overview. Artif. Intell., 175(7-8):1037–1052.
Doyle, J. (2004). Prospects for preferences. Computational Intelligence, 20(2):111–136.
Dubois, D., Fargier, H., and Bonnefon, J.-F. (2008). On the qualitative comparison of
decisions having positive and negative features. J. Artif. Intell. Res. (JAIR), 32:385–
417.
Fishburn, P. C. (1999). Preference structures and their numerical representations. Theor.
Comput. Sci., 217(2):359–383.
Gelfond, M. and Lifschitz, V. (1991). Classical negation in logic programs and disjunctive
databases. New Generation Comput., 9(3/4):365–386.
Kakas, A. C. and Mancarella, P. (1990). Database updates through abduction. In McLeod,
D., Sacks-Davis, R., and Schek, H.-J., editors, VLDB, pages 650–661. Morgan Kaufmann.
Krümpelmann, P. and Kern-Isberner, G. (2012). Belief base change operations for answer set programming. In Proceedings of the 13th European conference on Logics in
Artificial Intelligence, JELIA’12, pages 294–306, Berlin, Heidelberg. Springer-Verlag.
Labreuche, C. (2011). A general framework for explaining the results of a multi-attribute
preference model. Artif. Intell., 175(7-8):1410–1448.
Lafage, C. and Lang, J. (2000). Logical representation of preferences for group decision
making. In Cohn, A. G., Giunchiglia, F., and Selman, B., editors, KR, pages 457–468.
Morgan Kaufmann.
Osborne, M. and Rubinstein, A. (1994). A Course in Game Theory. MIT Press.
Sá, S. and Alcântara, J. (2012). Cooperative dialogues with conditional arguments. In
van der Hoek, W., Padgham, L., Conitzer, V., and Winikoff, M., editors, AAMAS,
pages 501–508. IFAAMAS.
Sá, S. and Alcântara, J. (2013). Preference handling for belief-based rational decisions. In
Cabalar, P. and Son, T. C., editors, LPNMR, volume 8148 of Lecture Notes in Computer
Science, pages 518–523. Springer.
Sakama, C. and Inoue, K. (2000). Prioritized logic programming and its application to
commonsense reasoning. Artif. Intell., 123(1-2):185–222.
Sakama, C. and Inoue, K. (2003). An abductive framework for computing knowledge
base updates. TPLP, 3(6):671–713.
Sakama, C. and Inoue, K. (2007). Negotiation by abduction and relaxation. In Durfee,
E. H., Yokoo, M., Huhns, M. N., and Shehory, O., editors, AAMAS, page 242. IFAAMAS.