120 likes | 214 Vues
This chapter delves into the concepts of probability theory, utility theory, and decision theory as they interact to guide an agent’s actions. Multiattribute utility functions, stochastic dominance, and decision networks are discussed, along with the value of information and decision-theoretic expert systems. The text introduces the value of perfect information, mutual preferential independence, and decision modeling techniques like augmented Bayesian networks. It also covers the evaluation of decision networks, sensitivity analysis, and the implications of different attributes on an agent’s preferences.
E N D
Chapter 16 March 25, 2004
Probability Theory: What an agent should believe based on the evidence • Utility Theory: What the agent wants • Decision Theory: Combines the above two theories to decide what the agent should do
16.4 Multiattribute Utility Functions • X = X1, … Xn • x = <x1, … xn> • By convention, higher values mean higher utilities • Strict Dominance, Figure 16.3 (no uncertainty) • Stochastic Dominance, Figure 16.4
If A stochastically dominates B, then for any monotonically nondecreasing utility function U(X), the expected utility of A is at least as high as the expected utility of B. • U(x1, … xn) = f [ f1(x1), …, fn(xn) ]
Definition: Two attributes X and Y are preferentially independent of a third attribute Z if the preference between outcomes <x, y, z> and <x’, y’, z> does not depend on z. • Definition: Mutual Preferential Independence (MPI)
Theorem: If attributes X1, …, Xn are MPI then the agent’s preference behavior can be described as maximizing the function V(x1, …, xn) = ∑ vi (xi) where each vi is a value function referring only to the attribute Xi
16.5 Decision Networks • Augmented Bayesian Networks • Figure 16.5 • Components • chance nodes (ovals) represent random variables • decision nodes (rectangles) represent choices of decision maker • utility nodes (diamonds) represent utility func.
Evaluation of Network • Set evidence variables for current state • For each possible value of decision node • Set decision node to value • Calculate posterior probability for parent nodes of utility node • Calculate resulting utility
16.6 The Value of Information • Typically, not everything is known. • Information Value Theory: Helps the agent decide what information to acquire • VPI: Value of Perfect Information
Information has value to the extent that it is likely to cause a change of plan to the extent that the new plan will be significantly better than the old plan • EU( | E) = is current best action • EU(Ej | E, Ej) = • VPIE(Ej) = Figure 16.7
Theorem: j, E { VPIE(Ej) >= 0}, i.e. the value of information is non-negative • Theorem: VPIE(Ej, Ek) = VPIE(Ej) + VPIE,Ej(Ek) = VPIE(Ek) + VPIE,Ek(Ej), i.e. collecting evidence is order independent • Figure 16.8, myopic information gathering agent
16.7 Decision Theoretic Expert Systems • Create causal model • Simplify (Figure 16.9) • Assign probabilities • Assign utilities • Verify and refine model • Perform sensitivity analysis