Uncertainty in Artificial Intelligence
First Name   Last Name   Password   Forgot Password   Log in!
    Proceedings         Authors   Author's Info   Article details         Search    
Possibilistic Conditioning and Propagation
Yen-Teh Hsia
We give an axiomatization of confidence transfer - a known conditioning scheme - from the perspective of expectation-based inference in the sense of Gardenfors and Makinson. Then, we use the notion of belief independence to "filter out" different proposal s of possibilistic conditioning rules, all are variations of confidence transfer. Among the three rules that we consider, only Dempster's rule of conditioning passes the test of supporting the notion of belief independence. With the use of this conditioning rule, we then show that we can use local computation for computing desired conditional marginal possibilities of the joint possibility satisfying the given constraints. It turns out that our local computation scheme is already proposed by Shenoy. However, our intuitions are completely different from that of Shenoy. While Shenoy just defines a local computation scheme that fits his framework of valuation-based systems, we derive that local computation scheme from II(,8) = tI(,8 I a) * II(a) and appropriate independence assumptions, just like how the Bayesians derive their local computation scheme.
Pages: 336-343
PS Link:
PDF Link: /papers/94/p336-hsia.pdf
AUTHOR = "Yen-Teh Hsia ",
TITLE = "Possibilistic Conditioning and Propagation",
BOOKTITLE = "Proceedings of the Tenth Conference Annual Conference on Uncertainty in Artificial Intelligence (UAI-94)",
PUBLISHER = "Morgan Kaufmann",
ADDRESS = "San Francisco, CA",
YEAR = "1994",
PAGES = "336--343"

hosted by DSL   •   site info   •   help