Get 20M+ Full-Text Papers For Less Than $1.50/day. Start a 14-Day Trial for You or Your Team.

Learn More →

Solving abduction by computing joint explanations

Solving abduction by computing joint explanations An extension of abduction is investigated where explanations are jointly computed by sets of interacting agents. On the one hand, agents are allowed to partially contribute to the reasoning task, so that joint explanations can be singled out even if each agent does not have enough knowledge for carrying out abduction on its own. On the other hand, agents maintain their autonomy in choosing explanations, each one being equipped with a weighting function reflecting its perception about the reliability of sets of hypotheses. Given that different agents may have different and possibly contrasting preferences on the hypotheses to be chosen, some reasonable notions of agents’ agreement are introduced, and their computational properties are thoroughly studied. As an example application of the framework discussed in the paper, it is shown how to handle data management issues in Peer-to-Peer systems and, specifically, how to provide a repair-based semantics to inconsistent ones. http://www.deepdyve.com/assets/images/DeepDyve-Logo-lg.png Annals of Mathematics and Artificial Intelligence Springer Journals

Solving abduction by computing joint explanations

Loading next page...
 
/lp/springer-journals/solving-abduction-by-computing-joint-explanations-7aQMLJT7tx

References (79)

Publisher
Springer Journals
Copyright
Copyright © 2007 by Springer Science+Business Media B.V.
Subject
Computer Science; Complexity; Computer Science, general ; Mathematics, general; Artificial Intelligence (incl. Robotics)
ISSN
1012-2443
eISSN
1573-7470
DOI
10.1007/s10472-007-9069-y
Publisher site
See Article on Publisher Site

Abstract

An extension of abduction is investigated where explanations are jointly computed by sets of interacting agents. On the one hand, agents are allowed to partially contribute to the reasoning task, so that joint explanations can be singled out even if each agent does not have enough knowledge for carrying out abduction on its own. On the other hand, agents maintain their autonomy in choosing explanations, each one being equipped with a weighting function reflecting its perception about the reliability of sets of hypotheses. Given that different agents may have different and possibly contrasting preferences on the hypotheses to be chosen, some reasonable notions of agents’ agreement are introduced, and their computational properties are thoroughly studied. As an example application of the framework discussed in the paper, it is shown how to handle data management issues in Peer-to-Peer systems and, specifically, how to provide a repair-based semantics to inconsistent ones.

Journal

Annals of Mathematics and Artificial IntelligenceSpringer Journals

Published: Jul 17, 2007

There are no references for this article.