
In this work, we present the design and implementation of a system for proof explanation in the Semantic Web, based on defeasible reasoning. Trust is a vital feature for Semantic Web. If users (humans and agents) are to use and integrate system answers, they must trust them. Thus, systems should be able to explain their actions, sources, and beliefs. Our system produces automatically proof explanations using a popular logic programming system (XSB), by interpreting the output from the proof's trace and converting it into a meaningful representation. It also supports an XML representation for agent communication, which is a common scenario in the Semantic Web. In this paper, we present the design and implementation of the system, a RuleML language extension for the representation of a proof explanation, and we give some examples of the system. The system in essence implements a proof layer for nonmonotonic rules on the Semantic Web.
Nonmonotonic rule systems, Proof explanation, 1802 Information Systems and Management, Defeasible reasoning, Semantic Web proof layer, 004
Nonmonotonic rule systems, Proof explanation, 1802 Information Systems and Management, Defeasible reasoning, Semantic Web proof layer, 004
| selected citations These citations are derived from selected sources. This is an alternative to the "Influence" indicator, which also reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | 14 | |
| popularity This indicator reflects the "current" impact/attention (the "hype") of an article in the research community at large, based on the underlying citation network. | Average | |
| influence This indicator reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | Top 10% | |
| impulse This indicator reflects the initial momentum of an article directly after its publication, based on the underlying citation network. | Top 10% |
