
arXiv: 1612.04868
handle: 10230/33295 , 10810/68986
User acceptance of artificial intelligence agents might depend on their ability to explain their reasoning, which requires adding an interpretability layer that fa- cilitates users to understand their behavior. This paper focuses on adding an in- terpretable layer on top of Semantic Textual Similarity (STS), which measures the degree of semantic equivalence between two sentences. The interpretability layer is formalized as the alignment between pairs of segments across the two sentences, where the relation between the segments is labeled with a relation type and a similarity score. We present a publicly available dataset of sentence pairs annotated following the formalization. We then develop a system trained on this dataset which, given a sentence pair, explains what is similar and different, in the form of graded and typed segment alignments. When evaluated on the dataset, the system performs better than an informed baseline, showing that the dataset and task are well-defined and feasible. Most importantly, two user studies show how the system output can be used to automatically produce explanations in natural language. Users performed better when having access to the explanations, pro- viding preliminary evidence that our dataset and method to automatically produce explanations is useful in real applications.
Preprint version, Knowledge-Based Systems (ISSN: 0950-7051). (2016)
Semantic textual similarity, FOS: Computer and information sciences, Computer Science - Machine Learning, semantic textual similarity, Computer Science - Computation and Language, tutoring systems, Computer Science - Artificial Intelligence, Natural language understanding, Machine Learning (cs.LG), Artificial Intelligence (cs.AI), Tutoring systems, natural language understanding, Interpretability, interpretability, Computation and Language (cs.CL)
Semantic textual similarity, FOS: Computer and information sciences, Computer Science - Machine Learning, semantic textual similarity, Computer Science - Computation and Language, tutoring systems, Computer Science - Artificial Intelligence, Natural language understanding, Machine Learning (cs.LG), Artificial Intelligence (cs.AI), Tutoring systems, natural language understanding, Interpretability, interpretability, Computation and Language (cs.CL)
| selected citations These citations are derived from selected sources. This is an alternative to the "Influence" indicator, which also reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | 36 | |
| popularity This indicator reflects the "current" impact/attention (the "hype") of an article in the research community at large, based on the underlying citation network. | Top 10% | |
| influence This indicator reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | Top 10% | |
| impulse This indicator reflects the initial momentum of an article directly after its publication, based on the underlying citation network. | Top 10% |
