
<script type="text/javascript">
<!--
document.write('<div id="oa_widget"></div>');
document.write('<script type="text/javascript" src="https://www.openaire.eu/index.php?option=com_openaire&view=widget&format=raw&projectId=undefined&type=result"></script>');
-->
</script>
pmid: 32915750
handle: 11380/1215137 , 11585/663866 , 2158/1356460
Attention is an increasingly popular mechanism used in a wide range of neural architectures. The mechanism itself has been realized in a variety of formats. However, because of the fast-paced advances in this domain, a systematic overview of attention is still missing. In this article, we define a unified model for attention architectures in natural language processing, with a focus on those designed to work with vector representations of the textual data. We propose a taxonomy of attention models according to four dimensions: the representation of the input, the compatibility function, the distribution function, and the multiplicity of the input and/or output. We present the examples of how prior information can be exploited in attention models and discuss ongoing research efforts and open challenges in the area, providing the first extensive categorization of the vast body of literature in this exciting domain.
18 pages, 8 figures
I.2, FOS: Computer and information sciences, Computer Science - Machine Learning, Computer Science - Computation and Language, Natural language processing (NLP); neural attention; neural networks; review; survey, Computation and Language Computer Science Artificial Intelligence Learning Machine Learning Natural language processing (NLP) neural attention neural networks review survey, Computer Science - Artificial Intelligence, Natural language processing (NLP); neural attention; neural networks; review; survey;, Machine Learning (stat.ML), I.7, Machine Learning (cs.LG), I.2; I.7, Artificial Intelligence (cs.AI), Statistics - Machine Learning, 68T50, 68T05, 68T07, Computation and Language (cs.CL)
I.2, FOS: Computer and information sciences, Computer Science - Machine Learning, Computer Science - Computation and Language, Natural language processing (NLP); neural attention; neural networks; review; survey, Computation and Language Computer Science Artificial Intelligence Learning Machine Learning Natural language processing (NLP) neural attention neural networks review survey, Computer Science - Artificial Intelligence, Natural language processing (NLP); neural attention; neural networks; review; survey;, Machine Learning (stat.ML), I.7, Machine Learning (cs.LG), I.2; I.7, Artificial Intelligence (cs.AI), Statistics - Machine Learning, 68T50, 68T05, 68T07, Computation and Language (cs.CL)
citations This is an alternative to the "Influence" indicator, which also reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | 343 | |
popularity This indicator reflects the "current" impact/attention (the "hype") of an article in the research community at large, based on the underlying citation network. | Top 0.1% | |
influence This indicator reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | Top 1% | |
impulse This indicator reflects the initial momentum of an article directly after its publication, based on the underlying citation network. | Top 0.1% |
views | 135 | |
downloads | 186 |