<script type="text/javascript">
<!--
document.write('<div id="oa_widget"></div>');
document.write('<script type="text/javascript" src="https://www.openaire.eu/index.php?option=com_openaire&view=widget&format=raw&projectId=undefined&type=result"></script>');
-->
</script>
handle: 11245/1.194131 , 2144/29381
While Kolmogorov complexity is the accepted absolute measure of information content of an individual finite object, a similarly absolute notion is needed for the relation between an individual data sample and an individual model summarizing the information in the data, for example, a finite set (or probability distribution) where the data sample typically came from. The statistical theory based on such relations between individual objects can be called algorithmic statistics, in contrast to classical statistical theory that deals with relations between probabilistic ensembles. We develop the algorithmic theory of statistic, sufficient statistic, and minimal sufficient statistic. This theory is based on two-part codes consisting of the code for the statistic (the model summarizing the regularity, the meaningful information, in the data) and the model-to-data code. In contrast to the situation in probabilistic statistical theory, the algorithmic relation of (minimal) sufficiency is an absolute relation between the individual model and the individual data sample. We distinguish implicit and explicit descriptions of the models. We give characterizations of algorithmic (Kolmogorov) minimal sufficient statistic for all data samples for both description modes--in the explicit mode under some constraints. We also strengthen and elaborate earlier results on the ``Kolmogorov structure function'' and ``absolutely non-stochastic objects''--those rare objects for which the simplest models that summarize their relevant information (minimal sufficient statistics) are at least as complex as the objects themselves. We demonstrate a close relation between the probabilistic notions and the algorithmic ones.
LaTeX, 22 pages, 1 figure, with correction to the published journal version
FOS: Computer and information sciences, Technology, Computer Science - Machine Learning, Minimal sufficient statistic (algorithmic), Mutual information (algorithmic), Kolmogorov complexity, Foundations of statistics, Algorithmic information theory, Statistical aspects of information-theoretic topics, 510, description format, Machine Learning (cs.LG), Engineering, algorithmic statistics, Foundations and philosophical topics in statistics, nonstochastic objects, minimal sufficient statistic, electrical & electronic, Nonstochastic objects, mutual information, measure of information, Description format (explicit, Electrical and electronic engineering, 004, Mathematics - Probability, Artificial intelligence and image processing, 62B05, 62B10, 68Q32, 68Q30, 60AXX, 68T04, Computer Science - Information Theory, FOS: Physical sciences, Mathematics - Statistics Theory, Statistics Theory (math.ST), information systems, Networking & telecommunications, FOS: Mathematics, implicit), algorithmic information theory, Sufficient statistic (algorithmic), Information Theory (cs.IT), Probability (math.PR), Two-part codes, Computer science, Information theory (general), Algorithmic information theory (Kolmogorov complexity, etc.), Physics - Data Analysis, Statistics and Probability, foundations of statistics, Science & technology, sufficient statistic, two-part codes, Communications technologies, Data Analysis, Statistics and Probability (physics.data-an)
FOS: Computer and information sciences, Technology, Computer Science - Machine Learning, Minimal sufficient statistic (algorithmic), Mutual information (algorithmic), Kolmogorov complexity, Foundations of statistics, Algorithmic information theory, Statistical aspects of information-theoretic topics, 510, description format, Machine Learning (cs.LG), Engineering, algorithmic statistics, Foundations and philosophical topics in statistics, nonstochastic objects, minimal sufficient statistic, electrical & electronic, Nonstochastic objects, mutual information, measure of information, Description format (explicit, Electrical and electronic engineering, 004, Mathematics - Probability, Artificial intelligence and image processing, 62B05, 62B10, 68Q32, 68Q30, 60AXX, 68T04, Computer Science - Information Theory, FOS: Physical sciences, Mathematics - Statistics Theory, Statistics Theory (math.ST), information systems, Networking & telecommunications, FOS: Mathematics, implicit), algorithmic information theory, Sufficient statistic (algorithmic), Information Theory (cs.IT), Probability (math.PR), Two-part codes, Computer science, Information theory (general), Algorithmic information theory (Kolmogorov complexity, etc.), Physics - Data Analysis, Statistics and Probability, foundations of statistics, Science & technology, sufficient statistic, two-part codes, Communications technologies, Data Analysis, Statistics and Probability (physics.data-an)
citations This is an alternative to the "Influence" indicator, which also reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | 79 | |
popularity This indicator reflects the "current" impact/attention (the "hype") of an article in the research community at large, based on the underlying citation network. | Top 10% | |
influence This indicator reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | Top 1% | |
impulse This indicator reflects the initial momentum of an article directly after its publication, based on the underlying citation network. | Top 10% |