
We present the Semantic Tolerance Law, the first statistical law for predicting dataset structural collapse in machine learning. Like other statistical laws (Zipf's law for language, Benford's law for numerical distributions), this law holds empirically with high probability rather than as an absolute deterministic rule. From the foundational axiom that learnable structure is captured by mutual information I(X;Y), we derive a composite metric Φ combining information-theoretic bounds and geometric properties. Validated across 5,885 datasets spanning 21 domains, the law predicts collapse with 98.4% success rate—higher reliability than most established statistical laws (Benford ~85%, Zipf ~90%). The framework combines: 40% rigorous theoretical foundation (established theorems) 30% theoretically motivated components (concentration principles) 25% empirical validation 5% engineering design choices Like thermodynamic laws (statistical mechanics) and information laws (Shannon's theorems), this is a probabilistic law with known boundary cases (<6% prevalence). It provides machine learning with its first systematic framework for pre-training quality assessment. Threshold Φ_c ≈ 0.007 shows domain-dependence (range 0.003-0.012) but the core principle appears universal across tested domains.
| selected citations These citations are derived from selected sources. This is an alternative to the "Influence" indicator, which also reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | 0 | |
| popularity This indicator reflects the "current" impact/attention (the "hype") of an article in the research community at large, based on the underlying citation network. | Average | |
| influence This indicator reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | Average | |
| impulse This indicator reflects the initial momentum of an article directly after its publication, based on the underlying citation network. | Average |
