
Structured pruning of neural networks operating on k-regular layer architectures requires enumerating all valid weight configurations that preserve full inter-layer connectivity. This problem is precisely the permanent of the N \times N bipartite adjacency matrix a #P-complete computation that has been considered intractable on commodity hardware, forcing practitioners to rely on stochastic heuristics with no guaranteed coverage of the solution space. We present a shared-nothing parallel implementation of Glynn's formula traversed by binary reflected Gray code, extended with arithmetic to support dense k-regular graphs (K \ge 7). The engine is validated against the Leibniz brute-force permanent across 20 test cases identity matrices, complete bipartite $K{N,N}$, random dense matrices, and k-regular bipartite graphs for K \in \{2,3,4\} and N \in \{4,...,10\}.
Glynn Formula, Matrix Permanent, k-Regular Bipartite Graphs, Structured Pruning, Neural Network Pruning, Edge AI
Glynn Formula, Matrix Permanent, k-Regular Bipartite Graphs, Structured Pruning, Neural Network Pruning, Edge AI
| selected citations These citations are derived from selected sources. This is an alternative to the "Influence" indicator, which also reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | 0 | |
| popularity This indicator reflects the "current" impact/attention (the "hype") of an article in the research community at large, based on the underlying citation network. | Average | |
| influence This indicator reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | Average | |
| impulse This indicator reflects the initial momentum of an article directly after its publication, based on the underlying citation network. | Average |
