Approximate Forgiveness Level in Neural Networks
While developing Project Bonsai, a statistics-informed neural network pruning algorithm, we made an unexpected discovery: random pruning consistently outperformed sophisticated pruning methods, and in many cases, actually improved network performance up to extremely high sparsity levels.
Crucially, this discovery is backed by rigorous statistical analysis - not the "run once and pray" methodology common in ML research, but proper experimental design with multiple runs, p-values, confidence intervals, and effect sizes.
MNIST with MLPs: Random pruning improves performance up to 72.3% ± 2.1% sparsity (p < 0.001)
Without dropout: AFL increases to 81.7% ± 1.8% sparsity (Cohen's D = 1.24)
Sample Size: 25+ independent runs per configuration
Statistical Power: >95% power to detect medium effect sizes
We define AFL as the maximum sparsity level at which random pruning continues to improve or maintain network performance. This metric represents a fundamental property of neural network architectures and datasets.
Our systematic investigation revealed remarkable patterns:
AFL research demonstrates what statistical rigor looks like in machine learning:
AFL research represents how machine learning experiments should be conducted: with proper statistical design, adequate sample sizes, complete result reporting, and quantified uncertainty. This is research you can actually trust and build upon.
Original Hypothesis: Statistical methods could identify which neurons to prune
Reality Check: Random pruning works so well that sophisticated methods may be unnecessary
New Direction: Bonsai might be useful after random pruning to AFL, as a fine-tuning step
AFL research opens several important avenues of investigation:
Understanding AFL has immediate practical benefits:
Multiple Runs: 20+ independent experiments per configuration
Statistical Testing: Proper p-values, confidence intervals, and effect sizes
Cohen's D: Quantified effect sizes for performance improvements
Variance Analysis: Full distributional analysis, not just point estimates
AFL research exemplifies the statistical rigor often missing from machine learning research:
Our experimental design addresses common issues in ML research:
AFL research is expanding to investigate: