The Price of Robustness: Stable Classifiers Need Overparameterization¶
Conference: ICLR 2026 arXiv: 2603.02806 Code: None Area: Learning Theory / Generalization Theory Keywords: overparameterization, robustness, stability, classifier, generalization bound, margin
TL;DR¶
This paper establishes stability-generalization bounds for discontinuous classifiers and proves a "law of robustness" for classification: any interpolating classifier with \(p \approx n\) parameters is necessarily unstable, and achieving high stability requires overparameterization on the order of \(p \approx nd\).
Background & Motivation¶
- The overparameterization paradox: Classical learning theory holds that more parameters lead to worse overfitting, yet modern overparameterized neural networks generalize well—a phenomenon known as double descent.
- Bubeck & Sellke 2021 robustness law: This work proved a smoothness–overparameterization trade-off for Lipschitz-continuous functions in the regression setting. However, the result relies on the Lipschitz assumption and does not apply to classifiers, whose discrete outputs are inherently discontinuous.
- Empirical findings: A large-scale study (Jiang et al. 2019) found that among 40+ complexity measures, margin (distance to the decision boundary) correlates most consistently with generalization, rather than norm-based measures.
- Core Problem: How can one build a stability-driven generalization theory for discontinuous classifiers?
Method¶
1. Class-Stability Definition¶
Definition 1 (Margin and Class-Stability): For a classifier \(f: \mathcal{X} \to \{-1, 1\}\):
Unsigned margin: $\(h_f(x) := |d_f(x)| = \inf\{\|x - z\|_2 : f(z) \neq f(x), z \in \mathcal{X}\}\)$
Class-stability (expected margin): $\(S(f) := \mathbb{E}[h_f]\)$
This measures the average robustness of the classifier's predictions to input perturbations under the data distribution.
2. Isoperimetric Assumption¶
The data distribution \(\mu\) is assumed to satisfy \(c\)-isoperimetry: for any bounded \(L\)-Lipschitz function \(f\) and \(t \geq 0\):
Gaussian distributions and the uniform distribution on the sphere satisfy this condition. Under the manifold hypothesis, \(d\) can be interpreted as the intrinsic manifold dimension.
3. Rademacher Bounds for Finite Function Classes¶
Theorem 4: Assume \(\min_{f \in \mathcal{F}} S(f) > S > 0\) and \(\log|\mathcal{F}| \geq n\):
Under regularity conditions, this can be improved to:
Key Insight: The factor \(1/S\) appears in front of \(\sqrt{\log|\mathcal{F}|}\)—stability reduces the effective complexity of the model class.
4. Law of Robustness for Classification¶
Corollary 6: Let \(p := \log|\mathcal{F}| \geq n\). Under appropriate conditions, with high probability:
Implications: - When \(p \approx n\), any interpolating classifier is necessarily unstable. - Simultaneously achieving low training error and high stability requires overparameterization of order \(p \approx nd\).
5. Extension to Infinite Function Classes¶
The paper introduces normalized co-stability: the output margin of the score function \(g_w\) in a classifier \(f = \text{sgn} \circ g_w\) is normalized accordingly. Combined with Lipschitz continuity of \(g_w\) with respect to both parameters and inputs, corresponding generalization bounds (Theorem 13) and a law of robustness (Corollary 15) are derived.
Key Experimental Results¶
MNIST Experiments¶
| Network Width | Test Accuracy | Class-Stability \(S(f)\) | Spectral Norm |
|---|---|---|---|
| Small | Low | Low | Irregular |
| Medium | Moderate | Moderate | Irregular |
| Large | High | High | Irregular |
CIFAR-10 Experiments¶
| Network Width | Test Accuracy | Normalized Co-Stability | Spectral Norm |
|---|---|---|---|
| Narrow | ~70% | Low | Inconsistent |
| Wide | ~85% | High | Inconsistent |
| Wider | ~90% | Higher | Inconsistent |
Key Findings¶
- Both stability and normalized co-stability monotonically increase with network width.
- Stability is positively correlated with test performance.
- Traditional norm-based measures (e.g., spectral norm) show no systematic association with generalization.
- Results validate the theoretical prediction: overparameterization → high stability → good generalization.
Highlights & Insights¶
- Extension to discontinuous functions: This is the first work to extend the law of robustness from Lipschitz regression to discontinuous classifiers.
- Direct analysis of the 0-1 loss: No Lipschitz loss assumption is required.
- Explaining overparameterization: Overparameterization is not the source of overfitting but rather a necessary condition for achieving robustness.
- Broad applicability: The framework covers inherently discontinuous models such as quantized neural networks and spiking neural networks.
- Special relevance to Transformers: Self-attention is generally not Lipschitz-continuous, making this framework more applicable than Lipschitz-based ones.
Limitations & Future Work¶
- The isoperimetric assumption may not hold for certain data distributions.
- Extending from finite to infinite function classes requires additional Lipschitz assumptions on the parameters.
- Generalization bounds may still be vacuous in practice, consistent with the critique of Nagarajan & Kolter 2021.
- The practical value of the theoretical dimension \(d\) is difficult to estimate precisely (extrinsic vs. intrinsic dimension).
- No direct connection to optimization dynamics (e.g., implicit bias) is established.
Related Work & Insights¶
- Law of robustness: Bubeck & Sellke 2021 (Lipschitz regression)
- Margin-based generalization bounds: Bartlett et al. 2017 (spectrally-normalized margin)
- Algorithmic stability: Bousquet & Elisseeff 2002
- Double descent: Belkin et al. 2019
Rating¶
- Novelty: ⭐⭐⭐⭐ — A natural and important extension of the law of robustness to classification.
- Technical Depth: ⭐⭐⭐⭐⭐ — Proof techniques are elegant, leveraging Lipschitz surrogates and signed distance representations.
- Experimental Thoroughness: ⭐⭐⭐ — Validated on MNIST and CIFAR-10; primarily qualitative.
- Value: ⭐⭐⭐⭐ — Provides theoretical grounding for understanding generalization in overparameterized models.