Abstract
We establish optimal convergence rates up to a factor for a class of deep neural networks in a classification setting under a restraint sometimes referred to as the Tsybakov noise condition. We construct classifiers based on empirical risk minimization in a general setting where the boundary of the Bayes rule can be approximated well by neural networks. Corresponding rates of convergence are proven with respect to the misclassification error using an additional condition that acts as a requirement for the “correct noise exponent”. It is then shown that these rates are optimal in the minimax sense. For other estimation procedures, similar convergence rates have been established. Our first main contribution is to prove that the rates are optimal under the additional condition. Secondly, our main theorem establishes almost optimal rates in a generalized setting. We use this to show optimal rates which circumvent the curse of dimensionality.
Acknowledgments
First and foremost, I would like to thank Enno Mammen for supporting me with some helpful comments and inspiring insights during the creation of this paper. Additionally, many thanks goes to Munir Hiabu for assisting with comments during the final stages of the working process.
Citation
Joseph T. Meyer. "Optimal convergence rates of deep neural networks in a classification setting." Electron. J. Statist. 17 (2) 3613 - 3659, 2023. https://doi.org/10.1214/23-EJS2187
Information