Statistical theory for image classification using deep convolutional neural networks with cross entropy loss

Michael Kohler, Sophie Langer*

*Corresponding author for this work

Research output: Working paperProfessional

2 Downloads (Pure)

Abstract

Convolutional neural networks learned by minimizing the cross-entropy loss are nowadays the standard for image classification. Till now, the statistical theory behind those networks is lacking. We analyze the rate of convergence of the misclassification risk of the estimates towards the optimal misclassification risk. Under suitable assumptions on the smoothness and structure of the aposteriori probability it is shown that these estimates achieve a rate of convergence which is independent of the dimension of the image. The study shed light on the good performance of CNNs learned by cross-entropy loss and partly explains their success in practical applications.
Original languageEnglish
PublisherarXiv.org
Number of pages27
Publication statusSubmitted - 27 Nov 2020
Externally publishedYes

Fingerprint

Dive into the research topics of 'Statistical theory for image classification using deep convolutional neural networks with cross entropy loss'. Together they form a unique fingerprint.

Cite this