Serviceeinschränkungen vom 12.-22.02.2026 - weitere Infos auf der UB-Homepage

Treffer: A model is worth tens of thousands of examples

Title:
A model is worth tens of thousands of examples
Contributors:
Technion - Israel Institute of Technology Haifa, CEntre de REcherches en MAthématiques de la DEcision (CEREMADE), Université Paris Dauphine-PSL, Université Paris Sciences et Lettres (PSL)-Université Paris Sciences et Lettres (PSL)-Centre National de la Recherche Scientifique (CNRS), Université Paris Sciences et Lettres (PSL), ANR-19-P3IA-0001,PRAIRIE,PaRis Artificial Intelligence Research InstitutE(2019)
Source:
Lecture Notes in Computer Science ; 9th International Conference on Scale Space and Variational Methods in Computer Vision (SSVM 2023) ; https://hal.science/hal-04271603 ; 9th International Conference on Scale Space and Variational Methods in Computer Vision (SSVM 2023), May 2023, Santa Margherita di Pula, Italy. ⟨10.1007/978-3-031-31975-4_17⟩
Publisher Information:
CCSD
Publication Year:
2023
Subject Geographic:
Document Type:
Konferenz conference object
Language:
English
DOI:
10.1007/978-3-031-31975-4_17
Rights:
info:eu-repo/semantics/OpenAccess
Accession Number:
edsbas.A5EE8141
Database:
BASE

Weitere Informationen

International audience ; Traditional signal processing methods relying on mathematical data generation models have been cast aside in favour of deep neural networks, which require vast amounts of data. Since the theoretical sample complexity is nearly impossible to evaluate, these amounts of examples are usually estimated with crude rules of thumb. However, these rules only suggest when the networks should work, but do not relate to the traditional methods. In particular, an interesting question is: how much data is required for neural networks to be on par or outperform, if possible, the traditional model-based methods? In this work, we empirically investigate this question in two simple examples, where the data is generated according to precisely defined mathematical models, and where well-understood optimal or state-of-the-art mathematical data-agnostic solutions are known. A first problem is deconvolving onedimensional Gaussian signals and a second one is estimating a circle's radius and location in random grayscale images of disks. By training various networks, either naive custom designed or well-established ones, with various amounts of training data, we find that networks require tens of thousands of examples in comparison to the traditional methods, whether the networks are trained from scratch or even with transferlearning or finetuning.