Sample complexity

The sample complexity of a machine learning algorithm represents the number of training-samples that it needs in order to successfully learn a target function. More precisely, the sample complexity is the number of training-samples that we need to supply to the algorithm, so that the function returned by the algorithm is within an arbitrarily small error of the best possible function, with probability arbitrarily close to 1. There are two variants of sample complexity:

Comment
enThe sample complexity of a machine learning algorithm represents the number of training-samples that it needs in order to successfully learn a target function. More precisely, the sample complexity is the number of training-samples that we need to supply to the algorithm, so that the function returned by the algorithm is within an arbitrarily small error of the best possible function, with probability arbitrarily close to 1. There are two variants of sample complexity:
Has abstract
enThe sample complexity of a machine learning algorithm represents the number of training-samples that it needs in order to successfully learn a target function. More precisely, the sample complexity is the number of training-samples that we need to supply to the algorithm, so that the function returned by the algorithm is within an arbitrarily small error of the best possible function, with probability arbitrarily close to 1. There are two variants of sample complexity: * The weak variant fixes a particular input-output distribution; * The strong variant takes the worst-case sample complexity over all input-output distributions. The No free lunch theorem, discussed below, proves that, in general, the strong sample complexity is infinite, i.e. that there is no algorithm that can learn the globally-optimal target function using a finite number of training samples. However, if we are only interested in a particular class of target functions (e.g, only linear functions) then the sample complexity is finite, and it depends linearly on the VC dimension on the class of target functions.
Is primary topic of
Sample complexity
Label
enSample complexity
Link from a Wikipage to another Wikipage
Active learning (machine learning)
Category:Machine learning
Dictionary learning
Empirical risk minimization
Glivenko-Cantelli class
Machine learning
Metric learning
Model-free (reinforcement learning)
Monte Carlo tree search
No free lunch in search and optimization
No free lunch theorem
Online machine learning
Probably approximately correct learning
Rademacher complexity
Random variable
Regularization (mathematics)
Reinforcement learning
Semi-supervised learning
Tikhonov regularization
Vapnik–Chervonenkis theory
VC dimension
SameAs
m.0114dpwp
mVCu
Q18354077
پیچیدگی نمونه
Subject
Category:Machine learning
WasDerivedFrom
Sample complexity?oldid=1068917677&ns=0
WikiPageLength
14205
Wikipage page ID
43269516
Wikipage revision ID
1068917677
WikiPageUsesTemplate
Template:Machine learning bar
Template:Reflist