https://blog.csdn.net/ChenVast/article/details/81385018
symbol | Meaning |
---|---|
Test samples | |
data set | |
Tag in the data set | |
The true mark | |
Training set to learn the model | |
Training set by the model of learned to predict output | |
Model of the desired predicted output |
variance
On a training set D model to predict the output of the test samples f x is f (x; D), then the learning algorithm for testing samples x f desired predicted as follows:
The above prediction is desirable for a different set of data D, f predictive value of whichever desired x (average prediction).
Use the same number of samples of different training sets generated variance:
deviation
Desired square error prediction and the real mark is referred to as a deviation (BIAS), for convenience, we take the direct deviation:
Generalization error
To return to the task, for example, squared prediction error learning algorithm is expected to:
Generalization error of the desired decomposition algorithm:
Make noise to zero , so zero red zone.
The last remaining , the result is a generalization error deviation = variance + noise +
Deviation, variance, noise
- Deviation: a measure of the expectations of the prediction model and the degree of deviation of actual results, characterizes the model itself fitting ability .
- Variance: measure the changes in learning performance variations of the same size as a result of the training set, that characterizes the disturbance caused by the impact of data .
- Noise: The expression of the desired lower bound on the generalization error of the current model can achieve any task, portray the difficulty of learning the problem itself .
Illustrates deviation and variance
Low variance | High variance | |
Low deviation | + Data points are concentrated on the data points fall prediction point | + Data not concentrated portion of the data points fall on point prediction (prediction accuracy is not high) |
High deviation | + Data points are concentrated there is a distance data point and the predicted point (inaccuracies) | Data points are not concentrated + data points do not fall substantially (inaccuracies) the predicted point |
Variance and deviation and fitting
Goodness of fit | variance | deviation | the reason | Solution |
Underfitting | Exorbitant | Inadequate training, bias leading generalization error | Integrated learning; deepen plus iterations; additional features; reduce regularization; | |
Overfitting | Exorbitant | Excessive training, leading generalization error variance | Reduce the complexity of the model; plus regular penalty term; plus training set; Save feature; improved regularization |
reference:
http://www.cnblogs.com/makefile/p/bias-var.html#fn2
Disclaimer: This article is a blogger original article, reproduced, please indicate the source at the beginning of the article (author + description link). https://blog.csdn.net/ChenVast/article/details/81385018