,b)=−mtest1i=1∑mtest[ytest(i)log(f(xtest(i)))+(1−ytest(i))log(1−f(xtest(i))]
2 cross-validation set
- split the training set into training set, cross-validation set and test set
- the cross-validation set is used to automatically choose the better model, and the test set is used to evaluate the model that chosed
3 bias and variance
- high bias:
J
t
r
a
i
n
J_{train}
Jtrain and
J
c
v
J_{cv}
Jcv is both high
- high variance:
J
t
r
a
i
n
J_{train}
Jtrain is low, but
J
c
v
J_{cv}
Jcv is high
- if high bias: get more training set is helpless
- if high variance: get more training set is helpful
4 regularization
- if
λ
lambda
λ is too small, it will lead to overfitting(high variance)
- if
λ
lambda
λ is too large, it will lead to underfitting(high bias)
5 method
- fix high variance:
- get more training set
- try smaller set of features
- reduce some of the higher-order terms
- increase
λ
lambda
λ
- fix high bias:
- get more addtional features
- add polynomial features
- decrease
λ
lambda
λ
6 neural network and bias variance
- a bigger network means a more complex model, so it will solve the high bias
- more data is helpful to solve high variance
- it turns out that a bigger(may be overfitting) and well regularized neural network is better than a small neural network
原文地址:https://blog.csdn.net/m0_65591847/article/details/135641692
本文来自互联网用户投稿,该文观点仅代表作者本人,不代表本站立场。本站仅提供信息存储空间服务,不拥有所有权,不承担相关法律责任。
如若转载,请注明出处:http://www.7code.cn/show_57830.html
如若内容造成侵权/违法违规/事实不符,请联系代码007邮箱:suwngjj01@126.com进行投诉反馈,一经查实,立即删除!