필터 지우기
필터 지우기

Overfitting and What is it ?

조회 수: 11 (최근 30일)
HAAAAH
HAAAAH 2019년 9월 12일
댓글: Image Analyst 2019년 9월 13일
I have a data set and will divide it into " training set" and " validation set" . I choose that divide 70% training and 30% validation. I have calculated the linear regression from " training set" , and then calculate data of validation set by using linear regression from "training set". My result is pretty small than the correct one. Explaination can be the concept of our linear regression which do not apply for the new data (in this case is validation data) . Or is there another reason?
"And a model’s ability to predict future outcomes in light of what you discovered here". What does the question mean ?
The formula for fitting error where y1 = y * e^(t) is the estimated value of y
E(error) = (1/n )* sqrt ( ( sum(y1 -y)^2) (i= 0 to n )
  댓글 수: 1
Torsten
Torsten 2019년 9월 12일
Explaination can be the concept of our linear regression which do not apply for the new data (in this case is validation data) . Or is there another reason?
How can we tell without knowing your data and the background of data gathering ?
"And a model’s ability to predict future outcomes in light of what you discovered here". What does the question mean ?
What question ?

댓글을 달려면 로그인하십시오.

답변 (1개)

Image Analyst
Image Analyst 2019년 9월 12일
You can't overfit a linear regression. Overfitting is basically where you have your model go through, or mostly through, your data points. For example if you had 4 data points and fit a cubic, that would be overfitting. If you have N data points and fit a polynomial with, oh I don't know, say, N/2 or something, then you might have over fitting. But that won't happen with a line since it won't go through all your data points unless you had only 2 points.
It could be that your data does not fit the model that was determined using the 70% of other points, but I doubt it since you probably chose them randomly. I'd think the 30% error was much worse than the other 70% just due to randomness. In general (usually) testing/validation will be worse error than the training points, but it could be better.
  댓글 수: 2
Bruno Luong
Bruno Luong 2019년 9월 12일
"You can't overfit a linear regression."
Of course you can
Linear regression means the model is linear wrt the parameters, it does not mean fitting necessary a line/plane.
You can fit
yj = sum on i ( A_i * f_i(x_j)) to some data and it is easy to overfit.
Spline fitting is in this formulation and poeple can overfit. Waba and al are notorious for studying regularization method to make a "right" fit.
Image Analyst
Image Analyst 2019년 9월 13일
You're right. What I meant was fitting a line through data, which is not the same as fitting a higher order polynomial via a sum of weighted x^n terms (i.e., linear regression).

댓글을 달려면 로그인하십시오.

Community Treasure Hunt

Find the treasures in MATLAB Central and discover how the community can help you!

Start Hunting!

Translated by