필터 지우기
필터 지우기

bootstraping for neural network

조회 수: 4 (최근 30일)
Mohamad
Mohamad 2013년 11월 21일
댓글: Mohamad 2013년 11월 24일
hi every body
I am using bootstrap sampling with neural network to select the best network architecture. For instance, if I use 30 bootstraping for each network, I would have 30 error for each training and test data related to them. To select the best model, should I use the variance of these errors and select the model with the lowest variance? and should the training and test error be considered separately for bootstrapping? Any other suggestion is welcomed.
  댓글 수: 1
Greg Heath
Greg Heath 2013년 11월 22일
There are several different ways to implement training/testing bootstrapping with N data points AND there are several different ways to implement training/validation/testing bootstrapping.
How are you implementing it while avoiding overtraining an over-fit net?

댓글을 달려면 로그인하십시오.

채택된 답변

Greg Heath
Greg Heath 2013년 11월 24일
편집: Greg Heath 2013년 11월 24일
Training errors are biased because the same data is used for creation and evaluation.
Attempts to mitigate the bias involve reducing the number of training equations, Ntrneq, by the number of estimated unknown weights, Nw, to obtain the number of estimation degrees of freedom, Ndof ( see Wikipedia/R-squared), that is used to obtain the adjusted mean-squared-error estimate:
[Ntrn O ] = size(ttrn) % size of the training target matrix
Ntrneq = prod(size(ttrn)) = Ntrn*O % Number of training equations
etrn = ttrn-ytrn; % Training set errors
SSEtrn = ssetrn(etrn) % Sum-squared-errors
MSEtrn = SSEtrn/Ntrneq % Mean-squared-error estimate (biased)
Ndof = Ntrneq-Nw
MSEtrna = SSEtrn/Ndof % "a"djusted mean-squared-error estimate
Although I would only rely on the test set estimates and standard deviations, I would also keep track of the estimates and standard deviations of MSEtrn and MSEtrna.
Hope this helps.
Thank you for formally accepting my answer
Greg
  댓글 수: 1
Mohamad
Mohamad 2013년 11월 24일
Hello Greg One of the methods that I use for avoiding over-fitting is to change the number of hidden nodes in a loop and plot the number of hidden nodes and MSe for the traing and test data set on the same figure. When the MSe for the test set started to increase in comparison with training set I would stop the process. Something like early stopping which is automatically implemented in Matlab. best

댓글을 달려면 로그인하십시오.

추가 답변 (0개)

카테고리

Help CenterFile Exchange에서 Deep Learning Toolbox에 대해 자세히 알아보기

Community Treasure Hunt

Find the treasures in MATLAB Central and discover how the community can help you!

Start Hunting!

Translated by