Early Stopping for Deep Networks

조회 수: 19 (최근 30일)
Roberto
Roberto 2019년 1월 15일
편집: Greg Heath 2019년 1월 19일
Hi everyone,
just a quick question.
How can I stop the training of a deep network (LSTM for instance) in order to have weights and biases set accordingly with the minimum of the validation loss?
In other words what's the reason of having a validation set if the final network is NOT the one that minimize the validation loss because it's overtrained in any case?
Validation Patience parameter is not useful in this sense because it stops the training when it's too late and setting it too small could result in being stuck in a local minimum.
The only way I found is repeating the training with max epochs set where the minimum of validation loss in the first training is reached but it's a crazy solution...
Any idea?
Thanks

답변 (1개)

Greg Heath
Greg Heath 2019년 1월 16일
편집: Greg Heath 2019년 1월 16일
It is not clear to me that, based on a random 15% of the data, this is a better choice. It would be interesting to make a formal comparison based on multiple runs with different random number seeds using multiple data sets.
I believe that a more important point is to try to minimize the number of hidden nodes subject to an upper bound on the training set error rate.
Hope this helps
Thank you for formally accepting tmy answer
Greg
  댓글 수: 2
Roberto
Roberto 2019년 1월 17일
Not sure to understand. You mean that L2 regularization could outperform early stopping?
In my opinion it's too related to the dataset to make a formal comparison, but in order to compare the two methods we still need a way to early stop the training in deep learning toolbox...
Greg Heath
Greg Heath 2019년 1월 19일
편집: Greg Heath 2019년 1월 19일
No. That is not what I meant.
HOWEVER
Any decent method will outperform others depending on the data set.
My shallow net double loop procedure (MANY examples in NEWSGROUP and ANSWERS) has been successfull for decades.
  1. Single hidden layer
  2. Outer loop over number of hidden nodes H = 0:dH:Hmax
  3. Inner loop over random intial weights
I have not tried it on deep nets but am interested if anyone else has.
Greg.

댓글을 달려면 로그인하십시오.

카테고리

Help CenterFile Exchange에서 Image Data Workflows에 대해 자세히 알아보기

제품


릴리스

R2018b

Community Treasure Hunt

Find the treasures in MATLAB Central and discover how the community can help you!

Start Hunting!

Translated by