How does NN prevent completion of training by Mu value?

조회 수: 31 (최근 30일)
Kyeonghwan Kim
Kyeonghwan Kim 2017년 8월 9일
댓글: Noor Thamir 2022년 5월 17일
When you do NN, the situation is that the training is completed by Mu value. How can I prevent it?
  댓글 수: 1
Noor Thamir
Noor Thamir 2022년 5월 17일
Hello Doctors:
please, I need clarification regarding the parameters in the attached image: performance, gradient, mu, and validation check.

댓글을 달려면 로그인하십시오.

채택된 답변

Amy
Amy 2017년 8월 11일
Hi Keonghwan,
The MU value is used to control the weights of the neurons updating process (back propagation) during training. If your training stops with the message "Maximum MU reached", it is a sign that additional training will not improve learning.
If you are reaching a maximum MU value too quickly, it might be because the data set you are using for training is too complex for the number of neurons in your network to successfully model. You can increase the number of neurons by increasing the value of 'hiddenLayerSize' in the script you used for training.
It is possible to set your own initial and maximum MU values for the 'trainlm' training function. See http://www.mathworks.com/help/nnet/ref/trainlm.html for information on the initial parameters and how they are updated.
You might also find the discussions of regularization and early stopping on this page helpful: http://www.mathworks.com/help/nnet/ug/improve-neural-network-generalization-and-avoid-overfitting.html
  댓글 수: 1
Chris P
Chris P 2020년 8월 9일
편집: Chris P 2020년 8월 9일
Finally, a helpful answer to a neural network-related question. A lot better than most of the answers. Not going to name any names but you all know who I'm talking about.

댓글을 달려면 로그인하십시오.

추가 답변 (0개)

카테고리

Help CenterFile Exchange에서 시계열, 시퀀스 및 텍스트에서의 딥러닝에 대해 자세히 알아보기

Community Treasure Hunt

Find the treasures in MATLAB Central and discover how the community can help you!

Start Hunting!