At what point does variation in neural network training become unreasonable?
조회 수: 4 (최근 30일)
이전 댓글 표시
I have a network trained in patternnet. By observing the confusion matrix after each training session, I am witnessing a variance of around 1.5% each time I retrain (From a 2% error rate to a 0% error rate). I feel like there is a big difference in these two percentages. Is there a way to reduce this variation or is it expected with this network?
댓글 수: 0
답변 (1개)
Greg Heath
2017년 6월 17일
Since initial weights are random, the variation in results can be considerable. That is why I routinely design 10 to 30 nets for each value of hidden nodes I consider. if you search in BOTH the NEWSGROUP and ANSWERS using
greg patternnet
you will find many examples. You can reduce the number by trying
greg patternnet tutorial
Hope this helps.
Thank you for formally accepting my answer
Greg
댓글 수: 0
참고 항목
카테고리
Help Center 및 File Exchange에서 Deep Learning Toolbox에 대해 자세히 알아보기
Community Treasure Hunt
Find the treasures in MATLAB Central and discover how the community can help you!
Start Hunting!