Question of Hyperparameter tuning of shallow neural network

조회 수: 4 (최근 30일)
Qiang Wang
Qiang Wang 2021년 2월 1일
댓글: Qiang Wang 2021년 2월 5일
Hi all,
I trained a shallow neuron network and tuning parameters following this post from @Greg Heath.
I only tuning the number of nuerons in one hidden layer and accept all default parameters.
here is comments from referee:
The machine learning models have numerous hyper-parameters, but the authors only optimise one, and did not even report the values of the others. It is not clear why they chosen only one, or why they chose the ones they did, but this is unacceptable. All independent hyper-parameters need to be optimised simultaneously.
how to answer this comment?
Thank you.

채택된 답변

Shashank Gupta
Shashank Gupta 2021년 2월 5일
Hi Qiang,
The points raised by the reviewers are valid, you can't just optimize one variable and leave the rest as it is. even if you are intent to optimize some of the variable, you must explain why it is done and must provide the values used for rest of the parameters. All the hyper parameters used in ML model have some significance and not conveying why it is not being optimized or even not been give the value is certainly no one will accept. My suggestion would be, if you not intent to optimize all parameter then you must convey the reason and give them the default set of values used for those parameters. You need to show clearity on the purpose.
I hope this helps.
Cheers.
  댓글 수: 3
Shashank Gupta
Shashank Gupta 2021년 2월 5일
Yes, its resonable to optimize, actually it is very important to optimize all of them. There are many ways to do it, MATLAB has a bayesian approach too. Check this out. It might help you.

댓글을 달려면 로그인하십시오.

추가 답변 (0개)

카테고리

Help CenterFile Exchange에서 Deep Learning Toolbox에 대해 자세히 알아보기

Community Treasure Hunt

Find the treasures in MATLAB Central and discover how the community can help you!

Start Hunting!

Translated by