Deep Neural LSTM Network Issues

조회 수: 6 (최근 30일)
Pappu Murthy
Pappu Murthy 2022년 8월 31일
댓글: Pappu Murthy 2023년 9월 11일
I am training a Deep Neural Net with a regression layer in the end. I have 20 inputs and a sequence of output with 10 steps. I tried using both LSTM and BiLISTM layers with something like 100 to 200 hidden units. I have also included two fully connected layers with some 100 hidden nodes or so. But no matter what I do and change learning rate etc.. the progress comes to a plateau very quickly like in a couple of epochs and remains so for the rest of the training. Tried changing learning rate, min batch size, Number of hidden units, adding relu and not adding relu etc etc.. you name it. But I can not improve the accracy to more than about 0.9 (validation MSE). Is there anything else I can try to improve?

답변 (1개)

Udit06
Udit06 2023년 9월 6일
Hi Pappu,
I understand that you aim to improve your validation set accuracy of the LSTM based deep neural network. You can try out the following approaches to improve your accuracy metric:
1) You can normalize your input features in the range [0, 1] to stabilize the training process and improve convergence.
2) You can use dropouts in your network to prevent overfitting and get a generalized model. You can refer to below MathWorks documentation to understand more about dropouts.
3) Instead of using a fixed learning rate, you can use Adam optimizer which utilizes adaptive learning rate based on the gradient magnitude. You can refer to below MathWorks documentation to understand more about Adam optimizer.
I hope this helps.
  댓글 수: 1
Pappu Murthy
Pappu Murthy 2023년 9월 11일
Tried those ideas that you have suggested already but not much help at all.

댓글을 달려면 로그인하십시오.

카테고리

Help CenterFile Exchange에서 Deep Learning Toolbox에 대해 자세히 알아보기

태그

제품


릴리스

R2021b

Community Treasure Hunt

Find the treasures in MATLAB Central and discover how the community can help you!

Start Hunting!

Translated by