Deep Neural LSTM Network Issues
조회 수: 6 (최근 30일)
이전 댓글 표시
I am training a Deep Neural Net with a regression layer in the end. I have 20 inputs and a sequence of output with 10 steps. I tried using both LSTM and BiLISTM layers with something like 100 to 200 hidden units. I have also included two fully connected layers with some 100 hidden nodes or so. But no matter what I do and change learning rate etc.. the progress comes to a plateau very quickly like in a couple of epochs and remains so for the rest of the training. Tried changing learning rate, min batch size, Number of hidden units, adding relu and not adding relu etc etc.. you name it. But I can not improve the accracy to more than about 0.9 (validation MSE). Is there anything else I can try to improve?
댓글 수: 0
답변 (1개)
Udit06
2023년 9월 6일
Hi Pappu,
I understand that you aim to improve your validation set accuracy of the LSTM based deep neural network. You can try out the following approaches to improve your accuracy metric:
1) You can normalize your input features in the range [0, 1] to stabilize the training process and improve convergence.
2) You can use dropouts in your network to prevent overfitting and get a generalized model. You can refer to below MathWorks documentation to understand more about dropouts.
3) Instead of using a fixed learning rate, you can use Adam optimizer which utilizes adaptive learning rate based on the gradient magnitude. You can refer to below MathWorks documentation to understand more about Adam optimizer.
I hope this helps.
참고 항목
카테고리
Help Center 및 File Exchange에서 Deep Learning Toolbox에 대해 자세히 알아보기
Community Treasure Hunt
Find the treasures in MATLAB Central and discover how the community can help you!
Start Hunting!