how to deal with out of memory error when training a large data set or large neural network narx

조회 수: 2 (최근 30일)
how to deal with out of memory error when training a large data set or large neural network stacked autoencoder ......just initiated to analyze deep neural network with time series predictions , what i guess is by using batch trainings but can any one help me in detail

답변 (2개)

Greg Heath
Greg Heath 2015년 4월 15일
What are
1. input and target
2. size(input), size(target)
3. Significant lags of
a. target autocorrelation function
b. input/target cross correlation function

SAM
SAM 2015년 4월 15일
편집: SAM 2015년 4월 15일
i am xtremly sorry i wrote narx but i actually tried it with stacked autoencoder . hidden layers [100 150] even on HPC server i was getting same out of memory error.
inputs are 4 *2795 and output 1*2795 .

카테고리

Help CenterFile Exchange에서 Sequence and Numeric Feature Data Workflows에 대해 자세히 알아보기

Community Treasure Hunt

Find the treasures in MATLAB Central and discover how the community can help you!

Start Hunting!

Translated by