L2 regularization in sparse stacked autoencoders not clear to me
조회 수: 1 (최근 30일)
이전 댓글 표시
Dear Matlab users,
on https://www.mathworks.com/help/nnet/ref/trainautoencoder.html , there is a bit of theory behind L2 regularization used in stacked autoencoders. However, the definition of L2 regularization is not clear to me. First, why is the sum running through all hidden layers l = 1..L but not through all neurons in each hidden layer? Second, I do not understand "k is the number of variables in the training data". Does it mean that k corresponds to the dimensionality of feature vectors?
Thanks for any clarification.
Regards, Lukas Vareka
댓글 수: 0
답변 (2개)
Sara Sharifzadeh
2020년 6월 6일
As this L2 regularisation aims to constraints all the weights, The first Sigma should sum all hidden layers+final layer and the two inner Sigma must count on the number of neurons in previous layer e.g. (l-1) and the current hidden layer (l) that in total will include all weights of the network to be constraint.
댓글 수: 0
BERGHOUT Tarek
2019년 4월 11일
i didnt undrestand your question , could make more clear , there is no k and L2 parameters in the link .
could you give us an example?.
any way i have made many types of autoencoders i upload some of them on my profile:
https://www.mathworks.com/matlabcentral/fileexchange/71115-denoising-autoencoders?s_tid=prof_contriblnk
댓글 수: 0
참고 항목
카테고리
Help Center 및 File Exchange에서 Sequence and Numeric Feature Data Workflows에 대해 자세히 알아보기
제품
Community Treasure Hunt
Find the treasures in MATLAB Central and discover how the community can help you!
Start Hunting!