Why sets Matlab automatically the activation functions for a neural network like this?

조회 수: 6 (최근 30일)
I am asking myself why chooses Matlab always automatically for the hidden layer tan-Sigmoid and for the output layer pureline as an activation function?
If it refers to a study, which discovers, that those activation functions are more efficient than the other, please let me know.

답변 (3개)

Greg Heath
Greg Heath 2019년 6월 29일
That is a standard configuation for a neural net. It's operation is explained in every elementary text.
Thank you for formally accepting my answer
Greg
  댓글 수: 3
Greg Heath
Greg Heath 2019년 7월 30일
편집: Greg Heath 2019년 7월 30일
Sorry, I lost all of my several hundred books via a moving van error..
See your library.
Greg
Greg Heath
Greg Heath 2019년 7월 30일
MAYBE I SHOULD HAVE SAID any ELEMENTARY TEXT ON NNs.
GREG

댓글을 달려면 로그인하십시오.


Greg Heath
Greg Heath 2019년 7월 30일
The simplest useful approximation is is a series of blocks with different heights and widths.
The simplest useful DIFFERENTIABLE approximation is is a series of ROUNDED blocks with different heights and lengths.
Combining sigmoids fits the bill!
GREG
  댓글 수: 1
Osama Tabbakh
Osama Tabbakh 2019년 7월 31일
Thank you for your answer, I do not need it just to understand it, but I need some references because I am writing right now my thesis.

댓글을 달려면 로그인하십시오.


Sai Bhargav Avula
Sai Bhargav Avula 2019년 8월 16일
As mentioned by others thats the default setup in MATLAB.
Coming to comparision between different activation functions.
It is generally recommended to use ReLU as the activation function. If your model suffers form dead neurons during training we should use leaky ReLu or Maxout function.
The Sigmoid and Tanh are generally not preferred as they suffer with vanishing Gradient Problem which causes a lots of problems to train,degrades the accuracy and performance of a deep Neural Network Model.
  댓글 수: 1
Osama Tabbakh
Osama Tabbakh 2019년 8월 18일
I can't use ReLU because I am not using a deep neural network. I need just a reason, why are set by Matlab like that?

댓글을 달려면 로그인하십시오.

카테고리

Help CenterFile Exchange에서 Sequence and Numeric Feature Data Workflows에 대해 자세히 알아보기

제품


릴리스

R2019a

Community Treasure Hunt

Find the treasures in MATLAB Central and discover how the community can help you!

Start Hunting!

Translated by