regarding Neural network toolbox

2 visualizaciones (últimos 30 días)
Kshitij Tripathi
Kshitij Tripathi el 15 de Dic. de 2015
Comentada: Walter Roberson el 16 de Dic. de 2015
Dear Sir I want to know that,whether we will implement different transfer function for different neurons on the same layer in matlab Neural Network toolbox

Respuesta aceptada

Greg Heath
Greg Heath el 15 de Dic. de 2015
If you do this you will have to write your own code. An equivalent, but easier alternative might be to use skip layer connections with each layer containing a different type of transfer function.
However, why in the world would you want to do this? A standard MATLAB function like FITNET or PATTERNNET with a single hidden layer with tansig transfer functions is a universal approximator.
Hope this helps.
Greg
  2 comentarios
Kshitij Tripathi
Kshitij Tripathi el 16 de Dic. de 2015
ok Thanks Greg Sir,but Just help ,give me the hint,how to write my own code in Matlab to do this.
Walter Roberson
Walter Roberson el 16 de Dic. de 2015
You can see the features you should support by looking at http://www.mathworks.com/help/nnet/ref/purelin.html
Your replacement transfer function should have the same functionality.

Iniciar sesión para comentar.

Más respuestas (1)

Walter Roberson
Walter Roberson el 15 de Dic. de 2015
You can if you want to but it would be unusual.
You can supply your own transfer function. That function can do its calculation any way you want, including being position dependent.
I would remind you that if this is not the input layer then you probably do not really know which neurons are participating in which calculation. Neural Network weights are calculated as "whatever works to give the desired result" not by partitioning the network into portions that calculate specific features.

Categorías

Más información sobre Sequence and Numeric Feature Data Workflows en Help Center y File Exchange.

Community Treasure Hunt

Find the treasures in MATLAB Central and discover how the community can help you!

Start Hunting!

Translated by