On whether the idea that âcranial nerves are step functions, so neural nets use sigmoidsâ is correct.
I think the fact that the brain works more like a step function means that cranial nerves are binary outputs that either pulse or do not pulse. This is a point that McCulloch and Pitts emphasized when they proposed formal neuron in 1943. formal neuron - Wikipedia
But 70 years have passed since then, and Dr. Kenji Doyaâs lecture notes, written in 1999, organizes it as follows
-
There are two possible ways: (1) firing rate expression (rate coding) and (2) timing expression (temporal coding).
-
In primary sensory neurons and primary motor neurons, the firing frequency uniquely corresponds to a physical quantity such as light intensity or muscle tension. The inter-spike interval (ISI) is also nearly constant for a constant stimulus.
-
The idea of temporal coding (timing representation), which assumes that there is information in the timing of firing⌠Experimental facts, for example, show synchronous firing in the visual cortex.
- firing rate expression
-
Spike firing rates that are considered to be encoding information.
- element representation
- combinatorial expression
- population representation
- Spatial Pattern Representation
-
There is also experimental and model support for the theory that the various representations described above take a âsparse representationâ in which the rate of actually active cells in the population is kept low.
-
- timing expression
-
In the firing rate representation, we focus on the average value of the frequency of neural spikes over an appropriate range of time and space, and do not concern ourselves with the timing at which each spike is emitted on the order of milliseconds. However, there is a position that it is precisely such subtle timing that reflects a wealth of information.
- time-delayed representation
- synchronous idle
- Ignition Pattern Expression
- synaptic plasticity
-
- firing rate expression
So the idea that âcranial nerves are step functionsâ has not been updated in almost 20 years of knowledge.
Also, âuse sigmoidâ was published in 1986, but as of 2018 it is considered:.
In 2011, Xavier Glorot et al. showed that using max(0, x) as the activation function of the hidden layer is an improvement over tanh or soft plus⌠In the world of neural networks it is called ReLU (English: Rectified Linear Unit, Rectifier, normalized linear function); a paper written by Yann LeCun and Jeffrey Hinton et al. in the journal Nature states that this is the best as of May 2015. Activation function - Wikipedia https://www.nature.com/articles/nature14539
Note that âas of May 2015, this is the best we can doâ is an interpretation, and the fact is that the great success in convolutional neural networks in 2012 (ImageNet Classification with Deep Convolutional Neural Networks) was attributed to ReLU.
This success came from the efficient use of GPUs, ReLUs, a new regularization technique called dropout, and techniques to generate more training examples by deforming the existing ones.
In other words, âcranial nerves are step functions, so neural nets use sigmoidsâ is incorrect in both the first half and the second half as of 2018.
This page is auto-translated from /nishio/čłăŻăšăăăé˘ć°ă using DeepL. If you looks something interesting but the auto-translated English is not good enough to understand it, feel free to let me know at @nishio_en. Iâm very happy to spread my thought to non-Japanese readers.