Brains work using gate like functions

Effectively we can prove that neurons are gate-like processors whose functions operate differently than digital gates but one can find the logical equivalent of neural processing to digital processes, inclusive of even learning! 

There are various hypotheses of neural spike codes where the complexity of the premise describes a neuron’s ion channels and how various spiking modes are produced. I am going to describe a much simpler and generalized coding scheme for neurons based on anatomical observations and in particular the auditory wiring of mammalian brains. Neurons exhibit firing ranges that directly relate to the degree of inputs they receive relative to some desensitizing of the neuron’s dendritic membrane. If we examine the cochlea and how the hairs stimulate neurons we can effectively see how nature is transforming mechanical energy into electrical signaling. A vital data point of the brain’s auditory system is the positioning of the neurons that are stimulated by the cochlea hairs, where the position of the neuron represents the auditory wavelength that a cochlea hair vibrates at. The second piece of information we can glean from the anatomy of the auditory circuits of the brain is the vibrating hair can control the degree or rate of neurons firing. So mammalian brains codify sound by:

1. The positioning of neurons along the cochlea for a particular wavelength.

2. The firing rate of a neuron indicates the energy or loudness of the auditory signal being heard.

The number of neurons associated to wavelengths along the cochlea improves sensitivity to those wavelengths, the more neurons applied to auditory ranges the more sensitive an animal’s hearing becomes. But the point that is being conveyed is neurons can encode by superimposing two pieces of information, symbolic representation by physical positioning, and two, the degree of signal strength of the symbolic representation. In other words, neurons produce fuzzy bits! Where an output of a neuron is more than true or false it is to a degree of truth represented by its firing rate. The truth table for neurons is dependent on its resistance to inputs or a weighted barrier, as shown in the truth table below:

The combination of the two input neurons A and B has to exceed the weighted barrier of neuron C’s dendritic membrane. With that said the question becomes can one build logic circuits similar to Boolean methods? Below is a diagram of a full adder using the neuron spike coding scheme: You’ll note that there are five neurons in the circuit, the truth table for the circuit is below:

Where the inhibitory neurons usually have 5 times the influence that Excitatory neurons have and where the D output is greater than 0 it is the equivalent to binary 1 and where E outputs are greater than zero are also the equivalent to binary 1. So effectively the representation of binary inputs are through neurons A and B but their outputs are effective only under certain firing rates! Where the minimum output from Neurons A and B have to exceed neuron D’s weighted barrier of 5 and neuron E is dependent on the inhibitory neuron C but is actually an Excitatory transmitter for neuron E. So both neurons A and B must exceed neuron C’s weighted barrier of 9 to force neuron E to output. From the truth table, we can see how neural logic can work, and the combinations of where the neurons operate as boolean adders is only under certain firing rates of the neural circuits. Meaning, unlike discrete logic elements that operate consistently overall input ranges(0 or 1) neural logic will only operate under critical firing rates that can overcome the weighted barriers. From a mere engineering perspective, neural logic might seem useless since there are states where the neurons just don’t work out the logic at all! But neurons have to operate under conditions that discrete logic could not work at all and that is a very noisy environment. You’ll notice that the diagram depicts glial cells and it has been observed that glial cells do discharge. One of the functions of glial cells is to act like little vacuum cleaners sopping up excess neural transmitters or ions of firing neurons. This absorption builds up potential within the glial cells and eventually causes a discharge of the electrical potential within the cell. At first, this may look like simple noise but if you realize that the glial cells absorb excess neural transmitters in direct proportion to the neighboring firing rate of neurons then we can see a very interesting effect from glial cells. Glial cells can actually provide feedback as in the form of a learning impetus that forces neurons to adjust their weights based on their own and surrounding neurons’ activity! With that said: From the neural logic example, the circuits will only operate under certain firing rate conditions, it does imply that neurons learn to ignore noise. This is advantageous since it doesn’t behoove a neuron to fire upon any input given to it. From a perspective of a neuron evolving to survive in a noisy environment firing randomly in response to noise would expend energy needlessly. A neuron adapting to ignore noise is conserving energy and is benefiting its host by being energy parsimonious. The additional benefit to ignoring noise is it forces the neuron to fire at signal strengths that improve the fidelity of information exchange between neurons. In addition to the mentioned advantages of energy conservation of neurons, each neuron will have its own set of inputs along with feedback from glial discharges and therefore uniquely derives its weight barrier in response to those influences. Realize that there really isn’t any way for a neuron to distinguish information from noise, so one could either view the individual neuron learning to ignore noise or is learning the information patterns from its inputs, I am asserting that they are one and the same.

With that said we can conclude that consciousness or mind is a product of emergence from the functional cortical systems of the brain and those systems need not be operating simultaneously, as proven through fMRI research. Below is an attempt at modeling consciousness on the premise of logical equivalency:

Concept model of consciousness.

The components like the “Memory Stream”, “Associations” and “Emotions” have been implemented in code. The software ability of free Association is done through an approach of partial feature fit of data that can be done with neural networks or other digital approaches.

While I’m not claiming that the model is complete or that all components have been developed it is a starting point. So far it is proving to be a daunting task as you can see from the diagram the memory stream is a very critical component and it is proving to take up a lot of resources. And there is a ton of preprocessing for all the sensors, which currently are just virtual sensors within a virtual body or pseudo bot.