Biological Plausibility of Neural Network Models

 

Irina Rabkina and Jessica Schroeder

 

Overview

Throughout the course of our neural networks class we learned about numerous different neural network models, all of which are useful in computer science applications such as classification problems and unsupervised learning.  However, as neural networks are often used to model the biological system off which they are based, we aimed to examine the actual biological plausibility of several of the models we discussed in class.  We researched a few controversial examples of such models, and present the arguments for and against their biological relevance and discuss the importance of these considerations on this page.

 

 

 

 

Main Questions

We had a few questions we wanted to keep in mind throughout our investigation:

 

How biologically relevant are current models?

We wanted to determine the extent to which ANY current models are really biologically relevant.

 

Are some models more biologically relevant than others?

We wanted to examine the difference in biological plausibility between different models.

 

How biologically relevant do models need to be in order to be trusted in scientific studies of the brain?

Humans consider experimenting on human brains immoral, so instead researchers use monkeys, rodents, and even insects as models.  Clearly, thereÕs a point at which we say Òwe know this isnÕt a perfect model of the human brain, but itÕs close enough to be helpful.Ó  Where is this point for artificial neural networks?  This is a question scientists will be debating for decades—weÕre not going to be able to answer it ourselves—but we thought it important to consider.

 

Is biological relevance of the model more important than the modelÕs performance?

If a researcher is faced with the choice between a biologically relevant network that does not produce consistent results and a less biologically relevant model that exhibits the desired behavior, does she need to use the biologically relevant model for biological studies?  Or is it legitimate to say that if itÕs performing well, it can give us insights into the behavior itÕs modeling?

 

 

 

 

Neural Network Models: General

 

Description: hemical_vs_Electrical.jpg

A diagram of a chemical synapse, the most prevalent type in the brain, which models fail to fully mimic (Clark, 2013).

 

 

Although connectionist neural network models often provide very good fits to experimental data, Sharkey and Sharkey (1998) argue that many aspects of such models are biologically problematic.  For one, the programmer chooses aspects such as the basic architecture, the learning technique, the learning rate, and even the data representations.  Because of this, models generally are tweaked to fit the data, instead of staying true to any aspects of human learning.  Another problem is the fact that in computational models, neurons are all the same, while the brain has many different types of neurons—each with a different combination of the many different neurotransmitters that exist in the brain—that work together to cause behavior and cognition.  Computational models also only model electrical synapses, which are rare; the brain is primarily made of chemical synapses, which can have longer-lasting effects than electrical synapses (Spencer, 2009).  In addition, biological neural networks have huge numbers of neurons, while computational ones have orders of magnitude fewer, and biological networks are able to perform much more quickly than artificial neural networks.  Finally, the difference between neurons and computer chips in terms of proneness to malfunction is approximately 1 to more than 109.  This makes sense—we donÕt want our personal computers malfunctioning—but the fact that our models fail to exhibit the randomness, and therefore the necessary robustness, that the brain must exhibit is a huge difference.  We therefore lose nuances of the brain when trying to use artificial neural network models to approximate it.

 

 

Description: https://lh4.googleusercontent.com/UcjufVI82AFY5bAmkIZKg4O21MUDO5grA3GcdSEXcfl5q511icKspn9gyT_vPdQE65WAqhnrQu1nG-78mPf6_D-AKaZEMaPNEbjbTmbGtbpFnlAzDynru--sJ1X7ho0K7gWK

Glial cells (green and red) and neurons (blue).  Schlesinger, 2011.

 

One last problem we want to point out with the concept of modeling the human brain is the degree to which we simply do not have the knowledge we need to accurately model it.  Even biological phenomena that scientists consider fairly well-characterized are not completely understood.  The best example of this lack of knowledge is the fact that we only ever model neurons, not glial cells.  Glial cells make up about 90% of the cells in the human brain.  In the image above, the green and red cells are glial cells; only the blue cells are neurons.  We donÕt talk much about glial cells simply because weÕre still not quite sure exactly what they do.  Their name literally means Òglue,Ó and came about because we thought their only function was structural support.  However, now know that theyÕre involved in the breakdown of certain neurotransmitters, and allow for transmission of action potentials over longer axons.  The fact that weÕre ignoring the existence of these cells in our models shows the extent to which we are NOT modeling the brain.

 

 

 

 

Hopfield Networks

 

Description: creen Shot 2014-04-13 at 10.27.06 AM.png Description: igure350_1.gif

 

    A Hebbian-Based Network Model of the Medial Temporal Lobe                                                            Image reference: Thebrainlabs.com

                         (Alvarez & Squire, 1994).

 

 

         Hopfield models are used fairly frequently in connectionist models of cognitive function.  For example, such networks have been used in models of memory disorder in diseases such as AlzheimerÕs disease (Zhao et al., 2010) and memory consolidation in the medial temporal lobe (Alvarez & Squire, 1994). Hopfield networks rely on Hebbian learning rules to create the connections between nodes of the network.  Hebbian learning rules are based off of biological phenomena such as associative learning: in associative learning, neurons that fire together have strengthened connectivity; the same holds true in Hebbian learning, as the connection between the postsynaptic and presynaptic neurons strengthens if they fire together.  However, many aspects of Hebbian learning rules, such as the requirement for symmetry, are biologically unfeasible (Mazzoni et al., 1991).  In addition, Hopfield networks simply fall into a stored pattern given an input; they never update.  As biological neural networks, especially those involved in memory processes, are continuously updating, a static model is not particularly biologically relevant.

 

 

 

 

Bolzmann Machines and Deep Learning

 

Description: creen Shot 2014-04-13 at 12.01.11 PM.png

    Boltzmann Machine (left) vs Hopfield Networks (right) (Barra et al., 2012).

 

 

Boltzmann Machines can be used to model many of the same probability distributions as Hopfield Networks, with a lower space requirement but slightly higher time requirement (Barra, Bernacchia, Santucci & Contucci 2012). From a neural modeling perspective, this is important as it gives a biologically more plausible alternative to Hopfield Networks.  Furthermore, O'Reilly (1998) claimed that ÒThe form of synaptic modification necessary to implement [Boltzmann Machines] is consistent with (though not directly validated by) known properties of biological synaptic modification mechanisms,Ó as Boltzmann Machines update stochastically.  In addition, in proposing a parallel search algorithm adapted from statistical mechanics, Ackley, Hinton, and Sejnowski suggested that the algorithm is biologically plausible as a model for classification in the brain because computing units are small and connections hold numeric values (1985). Restricted Boltzmann Machines, which build on the Boltzmann Machine, are now commonly used in deep learning--a process that has been likened to various types of learning in the biological brain (Utgoff and Stracuzzi, 2002). All this suggests that Boltzmann Machines are more biologically plausible than Hopfield Networks. However, Boltzmann Machines do also have symmetric weights, so they are far from perfect.

 

 

 

 

Self-Organizing Maps

 

 

Description: https://lh6.googleusercontent.com/8vuCDbxlxl-vMlnoTe2D-oQyES4nf3Ue-203OCpEVsyTvBmS2Zs2Mi_TfZ_vWVInyzOhE0acud_BxJAp8h2MuM87-vxdRQAZxazEXm11TgitAMf6r2D4fXbLnw-vubiTMZeK                                                   

                       

Self-Organizing Map based on Color                                                                          

(Chestnut, 2004).                                                                                                                                           

 

        

Self-Organizing Maps are by far the most biologically relevant models we examined in our investigation.  Self-organizing maps have been proposed as bases for a number of neurological phenomena.  They have been used to model everything from lower-level mechanisms such as the organization of the striate cortex (von der Malsburg, 1973) to higher-level concepts such as pathology found in autism (Noriega, 2007).  However, some dispute the biological plausibility on concepts such as Euclidian distance and the global supervision necessary to select neighborhoods (Miikkulainen, 1991).  In addition, many self-organizing maps have no concept of time.  They simply take all the inputs and create the maps in accordance with space.  Chappell and Taylor (1993) argue that for a self-organizing map to be at all biologically relevant it must be temporally sensitive as well, because human learning depends so much on time. It is important to note, however, that much of human memory is Ôself-organizing;Õ for example, grid cells within the entorhinal cortex organize into hexagonal patterns as one learns about spatial navigation (Mhatre et al., 2012), and language acquisition can be modeled using self-organized neural networks (Li et al., 2004).  They therefore seem to have some biological relevance, especially when augmented with a temporal sensitivity.

 

In addition, the basic features of Self-Organizing Maps are more plausible than many other models.  For example, they have a high degree of connectivity, and they follow Hebbian learning rules—neurons that fire together get stronger synaptic strength between them (Spitzer, 1995).   This is a phenomenon observed in human neuroscience and is an important aspect in any model of the brain.  In addition, two inputs that are similar will be closer together than two inputs that are very different.  As the human homunculus is organized in a similar matter—with the area of the brain that encodes signal for the fingers on the left hand being located near the area of the brain that encodes signal for the left palm, for example—this seems to be very biologically consistent.  Lastly and most importantly, Self-Organizing Maps use unlabeled data, whereas many neural network models rely on labeled data.  The human brain does sometimes get feedback when it makes a mistake, but the feedback is often much more vague and delayed than feedback in models that use labeled data.  The brain rarely gets any feedback so clear-cut as ÒI outputted 00101 and should have outputted 00110, so I should now account for that error,Ó and models that donÕt rely on such feedback are more relevant than models that do.

 

 

 

 

Ongoing Projects

Several major projects are looking to model the human brain more closely and/or understand it well enough to create such models. A few of these are described below.

 

NeuroGrid

NeuroGrid is a neuromorphic circuit built by Benjamin et al. (2014) that seeks to model processes of entire brain. It can perform many of the same computations as the brain, but it is much less efficient in terms of time, memory, and power. It is also much larger than the human brain. The scientists are seeking to update the circuit with newer technology in order to improve its performance; however, new technology may not be enough—we simply do not understand what makes our brains as efficient as they are, so recreating this efficiency at this point in time is nearly impossible.

 

The Virtual Brain

The Virtual Brain is an open-source project that allows users to model structural and functional connectivity. The connectivity is based off of DTI and fMRI data from real human participants. This is, however, very much a work in progress and is not yet particularly accurate (Jirsa et al., 2010).

 

 

The Human Connectome Project

The Human Connectome Project is a long-term collaboration between several universities in the United States and abroad that seeks to accurately model the structure and function of the human brain. They have collected structural data using HARDI, and R-fMRI and functional data using fMRI, EEG, MEG for models of function from hundreds of participants (Toga et al., 2012). This data will be combined to learn more about the brain; hopefully this will allow scientists to create a full working model in the near future.

 

 

 

 

References

á      Ackley, D. H., Hinton, G. E. and Sejnowski, T. J. (1985), A Learning Algorithm for Boltzmann Machines. Cognitive Science, 9: 147–169.Alvarez, P. & Squire, L. (1994). Memory consolidation and the medial temporal lobe: A simple network model. PNAS. 91. p.7041-7045.

á      Barra, A., Bernacchia, A., Santucci,E. & Contucci, P. (2012).On the equivalence of Hopfield networks and Boltzmann Machines.Neural Networks, 34.

á      Benjamin, B.V.; Gao, P.; McQuinn, E.; Choudhary, S.; Chandrasekaran, A.R.; Bussat, J.-M.; Alvarez-Icaza, R.; Arthur, J.V.; Merolla, P.A.; Boahen, K., (2014) "Neurogrid: A Mixed-Analog-Digital Multichip System for Large-Scale Neural Simulations," Proceedings of the IEEE , vol.PP, no.99, pp.1,18

á      Chappell, G. J., & Taylor, J. G. (1993). The temporal Koh¿nen map. Neural networks, 6(3), 441-445.

á      Chesnut, Casey. "Self Organizing Map AI for Pictures." (2004).

á      Clark, Sean.  ÒKIN-450 Neurophysiology Wiki.Ó  2013.  Accessed 20 April 2014.

á      Dyer, Michael G. "The promise and problems of connectionism." Behavioral and Brain Sciences 11.01 (1988): 32-33.

á      Hoffman RE (1997) Neural Network Simulations, Cortical Connectivity, and Schizophrenic Psychosis MD Computing 14:200-208

á      Jirsa, V. K., Sporns, O., Breakspear, M., Deco, G., & McIntosh, A. R. (2010). Towards the virtual brain: network modeling of the intact and the damaged brain. Arch. Ital. Biol, 148(3), 189-205.

á      Li, P., Farkas, I., & MacWhinney, B. (2004). Early lexical development in a self-organizing neural network. Neural Networks, 17(8), 1345-1362.

á      Mazzoni, P., Andersen, R. A. & Jordan, M. I. (1991). A more biologically plausible learning rule for neural networks. Proceedings of the National Academy of Sciences, 88(10), 4433-4437.

á      Mhatre, H., Gorchetchnikov, A., & Grossberg, S. (2012). Grid cell hexagonal patterns formed by fast selforganized learning within entorhinal cortex. Hippocampus, 22(2), 320-334.

á      Miikkulainen, Risto. "Self-organizing process based on lateral inhibition and synaptic resource redistribution." Proceedings of the International Conference on Artificial Neural Networks. 1991. 

á      Noriega, Gerardo. "Self-organizing maps as a model of brain mechanisms potentially linked to autism." Neural Systems and Rehabilitation Engineering, IEEE Transactions on 15.2 (2007): 217-226.

á      O'Reilly, R. C. (1998). Six principles for biologically based computational models of cortical cognition. Trends in cognitive sciences, 2(11), 455-462.

á      Sharkey, A., & Sharkey, N. (1998). Cognitive modeling: Psychology and connectionism. In M. Arbib (Ed.),The Handbook of Brain Theory and Neural Networks (1 ed., pp. 200-203). Cambridge, Massachusetts: MIT Press.

á      Schlesinger, Mark (2011).  The role of Glial Cells and Neuroplasticity in Pain Reception.  Schlesinger Pain Center.

á      Spencer KM (2009) The Functional Consequences of Cortical Circuit Abnormalities on Gamma Oscillations in Schizophrenia: Insights from Computational Modeling.  Front Hum Neurosci. 3:33.

á      Spitzer, M. (1995). A neurocomputational approach to delusions.Comprehensive psychiatry, 36(2), 83-105.

á      Toga, A. W., Clark, K. A., Thompson, P. M., Shattuck, D. W., & Van Horn, J. D. (2012). Mapping the human connectome. Neurosurgery, 71(1), 1.

á      Utgoff, P.E. and Stracuzzi, D.J. ÒMany-layered learning.Ó Neural Computation 14.10 (2002).

á      von der Malsburg, Chr. "Self-organization of orientation sensitive cells in the striate cortex." Kybernetik 14.2 (1973): 85-100.

á      Zhao, Wangxiong, Qingli Qiao, and Dan Wang. "A Hopfield-like hippocampal CA3 neural network model for studying associative memory in Alzheimer's disease." Neural Regeneration Research 5.22 (2010).