Skip Nav Destination
Close Modal
Update search
NARROW
Format
Journal
TocHeadingTitle
Date
Availability
1-4 of 4
Michael L. Hines
Close
Follow your search
Access your saved searches in your account
Would you like to receive an alert when new items match your search?
Sort by
Journal Articles
William W. Lytton, Alexandra H. Seidenstein, Salvador Dura-Bernal, Robert A. McDougal, Felix Schürmann ...
Publisher: Journals Gateway
Neural Computation (2016) 28 (10): 2063–2090.
Published: 01 October 2016
FIGURES
| View All (4)
Abstract
View article
PDF
Large multiscale neuronal network simulations are of increasing value as more big data are gathered about brain wiring and organization under the auspices of a current major research initiative, such as Brain Research through Advancing Innovative Neurotechnologies. The development of these models requires new simulation technologies. We describe here the current use of the NEURON simulator with message passing interface (MPI) for simulation in the domain of moderately large networks on commonly available high-performance computers (HPCs). We discuss the basic layout of such simulations, including the methods of simulation setup, the run-time spike-passing paradigm, and postsimulation data storage and data management approaches. Using the Neuroscience Gateway, a portal for computational neuroscience that provides access to large HPCs, we benchmark simulations of neuronal networks of different sizes (500–100,000 cells), and using different numbers of nodes (1–256). We compare three types of networks, composed of either Izhikevich integrate-and-fire neurons (I&F), single-compartment Hodgkin-Huxley (HH) cells, or a hybrid network with half of each. Results show simulation run time increased approximately linearly with network size and decreased almost linearly with the number of nodes. Networks with I&F neurons were faster than HH networks, although differences were small since all tested cells were point neurons with a single compartment.
Journal Articles
Samuel A. Neymotin, Robert A. McDougal, Mohamed A. Sherif, Christopher P. Fall, Michael L. Hines ...
Publisher: Journals Gateway
Neural Computation (2015) 27 (4): 898–924.
Published: 01 April 2015
FIGURES
| View All (114)
Abstract
View article
PDF
Calcium ( ) waves provide a complement to neuronal electrical signaling, forming a key part of a neuron’s second messenger system. We developed a reaction-diffusion model of an apical dendrite with diffusible inositol triphosphate ( ), diffusible , receptors ( s), endoplasmic reticulum (ER) leak, and ER pump (SERCA) on ER. is released from ER stores via s upon binding of and . This results in -induced- -release (CICR) and increases spread. At least two modes of wave spread have been suggested: a continuous mode based on presumed relative homogeneity of ER within the cell and a pseudo-saltatory model where regeneration occurs at discrete points with diffusion between them. We compared the effects of three patterns of hypothesized distribution: (1) continuous homogeneous ER, (2) hotspots with increased density ( hotspots), and (3) areas of increased ER density (ER stacks). All three modes produced waves with velocities similar to those measured in vitro (approximately 50–90 m /sec). Continuous ER showed high sensitivity to density increases, with time to onset reduced and speed increased. Increases in SERCA density resulted in opposite effects. The measures were sensitive to changes in density and spacing of hotspots and stacks. Increasing the apparent diffusion coefficient of substantially increased wave speed. An extended electrochemical model, including voltage-gated calcium channels and AMPA synapses, demonstrated that membrane priming via AMPA stimulation enhances subsequent wave amplitude and duration. Our modeling suggests that pharmacological targeting of s and SERCA could allow modulation of wave propagation in diseases where dysregulation has been implicated.
Journal Articles
Publisher: Journals Gateway
Neural Computation (2008) 20 (11): 2745–2756.
Published: 01 November 2008
Abstract
View article
PDF
The scale of large neuronal network simulations is memory limited due to the need to store connectivity information: connectivity storage grows as the square of neuron number up to anatomically relevant limits. Using the NEURON simulator as a discrete-event simulator (no integration), we explored the consequences of avoiding the space costs of connectivity through regenerating connectivity parameters when needed: just in time after a presynaptic cell fires. We explored various strategies for automated generation of one or more of the basic static connectivity parameters: delays, postsynaptic cell identities, and weights, as well as run-time connectivity state: the event queue. Comparison of the JitCon implementation to NEURON's standard NetCon connectivity method showed substantial space savings, with associated run-time penalty. Although JitCon saved space by eliminating connectivity parameters, larger simulations were still memory limited due to growth of the synaptic event queue. We therefore designed a JitEvent algorithm that added items to the queue only when required: instead of alerting multiple postsynaptic cells, a spiking presynaptic cell posted a callback event at the shortest synaptic delay time. At the time of the callback, this same presynaptic cell directly notified the first postsynaptic cell and generated another self-callback for the next delay time. The JitEvent implementation yielded substantial additional time and space savings. We conclude that just-in-time strategies are necessary for very large network simulations but that a variety of alternative strategies should be considered whose optimality will depend on the characteristics of the simulation to be run.
Journal Articles
Publisher: Journals Gateway
Neural Computation (2005) 17 (4): 903–921.
Published: 01 April 2005
Abstract
View article
PDF
Realistic neural networks involve the coexistence of stiff, coupled, continuous differential equations arising from the integrations of individual neurons, with the discrete events with delays used for modeling synaptic connections. We present here an integration method, the local variable time-step method ( lvardt ), that uses separate variable-step integrators for individual neurons in the network. Cells that are undergoing excitation tend to have small time steps, and cells that are at rest with little synaptic input tend to have large time steps. A synaptic input to a cell causes reinitialization of only that cell's integrator without affecting the integration of other cells. We illustrated the use of lvardt on three models: a worst-case synchronizing mutual-inhibition model, a best-case synfire chain model, and a more realistic thalamocortical network model.