Do you want to buy antibiotics online without prescription? https://buyantibiotics-24h.com/ - This is pharmacy online for you!

## Bcl.hamilton.ie

Time-Skew Hebb Rule in a Nonisopotential Neuron

**Barak A. Pearlmutter**
To appear (1995) in

*Neural Computation, 7*(4) 706–712

**Abstract**
In an isopotential neuron with rapid response, it has been shown that the receptive fields
formed by Hebbian synaptic modulation depend on the principal eigenspace of ✁ (0), the input
autocorrelation matrix, where

*Qij*(τ) = ξ

*i*(

*t*) ξ

*j*(

*t *τ) and ξ

*i*(

*t*) is the input to synapse

*i *at time

*t*
(Oja 1982). We relax the assumption of isopotentiality, introduce a time-skewed Hebb rule, andfind that the dynamics of synaptic evolution are determined by the principal eigenspace of ✁ .

(

*Qij *ψ

*i*)(τ)

*Kij*(τ)

*d*τ, where

*Kij*(τ) is the neuron’s voltage
response to a unit current injection at synapse

*j *as measured τ seconds later at synapse

*i*, and
ψ

*i*(τ) is the time course of the opportunity for modulation of synapse

*i *following the arrival of

**1 Introduction**
Hebbian synaptic modification involves the enhancement of synaptic efficacy in response to simul-taneous pre- and post-synaptic activity. This form of learning has taken on particular importancewith its discovery in various parts of the brain, a prominent example being long-term potentiation(LTP) in the hippocampus (Bliss and Collingridge 1993).

The classic Oja (1982) analysis of the temporal evolution of Hebbian synapses assumes depen-
dence on instantaneous conjunction of pre- and postsynaptic activity and a linear neuron with anisopotential membrane which responds quickly compared to the time course of its inputs. How-ever, it is believed that the trans-membrane potential in cortical neurons can vary widely across thedendritic tree on a time scale important to LTP (Koch

*et al. *1982; Rall 1977; Zador

*et al. *1995).

It also appears that LTP occurs when postsynaptic voltage is elevated during a finite time windowfollowing presynaptic activation (Bliss and Collingridge 1993). This paper analyzes the effects ofHebbian learning under these conditions, and grew out of an attempt to understand the clusteredspatial structure of facilitated Hebbian synapses seen in the model nonisopotential neurons of Brown

*et al. *(1991) and Mel (1992). Even though linearity in Hebbian systems can be less restrictive thanone might suppose (Miller 1990), it is important to emphasize that the analysis here concerns aneuron that is nonisopotential but nonetheless linear. Nonlinear channels and saturation1 are notconsidered, although they likely play an important computational role (Mel 1993).

Siemens Corporate Research, 755 College Road East, Princeton, NJ 08540, bap@learning.scr.siemens.com
1We model synaptic events as current injections,even though it would be more faithful to the biology to use conductance
changes. Unfortunately, using conductancechances instead of current injection introduces the solution of a general causallinear integral equation into the calculation of ✁ .

**2 Results**
In Oja (1982) and its extensions, the post-synaptic potential

*V*(

*t*) is shared across all synapses,and the input signal ξ(

*t*) plays two roles: that of presynaptic activity in

*V*(

*t*) = ☛

*j wj *ξ

*j*(

*t*), where

*wj *is the efficacy of synapse

*j*, and that of the window of opportunity for synaptic facilitation in

*dwi dt *= ηξ

*i*(

*t*)

*V*(

*t*) decay, where the learning rate η > 0 is a constant of proportionality. Taking
the expected value, this leads to an expected dynamics of

*d*
**decay **, where

*Qij*(τ) = ξ

*i*(

*t*) ξ

*j*(

*t *τ) . These expected dynamics approximate the true dynamics well when η is
small. The synaptic evolution is therefore governed2 by the principal eigenvectors of ✁ (0).

Here, we allow the neuron to be nonisopotential with
ξ

*j*(

*t *τ)

*Kij*(τ)

*d*τ =
where

*Vi*(

*t*) is the post-synaptic potential at synapse

*i*,

*Kij*(τ) gives the voltage response of the
neuron to a unit current injection at synapse

*j *as measured τ seconds later at synapse

*i*, andfunctional convolution is defined by (

*f g*)(

*t*) =
τ)

*g*(τ)

*d*τ. The transfer impedence matrix
(Koch

*et al. *1982) is easy to calculate and reason about in the temporal frequency domain (Zador
We also generalize the Hebb rule by distinguishing between ξ

*i*(

*t*), the presynapic input3 to
synapse

*i *at time

*t*, and ψ

*i*(τ), the window of opportunity for synaptic modulation following
presynaptic activity at synapse

*i*. The synaptic modulation equation thus becomes

*i*(τ)

*d*τ

*Vi*(

*t*)
= η (ξ

*i *ψ

*i*)(

*t*)

*Vi*(

*t*) decay

*Kij*(τ) (

*Qij *ψ

*i*)(τ)

*d*τ

*Kij*(τ) (

*Qij *ψ

*i*)(τ)

*d*τ
then, in matrix notation, we obtain the familiar
2When using a simple Hebb rule in a linear neuron, it is necessary to invoke an additional mechanism to prevent
runaway synaptic facilitation. The particular decay term used has a strong effect, an issue explored in Oja (1982), Miller

and MacKay (1994), and Goodhill and Barrow (1994). In essence, with a multipicative decay term,

**w **evolves to the

principal eigenvector of

**Q**(0) regardless of its initial value, whereas subtractive decay terms cause

**w **to evolve towards

principal eigenvectors of

**Q**(0) but also tend towards the corners of the space.

3To speak of it concretely, ξ

*i*(

*t*) typically consists of a series of delta pulses of constant magnitude representing the
arrival of action potentials at the presynaptic terminal. Presynaptic changes do not affect ξ

*i*(

*t*)—rather,

*wi *encapsulatesall the synaptic efficacy, regardless of whether changes to that efficacy are implemented by pre- or post-synaptic physical
where ✁ has replaced the usual instantaneous autocorrelation matrix ✁ (0). Thus, the evolution
is determined by the principal eigenspace of ✁ in exactly the same way that, in the classic
case, its evolution is determined by the the principal eigenspace of ✁ (0). Note that unlike ✁ (0), ingeneral ✁ is not symmetric. However, ✁ takes on simplified forms under important special cases,
such as when ψ

*i *is a simple delay, when the neuron is isopotential, or when the neuron’s response
time is fast compared to the time course of its input.

**3 Application**
Given known electrotonic structure, known correlational structure of the input, and known conditionsunder which synapses are modulated, one can calculate ✁ , and thus it’s principal components, and
thereby predict the receptive field patterns that will be stabilized. When the input and electrotonicstructure are particularly simple, it becomes possible to calculate the principal components of ✁✆analytically. (Broad-band synaptic input with 1

*f d *correlational structure in both time and space
impinging on either an infinite cable or an infinite sheet makes ✁ ’s principal eigenvectors particularly
tractable.) Thus one can predict the formation of clusters, and their length scale, in a manner similarthat used by Miller, Keller, and Stryker (1989) to account for receptive fields in the visual system,and precisely analogous to that used by Chernjavski and Moody (1990) to predict the length scale ofcortical columns. This is similar in spirit to Mainen

*et al.*’s (1991) account of synapse segregationon nonisopotential neurons, but operates analytically, without laborious simulation. Unfortunately,by tuning within the region of physiological parameter space consistent with experimental data, itis possible to obtain almost any length scale.

We will therefore apply the technique to a simpler situation, where it predicts a robust qualitative
effect. In figure 1 a simple model neuron is constructed. The two synapses are given uncorrelatedspike trains for input. We let ψ1 = ψ2 be a square wave of duration comparable to the inter-
spike interval. The elements

*Qij *ψ are therefore all equal. This means that, up to a constant,
(τ)

*d*τ. This matrix is trivially computed, since

*Qij *is simply the steady-state voltage
at synapse

*i *in response to a constant unit current injected at synapse

*j*. The two elements of theprincipal eigenvector of this matrix predict the equilibrium values of the two synaptic weights. Inplotting these two synaptic weights as a function of the diameter of the soma (figure 2) we noticean interesting effect: when the soma is absent (

*D *= 0) the symmetry of the situation is unbroken,so the two synaptic weights converge to the same value. As

*D *is raised, the soma begins to actas a current sink. This sink is more effective at the more proximal synapse, where the Hebb rulebecomes less effective due to the lowered postsynaptic potential. This results in a strengthening ofthe distal synapse at the expense of the proximal one.

The effect levels off, because even were the soma a perfect current sink, the dendritic resistance

*R*23 holds potential in the proximal dendritic compartment long enough for some synaptic modulation

**4 Summary and Conclusion**
We have shown that passive nonisopotential time-skewed Hebbian learning is mathematically anal-ogous to the isopotential instantaneous case, in that both take the form of an iterated linear operator,plus a decay term,

*d*
**decay **. The destiny of is thus determined by the principal

eigenspace of ✁ and the nature of the decay term, and techniques that have been applied to the
classic case can now be applied to the somewhat more realistic neurons and Hebb rule discussedhere.

**Acknowledgments**
I would like to thank Ted Carnevale, Uzi Levin, Zach Mainen, Bartlett Mel, John Moody, andespecially Tony Zador and Ken Miller for helpful comments and suggestions.

submission of this paper, I found that results similar to those in Pearlmutter and Brown (1992), apreliminary abstract of this paper, had been independently discovered by John E. Moody and AlexChernjavski in 1989, but were not published. Portions of this work were supported by grant ONR067-32123-332 to Thomas H. Brown and by grants NSF ECS-9114333 and ONR N00014-92-J-4062to John Moody.

**References**
Bliss, T. V. and Collingridge, G. L. (1993). A Synaptic Model of Memory: Long-Term Potentiation
in the Hippocampus.

*Nature*,

*361*(6407), 31–39.

Brown, T. H., Mainen, Z. F., Zador, A. M., and Claiborne, B. J. (1991). Self-Organization of Hebbian
Synapses in Hippocampal Neurons. In Lippmann, R. P., Moody, J. E., and Touretzky, D. S.

(Eds.),

*Advances in Neural Information Processing Systems 3*, pp. 39–45. Morgan Kaufmann.

Chernjavski, A. and Moody, J. E. (1990). Spontaneous Development of Modularity in Simple
Cortical Models.

*Neural Computation*,

*2*(3), 334–354.

Goodhill, G. J. and Barrow, H. G. (1994). The Role of Weight Normalization in Competitive
Learning.

*Neural Computation*,

*6*(2), 255–269.

Koch, C., Poggio, T., and Torre, V. (1982). Retinal Ganglion Cells: a Functional Interpretation of
Dendritic Morphology.

*Proc. of the Royal Soc. of London B*,

*298*, 227–264.

Mainen, Z. F., Claiborne, B. J., and Brown, T. H. (1991). A Novel Role for Synaptic Competition
in the Development of Cortical Lamination.

*Society for Neuroscience Abstracts*,

*17*(303.6),759.

Mel, B. W. (1993). Synaptic Integration in an Excitable Dendritic Tree.

*Journal of Neurophysiology*,
Mel, B. W. (1992). NMDA-Based Pattern Discrimination in a Modeled Cortical Neuron.

*Neural*
*Computation*,

*4*(4), 502–516.

Miller, K. D., Keller, J. B., and Stryker, M. P. (1989). Ocular Dominance Column Development:
Analysis and Simulation.

*Science*,

*245*, 605–615.

Miller, K. D. (1990). Derivation of Linear Hebbian Equations from a Nonlinear Hebbian Model of
Synaptic Plasticity.

*Neural Computation*,

*2*(3), 321–333.

Miller, K. D. and MacKay, D. J. C. (1994). The Role of Constraints in Hebbian Learning.

*Neural*
*Computation*,

*6*(1), 100–126.

Oja, E. (1982). A Simplified Neuron Model as a Principal Component Analyzer.

*Journal of*
*Mathematical Biology*,

*15*, 267–273.

Pearlmutter, B. A. and Brown, T. H. (1992). Hebbian Learning is Jointly Controlled by Electrotonic
and Input Structure.

*Society for Neuroscience Abstracts*,

*18*(567.23).

Rall, W. (1977). Core Conductor Theory and Cable Properties of Neurons. In Kandel, E. R. (Ed.),

*Handbook of Physiology*, Vol. 1, pp. 39–98. Amer. Physiol. Soc., Bethesda.

Zador, A. M., Claiborne, B. J., and Brown, T. H. (1991). Attenuation Transforms of Hippocampal
Neurons.

*Society for Neuroscience Abstracts*,

*17*(605.6), 1515.

Zador, A. M., Hagai, A.-S., and Segev, I. (1995). The Morphoelectrotonic Transform: A Graphical
Approach to Dendritic Function.

*Journal of Neuroscience*,

*15*(4), 1669–1682.

Zador, A. M. and Pearlmutter, B. A. (1993). Efficient Computation of Sparse Elements of the In-
verse of a Sparse Near-Tridiagonal Matrix with Application to the Nerve Equation. Tech. rep.

OGI-CSE-93-003, Oregon Graduate Institute of Science & Technology, Department of Com-puter Science and Engineering, Portland, OR. Ftp cse.ogi.edu: /pub/tech-reports/1993/93-003.ps.gz.

Figure 1: A three-compartment model neuron with two synaptic inputs. The equivalent circuitparameters are therefore

*C*1 =

*C*2 = π

*dLCm*,

*C*3 = π

*D*2

*Cm*,

*R*1 =

*R*2 =

*Rm *π

*dL*,

*R*3 =

*Rm *π

*D*2,

*R*12 =

*R*23 = 4

*RaL *π

*D*2. Membrane parameters used were

*Cm *= 1µ

*F *cm2,

*Rm *= 50

*k*Ωcm2,

*Ra *= 200Ωcm,

*d *= 2µ

*m*,

*L *= 100µ

*m*.

*D *was allowed to vary.

Figure 2: The distal (upper) and proximal (lower) final synaptic weights as a function of

*D*, thesomatic diameter in cm. These predicted final weights exactly match the final weights evolved insimulations of this process, which were conducted at

*D *= 0

*, *0

*.*002

*, *0

*.*004

*, *0

*.*01cm.

Source: http://www.bcl.hamilton.ie/~barak/papers/nc-nonisohebb.pdf

Oral Antibiotics Doxycycline, Minocycline, Tetracycline Oral antibiotics for acne have many more potential side effects. Our goal is to limit the dose and duration of oral antibiotic therapy as much as possible. Oral antibiotics used for acne can cause sunburn, antibiotic resistance, brain swelling and may also discolor baby’s teeth if used during pregnancy.

6-in-1 digital video camera able to capture true 5-megapixel images The Take-it® MV500 digital video camera makes digital video easy and affordable. Ideal for beginners, the MV500 is a full-featured digital camera, incorporating functions of digital camcorder, audio recorder, digital still camera, webcam, mass storage function and MP3 player. It is able to capture 5-megapixel (and up to 12-m