Skip to main content
Erschienen in: Cognitive Neurodynamics 2/2009

Open Access 01.06.2009 | Research Article

Working memory dynamics and spontaneous activity in a flip-flop oscillations network model with a Milnor attractor

verfasst von: David Colliaux, Colin Molter, Yoko Yamaguchi

Erschienen in: Cognitive Neurodynamics | Ausgabe 2/2009

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

Many cognitive tasks require the ability to maintain and manipulate simultaneously several chunks of information. Numerous neurobiological observations have reported that this ability, known as the working memory, is associated with both a slow oscillation (leading to the up and down states) and the presence of the theta rhythm. Furthermore, during resting state, the spontaneous activity of the cortex exhibits exquisite spatiotemporal patterns sharing similar features with the ones observed during specific memory tasks. Here to enlighten neural implication of working memory under these complicated dynamics, we propose a phenomenological network model with biologically plausible neural dynamics and recurrent connections. Each unit embeds an internal oscillation at the theta rhythm which can be triggered during up-state of the membrane potential. As a result, the resting state of a single unit is no longer a classical fixed point attractor but rather the Milnor attractor, and multiple oscillations appear in the dynamics of a coupled system. In conclusion, the interplay between the up and down states and theta rhythm endows high potential in working memory operation associated with complexity in spontaneous activities.
Hinweise
David Colliaux and Colin Molter contributed equally to this work.

Introduction

During the past 60 years, despite seminal observations suggesting the existence and the importance of complex dynamics in the brain (Nicolis and Tsuda 1985; Skarda and Freeman 1987; Babloyantz and Destexhe 1986), fixed point dynamics has been the predominant regime used to describe brain information processing and more precisely to code associative memories (Amari 1977; Hopfield 1982; Grossberg 1992). More recently, the increasing power of computers and the development of new statistical mathematics demonstrated less equivocally the necessity to rely on more complex dynamics (e.g. Varela et al. 2001; Kenet et al. 2003; Buzsaki and Draguhn 2004). In that view, by extending classical Hopfield networks to encode cyclic attractors, the authors demonstrated that cyclic and chaotic dynamics could encompass several limitations of fixed point dynamics (Molter et al. 2007a, b).
During working memory tasks, tasks requiring the ability to maintain and manipulate simultaneously several chunks of information for central execution (Baddeley 1986), human scalp EEG (Mizuhara and Yamaguchi 2007; Onton et al. 2005) and neural firing in monkeys (Tsujimoto et al. 2003; Rainer et al. 2004) suggested that the theta rhythm (4–8 Hz) plays an important role. The neural basis of the working memory has been widely investigated in primates by using delay to matching tasks. In these tasks, the primate has to retain specific information during a short period of time to guide a forthcoming response. Single cell recordings has shown that during this delay period, some cells located in specific brain areas had increased firing rates (Fuster and Alexander 1971; Rainer et al. 1998).
How these cells can maintain sustained activity is not solved. However, while the coding of information in long-term memory is mediated by synaptic plasticity (e.g. Whitlock et al. 2006), it seems that working memory and more generally short-term memory relies on different mechanisms, potential candidates being that the information is maintained in the network’s dynamics itself (Goldman-Rakic 1995). Recent related observations reported the existence of slow oscillations in the cortex associated with ‘flip-flop’ transitions between bistable up- and down-states of the membrane potential; Up-states, or depolarized states, being associated with high firing rates of the cells during seconds or more. The transition between up and down states as well as the maintenance of high activity during up states could be the result of network interactions, and the neural basis for working memory (McCormick 2005).
Different models of working memory have been proposed. According to the classification proposed by Durstewitz et al. (2000), one type of model is based on a cellular mechanism of bistability (e.g. Lisman and Idiart 1995). Another type follows to the tradition of classical rate coding scheme proposed by (Amari 1977; Hopfield 1982; Grossberg 1992). In these models (e.g. Brunel et al. 2004; Mongillo et al. 2005), the associative memories are coded by cell assemblies in a recurrent network where the synaptic weights are usually pre-encoded according to a Hebbian rule (Hebb 1949). As a result, the recovery of a memory, i.e. of a cell assembly, leads to persistent activity of that cell assembly through recurrent excitation. This could be described as an up-state, while the attractor to the resting state would be the down-state. Flip-flop transitions between up- and down-states are regulated by applying transient inputs.
Remarkably, these models neglect the presence, and accordingly the possible contribution, of the theta rhythm observed during working memory tasks. However, many physics studies have demonstrated the powerful impact of rhythms and oscillations on synchronization which in turn could play an important role during working memory tasks. Here, to conciliate the cell assembly theory with the presence of brain rhythms, we hypothesize that during up-states the firing rate is temporally modulated by an intrinsic cellular theta rhythm. Synchronization among cellular rhythms leads to dynamical cell assembly formation in agreement with observation of EEG rhythms. Grounding on two previous reports (Colliaux et al. 2007; Molter et al. 2008), we propose here the “flip-flop oscillation network” characterized by two different temporal and spatial scale. First, at the unit scale, a theta oscillation is implemented, second, at the network scale, cell assemblies are implemented in the recurrent connections. We demonstrate how the intrinsic cellular oscillation implemented at the unit level can enrich, at the network level, the dynamics of the flip-flop associative memory.
In the following section, we first formulate the flip-flop oscillation network model. Then, we show that the dynamics for a single cell network has an interesting attractor in term of the Milnor attractor (Milnor 1985). Following that, network dynamics in two coupled system is further analyzed. Finally, associative memory networks are elucidated by focusing on possible temporal coding of working memory.

Model

To realize up- and down-states where up-states are associated with oscillations, two phenomenological models are combined. First, in tradition of Hopfield networks and the theory of associative memories, each cell i is characterized by its membrane potential S i , and is modulated by the activity of other cells through recurrent connections (Hopfield 1982). Second, to account for the presence of an intrinsic oscillation, each cell is characterized by a phase ϕ i . The phase follows a simple phase equation defining two stable states: a resting state and a periodic motion (Yamaguchi 2003; Kaneko 2002; Molter et al. 2007c). These two variables are coupled such that, first, an oscillation component cos ϕ i produces intrinsic oscillation of the membrane potential, second, the evolution of the phase depends on the level of depolarization. As a result, the cell’s dynamics results from the non linear coupling between two simpler dynamics having different time constant. Thus, in a network of N units, the state of each cell is defined by \(\{S_i, \phi_i\} \in \Re \times [0,2\pi[\) (\(i \in [1,N]\)) and evolves according to the dynamics:
$$ \left\{ \begin{array}{l}\frac{dS_i}{dt} = -S_i + \sum_{j=1}^N w_{ij} R(S_j) +\Upgamma(\phi_i)+I_i\\ \frac{d\phi_i}{dt} = \omega + (\beta -\Lambda(S_i)) \sin\phi_i \end{array} \right. $$
(1)
with w ij , the synaptic weight between cells i and j, R(S j ), the spike density of the cell j, and I i represents the driving stimulus which enables to selectively activate a cell. In the second equation, ω and β are respectively the frequency and the stabilization coefficient of the internal oscillation.
The spike density is defined by a sigmoid function:
$$ R(x) = \frac{1}{2} \left(\tanh\left( g ( x - 0.5 ) \right) + 1 \right), $$
(2)
The couplings between the two equations, Γ and Λ appear as follows:
$$ \left\{ \begin{array}{l}\Upgamma(\phi_i)= \sigma (\cos\phi_i-\cos\phi_0)\\ \Uplambda(S_i) = \rho{S_i} \end{array} \right. $$
(3)
where ρ and σ modulates the coupling between the internal oscillation and the membrane potential, and ϕ0 is the equilibrium phase obtained when all cells are silent (S i  = 0); i.e. \(\phi_0=\arcsin(-\omega/\beta)\).
The following set of parameters was used: ω = 1, β = 1.2 and g = 10. Accordingly, \(\cos\phi_0\approx-0.55\). ρ, σ and w ij are adjusted in each simulations. A C++ Runge–Kutta Gill integration algorithm is used with a time step set to h = 0.01 in first simulations and to h = 0.1 in Section “Working memory of cell assemblies” simulations (to fasten the computation time). The variation of h leaded to no visible dynamical change. If we consider that ω = 1 represents the 8 Hz theta oscillation, one computational time step represents \(0.01/8/2\pi\sim0.199\,\hbox {ms}\) in first simulations and ∼1.99 ms in last section.

One unit

For one unit, the dynamics in Eq. 1 simplifies to:
$$ \left\{ \begin{array}{l}\frac{dS}{dt}=-S + \sigma (\cos\phi-\cos\phi_0) +I\\\frac{d\phi}{dt} = \omega + (\beta - \rho S) \sin\phi\\\end{array} \right. $$
(4)
Next paragraph analyzes the dynamics in absence of any external input (I = 0). Then, to understand how information is processed by a neural unit, we observe the dynamics in response to different patterns of stimulation.

Fixed point attractors

When I = 0, since ω < β, the dynamics defined by Eq. 1 has two fixed points, M 0 = (0, ϕ0) and M 1 = (S1, ϕ1). The linear stability of M 0 is analyzed by developing the Jacobian of the coupled Eq. 4 at M 0:
$$ DF|_{M_0}=\left(\begin{array}{ll}-1 & -\sigma \sin(\phi_0)\\ -\rho \sin(\phi_0) & \beta\cos(\phi_0) \cr\end{array}\right) $$
(5)
The eigenvalues are given by:
$$ \lambda_{1,2}=\frac{1}{2}\left(\eta-1\pm\sqrt{(\eta-1)^2+4(\rho\sigma\sin^2\phi_0+\eta)}\right)$$
(6)
where \(\eta=\beta\cos\phi_0.\) The linear stability of the fixed point M0 is a function of the internal coupling between S and ϕ, μ = ρσ. With our choice of ω = 1 and β = 1.2, one eigenvalue becomes positive and M0 becomes unstable for \(\mu > \mu_c=0.96.\) Since μ is the crucial parameter for the stability of the resting state, we fixed here ρ = 1 and we analyzed the dynamics according to σ.
Figure 1a shows trajectories and nullclines in the three possible scenario (μ < μ c , μ = μ c and μ > μ c ). The two fixed points M 0 and M 1 (defined by the intersection of the two nullclines) merge for μ = μ c . To have a better understanding of the dynamics, Fig. 1b shows the evolution of the Lyapunov exponents at M 0 and M 1 when σ1 is varied. Figure 1c shows how the two fixed points exchange their stability at criticality. This corresponds to a transcritical bifurcation. Around the bifurcation our two equations (Eq. 1) can be reduced to the normal form of the bifurcation:
$$ \begin{array}{l}\frac{dx_1}{dt}=a x_1^2 + \lambda_1 x_1\\ \frac{dx_2}{dt}= \lambda_2x_2\\ \end{array} $$
(7)
Figure 2 shows how the fixed points changed stability according to the variation of λ1.
For the fixed point M 0:
  • μ < μ c : both Lyapunov exponents are negative. All trajectories converge to the trajectories M 0 which is accordingly stable. The other fixed point, M 1, is unstable with ϕ1 < ϕ0 and S 1 > S 0;
  • μ > μ c : one Lyapunov exponent is negative while the other is positive. It corresponds to the co-existence of both attracting and diverging trajectories. Trajectories attracted to M 0 escape directly from it. M 0 corresponds to an unstable fixed point. M 1 is stable with ϕ1 > ϕ0 and S 1 < S 0;
  • μ = μ c one Lyapunov exponent is negative. Accordingly, the basin of attraction of M 0 has a positive measure and many trajectories are attracted to it. However, since the other Lyapunov exponent equals zero, it exists an ‘unstable’ direction along which the dynamics can escape due to infinitesimal perturbation. Accordingly, M 0 does not attract all trajectories from an open neighborhood and does not define a classical fixed point attractor. However, it is still an attractor if we consider Milnor’s extended definition of attractors (Milnor 1985). At M 0, the two nullclines becomes tangent and the two fixed points merge, M 0 = M 1.
The particularity and interest of the Milnor attractor is the following. For all values of μ, any initial state of our system is attracted and converges to a fixed point of the system (M 0 or M 1). To escape from that fixed point, and to perform an oscillation, a given amount of energy has to be provided, e.g. as a transient input. The more \(\mu\ll\mu_c\) or \(\mu\gg\mu_c,\) the more the amount of energy required to escape from that fixed point is important. By contrary, at the Milnor attractor (or for \(\mu\approx\mu_c,\)) the dynamics becomes very receptive and an infinitesimal perturbation can push the dynamics to iterate through an oscillation.

Response to a constant input

Under constant inputs, the dynamics of the system 4 can either converge to a fixed point, either converge to a limit cycle. To obtain a fixed point \((S_*,\phi_*),\) dϕ/dt must be equal to zero in Eq. 4, which requires:
$$ \left| \frac{\omega}{\beta-S_*} \right| < 1$$
(8)
and Eq. 4 becomes:
$$ \begin{array}{l}S_*=\sigma (\cos \phi_* - \cos \phi_0)+I\\ 0 =\omega + (\beta -\sigma(\cos\phi_* -\cos\phi_0) - I)\sin\phi_* \end{array} $$
(9)
Figures 3 and 4a show the evolution of the dynamics for various levels of input. Fixed points are obtained either for small values of I (S < (β−ω) in Condition 8), either for large values of I (S > β in Condition 8). In between, oscillatory dynamics occurs. The period of the oscillation can be approximated by identifying S with its temporal average, \(\overline{S}\) and by solving the integral \(\int_0^{2 \pi} \frac{d\phi} {\omega + (\beta-\overline{S})sin(\phi)}.\) This approximation gives an oscillation at frequency \(\omega'=\sqrt{\omega^2-(\beta-\overline{S})^2},\) which is in good agreement with computer experiments (Fig. 3).
In Fig. 4a, the dynamics of our system of two equations is compared with the dynamics obtained for σ = 0 (i.e. for a simple integrating unit as in the rate model) and for the current output from the phase model (\(\cos\phi\)). As expected from Fig. 3, the coupling of the phase adds oscillation in a large range of inputs, and, as a first order approximation, our unit’s dynamics appears as a non-linear combination of both dynamics.

Response to oscillatory input

Since the output of a unit can become the input to other units and can contain oscillations, it is interesting to see how a unit reacts to oscillatory inputs. Figure 4b shows the dynamics obtained when considering an oscillating currents having an increasing frequency. The input current follows the ZAP function, as proposed in (Muresan and Savin 2007): \(I=\sin(\alpha t^{\beta}),\) with \(\alpha=2 \pi 10^{-5}\) and β = 3.
Again the two limit cases are considered. First, when σ = 0 (third row), the unit becomes an integrating unit and the membrane potential oscillates at the input frequency but with exponential attenuation in amplitude. Second, for the phase impact (fourth row), at slow input frequency (lower than ω), one or several oscillations are nested in the up-state. When the input frequency is larger than ω (t > 650 ms), oscillations follows the input frequency which leads to faster oscillations with attenuation.
When the two dynamics are coupled (second row), at very low frequencies (lower than 2 Hz), the dynamics goes from fixed point dynamics to limit-cycle dynamics during up-states. We propose that these oscillations at the theta frequency occurring for positive currents could model the increase of activity during ‘up-states’. For faster frequencies of the input signal, but still slower than the phase frequency ω, one oscillation occurs during the up-state, leading to higher depolarization of the membrane potential, and an attenuation function of the frequency appears as a result of the integration. At frequencies faster that the phase frequency ω, interference occur and generate higher mode resonances.

Two coupled units

For sake of simplicity, we will discuss results obtained for symmetrical connections (w 12 = w 21 = w). In the two following simulations, the dynamics of the network is analyzed after a transient excitation is applied to one unit. Figure 5 analyzes the dynamics by looking at the extrema of the membrane potential and by quantifying the phase difference between the two units. The measure of synchrony is obtained by dividing the time between the maximum of the membrane potential of the two units, \(\delta \psi=t(S_1^{\max})-t(S_2^{\max}),\) by the period of the signal, T. Accordingly, \(\delta \psi/ T=0\) corresponds to the two units oscillating at same phase (phase locked), and \(\delta \psi/ T=0.5\) corresponds to the two units being antiphase locked.
The impact of different type of inputs is beyond the scope of this paper and only the impact of a transient input applied to one unit is considered.

From an Hopfield network to a flip flop network

For σ = 0, the membrane potential is not influenced by the phase and the system of two equations simplifies to a classical recurrent network. Since Hopfield, that system is well known for its ability to encode patterns as associative memories in fixed point attractors (Hopfield 1982; Amit 1989; Durstewitz et al. 2000). In our two units network, if the coupling strength w is strong enough (w larger than 0.68), the dynamics is characterized by two stable fixed points, a down state (S 1 = S 2 = 0) and an up-state (visible in Fig. 5a with S 1 ∼ 0.65). Applying a positive transient input to one of the two cells suffices to move the dynamics from the down-state to the up-state. In return, a negative input can bring back the dynamics to the down state.
When increasing σ, a small oscillation characterizes the up-state and associative memory properties (storage and completion) are preserved. The two cells oscillate at nearly identical phases.
For large values of σ (i.e. large internal coupling), the membrane potential saturates. The region of interest lies before saturation when the membrane returns transiently near the resting state and accordingly not so far from the Milnor attractor. This leads to complex dynamics as shown by the fractal structure of the local minima of S 1 (lower part of Fig. 5a). In that situation, a transient input applied simultaneously to the two units would synchronize the two units’ activity and after one oscillation, when the membrane potentials cross the resting state, they stop their activity and remain in the fixed point attractor (not shown here).

Influence of the synaptic weight

Figures 5b and 6 show the impact of varying the synaptic weight on the dynamics of the two units after transient stimulation of one unit.
For small synaptic weights, the two units oscillate anti-phase (Fig. 6a) and their membrane potentials periodically visit the resting state for a long time. The frequency of this oscillation increases with the synaptic weights and for strong synaptic weights (Fig. 6c), rapid in-phase oscillation appear. In that situation, the oscillation does not cross anymore the fixed point attractor, the dynamics remains in up-state. For intermediate coupling strength, an intermediate cycle is observed and more complex dynamics occur for a small range of weights (0.58 < w < 0.78 with our parameters, Fig. 5b). The bifurcation diagram shows multiple roads to chaos through period doubling (Fig. 6b).

Working memory of cell assemblies

In this section, a network of N = 80 cells containing eight strongly overlapped cell assemblies is simulated. Each cell assembly contained 10 cells among which seven were overlapping with other cell assemblies. The intersection between two cell assemblies was limited to two cells. The connectivity of the network was chosen bimodal: The synaptic weights between cells lying in the same cell assembly were chosen from the normal distribution (μ = 0.8 and σ = 0.15), while the other weights were chosen from the normal distribution (μ = 0.2 and σ = 0.1).
To avoid saturation, a global inhibition unit has been added to the network. This unit computes the total activity of the network \((A={\sum}R(S_i))\) and inhibits all cell with the following negative current:
$$ \Updelta\left(\gamma\left(A-\kappa N\right)\right) $$
(10)
where Δ(x) = −x for x > 0 and 0 elsewhere; γ defines the strength of the inhibitory cell (here 0.1) and κ (in %) defines a threshold triggering the inhibition (here 0.03, meaning that inhibition starts when more than 3% of cells are activated).

Spontaneous activity

Figure 7 shows the spontaneous activity of the network, i.e. in absence of any external stimuli. The upper part of the figure shows the evolution of the membrane potential for the 80 cells (each cell is represented by a different color). The membrane potential is maintained around the Milnor attractor (nearby the stable fixed point). The middle part in Fig. 7 shows rasterplot of cells activity. Global inhibition was tuned to prevent the simultaneous activation of multiple cell assemblies and each period is associated with the activation of a specific subset of cells. Lower part in Fig. 7 quantifies the proportion of each cell assembly activated at each time step. During the observed period, all cell assemblies (characterized in the figure by a different color and letter) are reactivated. It has to be noted that since cell assemblies are overlapping, the total reactivation of a cell assembly is necessarily associated with the partial reactivation of other cell assemblies.
The activity of the network can be explained in the following way: when one cell is activated, it tends to activate the cells belonging to the same cell assembly and the activated cells undergo one oscillation before going back to the resting state, nearby to the Milnor attractor. At the Milnor attractor, any cell can be kicked out of the attractor leading to the activation of a different or the same cell assembly. As a result, the spontaneous activity of the system is characterized by the Milnor attractor which provides a kind of reset of the entire network, enabling the activation of the different stored memories. The system itinerates among the previously stored memories, with the passage from one memory to a new one characterized by the “ruin of the Milnor attractor” leading to a kind of “chaotic itinerancy” (Tsuda 1992; Kaneko 1992).
The overlapping structure of the cell assemblies can influence the sequence of reactivation: The reactivation of one specific cell assembly will tend to reactivate the cell assemblies sharing common cells.

Working memory

In this section, after applying external stimuli during short time periods to part of one or several cell assemblies embedded in the network, we observe the network’s ability to sustain the activity of these cell assemblies in agreement with the theory of working memory. To reflect biologically plausible conditions, cell assemblies were sequentially activated during 100 computational time steps; i.e. approximately 200 ms.
In the previous section, we saw that after the reactivation of one cell assembly, all cell assemblies are equally probable to be reactivated. To modify this probability and to increase the probability of selected cell assemblies to be reactivated, as previously proposed (Molter et al. 2008), we simulated a transient period of Hebbian synaptic plasticity during the presentation of the stimuli. Practically, at each computational step, weights between cells reactivated during that period were increased by 0.01. We believe that the proposed mechanism can be related with growing evidence showing that attention-like processes are associated with period of short term plasticity (van Swinderen 2007; Jaaskelainen et al. 2007).
In Fig. 8a, part (40%) of one cell assembly was stimulated. In Fig. 8b, part (40%) of three cell assemblies were sequentially stimulated. First, we observe that in agreement with the theory of associative memories, in both scenario, rasterplot figures indicate that the stimulation of 40% of a cell assembly results in its complete reactivation. Second, it appears that in both cases, the transient stimulation of a cell assembly leads to its preferential reactivation and to a form of sustained activity.
The explanation is the following. After external stimulation, the network is attracted to its resting state, and again the dynamics reflects the presence of diverging orbits crossing attracting orbits. When one cell assembly was reactivated (Fig. 8a), its recurrent connections were increased during the attention-like reactivation, and that cell assembly is now more likely to win the competition and to be reactivated. Since the rapid weight increase due to attention is not balanced by a slow decrease, no dynamics transition is expected and short term memory is expected to last forever (as in Brunel et al. 2004; Mongillo et al. 2008). To implement a slow extinction of the working memory (after dozens of seconds), a mechanism of slow weight decrease could be implemented.
When three cell assemblies were reactivated (Fig. 8b), the balance between excitation and inhibition results in the competition of these three cell assemblies which leads to complex patterns of activity. As a result, the three cell assemblies show clear sustained activity, and these reactivations occur at different times. This result appears important since their simultaneous reactivation would mix them and would prevent any possibility to decode the utile information embedded in the network. Together, these results confirm that our model satisfies one important defining feature of the working memory: The ability to maintain simultaneously multiple chunks of information in short term memory.

Discussion

Recent neuroscience emphasized the importance of the brain dynamics to achieve higher cognitive functions. From a computational perspective, it gave straightforward directions for the development of new type of models where dynamics played more central roles (Kozma and Freeman 2003; Molter et al. 2007). In this paper, to conciliate the classical Hopfield network (Hopfield 1982) with the need of more complex dynamics, we propose a new biologically motivated network model, called “flip-flop oscillation network”.
In this model, each cell is embedded in a recurrent network and is characterized by two internally coupled variables, the cell’s membrane potential and the cell’s phase of firing activity relatively to the theta local field potential.
In a first study, we demonstrate theoretically that in the cylinder space, the Milnor attractor (Milnor 1985) appears at a critical condition through forward and reverse saddle-node bifurcations for a one-cell network (Figs. 1, 2). Near the critical condition, the pair of saddle and node constructs a pseudo-attractor, which leads to Milnor attractor-like properties in computer experiments. Nearby the attractor, the dynamics of the cell is characterized by high sensitivity: infinitesimal transient perturbation can activate an oscillation of the membrane potential.
Simulations of a two-cell network revealed the presence of numerous complex dynamics. We observed that semi-stability of the Milnor attractor dynamics characterizing one cell dynamics, combined in a two cell-network leads to oscillations and chaotic dynamics through period doubling roads (Fig. 5). The important role played by the Milnor attractor for the apparition of the chaotic attractor suggests chaotic itinerancy (Tsuda 2001).
Finally, we tested our model during spontaneous activity (Fig. 7) and for selective maintenance of previously stored information (Fig. 8). In agreement with the cell assembly theory (Hebb 1949), multiple overlapping cell assemblies were phenomenologically embedded in the network to simulate chunks of information.
During spontaneous activity, the network is dynamically switching between the different cell assemblies. After the reactivation of a specific cell assembly, the quasi-stable state of the Milnor attractor provides receptivity to second order dynamics of the internal state and to external stimuli. During that “receptive” or “attention-like” state, a different information can be reactivated. This is supported by recent biological reports; for example, the spatiotemporal pattern of activity of cortical neurons observed during thalamically triggered events are similar to the ones observed during spontaneous events (MacLean et al. 2005). From a dynamics perspective, the presence of the Milnor attractor prevents the network to be “occulted” by one information and leads to chaotic itinerancy between the previously stored cell assemblies.
To simulate working memory tasks, we analyzed if the network could transiently hold specific memories triggered by external stimuli. To reliably and robustly enforce the maintenance of triggered cell assemblies, short period of synaptic plasticity was simulated, reproducing attention-like processes (van Swinderen 2007; Jaaskelainen et al. 2007). As a result, one or multiple cell assemblies could selectively be maintained. An important feature brought by the addition of the theta rhythm to the classical Hopfield network is that different cell assemblies are separately reactivated at different theta oscillations. At each theta cycle, cells from a unique cell assembly are synchronously reactivated. That enables our network to solve the binding problem by reactivating at different phases different overlapping cell assemblies.
If the addition of an internal oscillation to classical rate coding model (e.g. Brunel et al. 2004; Mongillo et al. 2008) can solve the binding problem, it can still not explain in an easy way the limited memory capacity of working memory [e.g. the magical number seven (Miller 1956)]. In that sense, it differs from the seminal paper from Idiart et al. (Lisman and Idiart 1995) where it was proposed that memories are stored in gamma cycles embedded in theta cycles, and that the magical number seven is explained by the number of gamma cycles which can be embedded in a theta cycle. Further studies should focus more deeply on the capacity problem. More precisely, the reactivation of multiple items at different phase of a same theta cycle should be tested.
To summarize, we are proposing a compact, effective, and powerful “flip-flop oscillations network” whose dynamical complexity can be of interest for further analysis of integrative brain dynamics. First attempts to solve working memory tasks gave promising results were different chunks of information were reactivated at different theta cycles. Finally, in our model, the information conveyed during spontaneous activity and working memory tasks appeared similar, reminding the famous quote from Rodolfo Llinas: “A person’s waking life is a dream modulated by the senses” (Llineas 2001).

Open Access

This article is distributed under the terms of the Creative Commons Attribution Noncommercial License which permits any noncommercial use, distribution, and reproduction in any medium, provided the original author(s) and source are credited.
Open AccessThis is an open access article distributed under the terms of the Creative Commons Attribution Noncommercial License (https://​creativecommons.​org/​licenses/​by-nc/​2.​0), which permits any noncommercial use, distribution, and reproduction in any medium, provided the original author(s) and source are credited.
Fußnoten
1
Lyapunov exponents aims to quantify the dependency of the dynamics to infinitesimal perturbation. They are computed here analytically as the eigenvalues of the Jacobian (see Ott 1993).
 
Literatur
Zurück zum Zitat Amit DJ (1989) Modelling brain function: the world of attractor networks. Cambridge University Press, New York Amit DJ (1989) Modelling brain function: the world of attractor networks. Cambridge University Press, New York
Zurück zum Zitat Babloyantz A, Destexhe A (1986) Low-dimensional chaos in an instance of epilepsy. Proc Natl Acad Sci USA 83:3513–3517 Babloyantz A, Destexhe A (1986) Low-dimensional chaos in an instance of epilepsy. Proc Natl Acad Sci USA 83:3513–3517
Zurück zum Zitat Baddeley AD (1986) Working memory. Oxford University Press, New York Baddeley AD (1986) Working memory. Oxford University Press, New York
Zurück zum Zitat Brunel N, Hakim V, Isope P, Nadal JP, Barbour B (2004) Optimal information storage and the distribution of synaptic weights; perceptron versus purkinje cell. Neuron 43:745–757PubMed Brunel N, Hakim V, Isope P, Nadal JP, Barbour B (2004) Optimal information storage and the distribution of synaptic weights; perceptron versus purkinje cell. Neuron 43:745–757PubMed
Zurück zum Zitat Buzsaki G, Draguhn A (2004) Neuronal oscillations in cortical networks. Science 304:1926–1929 Buzsaki G, Draguhn A (2004) Neuronal oscillations in cortical networks. Science 304:1926–1929
Zurück zum Zitat Colliaux D, Yamaguchi Y, Molter C, Wagatsuma H (2007) Working memory dynamics in a flip-flop oscillations network model with milnor attractor. In: Proceedings of ICONIP, Kyoto, Japan Colliaux D, Yamaguchi Y, Molter C, Wagatsuma H (2007) Working memory dynamics in a flip-flop oscillations network model with milnor attractor. In: Proceedings of ICONIP, Kyoto, Japan
Zurück zum Zitat Durstewitz D, Seamans JK, Sejnowski TJ (2000) Neurocomputational models of working memory. Nat Neurosci 3:1184–1191PubMedCrossRef Durstewitz D, Seamans JK, Sejnowski TJ (2000) Neurocomputational models of working memory. Nat Neurosci 3:1184–1191PubMedCrossRef
Zurück zum Zitat Fuster JM, Alexander GE (1971) Neuron activity related to short-term memory. Science 173(997):652–654PubMedCrossRef Fuster JM, Alexander GE (1971) Neuron activity related to short-term memory. Science 173(997):652–654PubMedCrossRef
Zurück zum Zitat Gianluigi M, Barak O, Tsodyks M (2008) Synaptic theory of working memory. Science 319(5869):1543–1546PubMedCrossRef Gianluigi M, Barak O, Tsodyks M (2008) Synaptic theory of working memory. Science 319(5869):1543–1546PubMedCrossRef
Zurück zum Zitat Goldman-Rakic PS (1995) Models of information processing in the basal ganglia. MIT Press, Cambridge, pp 131–148 Goldman-Rakic PS (1995) Models of information processing in the basal ganglia. MIT Press, Cambridge, pp 131–148
Zurück zum Zitat Grossberg S (1992) Neural networks and natural intelligence. MIT Press, Cambridge Grossberg S (1992) Neural networks and natural intelligence. MIT Press, Cambridge
Zurück zum Zitat Hebb DO (1949) The organization of behavior; a neuropsychological theory. Wiley, New York Hebb DO (1949) The organization of behavior; a neuropsychological theory. Wiley, New York
Zurück zum Zitat Hopfield JJ (1982) Neural networks and physical systems with emergent collective computational abilities. Proc Natl Acad Sci USA 79:2554–2558PubMedCrossRef Hopfield JJ (1982) Neural networks and physical systems with emergent collective computational abilities. Proc Natl Acad Sci USA 79:2554–2558PubMedCrossRef
Zurück zum Zitat Jaaskelainen IP, Ahveninen J, Belliveau JW, Raij T, Sams M (2007) Short-term plasticity in auditory cognition. Trends Neurosci 30(12):653–661PubMed Jaaskelainen IP, Ahveninen J, Belliveau JW, Raij T, Sams M (2007) Short-term plasticity in auditory cognition. Trends Neurosci 30(12):653–661PubMed
Zurück zum Zitat Kaneko K (1992) Pattern dynamics in spatiotemporal chaos. Physica D 34:141CrossRef Kaneko K (1992) Pattern dynamics in spatiotemporal chaos. Physica D 34:141CrossRef
Zurück zum Zitat Kaneko K (2002) Dominance of milnor attractors in globally coupled dynamical systems with more than 7 ± 2 degrees of freedom. Phys Rev E 66:055201CrossRef Kaneko K (2002) Dominance of milnor attractors in globally coupled dynamical systems with more than 7 ± 2 degrees of freedom. Phys Rev E 66:055201CrossRef
Zurück zum Zitat Kenet T, Bibitchkov D, Tsodyks M, Grinvald A, Arieli A (2003) Spontaneously emerging cortical representations of visual attributes. Nature 425:954–956PubMedCrossRef Kenet T, Bibitchkov D, Tsodyks M, Grinvald A, Arieli A (2003) Spontaneously emerging cortical representations of visual attributes. Nature 425:954–956PubMedCrossRef
Zurück zum Zitat Kozma R, Freeman WJ (2003) Basic principles of the kiv model and its application to the navigation problem. J Integr Neurosci 2(1):125–145PubMedCrossRef Kozma R, Freeman WJ (2003) Basic principles of the kiv model and its application to the navigation problem. J Integr Neurosci 2(1):125–145PubMedCrossRef
Zurück zum Zitat Lisman JE, Idiart MA (1995) Storage of 7 ± 2 short-term memories in oscillatory subcycles. Science 267:1512–1516PubMedCrossRef Lisman JE, Idiart MA (1995) Storage of 7 ± 2 short-term memories in oscillatory subcycles. Science 267:1512–1516PubMedCrossRef
Zurück zum Zitat Llineas R (2001) I of the vortex: from neurons to self. MIT Press, Cambridge Llineas R (2001) I of the vortex: from neurons to self. MIT Press, Cambridge
Zurück zum Zitat MacLean JN, Watson BO, Aaron GB, Yuste R (2005) Internal dynamics determine the cortical response to thalamic stimulation. Neuron 48:811–823PubMedCrossRef MacLean JN, Watson BO, Aaron GB, Yuste R (2005) Internal dynamics determine the cortical response to thalamic stimulation. Neuron 48:811–823PubMedCrossRef
Zurück zum Zitat McCormick DA (2005) Neuronal networks: flip-flops in the brain. Curr Biol 15(8):294–296CrossRef McCormick DA (2005) Neuronal networks: flip-flops in the brain. Curr Biol 15(8):294–296CrossRef
Zurück zum Zitat Milnor J (1985) On the concept of attractor. Commun Math Phys 99(102):177–195CrossRef Milnor J (1985) On the concept of attractor. Commun Math Phys 99(102):177–195CrossRef
Zurück zum Zitat Mizuhara H, Yamaguchi Y (2007) Human cortical circuits for central executive function emerge by theta phase synchronization. Neuroimage 36(361):232–244PubMedCrossRef Mizuhara H, Yamaguchi Y (2007) Human cortical circuits for central executive function emerge by theta phase synchronization. Neuroimage 36(361):232–244PubMedCrossRef
Zurück zum Zitat Molter C, Salihoglu U, Bersini H (2007a) The road to chaos by time asymmetric hebbian learning in recurrent neural networks. Neural Comput 19(1):100CrossRef Molter C, Salihoglu U, Bersini H (2007a) The road to chaos by time asymmetric hebbian learning in recurrent neural networks. Neural Comput 19(1):100CrossRef
Zurück zum Zitat Molter C, Salihoglu U, and Bersini H (2007b) Neurodynamics of higher level cognition and consciousness. In: Giving meaning to cycles to go beyond the limitations of fixed point attractors. Springer Verlag Molter C, Salihoglu U, and Bersini H (2007b) Neurodynamics of higher level cognition and consciousness. In: Giving meaning to cycles to go beyond the limitations of fixed point attractors. Springer Verlag
Zurück zum Zitat Molter C, Sato N, Yamaguchi Y (2007c) Reactivation of behavioral activity during sharp waves: a computational model for two stage hippocampal dynamics. Hippocampus 17(3):201–209PubMedCrossRef Molter C, Sato N, Yamaguchi Y (2007c) Reactivation of behavioral activity during sharp waves: a computational model for two stage hippocampal dynamics. Hippocampus 17(3):201–209PubMedCrossRef
Zurück zum Zitat Molter C, Colliaux D, Yamaguchi Y (2008) Working memory and spontaneous activity of cell assemblies. A biologically motivated computational model. In: Proceedings of the IJCNN conference. IEEE Press, Hong Kong, pp 3069–3076 Molter C, Colliaux D, Yamaguchi Y (2008) Working memory and spontaneous activity of cell assemblies. A biologically motivated computational model. In: Proceedings of the IJCNN conference. IEEE Press, Hong Kong, pp 3069–3076
Zurück zum Zitat Miller GA (1956) The magical number seven, plus or minus two: some limits on our capacity for processing information. Psychol Rev 63:81–97PubMedCrossRef Miller GA (1956) The magical number seven, plus or minus two: some limits on our capacity for processing information. Psychol Rev 63:81–97PubMedCrossRef
Zurück zum Zitat Muresan RC, Savin C (2007) Resonance or integration? Self-sustained dynamics and excitability of neural microcircuits. J Neurophysiol 97:1911–1930PubMedCrossRef Muresan RC, Savin C (2007) Resonance or integration? Self-sustained dynamics and excitability of neural microcircuits. J Neurophysiol 97:1911–1930PubMedCrossRef
Zurück zum Zitat Nicolis J, Tsuda I (1985) Chaotic dynamics of information processing: the “magic number seven plus-minus two” revisited. Bull Math Biol 47:343–365PubMed Nicolis J, Tsuda I (1985) Chaotic dynamics of information processing: the “magic number seven plus-minus two” revisited. Bull Math Biol 47:343–365PubMed
Zurück zum Zitat Onton J, Delorme A, Makeig S (2005) Frontal midline eeg dynamics during working memory. Neuroimage 27(2):341–356PubMedCrossRef Onton J, Delorme A, Makeig S (2005) Frontal midline eeg dynamics during working memory. Neuroimage 27(2):341–356PubMedCrossRef
Zurück zum Zitat Ott E (1993) Chaos in dynamical systems. Cambridge University Press, Cambridge. Ott E (1993) Chaos in dynamical systems. Cambridge University Press, Cambridge.
Zurück zum Zitat Rainer G, Asaad WF, Miller EK (1998) Selective representation of relevant information by neurons in the primate prefrontal cortex. Nature 393:577–579PubMedCrossRef Rainer G, Asaad WF, Miller EK (1998) Selective representation of relevant information by neurons in the primate prefrontal cortex. Nature 393:577–579PubMedCrossRef
Zurück zum Zitat Rainer G, Lee H, Simpson GV, Logothetis NK (2004) Working-memory related theta (4–7 Hz) frequency oscillations observed in monkey extrastriate visual cortex. Neurocomputing 58–60:965–969CrossRef Rainer G, Lee H, Simpson GV, Logothetis NK (2004) Working-memory related theta (4–7 Hz) frequency oscillations observed in monkey extrastriate visual cortex. Neurocomputing 58–60:965–969CrossRef
Zurück zum Zitat Skarda CA, Freeman W (1987) How brains make chaos in order to make sense of the world. Behav Brain Sci 10:161–195CrossRef Skarda CA, Freeman W (1987) How brains make chaos in order to make sense of the world. Behav Brain Sci 10:161–195CrossRef
Zurück zum Zitat Tsuda I (1992) Dynamic link of memorychaotic memory map in nonequilibrium neural networks. Neural Netw 5:313–326CrossRef Tsuda I (1992) Dynamic link of memorychaotic memory map in nonequilibrium neural networks. Neural Netw 5:313–326CrossRef
Zurück zum Zitat Tsuda I (2001) Towards an interpretation of dynamic neural activity in terms of chaotic dynamical systems. Behav Brain Sci 24:793–810 Tsuda I (2001) Towards an interpretation of dynamic neural activity in terms of chaotic dynamical systems. Behav Brain Sci 24:793–810
Zurück zum Zitat Tsujimoto T, Shimazu H, Isomura Y, Sasaki K (2003) Prefrontal theta oscillations associated with hand movements triggered by warning and imperative stimuli in the monkey. Neurosci Lett 351(2):103–106PubMedCrossRef Tsujimoto T, Shimazu H, Isomura Y, Sasaki K (2003) Prefrontal theta oscillations associated with hand movements triggered by warning and imperative stimuli in the monkey. Neurosci Lett 351(2):103–106PubMedCrossRef
Zurück zum Zitat van Swinderen B (2007) Attention-like processes in drosophila require short-term memory genes. Science 315(5818):1590–1593PubMedCrossRef van Swinderen B (2007) Attention-like processes in drosophila require short-term memory genes. Science 315(5818):1590–1593PubMedCrossRef
Zurück zum Zitat Varela F, Lachaux J-P, Rodriguez E, Martinerie J (2001) The brainweb: phase synchronization and large-scale integration. Nat Rev Neurosci 2:229–239 Varela F, Lachaux J-P, Rodriguez E, Martinerie J (2001) The brainweb: phase synchronization and large-scale integration. Nat Rev Neurosci 2:229–239
Zurück zum Zitat Whitlock JR, Heynen AJ, Shuler MG, Bear MF (2006) Learning induces long-term potentiation in the hippocampus. Science 313(5790):1093–1097PubMedCrossRef Whitlock JR, Heynen AJ, Shuler MG, Bear MF (2006) Learning induces long-term potentiation in the hippocampus. Science 313(5790):1093–1097PubMedCrossRef
Zurück zum Zitat Yamaguchi Y (2003) A theory of hippocampal memory based on theta phase precession. Biol Cybern 89:1–9PubMed Yamaguchi Y (2003) A theory of hippocampal memory based on theta phase precession. Biol Cybern 89:1–9PubMed
Metadaten
Titel
Working memory dynamics and spontaneous activity in a flip-flop oscillations network model with a Milnor attractor
verfasst von
David Colliaux
Colin Molter
Yoko Yamaguchi
Publikationsdatum
01.06.2009
Verlag
Springer Netherlands
Erschienen in
Cognitive Neurodynamics / Ausgabe 2/2009
Print ISSN: 1871-4080
Elektronische ISSN: 1871-4099
DOI
https://doi.org/10.1007/s11571-009-9078-0

Weitere Artikel der Ausgabe 2/2009

Cognitive Neurodynamics 2/2009 Zur Ausgabe