Skip to main content
Philosophical Transactions of the Royal Society B: Biological Sciences logoLink to Philosophical Transactions of the Royal Society B: Biological Sciences
. 2018 Feb 26;373(1744):20170275. doi: 10.1098/rstb.2017.0275

A network model of behavioural performance in a rule learning task

Michael E Hasselmo 1,, Chantal E Stern 1
PMCID: PMC5832697  PMID: 29483357

Abstract

Humans demonstrate differences in performance on cognitive rule learning tasks which could involve differences in properties of neural circuits. An example model is presented to show how gating of the spread of neural activity could underlie rule learning and the generalization of rules to previously unseen stimuli. This model uses the activity of gating units to regulate the pattern of connectivity between neurons responding to sensory input and subsequent gating units or output units. This model allows analysis of network parameters that could contribute to differences in cognitive rule learning. These network parameters include differences in the parameters of synaptic modification and presynaptic inhibition of synaptic transmission that could be regulated by neuromodulatory influences on neural circuits. Neuromodulatory receptors play an important role in cognitive function, as demonstrated by the fact that drugs that block cholinergic muscarinic receptors can cause cognitive impairments. In discussions of the links between neuromodulatory systems and biologically based traits, the issue of mechanisms through which these linkages are realized is often missing. This model demonstrates potential roles of neural circuit parameters regulated by acetylcholine in learning context-dependent rules, and demonstrates the potential contribution of variation in neural circuit properties and neuromodulatory function to individual differences in cognitive function.

This article is part of the theme issue ‘Diverse perspectives on diversity: multi-disciplinary approaches to taxonomies of individual differences’.

Keywords: neocortex, rule learning, muscarinic receptors, acetylcholine

1. Introduction

Humans demonstrate important cognitive capacities for rule learning and symbolic processing. Rule learning includes the ability to generalize from previous experience, to allow previously unseen sensory input to be interpreted in terms of a previously learned rule to generate the correct response [1]. Humans show a range of capabilities for learning and detecting rules and applying them to new sensory input, as shown for example by the differences in rule learning performance on the Raven's progressive matrices task [24] that are a reflection of probabilistic abilities, and on other rule learning tasks such as the context-dependent association task described here [5].

The mechanisms of rule learning involve neural circuits in cortical structures such as the prefrontal cortex [6,7], but the detailed circuitry for rule learning has not been determined. The ability to flexibly apply rules to new input requires circuits for symbolic processing. Here we focus on symbolic processing as the capacity for a specific cognitive role in a rule to be flexibly filled by different sensory inputs. This flexible linking of roles to different fillers contributes to the productivity, systematicity and compositionality of cognitive function [1]. This flexible application of rules to different fillers has been modelled in different ways, including the synchrony of oscillations [810], the use of compression operators to link codes [11,12] or the flexible gating between different cortical working memory buffers by the basal ganglia [13,14] consistent with previous theories of flexible routing in prefrontal cortex [6]. The model presented in this paper demonstrates how the concept of flexible gating could be implemented by circuits within neocortex to allow learning and generalization of rules. This builds on previous models using interacting populations of neurons to gate selection of motor actions [15,16] or selection of memory actions loading information into working memory or episodic memory for solving behavioural tasks [15,17,18].

The fundamental feature of this model is the direct influence of gating neurons on the synaptic spread of activity between other neurons, which could arise via different possible mechanisms. This gating could involve synapses on the dendritic tree of pyramidal cells that interact via voltage-dependent N-methyl-d-aspartate (NMDA) conductances, such that a conjunction of nearby synaptic inputs is necessary to generate an output [19,20]. This model is supported by the fact that neurons receive multiple different functional synaptic inputs, but commonly only respond to a subset of these inputs [21,22]. Alternately, it could involve axo-axonic inhibitory interneurons that can directly regulate spiking output [23,24].

The range of human cognitive function could reflect a range of neuromodulatory influences in different individuals [25,26]. This could involve different neuromodulatory systems including dopamine, norepinephrine, serotonin and acetylcholine, which could regulate different aspects of cognitive function [25]. This paper will focus on the potential neuromodulatory role of acetylcholine in cortical circuits. Extensive data indicate an important role of acetylcholine in regulating cortical cognitive function. A broad range of medical anecdotes indicate that blockade of muscarinic acetylcholine receptors by high doses of drugs such as scopolamine or atropine (present in some plant species) causes a breakdown of cognitive function resulting in delirium and hallucinations [27], and this has been proposed to contribute to cognitive dysfunction in diseases such as Lewy body dementia [28]. The essential role for muscarinic receptors in cognition is further supported by studies showing more selective cognitive impairments including impaired encoding of lists of words and impaired working memory with lower levels of muscarinic blockade [2931].

The cognitive effects of muscarinic receptor blockade appear to result from blockade of specific physiological effects at the cellular and circuit level within cortical structures. These muscarinic effects are present in most cortical regions that have been tested including hippocampus, entorhinal cortex and prefrontal cortex [32]. Nicotinic receptors also contribute to the effects of acetylcholine, but blockade of nicotine receptors has weaker effects on cognitive function. The model presented here contains features of neural function that are inspired by data on physiological effects of muscarinic cholinergic receptors [32,33]. Data show that acetylcholine is released during waking in cortical structures including the neocortex and hippocampus [34,35]. This release of acetylcholine causes a number of strong physiological effects that could be important for cognitive function.

Cholinergic modulation strongly influences the spiking activity of cortical neurons by regulating different conductances. Muscarinic receptors cause a direct depolarization of neurons through the block of a leak potassium current [36]. In addition, by blocking calcium-dependent potassium currents [3739] and activating a calcium-sensitive non-specific cation current (CAN current), acetylcholine can cause persistent spiking in a subset of neurons when spiking activity causes calcium influx that activates the calcium-sensitive cation current [4043]. This persistent spiking could provide a means for maintaining working memory of prior activity [44], which could allow gating neurons to bridge the sequential gaps between inputs. In vivo studies show that acetylcholine increases the firing rate and reliability of the spiking response of cortical neurons to sensory input [4547]. Variation in cognitive function between individuals could result from genetic variation in the structure of muscarinic receptor proteins. For example, it has been shown that a mutation in a potassium channel regulated by muscarinic receptors (KCNQ) can cause an inherited epilepsy syndrome [48].

Cholinergic modulation also regulates the induction and expression of synaptic modification, as shown in studies in which acetylcholine enhances long-term potentiation [4951] and the NMDA conductances that could contribute to long-term potentiation [52]. At the same time that cholinergic modulation enhances synaptic modification, the cholinergic activation of muscarinic receptors also causes strong presynaptic inhibition of glutamatergic synaptic transmission which is stronger at excitatory feedback synapses compared to afferent input [32,33,53] and depends upon M4 muscarinic receptors [54]. This presynaptic inhibition has been shown to be stronger at excitatory intrinsic synapses within neocortex [55], in contrast to the afferent input which often undergoes nicotinic enhancement [56,57]. Modelling shows how increases of acetylcholine would cause muscarinic presynaptic inhibition of recurrent synapses that would prevent retrieval of previously stored associations from interfering with encoding of new patterns [32,58], whereas lower levels of acetylcholine could allow stronger synaptic feedback to mediate consolidation [33], consistent with the enhancement of consolidation seen with quiet rest after encoding [59,60].

In this paper, we present a model of cognitive rule learning that shows the capacity to learn a rule across a large number of stimulus examples, and to generalize this rule to previously unseen examples. The model also shows how variation in cortical parameters could alter the performance of the network to effectively learn the rule, and could thereby underlie the large individual variation seen in data from human performance on this task [5]. The variations in parameters that could influence performance include: (i) variation in the amount of excitatory noise added to the network, which influences the capacity to explore different representations (equation (2.1)); (ii) variation in the efficacy of synaptic modification based on correct output (equations (2.3) and (2.4)); and (iii) variation in the presynaptic inhibition of previously modified synapses to allow incorrect responses to update the pattern of network connectivity.

2. Methods

(a). Overview of task

The task being modelled requires behavioural go or no-go responses to easily identifiable pairs of items, as illustrated in figure 1. The correct response on each trial is determined by the rules determined by the current pair of items and the context of spatial location. On each trial, human participants are presented with two different item stimuli in the same spatial context (screen location) separated by a short memory delay. Participants make a keyboard response based on whether the presented items are paired in a ‘Go’ or ‘NoGo’ configuration. The stimulus pairings for correct ‘Go’ responses are summarized in figure 1, where the first stimuli are indicated by A,B,C,D and the second stimulus by X or Y. Feedback indicates whether the response is correct or incorrect for each response. Importantly, participants are naive to the fact that pairings are made based on spatial location. For example, in context 1, item cue A followed by cue X requires a correct Go response, but in context 2 item cue A followed by X would require a NoGo response, whereas in context 2 item cue A followed by Y would require a Go response. These form some of the context-dependent rules that participants must learn. Participants must learn the correct behavioural responses and rules for each pair of items using the feedback provided. In the behavioural data [5], human subjects show a wide variety of capacities for learning this task. Some subjects show rapid learning over a small set of trials that indicates rapid learning and generalization of the basic rule structure of the task. Other subjects do not learn the correct rule structure at all over an extensive period of training, whereas other subjects correctly learn the rule structure at an intermediate stage during learning.

Figure 1.

Figure 1.

(a) Diagram of the behavioural task. Participants learn associations between four items A,B,C,D and two other items X,Y. The association rules change with spatial quadrant, but pairs of quadrants have the same associations (e.g. 1 and 4, 2 and 3), allowing participants to extract a common higher order rule. For example, item A followed by item X in quadrant 1 requires a Go response to be correct. There are 16 Go pairings and 16 NoGo pairings. (b) Task with probe stimuli testing generalization. The task is the same as A, except that in each quadrant, about half of the cue combinations are not shown during training (indicated by the question mark ‘?’). All examples of A and C are shown, but only a single example for B and D are shown. After training, simulations are presented with the missing cue and have to infer the correct paired associate (X, Y) from experience in the equivalent quadrant.

(b). Overview of model

To understand how these rules could be learned and generalized, and to understand the range across human subjects in capacity for learning the rules, the performance of the task has been modelled within a neural network structure that uses gating of weights. Gating of weights refers to a subset of active units in the model that determine the nature of synaptic interactions between other units in the model.

The basic function of the model is similar to most neural network models, in which the activity ai(t) of neuron population 1 at time t influences the activity aj(t + 1) of neuron population 2 at time t + 1. The activity is influenced via a set of synaptic weights with values represented by the matrix Wij at time t, as follows:

(b). 2.1

Note that this activation equation also includes an added noise term where μ is 0.1 and σ is the rand function from MATLAB which produces random numbers drawn from a uniform distribution between 0 and 1. This noise gives the network flexibility for exploring different activity patterns. The full activity vector summarized in figure 2a contains input units (indices 1–10) responding to the first item (A,B,C,D), the second item (X,Y) and the location (1,2,3,4). The vector also includes gating units (with indices 11–17 in most simulations) and two output units representing the Go or NoGo response (with indices 18–19).

Figure 2.

Figure 2.

Network structure. (a) The activity array used in simulations. Units are assigned to the first items A,B,C,D, second items X,Y, locations 1,2,3,4 and gates 1,2,3,4,5,6,7 and output Go, NoGo. (b) The network connectivity of gating in the model. Gate 1 regulates gated weights Gij1 between first patterns A,B,C or D and gates 2,3,4,5. Black squares indicate strengthened connections in a specific example of function also shown in figure 3. Gates 2,3,4,5 then regulate gated weights Gij2, Gij3, Gij4, Gij5 between second patterns X or Y and gates 6 or 7. Gates 6 or 7 then regulate gated weights Gij6 and Gij7 between location inputs 1,2,3 or 4 and outputs Go (G) or NoGo (N). (c) The same network connectivity as in b, but shown with circles representing units and lines (instead of matrices) representing weights. Filled circles represent active units for the examples described in the main text. (d) Example of a network without gating units or gated weights. Synapses between neurons representing patterns are strengthening with Hebbian modification only. This results in overlap for the association of pattern A with pattern X in location 1 versus pattern C with pattern X in location 1. The need for two different outputs cannot be accommodated in this structure.

Figure 2b,c summarizes the overall network connectivity. The simulations use three functional sets of weights that link the input arriving at different time points in the task to different sets of gating units. The gating units then regulate gating matrices that determine the currently functional set of connections within the network. Each trial starts with gate 1 active (gate 1 has index 11 within the overall activity vector because the vector includes 10 input neurons with indices 1–10). Gate 1 regulates the first set of gating weights that link input from neurons 1–4 that respond to the first item (A,B,C or D in figure 1) and send output to update activity in another set of gating units (numbered 2–5 in figure 2, and corresponding to indices 12–15 in the overall vector). Gates 2–5 regulate the second set of weights that link inputs from neurons 5–6 that respond to the second item (X or Y in figures 1 and 2) and send output to update activity in a second set of gates (numbered 6–7 in the figure, and corresponding to indices 16–17 in the overall vector). The third set of weights links input from neurons 7–10 that represent the spatial location 1,2,3,4 of the input (upper left, upper right, lower left, lower right) and send output to neurons representing the selection of a Go or NoGo output. For ease of programming, these three sets of functional weights were all implemented with single gating matrices in which the functional weights are just subsets of all possible connections between elements of the activity vector. Thus, in figure 3, the grey scale plots of weight matrices show all possible connections between units 1–19 and the same set of units 1–19, but only a subset of weights are modified during the performance of the task.

Figure 3.

Figure 3.

Simulation of task after training. (a) On each trial, the full network was presented with a sequence of binary patterns with active units (black rectangles) representing one item from A,B,C,D (top row, ‘C’) followed by one item from X,Y (second row, ‘X’) followed by one of 4 location cues (third row, ‘Loc3’). Each trial started with activity in gate 1 (index 11), and after learning cue C activates gate 5 (index 15), cue X activates gate 7 (index 17), and Loc3 activates the Go response via weights shown in part b. These network weights were previously strengthened when the network was rewarded for a correct Go response or NoGo based on the combinations shown in figure 1. The network learned to correctly respond to all combinations. (b) After learning, gate 1 has a weight matrix Gij1 with four strong weights (black squares) linking the four first cues to specific gates, so input ‘C’ in column 3 activates output row corresponding to gate 5 (index 15). The grey scale values represent the random initial weights that were not modified due to the absence of pre- or post-synaptic activity. Darker grey represents stronger weights. (c) The next four gates link the second cue to a different set of gates. In this example, active gate 5 (index 15) allows input X to activate gate 7. (d) The next two gates link the location context to a correct response. In this example, active gate 7 (index 17) allows Loc3 to activate Go. (e) Left: Performance of a single network over 3000 training trials, showing relatively rapid increase in performance from random (50%) to 100% correct around trial 800. Right: Performance change averaged over 20 networks, showing that all networks converge over time to 100% correct responses on all 32 trial types.

(c). Gating of weights

The regulation of the weights W in this model is somewhat different from most neural network models. Most neural network models update a single matrix W based on Hebbian learning rules or backpropagation of error. By contrast, the essential novel feature of this model is the gating of the spread of activity in the network. This corresponds to an extensive connectivity pattern W that is selectively gated by individual gating units that activate gating matrices G that regulate the influence of component synapses in W, potentially through multiplicative interactions of the gating neuron input with adjacent synapses in W. Neurophysiologically, this could involve nonlinear interactions between synapses on a dendritic tree such that one synapse only has a significant effect on spiking when accompanied by input at a nearby synapse on the same dendritic tree. This was modelled in previous studies in the Hasselmo laboratory [61], based on earlier models of nonlinear dendritic interactions involving NMDA receptors [19,20].

This property is here represented by allowing the pattern of synaptic weights to be directly regulated by the effect of activity in a set of gating units that are part of the vector of activity. Thus, as shown in figure 2, the activity vector ai is divided up into different sets of units ain (input units), aout (output units) and ag (gating units). Note that for simulations, all of these separate components are combined in a single vector, so the indices in, out and g all fall within the larger vector index i in equation (2.1).

The gating of weights is represented by a gating matrix of weights Gijg that map activity of the gating units ag to weights Wij as follows:

(c). 2.2

At each time step, the active gating units ag determine how each gating matrix Gijg contributes to the pattern of weights Wij within the network. In this notation, the subscripted index ijg indicates the indices of the interaction from an individual active gating unit ag with index g to the gated weights ij associated with that index. As noted above, the units ag used to influence gating are a separate set of gating units within the overall population of units. The notation with square brackets and the plus sign indicates a threshold linear function with output of zero if the function within the square brackets is negative (i.e. when ag(t) falls below threshold θ) and linear when ag(t) is above threshold θ. The maintenance of the activity of gating units is proposed to involve muscarinic activation of the mechanisms of intrinsic persistent spiking within individual neurons [42,62].

The effect of gating units on the matrix Wij is rapid in equation (2.2), but the matrix gating weights Gijg change slowly over time depending on learning as described below. The learning occurs initially due to noise causing random activity of units within the network that randomly generates correct responses, which result in the modification of the gating weights. Learning of the rule progresses to asymptotic performance, during which the gating depends upon the updating or control of W at each time step by a subset of active units within a(t) that act on W via weights Gijg. The essential learning within the network concerns the modification of weights Gijg so that the appropriate sets of units get activated to implement the appropriate gating of interactions between specific roles within the rule and the associated fillers such as the stimuli. For example, a specific active gating unit would activate the gating weights that link a specific stimulus location (e.g. 2′) to a specific output (Go).

(d). Updating of internal weights

The update of the gating weights G depends upon the generation of correct responses in the network, which initially occurs due to noise causing random selection of activity of gating units. If a correct response is generated randomly, the gating weights are updated. The updating based on correct responses is proposed to involve the release of a neuromodulator such as dopamine or acetylcholine, which has been shown to enhance mechanisms of long-term potentiation [63].

The gating weight matrices Gijg all start out with random initial weights. Sections 1 and 2 below describe the spread of activity as described in equations (2.1) and (2.2) above, and then sections 3–5 describe how the activity results in the modification of weights G in equations (2.3)–(2.5) below.

(1) As shown in equation (2.1) above, presynaptic activity spreads across the current synaptic weight matrix Wij to influence postsynaptic activity with added noise. Because of the restrictions on connectivity described above, the postsynaptic activity on each time step is distributed across the gating units. The network then selects the gating unit with the highest activity (winner-take-all).

(2) As shown in equation (2.2) above, on a given trial, the current weight matrix Wij depends upon the sum of the gating matrices Gijg that are associated with currently active gating units g. These are summarized in figure 2. To provide an initial gate, the gating unit 1 (index 11) is set to be active at the start of each trial. In the simulations shown here, the weight matrix for gate 1 is restricted to only allow connections from units 1–4 representing cues A,B,C,D to the four gates 2,3,4,5 (indices 12–15). Gates 2–5 are restricted to regulate weights between input units with indices 5, 6 representing cues X,Y to the two gates 6,7 (indices 16–17). The gates 6 and 7 are restricted to regulate weights between units with indices 7–10 representing the four quadrant locations (Location 1: upper left, 2: upper right, 3: lower left, 4: lower right), with postsynaptic targets being the output units (indices 18,19) representing the Go or NoGo responses.

(3) The network then computes Hebbian learning based on the association of the currently active presynaptic input unit aj(t) and the currently active postsynaptic gating unit selected by winner-take-all ag(t) (or output unit representing Go or NoGo). Note that the change of Gijg(t)(t) is associated with the index of the active gating unit g(t) determined by the current time step t as seen in the index. This Hebbian learning is not immediately implemented, but is stored as a possible change in the manner described as synaptic tagging [64,65].

(d). 2.3

(4) The strengthening of tagged synapses is determined by the match of network output to desired output. The network output activity (Go or NoGo) in the final layer of the network is determined either by the gated weights or by random input. This output activity ao(t) is then compared with the desired output activity o(t) at that final stage. The desired output o(t) is determined by the reward structure of the task, so in the desired output vector o(t), the Go or NoGo unit is set to one as defined by the behavioural rule shown in figure 1. If the network is correct, then the release of a neuromodulator (possibly dopamine) updates the gates G based on the Hebbian learning stored previously. Note that this process updates the weight matrices for each of the gating units for the three steps of update. Thus, the gating unit active on the first time step (gate 1, index 11) updates based on the synaptic tag from that time step (t = 1) and so on.

(d). 2.4

(5) In addition to the synaptic modification, there is also postsynaptic normalization of the weights so that the sum of the row of weights entering a particular postsynaptic gate or output does not change. This uses the following equation:

(d). 2.5

This normalization ensures that weights do not grow explosively and also causes heterosynaptic depression of other weights that are not associated with the current postsynaptic activity.

(6) Because the postsynaptic activity is being chosen randomly, the network has the potential to get stuck in incorrect response structures, even though the neuromodulatory updating of weights depends upon correct output. Because of this, the network has an additional feature that incorrect responses are associated with a small probability of resetting the gating matrices associated with active gates on that trial. On incorrect trials if a uniformly distributed random number is r < 0.2, then the matrices Gijg of all active gating units ag(t) are set to random initial connectivity. This is motivated by the idea that an incorrect response causes release of acetylcholine. The acetylcholine would then cause a diffuse presynaptic inhibition that could reduce synaptic transmission at all previously strengthened synapses (specifically influencing the gating matrices associated with currently active gating units) and thereby allow a change in postsynaptic activity and an alternation in the pattern of connectivity.

3. Results

The network model described in the Methods section (figure 2) can be successfully trained on the behavioural task described in the same section (figure 1). As shown in figure 3, after training the network can generate the correct Go or NoGo response to each of the 32 pattern combinations. In addition, as shown in figure 4, the network can also generalize to previously unseen pattern combinations that fit the same rule learned by the network. The subsequent sections of the results below will describe the detailed activity and gating weights associated with correct function of the network.

Figure 4.

Figure 4.

More examples of neural networks trained on task. (a) A different example, after different training. In this case, gate 1 (index 11) links input cue ‘A’ to gate 3 (index 13). Gate 3 links cue ‘Y’ to gate 7 (index 17). Gate 7 links ‘Loc2’ to Go response which is correct. (b) Performance of the network for different levels of noise (0.0 to 1.0) when trained on all 32 combinations. The network does not perform well with zero noise, because it does not explore enough gating patterns, but does perform well for noise of 0.1 and 0.2. Higher levels of noise cause a decrease in correct steady state responding. (c) Example of generalization from average across 20 networks, showing that when presented with only 18 of the 32 combinations, the networks attain about 75% performance. Then on the last 400 trials, most of the networks (19 out of 20) can correctly generalize to 100% correct responses when presented with all 32 pattern combinations (without further learning). (d) With noise set to zero, the average performance of 20 networks does not learn the rules and remains at 50% performance and does not generalize, indicating sensitivity to specific network parameters. (e) Performance of the network for different numbers of gating units, showing that insufficient numbers prevent good performance. Best performance is attained first with seven units, which allows four gating units to be activated by the four input items A,B,C,D. (f) With purely Hebbian connectivity without gating units and gated weights, a network cannot learn to attain better than 50% performance, due to incompatible responses that must be routed through the same neuron.

Figure 2 illustrates how the gating units provide a mechanism for learning and applying the rule across all 32 pattern combinations. Figure 2b illustrates the pattern of weights that were learned in one instance of training. These patterns can be used to trace the generation of correct responses for the example of this specific set of learned gating weights, as summarized in figure 2c. The first gate 1 is always active at the start of the trial, and has weight Gij1 in figure 2c(i). In this example, this gating matrix allows input item A to activate gate 3. Then gating matrix Gij3 allows input X to activate gate 6. Then gating matrix Gij6 allows input location 1 to activate output G (Go). Figure 2c(ii) shows a different set of inputs, illustrating that gate 1 allows input item C to activate gate 5. Gate 5 allows input X to activate gate 7 (instead of 6). This then allows input location 1 to activate gate N (NoGo). So the gating matrices allow different initial items (A versus C) to activate nonoverlapping representations to generate distinct output responses, even though the intervening stimuli are the same (X and location 1).

By contrast, figure 2d illustrates that a purely Hebbian network without gating cannot learn these distinct context-dependent associations. The Hebbian network can learn associations between input pattern A to pattern X, from pattern X to location 1 and from location 1 to output G. However, when input pattern C is presented, the overlapping components (X and location 1) prevent location 1 from being associated with an unambiguous correct response. During learning, location 1 gets associated with both Go and NoGo responses, and the network will not be able to attain better than 50% correct responses. The performance of the simulation for the Hebbian case is shown in figure 4f, and the performance across different numbers of gating units is shown in figure 4e (fewer gates result in poor performance).

Figure 3 shows a specific example trial after training on all 32 pattern combinations, as well as showing the summary of performance over trials. In figure 3a, the activity of the network is shown for each time step, with each row showing the vector of activity for a single time step. At the start of the example trial shown here, the network receives an input representing the pattern ‘C’ on the first time step (each trial starts with one of the four cues A,B,C,D). In addition, at the start of all trials, the network starts with gate 1 active at the start of the trial (seen as ‘gate 1’ in the figure). Gate 1 activates the gating matrix shown in figure 3b which becomes the connectivity Wij of the network. As can be seen in this gating matrix in figure 3b, the previous learning strengthened connections that link each of the input cues A,B,C,D with one of the gating units. When this matrix is multiplied by the matrix containing the cue ‘C’, it generates the output that activates gate 5. Note that input ‘C’ corresponds to column 3, which has a strengthened weight (black square) at column 3 and row 15 (gate 5), so that the activity spreads from input 3 to gate 5.

Now that gate 5 is active, it can be seen on row 2 of figure 3a, and it activates the gating matrix shown in figure 3c (lower right). In addition, as seen in row 2 of figure 3a, on time step 2 the network receives the input representing cue ‘X’. This input cue then spreads across the element of the weight matrix. As can be seen in the weight matrix in figure 3c (gate 5), the gating matrix contains a strengthened synapse that will gate the input from column 5 (representing the cue ‘X’) to output row 17 (activating postsynaptic gating unit 7).

Now that gate 7 is active, this can be seen on row 3 (time step 3) of figure 3a. This activates the gating matrix shown in figure 3d (right). In addition, as seen in row 3 of figure 3a, on time step 3, the network receives the input representing the spatial location in the lower left (labelled as ‘Loc3’ on row 3). This input activity spreads across the weight matrix shown for gate 7 in figure 3d. The input at column 9 spreads across the strengthened weight in column 9 row 18 (black square) to cause postsynaptic activity in the output unit representing a ‘Go’ response. As can be seen, this corresponds to the correct desired output for this combination of input cues and location. Thus, for this particular example trial, the response is correct. This network is able to also generate the correct ‘Go’ responses for the full set of 16 combinations requiring a ‘Go’ response, as summarized in figure 1. In addition, the network generates the correct NoGo responses for the full set of 16 combinations requiring a NoGo response, as summarized in figure 1.

As shown in figure 3e, during training a single network (left) initially performs at around 50% (due to random selection of Go or NoGo responses), but then at one point relatively rapidly increases to 100% performance. This is consistent with the rapid improvement of performance on the task when some individuals discover the rule [5]. Consistent with the data, many of the example networks achieve 100% more rapidly, but this example was selected to show that performance can increase at a later stage. The network activates random gates that form associations between the relevant inputs and gating outputs at each time step. In the case shown here, the initial randomly selected gates do not give good performance and performance only improves when the correct output causes strengthening of weights in the gating matrices to select specific gates.

The plot on the right side of figure 3e shows the performance averaged over 20 different networks with different initial conditions. Each of the individual networks achieves good performance relatively rapidly at different times, but the average shows a smooth approach to 100% performance. The full set of 32 combinations is presented during training, and every single network increases to 100% performance over the 3000 training trials. The noise component of the activation rule is turned off for the last 400 trials, so that the incorrect responses generated by the noise no longer perturb the performance, causing a smoother line at the end. If noise is kept large throughout the simulations, then the noise will perturb the performance as shown in the plot in figure 4b, which shows performance over the last 100 trials for different levels of noise when all patterns are trained. Note that as noise increases, the steady state of the network decreases. However, the network cannot function well with the absence of noise, because noise assists with the selection of different gating units that allow the random generation of correct responses that allow effective learning of the full set of rules. Thus, when noise is set to 0, the network does not attain good performance as shown on the left side of the plot in figure 4b.

As shown in figure 4, the network can effectively generalize to unseen patterns. Figure 4a shows an example of a sequence of patterns that was not seen during training, but generates a correct response due to generalization. Figure 4c shows the performance of networks tested for generalization, using the set of patterns summarized in figure 1b. In this test, the network has seen all examples for pattern A and pattern C, but is only shown a single example of pattern B and pattern D. Because the network received only about 50% of the input patterns during training on time step 1 to 4600, its initial performance during this period is around 75% because it is only being presented with about 50% of the patterns, and then it guesses correctly on 50% of the remaining trials, but it is being evaluated relative to the full set of 32 patterns. Then once the full set of 32 patterns is presented during the final 400 trials, it demonstrates effective generalization to respond correctly to all of the patterns. Figure 4c shows the average performance across 20 different networks. Nineteen out of 20 networks correctly generalize to all 32 patterns at the end of training, even though they were only trained on 18 pattern combinations (as shown in figure 1b). Thus, the network can generalize to 14 previously unseen stimuli based on the rule that it has extracted. Effective generalization was also seen with many other configurations of learned stimuli and stimuli not seen during learning.

For correct performance, the network depends upon a number of specific parameters. In particular, if the noise term μ is set to zero, the network does not explore enough of the activation space and does not effectively create enough new gating units and therefore cannot effectively learn the rules. In this case, activity stays at about 50% as shown in the average across 20 networks in figure 4d. The importance of the potential modulatory reset of weights was shown in other simulations that are not illustrated. When the code for the network was modified to remove the capacity for reset of weights, then performance stayed at about 80% during training because there were more cases of failure to learn the full rule.

The network needed gating units to perform effectively, as already illustrated schematically in figure 2d. Simulations shown in figure 4f demonstrate that when there was only Hebbian modification of connections between layers responding to sensory input, then the network could not perform better than about 50% correct. The performance of the network with different numbers of gating units is shown in figure 4e for numbers ranging from 4 gating units (one for each set of connections in figure 2b) to 19 gating units. The number of gating units was specifically varied for the gating units activated by the first input items A,B,C,D. Note that the best performance first appears for seven gating units, which is the number of gating units shown in figure 2b,c. This allows four gating units to be activated by each of the four input items A,B,C,D. As the number is increased, the performance remains high, as a subset of four gating units for the first item input are selected by each network. However, increasing the number of second input gating units beyond two causes a decrease in performance.

4. Discussion

The simulations presented here provide a general framework for the role of neural gating in the learning of rules shown in figure 1. As shown in figures 2 and 3, the neural network model can learn complex rules and generate correct responses based on this rule across a large number of combinations of patterns. As shown in figure 4, the network can perform generalization based on learning of about half of the training set to generate correct answers for the other half of the learned combinations. More complex combinations of gating could allow learning of more sophisticated hierarchical rule structures.

The network model also indicates how differences in neural circuit parameters could influence the cognitive capabilities of different individuals, consistent with the wide variety of learning capabilities demonstrated in behavioural tests using the training set illustrated in figure 1. Experimental data shows that when trained on this set of stimuli, human subjects vary from immediate learning and application of the rule, to delayed acquisition of the rule, to complete inability to implement the rule across the full training set [5]. As shown in figures 3 and 4, the network presented here shows a similar diversity of behaviour that can arise from a number of different sources. One source is the level of noise in equation (2.1). The noise is necessary for the network to explore a range of different gating representations through noisy postsynaptic activation that is necessary to set up distinct rules. With noise set at 0.1, the network can achieve 100% performance (figure 3) and generalize the rule to new patterns (figure 4). With noise set at 0, the network does not learn and only achieves 75% performance due to random guessing of responses (figure 4e). Even when parameters are kept at the same value, the influence of noise results in different runs of the same network achieving correct performance at different training times. This illustrates how a relatively simple parameter of neural function could influence behaviour on a broad scale. The noise is necessary for the network to learn, and could be considered as a component of arousal allowing exploration of different internal representations [25].

Another aspect of performance depends on the ability of the network to respond to incorrect responses by weakening the influence of previously modified weights, to allow the activation of new alternative representations. This could correspond to the presynaptic inhibition of synaptic transmission in cortical structures by acetylcholine [32,33] or norepinephrine [55,66] that might be activated during pulses of cholinergic or noradrenergic activity evoked by errors during performance; or it could involve presynaptic inhibition by GABAB receptors [67,68], or metabotropic glutamate receptors [69]. Another component that is important to the function of the network is the modulatory expression of synaptic modification at tagged synapses dependent upon the reward associated with a correct response. This could correspond to the cholinergic enhancement of long-term potentiation shown in a number of studies. Thus, the model presented in this paper demonstrates that the diversity of cognitive rule learning function across individuals could arise from variation in the impact of different neuromodulators across individuals.

Currently, we do not have an accepted scientific framework for modelling how flexible symbolic processing allows generalization of performance within cognitive tasks. If the neural mechanisms for human rule learning can be modelled computationally in terms of basic elements of cognitive function that are combined into higher order cognitive functions, this could then form the basis of a systematic taxonomy of cognitive function. As a distantly related analogy, one could think of the breakthrough in chemistry and physics when chemical elements were described and shown to systematically combine with different stoichiometries into different chemical compounds. Cognitive science does not have such a systematic framework for adding to knowledge of the elements of cognitive function, but the interaction of individual cognitive gates could involve different internal processing functions that could be characterized as different elements of cognitive function. As a simple first step, the analysis of learning and performance patterns in models such as the one presented here could indicate the behavioural manifestation of differences in parameters of cognitive function (for example, the different properties described here as noise in response selection and as weakening of previously modified weights). The manifestation of these behavioural differences in rule learning and generalization could allow analysis of human cognitive behaviour in terms of the potential parameters of neural circuit elements.

Acknowledgements

We appreciate the work of Allen Chang on behavioural task development.

Data accessibility

This article has no additional data.

Competing interests

The authors have no conflicts of interest.

Funding

This work supported by the National Institutes of Health, grant numbers R01 MH60013, R01 MH61492 and by the US Office of Naval Research MURI award number N00014-16-1-2832.

References

  • 1.Fodor JA, Pylyshyn ZW. 1988. Connectionism and cognitive architecture: a critical analysis. Cognition 8, 3–71. ( 10.1016/0010-0277(88)90031-5) [DOI] [PubMed] [Google Scholar]
  • 2.Raven J. 2000. The Raven's progressive matrices: change and stability over culture and time. Cogn. Psychol. 41, 1–48. ( 10.1006/cogp.1999.0735) [DOI] [PubMed] [Google Scholar]
  • 3.Carpenter PA, Just MA, Shell P. 1990. What one intelligence test measures: a theoretical account of the processing in the Raven progressive matrices test. Psychol. Rev. 97, 404–431. ( 10.1037/0033-295X.97.3.404) [DOI] [PubMed] [Google Scholar]
  • 4.Rasmussen D, Eliasmith C. 2011. A neural model of rule generation in inductive reasoning. Topic. Cogn. Sci. 3, 140–153. ( 10.1111/j.1756-8765.2010.01127.x) [DOI] [PubMed] [Google Scholar]
  • 5.Chang AE, Ren Y, Whiteman AS, Stern CE.. 2017. Evaluating the relationship between activity in frontostriatal regions and successful and unsuccessful performance of a context-dependent rule learning task. Program no. 619.19, Society for Neuroscience Meeting Planner. San Diego, CA: Society for Neuroscience. Online.
  • 6.Miller EK, Cohen JD. 2001. An integrative theory of prefrontal cortex function. Ann. Rev. Neurosci. 24, 167–202. ( 10.1146/annurev.neuro.24.1.167) [DOI] [PubMed] [Google Scholar]
  • 7.Wallis JD, Anderson KC, Miller EK. 2001. Single neurons in prefrontal cortex encode abstract rules. Nature 411, 953–956. ( 10.1038/35082081) [DOI] [PubMed] [Google Scholar]
  • 8.Hummel JE, Biederman I. 1992. Dynamic binding in a neural network for shape recognition. Psychol. Rev. 99, 480–517. ( 10.1037/0033-295X.99.3.480) [DOI] [PubMed] [Google Scholar]
  • 9.von der Malsburg C. 1995. Binding in models of perception and brain function. Curr. Opin. Neurobiol. 5, 520–526. ( 10.1016/0959-4388(95)80014-X) [DOI] [PubMed] [Google Scholar]
  • 10.Siegel M, Warden MR, Miller EK. 2009. Phase-dependent neuronal coding of objects in short-term memory. Proc. Natl Acad. Sci. USA 106, 21 341–21 346. ( 10.1073/pnas.0908193106) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.Eliasmith C, Stewart TC, Choo X, Bekolay T, DeWolf T, Tang Y. 2012. A large-scale model of the functioning brain. Science 338, 1202–1205. ( 10.1126/science.1225266) [DOI] [PubMed] [Google Scholar]
  • 12.Eliasmith C. 2013. How to build a brain: a neural architecture for biological cognition. New York, NY: Oxford University Press. [Google Scholar]
  • 13.Kriete T, Noelle DC, Cohen JD, O'Reilly RC. 2013. Indirection and symbol-like processing in the prefrontal cortex and basal ganglia. Proc. Natl Acad. Sci. USA 110, 16 390–16 395. ( 10.1073/pnas.1303547110) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.O'Reilly RC, Frank MJ. 2006. Making working memory work: a computational model of learning in the prefrontal cortex and basal ganglia. Neural Comput. 18, 283–328. ( 10.1162/089976606775093909) [DOI] [PubMed] [Google Scholar]
  • 15.Hasselmo ME. 2005. A model of prefrontal cortical mechanisms for goal-directed behavior. J. Cogn. Neurosci. 17, 1115–1129. ( 10.1162/0898929054475190) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Koene RA, Hasselmo ME. 2005. An integrate-and-fire model of prefrontal cortex neuronal activity during performance of goal-directed decision making. Cereb. Cortex 15, 1964–1981. ( 10.1093/cercor/bhi072) [DOI] [PubMed] [Google Scholar]
  • 17.Hasselmo ME, Eichenbaum H. 2005. Hippocampal mechanisms for the context-dependent retrieval of episodes. Neural Netw. 15, 689–707. ( 10.1016/S0893-6080(02)00057-6) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Zilli EA, Hasselmo ME. 2008. Modeling the role of working memory and episodic memory in behavioral tasks. Hippocampus 18, 193–209. ( 10.1002/hipo.20382) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Mel BW. 1993. Synaptic integration in an excitable dendritic tree. J. Neurophysiol. 70, 1086–1101. ( 10.1152/jn.1993.70.3.1086) [DOI] [PubMed] [Google Scholar]
  • 20.Poirazi P, Brannon T, Mel BW. 2003. Arithmetic of subthreshold synaptic summation in a model CA1 pyramidal cell. Neuron 37, 977–987. ( 10.1016/S0896-6273(03)00148-X) [DOI] [PubMed] [Google Scholar]
  • 21.Chen TW, et al. 2013. Ultrasensitive fluorescent proteins for imaging neuronal activity. Nature 499, 295–300. ( 10.1038/nature12354) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22.Chen X, Leischner U, Rochefort NL, Nelken I, Konnerth A. 2011. Functional mapping of single spines in cortical neurons in vivo. Nature 475, 501–505. ( 10.1038/nature10193) [DOI] [PubMed] [Google Scholar]
  • 23.Klausberger T, Magill PJ, Marton LF, Roberts JD, Cobden PM, Buzsaki G, Somogyi P. 2003. Brain-state- and cell-type-specific firing of hippocampal interneurons in vivo. Nature 421, 844–848. ( 10.1038/nature01374) [DOI] [PubMed] [Google Scholar]
  • 24.Cutsuridis V, Hasselmo M. 2012. GABAergic contributions to gating, timing, and phase precession of hippocampal neuronal activity during theta oscillations. Hippocampus 22, 1597–1621. ( 10.1002/hipo.21002) [DOI] [PubMed] [Google Scholar]
  • 25.Trofimova I, Robbins TW. 2016. Temperament and arousal systems: a new synthesis of differential psychology and functional neurochemistry. Neurosci. Biobehav. Rev. 64, 382–402. ( 10.1016/j.neubiorev.2016.03.008) [DOI] [PubMed] [Google Scholar]
  • 26.Trofimova I. 2016. The interlocking between functional aspects of activities and a neurochemical model of adult temperament. In Temperaments: individual differences, social and environmental influences and impact on quality of life (ed. Arnold MC.), pp. 77–147. New York, NY: Nova Sciences Publishers. [Google Scholar]
  • 27.Lee MR. 2007. Solanaceae IV: Atropa belladonna, deadly nightshade. J. R. Coll. Physicians Edinb. 37, 77–84. [PubMed] [Google Scholar]
  • 28.Perry EK, Perry RH. 1995. Acetylcholine and hallucinations: disease-related compared to drug-induced alterations in human consciousness. Brain Cogn. 28, 240–258. ( 10.1006/brcg.1995.1255) [DOI] [PubMed] [Google Scholar]
  • 29.Robbins TW, Semple J, Kumar R, Truman MI, Shorter J, Ferraro A, Fox B, McKay G, Matthews K. 1997. Effects of scopolamine on delayed-matching-to-sample and paired associates tests of visual memory and learning in human subjects: comparison with diazepam and implications for dementia. Psychopharmacology 134, 95–106. ( 10.1007/s002130050430) [DOI] [PubMed] [Google Scholar]
  • 30.Ghoneim MM, Mewaldt SP. 1977. Studies on human memory: the interactions of diazepam, scopolamine, and physostigmine. Psychopharmacology 52, 1–6. ( 10.1007/BF00426592) [DOI] [PubMed] [Google Scholar]
  • 31.Atri A, et al. 2004. Blockade of central cholinergic receptors impairs new learning and increases proactive interference in a word paired-associate memory task. Behav. Neurosci. 118, 223–236. ( 10.1037/0735-7044.118.1.223) [DOI] [PubMed] [Google Scholar]
  • 32.Hasselmo ME. 2006. The role of acetylcholine in learning and memory. Curr. Opin. Neurobiol. 16, 710–715. ( 10.1016/j.conb.2006.09.002) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33.Hasselmo ME. 1999. Neuromodulation: acetylcholine and memory consolidation. Trends Cogn. Sci. 3, 351–359. ( 10.1016/S1364-6613(99)01365-0) [DOI] [PubMed] [Google Scholar]
  • 34.Marrosu F, Portas C, Mascia MS, Casu MA, Fà M, Giagheddu M, Imperato A, Gessa GL. 1995. Microdialysis measurement of cortical and hippocampal acetylcholine release during sleep-wake cycle in freely moving cats. Brain Res. 671, 329–332. ( 10.1016/0006-8993(94)01399-3) [DOI] [PubMed] [Google Scholar]
  • 35.Parikh V, Kozak R, Martinez V, Sarter M. 2007. Prefrontal acetylcholine release controls cue detection on multiple timescales. Neuron 56, 141–154. ( 10.1016/j.neuron.2007.08.025) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 36.Cole AE, Nicoll RA. 1984. Characterization of a slow cholinergic postsynaptic potential recorded in vitro from rat hippocampal pyramidal cells. J. Physiol. 352, 173–188. ( 10.1113/jphysiol.1984.sp015285) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 37.Madison DV, Nicoll RA. 1984. Control of the repetitive discharge of rat CA 1 pyramidal neurones in vitro. J. Physiol. 354, 319–331. ( 10.1113/jphysiol.1984.sp015378) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 38.Yoshida M, Jochems A, Hasselmo ME. 2013. Comparison of properties of medial entorhinal cortex layer II neurons in two anatomical dimensions with and without cholinergic activation. PLoS ONE 8, e73904 ( 10.1371/journal.pone.0073904) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 39.Tang AC, Bartels AM, Sejnowski TJ. 1997. Effects of cholinergic modulation on responses of neocortical neurons to fluctuating input. Cereb. Cortex 7, 502–509. ( 10.1093/cercor/7.6.502) [DOI] [PubMed] [Google Scholar]
  • 40.Knauer B, Jochems A, Valero-Aracama MJ, Yoshida M. 2013. Long-lasting intrinsic persistent firing in rat CA1 pyramidal cells: a possible mechanism for active maintenance of memory. Hippocampus 23, 820–831. ( 10.1002/hipo.22136) [DOI] [PubMed] [Google Scholar]
  • 41.Jochems A, Yoshida M. 2013. Persistent firing supported by an intrinsic cellular mechanism in hippocampal CA3 pyramidal cells. Eur. J. Neurosci. 38, 2250–2259. ( 10.1111/ejn.12236) [DOI] [PubMed] [Google Scholar]
  • 42.Yoshida M, Hasselmo ME. 2009. Persistent firing supported by an intrinsic cellular mechanism in a component of the head direction system. J. Neurosci. 29, 4945–4952. ( 10.1523/JNEUROSCI.5154-08.2009) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 43.Alonso A, Klink R. 1993. Differential electroresponsiveness of stellate and pyramidal-like cells of medial entorhinal cortex layer II. J. Neurophysiol. 70, 128–143. ( 10.1152/jn.1993.70.1.128) [DOI] [PubMed] [Google Scholar]
  • 44.Hasselmo ME, Stern CE. 2006. Mechanisms underlying working memory for novel information. Trends Cogn. Sci. 10, 487–493. ( 10.1016/j.tics.2006.09.005) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 45.Avery MC, Dutt N, Krichmar JL. 2014. Mechanisms underlying the basal forebrain enhancement of top-down and bottom-up attention. Eur. J. Neurosci. 39, 852–865. ( 10.1111/ejn.12433) [DOI] [PubMed] [Google Scholar]
  • 46.Herrero JL, Roberts MJ, Delicato LS, Gieselmann MA, Dayan P, Thiele A. 2008. Acetylcholine contributes through muscarinic receptors to attentional modulation in V1. Nature 454, 1110–1114. ( 10.1038/nature07141) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 47.Goard M, Dan Y. 2009. Basal forebrain activation enhances cortical coding of natural scenes. Nat. Neurosci. 12, 1444–1449. ( 10.1038/nn.2402) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 48.Peters HC, Hu H, Pongs O, Storm JF, Isbrandt D. 2005. Conditional transgenic suppression of M channels in mouse brain reveals functions in neuronal excitability, resonance and behavior. Nat. Neurosci. 8, 51–60. ( 10.1038/nn1375) [DOI] [PubMed] [Google Scholar]
  • 49.Blitzer RD, Gil O, Landau EM. 1990. Cholinergic stimulation enhances long-term potentiation in the CA1 region of rat hippocampus. Neurosci. Lett. 119, 207–210. ( 10.1016/0304-3940(90)90835-W) [DOI] [PubMed] [Google Scholar]
  • 50.Adams SV, Winterer J, Muller W. 2004. Muscarinic signaling is required for spike-pairing induction of long-term potentiation at rat Schaffer collateral-CA1 synapses. Hippocampus 14, 413–416. ( 10.1002/hipo.10197) [DOI] [PubMed] [Google Scholar]
  • 51.Brocher S, Artola A, Singer W. 1992. Agonists of cholinergic and noradrenergic receptors facilitate synergistically the induction of long-term potentiation in slices of rat visual cortex. Brain Res. 573, 27–36. ( 10.1016/0006-8993(92)90110-U) [DOI] [PubMed] [Google Scholar]
  • 52.Markram H, Segal M. 1990. Acetylcholine potentiates responses to N-methyl-d-aspartate in the rat hippocampus. Neurosci. Lett. 113, 62–65. ( 10.1016/0304-3940(90)90495-U) [DOI] [PubMed] [Google Scholar]
  • 53.Hasselmo ME, Schnell E, Barkai E. 1995. Dynamics of learning and recall at excitatory recurrent synapses and cholinergic modulation in rat hippocampal region CA3. J. Neurosci. 15, 5249–5262. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 54.Dasari S, Gulledge AT. 2011. M1 and M4 receptors modulate hippocampal pyramidal neurons. J. Neurophysiol. 105, 779–792. ( 10.1152/jn.00686.2010) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 55.Hasselmo ME, Cekic M. 1996. Suppression of synaptic transmission may allow combination of associative feedback and self-organizing feedforward connections in the neocortex. Behav. Brain Res. 79, 153–161. ( 10.1016/0166-4328(96)00010-1) [DOI] [PubMed] [Google Scholar]
  • 56.Gil Z, Conners BW, Amitai Y. 1997. Differential regulation of neocortical synapses by neuromodulators and activity. Neuron 19, 679–686. ( 10.1016/S0896-6273(00)80380-3) [DOI] [PubMed] [Google Scholar]
  • 57.Hsieh CY, Cruikshank SJ, Metherate R. 2000. Differential modulation of auditory thalamocortical and intracortical synaptic transmission by cholinergic agonist. Brain Res. 880, 51–64. ( 10.1016/S0006-8993(00)02766-9) [DOI] [PubMed] [Google Scholar]
  • 58.Hasselmo ME, Anderson BP, Bower JM. 1992. Cholinergic modulation of cortical associative memory function. J. Neurophysiol. 67, 1230–1246. ( 10.1152/jn.1992.67.5.1230) [DOI] [PubMed] [Google Scholar]
  • 59.Dewar M, Alber J, Cowan N, Della Sala S. 2014. Boosting long-term memory via wakeful rest: intentional rehearsal is not necessary, consolidation is sufficient. PLoS ONE 9, e109542 ( 10.1371/journal.pone.0109542) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 60.Craig M, Dewar M, Harris MA, Della Sala S, Wolbers T. 2016. Wakeful rest promotes the integration of spatial memories into accurate cognitive maps. Hippocampus 26, 185–193. ( 10.1002/hipo.22502) [DOI] [PubMed] [Google Scholar]
  • 61.Katz Y, Kath WL, Spruston N, Hasselmo ME. 2007. Coincidence detection of place and temporal context in a network model of spiking hippocampal neurons. PLoS Comput. Biol. 3, e234 ( 10.1371/journal.pcbi.0030234) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 62.Fransen E, Alonso AA, Hasselmo ME. 2002. Simulations of the role of the muscarinic-activated calcium-sensitive nonspecific cation current INCM in entorhinal neuronal activity during delayed matching tasks. J. Neurosci. 22, 1081–1097. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 63.Markram H, Segal M. 1990. Long-lasting facilitation of excitatory postsynaptic potentials in the rat hippocampus by acetylcholine. J. Physiol. 427, 381–393. ( 10.1113/jphysiol.1990.sp018177) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 64.Frey U, Morris RG. 1997. Synaptic tagging and long-term potentiation. Nature 385, 533–536. ( 10.1038/385533a0) [DOI] [PubMed] [Google Scholar]
  • 65.Frey U, Morris RG. 1998. Synaptic tagging: implications for late maintenance of hippocampal long-term potentiation. Trends Neurosci. 21, 181–188. ( 10.1016/S0166-2236(97)01189-2) [DOI] [PubMed] [Google Scholar]
  • 66.Hasselmo ME, Linster C, Patil M, Ma D, Cekic M. 1997. Noradrenergic suppression of synaptic transmission may influence cortical signal-to-noise ratio. J. Neurophysiol. 77, 3326–3339. ( 10.1152/jn.1997.77.6.3326) [DOI] [PubMed] [Google Scholar]
  • 67.Tang AC, Hasselmo ME. 1994. Selective suppression of intrinsic but not afferent fiber synaptic transmission by baclofen in the piriform (olfactory) cortex. Brain Res. 659, 75–81. ( 10.1016/0006-8993(94)90865-6) [DOI] [PubMed] [Google Scholar]
  • 68.Molyneaux BJ, Hasselmo ME. 2002. GABA(B) presynaptic inhibition has an in vivo time constant sufficiently rapid to allow modulation at theta frequency. J. Neurophysiol. 87, 1196–1205. ( 10.1152/jn.00077.2001) [DOI] [PubMed] [Google Scholar]
  • 69.Giocomo LM, Hasselmo ME. 2006. Difference in time course of modulation of synaptic transmission by group II versus group III metabotropic glutamate receptors in region CA1 of the hippocampus. Hippocampus 16, 1004–1016. ( 10.1002/hipo.20231) [DOI] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

This article has no additional data.


Articles from Philosophical Transactions of the Royal Society B: Biological Sciences are provided here courtesy of The Royal Society

RESOURCES