Summary
The cerebellum is thought to detect and correct errors between intended and executed commands1,2 and is critical for social behaviors, cognition and emotion3–6. Computations for motor control must be performed quickly to correct errors in real time and should be sensitive to small differences between patterns for fine error correction while being resilient to noise7. Influential theories of cerebellar information processing have largely assumed random network connectivity, which increases the encoding capacity of the network’s first layer8–13. However, maximizing encoding capacity reduces resiliency to noise7. To understand how neuronal circuits address this fundamental tradeoff, we mapped the feedforward connectivity in the mouse cerebellar cortex using automated large-scale transmission electron microscopy (EM) and convolutional neural network-based image segmentation. We found that both the input and output layers of the circuit exhibit redundant and selective connectivity motifs, which contrast with prevailing models. Numerical simulations suggest these redundant, non-random connectivity motifs increase resilience to noise at a negligible cost to overall encoding capacity. This work reveals how neuronal network structure can support a trade-off between encoding capacity and redundancy, unveiling principles of biological network architecture with implications for artificial neural network design.
Seminal work in the 1960s and 70s14,15 provided the basis for influential theories of cerebellar information processing8,9. At the input layer, a relatively small number of mossy fibers (MFs, input neurons) are hypothesized to be randomly sampled by a massive numbers of granule cells (GrCs, parallel fibers) resulting in expansion recoding (Fig. 1a,b). Through non-linear activation of the GrCs, expansion recoding projects MF sensory and motor information into a high dimensional representation facilitating pattern separation10–13,16 (Fig. 1c). At the output layer, each Purkinje cell (PC, readout neuron) integrates tens of thousands of GrC inputs to form appropriate associations8,9,17 and are the sole output of the cerebellar cortex (Fig. 1a,b). Association learning is hypothesized to occur through linear decoding17 and reinforcement learning rules6,18,19 and may help generate subnetworks of PCs suitable for ensemble learning20,21. This network architecture of expansion and re-convergence is also seen in the mushroom body of fly and the electrosensory lobe of electric fish13,22–24, suggesting that cerebellar architecture exemplifies a general principle of neuronal circuit organization critical for separating neuronal activity patterns before associative learning. Indeed, these input and output layers are similar to convolutional and linear output layers in artificial deep neural network architectures25 (Extended Data Fig. 1).
Information theoretical models usually assume that MF→GrC connectivity is random in order to maximize the encoding capacity of the input layer of the network8–11,13. However, maximizing encoding capacity makes networks less resilient to noise7. To understand how the cerebellar cortex balances the tradeoff between encoding capacity and resilience, we mapped the precise synaptic connectivity across both input (MF→GrC) and output (GrC→PC) layers of the feedforward circuit in the cerebellar cortex.
3D EM reconstruction
We generated a synapse-resolution EM dataset that contains a proximal portion of lobule V from the adult mouse vermis (Fig. 1d, Extended Data Fig. 2a). The sample was cut into 1176 serial parasagittal sections, each ~45 nm thick and imaged at a resolution of 4 × 4 nm2 per pixel. The dataset was aligned into a 37.2 trillion voxel volume spanning ~28 million μm3 (Supplementary Video 1). To facilitate analysis, we densely segmented neurons using artificial neural networks (Fig. 1e, Extended Data Fig. 2). To process a large EM dataset efficiently, we developed a scalable framework to process the dataset in parallel (Extended Data Fig. 2c) and a proofreading platform (Extended Data Fig. 2d) for targeted neuron reconstruction (Extended Data Fig. 2e). We predicted synapse locations and their weights using a separate artificial neural network trained to identify synapses based on their ultrastructural features (Extended Data Fig. 2f and 4a,b) with high accuracy (precision: 95.4%, recall: 92.2%, f-score: 93.8%). In total, we reconstructed >5,500 neurons (Fig. 1a,e) and analyzed >150,000 synapses (116,571 MF bouton→GrC and 34,932 GrC→PC) constituting >36,000 unitary connections (13,385 MF bouton→GrC and 23,365 GrC→PC).
Structured MF→GrC connectivity
We first tested the hypothesis that feedforward connectivity in the input layer (MFs→GrCs) was random. To do so, we densely reconstructed neurons in the GrC layer and identified MFs and GrCs based on morphology15 (Fig. 2a,b, Extended Data Fig. 3a,b, Supplementary Data 1-2, Video 2). The basic properties of GrCs and MFs were consistent with prior findings11,26. GrCs had 4.38 ± 1.00 (mean ± SD) dendrites (Extended Data Fig. 3c), each of which ended in a “claw” that wrapped around and received 10.4 ± 5.6 (mean ± SD) synapses from a single MF bouton (Extended Data Fig. 4). GrC dendrites measured 21.7 ± 8.5 μm (mean ± SD) from the cell body center to the center of the presynaptic MF bouton (Extended Data Fig. 3d,e). Individual MF boutons were sampled by 14.8 ± 8.4 (mean ± SD) GrCs. Though MF collaterals give rise to multiple boutons in the GrC layer, MF boutons are typically spaced such that the likelihood of individual GrCs innervated by multiple boutons from the same MF is low11,15,27. While a subset of MF boutons in the dataset belonged to the same MF axons (n = 182 out of 784 boutons), we found that GrCs rarely received inputs from different boutons of the same MF axon (n = 20 out of 4400) in the dataset. Notably, our subsequent analyses did not significantly change when we considered MF boutons independently or combined MF boutons arising from the same reconstructed axons (data not shown), so for the remainder of our analysis we focus on MF boutons. Additionally, we found a small fraction (8.8%) of MF boutons have axon collaterals that each make one or two synapses onto dendrites and claws of GrCs; however, the vast majority of synapses were from the main MF boutons (Extended Data Fig. 5).
Recent theoretical work has assumed unstructured MF→GrC connectivity, which implies low levels of shared MF bouton inputs between GrCs12,13,28 even when considering the limited reach of their dendrites11,26. We found GrC pairs shared 2 or more MF bouton inputs more often than predicted by anatomically-constrained random connectivity11 (reconstructed: 4.43 ± 3.63, shuffled: 2.34 ± 2.08, mean ± SD; p = 3.9 × 10−12, n = 211, two-sided Wilcoxon rank-sum test, Fig. 2c,d, Extended Data Fig. 3f–h). The distribution of sharing is consistently higher across different numbers of shared inputs and synapse prediction accuracies (Extended Data Fig. 3i–m, Extended Data Fig. 6a, Extended Data Fig. 7a). Oversharing of inputs among GrCs implies an overconvergence of MF boutons (Extended Data Fig. 6b), and suggests that GrC activity is more correlated than expected, in opposition to the prediction that MF→GrC encoding minimizes population correlations to maximize encoding capacity7–9,12,13.
Furthermore, we found that MF boutons were not sampled uniformly: the most connected third of MF boutons had 4 times as many postsynaptic GrCs as the least connected third of MF boutons (Extended Data Fig. 3n). To rule out the possibility that MF boutons simply had high variability in the number of postsynaptic partners (σ2: 0.522 vs. 0.297 (reconstructed vs. random), p = 0.001, permutation test, n = 62), we quantified the overrepresentation of MF boutons by GrCs with their kurtosis and skew measuring prevalence of outliers and asymmetry of the distributions respectively and found each to be significantly greater than those of random connectivity models (kurtosis: 1.49 vs. −0.01 (reconstructed vs. random), p = 0.0146; skew: 0.963 vs 0.233 (reconstructed vs. random), p = 0.0178; permutation tests, Fig. 2e, Extended Data Fig. 3f,o, Extended Data Fig. 7b). By overrepresenting inputs from select MF boutons, these inputs are redundantly encoded by more GrCs, influencing more of the GrC coding space (Fig. 2f). Select MF boutons may be more informative or reliable, potentially enhancing noise resiliency of specific inputs. In summary, we found oversharing and overrepresentation to be two redundant wiring motifs that permit expansion coding in the input layer to be more robust to noise and may partially explain low dimensional GrC activity recently reported in vivo29–31.
Selective GrC→PC connectivity
In the output layer, given that adjacent PC dendrites share many potential GrC axon inputs (Fig. 1e, Fig. 3a), we hypothesized that there are redundant and non-random motifs within the GrC→PC connectivity that could contribute to PC synchrony32–34. Given the convergence of PCs to downstream deep cerebellar nuclei (DCN), we propose that structured GrC→PC connectivity can give rise to groups of partially redundant PCs to produce more accurate pattern associations. To test these hypotheses, we densely reconstructed the connectivity between GrC axons and PCs (Supplementary Data 2-3 and Videos 2-3). Reconstructed GrC axons were categorized as either “local” or “non-local”. Local GrC axons (n = 541) had cell bodies within the EM volume, while non-local GrC axons (n = 4,439) were parallel fibers without a cell body identified within the dataset. Dendrites of reconstructed PCs (n = 153 total, n = 28 with somas) ramified across an area of 208 ± 5.2 μm by 158 ± 19.5 μm (mean ± SD) in height and width (Supplementary Data 3), consistent with light microscopy15. A common assumption used in computational models is that all GrC-PC pairs that contact each other make synapses13,17. However, we found the connectivity rate was 49 ± 4.4% (mean ± SD) (Fig. 3a,c, Extended Data Fig. 8a), with each GrC→PC connection typically comprising one or two synapses (Extended Data Fig. 4c), consistent with prior work35–37. The lack of all-to-all GrC→PC connectivity raised the question of how selective these unitary connections are and how this selectivity constrains information processing and plasticity mechanisms in the circuit.
To examine selectivity in GrC→PC connectivity, we compared pairs of PCs by using Hamming similarity as a metric to quantify the similarity in their GrC input populations, which accounts for both connected and non-connected GrC-PC pairs (Fig. 3b). PC input patterns from local and non-local GrC axons were significantly more similar than shuffled controls, and local GrCs have significantly more similar connection patterns onto PCs than non-local GrC axons (Fig. 3d; local: 0.567 ± 0.071, non-local: 0.513 ± 0.065, shuffled: 0.503 ± 0.056 (mean ± SD); local GrC axons vs. shuffle, p = 4.9 × 10−61; non-local GrC axons vs. shuffle, p = 0.0043; non-local GrC axons vs. local GrC axons; p = 9.1 × 10−32; Kruskal-Wallis and Dunn’s post hoc tests, Bonferroni corrected for multiple comparisons), suggesting that PCs exhibit input selectivity. Then, when comparing pairs of GrCs based on the PCs to which they provide input, we detected a small yet significant trend suggesting that sharing MF inputs influences GrC→PC connectivity (Extended Data Fig. 8b, p = 1.1 × 10−4, Kruskal-Wallis test). The difference between local and non-local inputs may be related to the differences in dimensionality recorded from localized GrC cell bodies30,31 versus GrC axons in the molecular layer38.
Modeling network capacity & resilience
To explore the functional implications of the observed connectivity, we modeled pattern separation, a hypothesized function of the expanded representation from MFs to GrCs8–13 (Fig. 1c). We used dimensionality12,13 to quantify network encoding capacity, and signal-to-noise ratio (SNR) to quantify pattern separation performance and robustness to noise (Fig. 4a). We computed the dimensionality of the GrC population for input patterns with different levels of variability to better understand how the system separates similar versus different input patterns. It has been suggested that random connectivity maximizes dimensionality7,13, which predicts the observed redundant and selective connectivity motifs would decrease dimensionality. Although the reconstructed network showed lower dimensionality than random connectivity when the variability between input patterns was large, overall the reconstructed network preserved high dimensionality, particularly for less variable input patterns (Fig. 4b, Extended Data Fig. 9a). We next investigated amplification of selected inputs and noise resiliency as two possible benefits of redundant and selective connectivity. Examining over- and under-represented MF inputs separately, we found that dimensionality for overrepresented MF inputs increased, allowing them to utilize more of the encoding space (Fig. 4c), and increasing their pattern separability performance (Extended Data Fig. 9b). This may help with discriminating similar input patterns and possibly increase resilience to noise7. Furthermore, we found redundant oversharing of inputs increased noise robustness for a subset of the GrCs (Extended Data Fig. 9c,d), which may help increase reliability. Finally, we explored how the observed selective GrC-PC subsampling (Fig. 3c,d) could influence decoding from the perspective of PCs. Selective GrC-PC subsampling (Extended Data Fig. 9e) substantially increases pattern separability compared to random-subsampling or all-to-all connectivity (Fig. 4d, Extended Data Fig. 9f). Interestingly, the observed absence of roughly half of potential GrC axon inputs only slightly decreased dimensionality (Extended Data Fig. 9g,h). This suggests a preservation of information across layers by the network that may be supported by the redundant and selective connectivity motifs of the input layer.
Discussion
In summary, we found two connectivity motifs in the input layer: redundant oversharing of inputs among GrCs (Fig. 2c), and selective overrepresentation of specific MF boutons (Fig. 2e). Computational analyses suggest these features support a trade-off between network capacity and robustness, possibly to combat input variability and synaptic noise37,39. These findings run counter to prevailing models of random connectivity where dimensionality is solely maximized8–13 and may partially explain recent findings of lower-than-expected dimensionality of in vivo GrC activity29–31. Selective overrepresentation of MF boutons, in addition to increasing redundancy of signals from more informative inputs, could also be a mechanism for differential integration of multimodal MFs40,41. Our findings are also compatible with the hypothesis of activity-dependent synaptic plasticity underlying learned representational changes in MF-GrC connectivity6,42. Future studies incorporating long-range connectivity40,41,43 and functional characterization30,31 may reveal why some MF boutons are oversampled relative to others.
At the output layer, we found the connectivity rate of GrC to PCs to be ~50% (Fig. 3c) and observed some input similarity between adjacent PCs, consistent with learning occurring in the output layer8,9 (Fig. 3d). For neighboring PCs, a balance of similar, but non-identical inputs may serve to ensure variability and non-convergent learning, enabling a trade-off between specificity and generality. This may allow downstream cells in the DCN to more effectively implement ensemble learning, a machine learning approach where many weak and redundant learners (i.e. PCs) are pooled to make a more accurate prediction20,21.
In conclusion, we examined the feedforward connectivity of the cerebellum at synapse resolution, unveiling the role of non-random redundancy within a circuit that was hypothesized to rely on randomness for optimal performance. This is consistent with non-random biases detected in the parallel fiber system of the fruit fly44,45, suggesting a conserved principle of circuit architecture across species. Beyond the feedforward network, we expect continued analysis of the dataset will enable the comprehensive examination of cell types making up the cerebellar network14–16,46–49. Considering its similarities to modern deep learning architectures (Extended Data Fig. 1), our findings of structured connectivity may help to improve artificial intelligence algorithms, build and constrain more complete circuit models, and further our understanding of learning rules6,18,19 underlying cerebellar motor control and cognitive function3,6,42,50.
Methods
Experimental animals.
Experimental procedures were approved by the Harvard Medical School Institutional Animal Care and Use Committee and performed in accordance with the Guide for Animal Care and Use of Laboratory Animals and the animal welfare guidelines of the National Institutes of Health. The mouse (Mus musculus) used in this study was a P40 male C57BL/6J (Jackson Laboratory) housed on a normal light–dark cycle with an ambient temperature of 18–23°C with 40–60% humidity.
Sample preparation.
The specimen was prepared for EM as described previously51,52. We used an enhanced rOTO protocol for heavy metal staining51,53 followed by dehydration in a graded ethanol series and embedding in LX112 resin (Ladd Research Industries). The sample was polymerized at 60°C for 3 days.
Sectioning and imaging.
The EM dataset of the cerebellum (Fig. 1d,e, Extended Data Fig. 2a) was collected and imaged using the GridTape pipeline for automated serial-section TEM52. In brief, embedded samples were trimmed (Trim 90, Diatome). Serial ~45 nm-thick sections were cut using a 35° diamond knife (Diatome) and collected onto LUXFilm-coated GridTape (Luxel Corporation). A total of 1176 sections were cut for a total of ~49.5 μm total sample thickness. There were 91 single-section losses, 20 instances of adjacent two-section losses, 5 instances of adjacent three-section losses, 2 instances of adjacent four-section losses, and 1 instance of an adjacent five-section loss. Folds, staining artifacts, and cracks occasionally occurred during section processing, but were typically isolated to section edges and therefore not problematic.
Sections were imaged on a JEOL 1200EX transmission electron microscope at 120 kV accelerating potential at 2,500× magnification using a reel-to-reel GridTape stage and a 2 × 2 array of sCMOS cameras (Zyla 4.2, Andor) at 4.3 nm/pixel52. Imaging required 521 hours of acquisition on a single tape-based transmission electron microscope with a camera array. The EM images were then stitched together and aligned into a continuous volume using the software AlignTK (https://mmbios.pitt.edu/software#aligntk).
The aligned EM sections were then first imported into CATMAID54, then to Zarr (github.com/zarr-developers/zarr-python) and N5 containers (github.com/saalfeldlab/n5) for visualization and segmentation.
Automated segmentation.
We adapted an automated segmentation workflow based on a segmentation pipeline for TEM data55,56. In brief, the pipeline consists of three steps: (1) affinity prediction, fragment extraction (over-segmentation), and agglomeration. The affinity prediction uses a 3D U-net convolutional neural network (CNN)57 to predict a “connected-ness” probability of each voxel to adjacent voxels. An over-segmentation graph is then produced through a watershed algorithm (2D per-section), and agglomeration is performed iteratively using the “mean affinity” metric58 set at 0.5 threshold (segments are only joined when their immediate boundary has affinity greater than 0.5). We modified the original code56 (without local shape descriptors) to optimize for runtime performance and to adapt to our supercomputing cluster environment (Harvard Medical School Research Computing). The input to the network is a two-fold downsampled in XY images (effective resolution 8 × 8 × 40 nm); we found the decrease in size and performance to outweigh a minor reduction in accuracy.
Affinity CNN training.
We trained the CNN using a two-step process: bootstrapped ground truth (GT), and subsequent refinement based on skeleton GT. A set of bootstrapped GT was first created by applying a trained CREMI network (https://cremi.org/) directly on several subvolumes of the dataset, then applying manual corrections using Armitage/BrainMaps (Google internal tool). Training on these bootstrapped GT volumes generated an over-segmenting network.
To further improve the accuracy of the network, we scaled up the GT coverage to include 2 regions in the GrC layer, 3 regions in the molecular layer, and 1 region in the Purkinje cell layer – each 6 × 6 × 6 μm (1536 × 1536 × 150 voxels) in size. To accurately and efficiently proofread these volumes, we developed a skeleton-based GT method where we used the connectivity of the manually annotated point-wise skeletons of neurons to correct for errors of the bootstrapped network output. First, human tracers produced dense skeletons of the GT volumes using CATMAID54. To correct for split errors, the bootstrapped outputs were agglomerated at very high thresholds (0.7–0.9) to produce an under-segmented graph. Then the GT skeletons are used to correct merge errors, producing an output that can be reviewed and used for improved network training. We iteratively performed the annotation-train-review loop several times per volume as needed (sometimes without further annotation) to converge on a set of high-quality GT volumes.
Using the above GT volumes and supplementary volumes (to be described later), we trained the network using gunpowder (https://github.com/funkey/gunpowder). We used the CREMI network architecture for TEM images55. In brief, the U-net had four levels of resolution with downsampling factors in xyz of (3,3,1), (3,3,1), and (3,3,3). The topmost level had 12 feature maps, with the subsequent levels increasing the number of feature maps by 5 fold. Each layer was composed of two convolution passes with kernel sizes of (3,3,3) followed by a rectified linear unit (ReLU) activation function. A final convolution of kernel size (1,1,1) with 3 feature maps and a sigmoid function produced the affinity probability map. The mini batch input size was 268 × 268 × 84 pixels, and the output size was 56 × 56 × 48 pixels. An Adam optimizer was used with learning rate = 0.5 × 10−4, beta1 = 0.95, beta2 = 0.999, epsilon = 10−8. Gunpowder’s built-in augmentations were used to enable more varied training samples – these include elastic deformation, rotation in the z-axis, artificial slip/shift misalignment, mirroring, transpose, and intensity scale and shift, and missing sections. We also wrote and added duplicated augmentation that randomly selects a section and duplicates it up to 5 times. This is to simulate missing sections where, during deployment, missing sections are replaced with adjacent non-missing data sections. We trained the network to 350,000 iterations.
To evaluate performance of the neural network on unseen data, we further densely traced GT skeletons in another 9 cutouts in various regions of the GrC layer - each spanning 6 × 6 × 6 μm (1536 × 1536 × 150 voxels) in size - and used custom code to evaluate split and merge errors. Compared with other metrics like VOI or RAND index, comparing the number of topological errors is a more direct assessment of the amount of time that proofreaders will take to correct split and merge errors. Plus, producing skeleton GT cutouts is faster, allowing us to make more GT in less time. Evaluating across different artificial neural network parameters, we found that our best segmentation network, at an agglomeration threshold of 0.5, produces an average of 2.33 false merges and 27 false splits per cutout (Extended Data Fig. 2e). Analyzing the results in depth, we found that most false splits belong to small axons at locations with poorly aligned or missing sections.
Supplementary ground truth sub-volumes.
Beyond the targeted performance evaluation with the nine GT cutouts as described above and before deploying the network across the entire dataset, we had to address problems with segmenting (1) myelin that spans across large regions of the dataset often resulting in merge errors, (2) broken blood vessels that resulted in merge errors with surrounding neurons, (3) darkly stained mitochondria that can make adjacent neuron membranes ambiguous, and (4) GrC axons that are sometimes darkly stained.
To address myelin, we produced 3 additional volumes of regions with dense myelin labeled and trained the network to predict low (i.e., zero) affinity for myelin-containing voxels; the fragment extraction step was then modified to remove fragments that have averaged affinity (average of x, y, z affinity values) less than 0.3. The removal of these fragments reduced myelin-related merge errors.
Blood vessels are relatively large structures in EM that are typically easy to segment. Broken capillaries (during serial sectioning), however, can produce artifacts with misalignments, blending with surrounding neurites and producing merge errors. To address this problem, we used Ilastik59 to train a machine learning model (random forest) on 32-fold XY downsampled EM images (effective resolution 128 × 128 × 40 nm) to detect blood vessel voxels and mask them out from the affinity prediction prior to the fragment extraction step.
Mitochondria are often darkly stained in the EM dataset. Because the network is trained to predict high affinities for electron-dense membranous structures including mitochondria, the network can under-predict neuron boundaries when mitochondria are right next to the boundaries creating conditions that can be ambiguous even to human experts. We sparsely annotated an additional GT cutout (3 × 6 × 3 μm in size) with dense mitochondria for training the segmentation network. To our surprise, this worked well without further masking agglomeration or other post processing steps.
Last, there were an infrequent number of GrC axons that were partially darkly stained. These would often be normally stained proximal to the soma but became more darkly stained distally before becoming normally stained again along the ascending portion of the axon. Human annotators can often trace these axons reliably, but they generated errors by the initially trained segmentation network. To mitigate this problem, we made a skeleton GT cutout specifically targeting a region dense with dark GrC axons and added it to the training procedure. Except for regions where multiple darkened axons come together to form a tight bundle of axons, this was effective for segmenting isolated dark axons, speeding up the bulk of the GrC axon reconstruction process.
Large-scale parallel processing deployment.
To efficiently run the trained CNN and other steps in the automated segmentation pipeline across a large dataset, we developed Daisy60, an open-source work scheduler that can divide the input dataset into spatial chunks and operate on them in parallel (Extended Data Fig. 2c). We developed Daisy with the following design requirements: (1) be scalable and efficient across thousands of workers, (2) be able to tolerate worker and scheduler failures and resumable across reboots, (3) be extensible to non-segmentation tasks, (4) be Python-native for ease of integration with the existing pipelines, and (5) be able to provide on-the-fly dependency calculation between computational stages. At the time of development, no existing tools met such requirements.
Targeted neuron reconstruction and proofreading.
Although split and merge errors exist in the automated segmentation, such errors are usually easy for humans to recognize in 3D visualizations of reconstructed neurons. Thus, proofreading automated segmentation results is typically faster than manual tracing. That said, merge errors – especially between big neurons and in regions not well-represented by the GT – are endemic for large EM volumes, requiring a low merge-threshold (which creates more split errors), sophisticated un-merge algorithms, and/or exorbitant person-hours to proofread and diagnose where the merge errors happen. These obstacles make using automated large-scale segmentation inaccessible to labs with less resources.
To combat the merge error problem, we developed merge-deferred segmentation (MD-Seg) proofreading method, a novel workflow that pre-agglomerates fragments only in local blocks (4 × 8 × 8 μm in this work) and defers inter-block merge decisions to proofreaders. In this manner, if there is a merge error, it is limited to the local block and not automatically propagated to hundreds of other blocks which may contain even more merge errors. While deferred, inter-block merge decisions of each neuron are still computed and on-line accessible to proofreaders through a hot key in the user interface, which is based on Neuroglancer (https://github.com/google/neuroglancer) (Extended Data Fig. 2d).
By deferring merge decisions across the dataset, MD-Seg allowed us to focus proofreading, targeting regions and cells of interest. Although conventional proofreading can also target specific neurons, because agglomeration is performed globally, a generally high-quality assurance (QA) is needed to be performed for all neurons prior to proofreading to minimize excessive errors among these neurons to the neurons of interest. This pre-proofreading QA burden can be excessive for projects with fewer GT and resources, preventing the start of proofreading. In contrast, in MD-Seg, errors between non-targeted neurons do not affect segmentation quality of targeted neurons in most cases. Furthermore, because local blocks are agglomerated separately, it is easier to run segmentation for parts of the dataset, or re-run and update segmentations for specific regions without affecting on-going proofreading progress.
To reconstruct neurons in MD-Seg, MFs, GrCs, and PCs were first identified manually based on their stereotypical morphological characteristics15 (Fig. 1f, Supplementary Data 1-3 and Videos 2-3). Proofreaders typically started by selecting a neuron fragment (contained within a single block), and sequentially ‘grew’ the neuron by adding computed merge decisions of fragments in adjacent blocks through a hot-key. During each growth step, the 3D morphology of the neuron was visualized and checked for errors. When merge errors occur, the blocks containing the merge are ‘frozen’ to prevent growth from the merged segment. When a neuron branch stops growing (has no continuations), the proofreader inspects the end of the branch to check for missed continuations (split errors). In this way, both split and merge errors can be corrected.
At any time during growth steps, the proofreader can associate annotations like “uncertain continuations” (potential split errors) as well as periodically save the partial progress of the reconstructed neurons into an on-line database. Upon saving, MD-Seg automatically checks if the neuron overlaps with any existing saved neurons, and reports a failure back to the proofreader, enforcing an invariance that any single fragments can only belong to one neuron object. For neurons with higher complexity and overlapping branches (such as a Purkinje cell), to decrease cognitive load, proofreaders saved the neuron as simpler sub-objects (e.g., purkinje_0.axon, purkinje_0.dendrite_0, …). Once a neuron is marked as completed, a reviewer then loads the neuron to check for possible split or merge errors in a second independent review. Gallery of example reconstructions of MFs, GrCs, and PCs are provided (Supplementary Data 1-3 and Videos 2-3).
Using this pipeline, we reconstructed 153 PCs (total; 28 with somas), 784 MF boutons and 2,397 GrCs based on morphology. 550 of the 784 MF boutons are from unique MFs, with rare instances of GrCs receiving input from different boutons belonging to the same MF (n = 20). 541 of the 2,397 GrCs had axons extending into the molecular layer. We also reconstructed 4,439 axons from GrCs whose cell bodies were not present in our volume.
Automated synapse prediction.
We adapted an artificial neural network for synapse prediction on mammalian brain datasets using synful61. We modified the algorithm to predict synaptic clefts from pointwise ground truth annotations and identified pre- and post-synaptic partners by applying the synapse directionality prediction on the cerebellum dataset. We used webKnossos62 to make rough synaptic cleft masks and trained the network to predict these cleft masks. We found that training the network to identify the synaptic clefts produced more generalizable results than either pre- or post-synaptic point predictions and removed ambiguity of one-to-many or many-to-one synapses if only the pre- or post-synaptic partners are predicted. For the network architecture, we used the “small” network size and the cross-entropy loss. We downsampled the GT cutouts by a factor of four in the x- and y-dimension for an effective resolution of 16 × 16 × 40 nm prior to training. Consequently, we reduced the downsampling factors of the U-Net to (2,2,1), (2,2,1), and (2,2,3).
To train and evaluate the synapse prediction network, we annotated nine GT cutouts of the GrC layer (GCL; previously used for evaluating the automated segmentation network, see above), two more of the molecular layer (ML), and one more of the PC layer (PCL) – all of which are 6 × 6 × 6 μm. Five of the GCL cutouts and one of the ML cutouts were used for training – the rest were used for evaluation. Because of the efficiency of pointwise GT generation, we were able to produce a large amount of GT for both training and evaluation. The synapse predictions exhibited excellent accuracy (precision: 95.4%, recall: 92.2%, F-score: 0.938; Extended Data Fig. 2f), exceeding that of other TEM datasets52,61 (F-score: ~0.60–0.75) and FIB-SEM datasets63 (precision/recall: ~60–93%). It must be noted, however, that other datasets are of Drosophila, which have smaller and more numerous synapses that lead to disagreements even among expert annotators61,63. Inspecting the synapse predictions, we found that prediction “errors” are typically ambiguous GCL synapses consistent with errors found evaluating synapse predictions elsewhere63. In contrast, there were no mispredictions in the ML evaluation dataset where synapses are both larger and clearer in EM (Extended Data Fig. 4).
To conduct a sensitivity analysis of the automated synapse prediction, we evaluated the robustness of our results with different prediction thresholds to artificially add false positive (FP) and false negative (FN) synapses to the predicted network connectivity graph (Extended Data Fig. 7). In our prediction framework, each synapse is scored by how confident the detected synapse is and how big it is (i.e., score is the sum of the confidence value of the connected components); synapses with scores above this threshold are recognized while those below this threshold are ignored. We can combine predictions from multiple thresholds together to simulate a model with specific accuracy values. For example, if threshold_a produced a graph with a 20% FP rate and threshold_b produced one with a 20% FN rate, to get an 80% accurate model that has 20% FP and 20% FN rates, we take the default model and subtract synapses that were not detected with threshold_a and add synapses that were falsely detected with threshold_b. We then tested models with FP/FN rates at 10%, 20%, …, 90% to represent models with 90%, 80%, …, 10% accuracy. Note, however, that since the lowest threshold possible can only produce 30% FP rate, models with less than 70% accuracy only had 30% FP rate.
Quantifying MF→GrC connectivity.
We used the reconstructed neurons and the automated synapse predictions to map the connectivity between MF boutons and GrCs. The center of mass of GrCs cell bodies were manually annotated. To accurately get the center of mass of MF boutons, we first collected all MF→GrC synapses (i.e., synapses onto non-GrC neurons were filtered out) associated with a single MF, clustered them into individual boutons through DBSCAN64 with parameters eps = 8 μm and min samples = 2, then averaged the synapse locations of each bouton to get the bouton’s center of mass.
MF bouton - GrC connectivity proofreading.
Beyond optimizing the neural network to minimize false positive (FP) and false negative (FN) rates, we also performed synapse count thresholding and targeted proofreading to more accurately determine binary connectivity between MF boutons and GrCs. Considering that each MF bouton to GrC connection has multiple synapses (Extended Data Fig. 4c) and that FPs often would only result in connections with single synapses, simple thresholding would make binary connectivity identification robust even with some FNs. However, we considered two additional factors. First, when MF boutons are at the edge of the volume, only a partial number of synapses are visible, increasing the need for a lower threshold. Second, MF boutons can make axon collaterals that make 1 to 2 synapses onto GrC dendrites either without “claws”; with claws connecting to different MF boutons than the collateral; or with claws connecting to the same MF (Extended Data Fig. 5). Considering these factors, we set the minimum automatic synapse threshold of MF bouton→GrC connections to 3, and then manually validated 2-synapse connections.
Spatial distribution of connectivity.
To get the spatial distribution of GrC dendrites in Extended Data Fig. 3e, for each GrC, we quantified the spatial displacement (Euclidean distance) of each MF bouton connected to it. To avoid edge effects, we defined margins where we removed from our analysis GrCs within 200 μm of the reconstruction boundaries in the X-axis. We further only counted the GrCs that are either at the top (or bottom) 10 μm of the dataset in Z, removed connections to MFs that are above (or below) the GrC centers of mass in Z, then combined the two distributions together. The Y margins do not need to be accounted for since they are natural boundaries. Since the dataset has a Z thickness of 49.5 μm, this method allows unbiased measurement of dendrite distribution of up to 200 μm in X and 39.5 μm in Z. Taken together, we found that GrCs preferentially connect with MF boutons spread in the dorsal-ventral axis rather than in the medial-lateral or the anterior-posterior axis, consistent with previous reports26.
Counting presynaptic and postsynaptic partners.
In Fig. 2e, we counted the number of postsynaptic GrCs connected to each MF bouton. To avoid edge effects, we defined margins to not include in the analysis MF boutons within 60 μm and 20 μm to the reconstruction boundaries in X- and Z-axis respectively (Y margins do not need to be accounted for since they are natural boundaries). As shown in Extended Data Fig. 3e, these margins capture the extent of the dendritic reach of GrC dendrites. For Fig. 2e, only connections from the centermost MF boutons are counted (n = 62, though we used all surrounding GrCs n = 4,400). Similarly, for Extended Data Fig. 3c,d and Fig. 2, we removed analysis GrCs that are within 60 μm and 20 μm of the reconstruction boundaries in X- and Z-axis respectively. Note that the reconstructed distribution is plotted as an empirical CDF with discrete numbers of connected neuron pairs, while the random distribution is plotted as CDFs with error shaded across randomly generated networks.
Counting input redundancy between GrCs.
In Fig. 2c, we analyzed the MF bouton input redundancy to each GrC by counting the number of other GrCs sharing 2 or more MF inputs. To avoid the edge effects, even though we computed input redundancy using the entire GrC (n = 4,400) and MF (n = 1,145) population, we only counted the sharings of the centermost n = 211 GrCs.
MF→GrC random connectivity models.
To compare with the observed connectivity in Fig. 2, we developed random connectivity models (Extended Data Fig. 3f). The “Radius” models are based on random spherical sampling connecting GrCs to MF boutons closest to a single given dendrite length11–13,26. “Radius-Average” uses the average dendrite length from our EM reconstructions. The “Radius-Distribution” model is similar, but it uses dendrite lengths drawn from the reconstructed distribution. In the “Vector-Shuffle” model, the dendrite targets of each GrC are drawn (with replacement) from the observed distribution of the spatial displacements between connected GrCs and MF boutons (Extended Data Fig. 3e). In all models, the locations of GrCs/MF boutons and the number of dendrites per GrC are maintained to be the same as the reconstructed graph, the margin of searching for MF boutons is set to 10 μm, and there are no double connections between any MF bouton→GrC pair.
MF axon collaterals → GrC proofreading.
To obtain the distributions in Extended Data Fig. 5c–e,h, we densely proofread 68 MF boutons within a 12,000 μm3 subvolume and determined whether a bouton made axon collaterals of at least 10 μm in length, how many boutons there are within a collateral axon, and whether they made synapses onto GrCs. We then quantified whether the synapses are located at the GrC dendrite claws or trunks, and whether the dendrite made a claw or not.
Quantification of GrC→PC connectivity.
Similar to MF→GrC connectivity analysis, we used automated segmentation and synapse detection to map GrC→PC connectivity.
GrC→PC connectivity proofreading.
Since the majority of GrC→PC connections consist of single synapses (Extended Data Fig. 4c), it was necessary to proofread GrC→PC synapses to minimize false negative errors. Virtually all false negatives consist of synapses from a GrC axon onto “orphaned” PC spines that are not connected to their PC dendrite. To correct these errors, we found all synapse locations between the reconstructed GrC axons to orphaned segments, and manually proofread them.
GrC axon to PC touches.
To find locations where GrC axons touch PCs and have the potential, but do not make synapses, we computed and utilized mesh representations of neuron segmentations, where neuron meshes consist of a reduced set of vertices that describe their boundaries. Touches between two neuron meshes were then determined through thresholding the shortest distance between the two sets of vertices; this threshold was set to 10 pixels (at 16 nm mesh resolution) to account for the mesh coordinate approximations. To reduce the number of all-to-all comparisons and improve performance, we simplified vertex locations through downsampling to make rough predictions of distances and prune vertices prior to calculating the exact distances.
GrC→PC random connectivity models.
To make the “Shuffle” model (i.e., the configuration model) in Fig. 3d, for each PC, we shuffled the “connected” status among all connected and touching GrCs preserving the total number of connected GrCs per PC. We shuffled only connected and touching neurons to preserve the spatial constraints on connectivity (i.e., PCs have limited dendrite spans and do not touch all GrC axons in the dataset).
Computing Hamming similarity.
To measure similarity between pairs of PCs in terms of their GrC input populations (Fig. 3), and pairs of GrCs in terms of the PCs to which they provide input (Extended Data Fig. 8b), we used Hamming similarity. We used this metric to give connections and the absence of connections equal influence in assessing the similarity of connectivity patterns. Hamming similarity is defined as the inverse of Hamming distance65 normalized to the length of the vectors:
For example, to compare the input similarity of two PCs which might have different sets of potential presynaptic GrC axons due to differences in the location of their dendrites, we first filtered out all axons that are not either connected to or touching both PCs, prior to computing the Hamming similarity score. Similarly for the comparing GrC pairs, we computed the Hamming similarity between the output vectors of two GrCs after filtering out all PCs that were neither connected to nor touched by both GrCs (Extended Data Fig. 8b). For Fig. 3d, we further filtered out PC pairs that had less than 30 common connected or touching GrC axons to minimize Poisson noise in the analyzed distributions.
Numerical analyses.
We developed a simulation framework to generate modeled input and output activity patterns of the MF→GrC layer given the reconstructed (and randomized) MF bouton→GrC connectivity. In all analyses in Fig. 4 and Extended Data Fig. 9 (with the exception of Extended Data Fig. 9c–d), to decrease edge effects where GrCs near the edge of the reconstruction boundaries showed fewer inputs (e.g., GrCs that had four dendrites, but only one is observable) and MFs near edges showed fewer outputs (e.g., boutons with only a portion contained in the EM volume), we used a subset of the reconstructed MFs and GrCs for our simulations. We used only the GrCs within the center 200 μm in the X dimension and 34 μm in the Z dimension, which totaled 1459 neurons, and MF boutons within the center 240 μm in the X dimension and 44 μm in the Z dimension totalling 630 boutons. While this graph is most representative of the EM sample, the overall MF input sharing among GrCs was not as high as observed in Fig. 2a and Extended Data Fig. 3g–m because of GrCs near the edges. To better test the high MF input sharing principle in Extended Data Fig. 9c–d we constructed a 120 × 80 × 120 μm MF-GrC graph by taking the GrCs and MF boutons in the center 120 × 80 × 40 um and replicate these node positions 3 fold along the Z (mediolateral) axis for a graph with 2541 GrCs and 771 MF boutons. With this graph we tested two models, one with high MF input sharing and one with normal (as expected from random wiring) MF input sharing. To achieve high input sharing while preserving the observed distributions of GrC dendrites (Extended Data Fig. 3c), GrC dendrite lengths (Extended Data Fig. 3d), and MF bouton sizes (Fig. 2e), we randomly added extra dendrite (26% more for this graph) to achieve the observed level of sharing (Fig. 2c and Extended Data Fig. 3l,m) then pruned away66 those that least contributed to MF sharing. On each iteration, we randomized the graph by shuffling the dendrite and bouton size distributions and randomly connecting GrCs to MF boutons with a distance sampled from the observed dendrite length distribution.
We assumed the MF input pattern to be binary with a mean activity level of 50%, consistent with prior work12,13. Each GrC integrates its inputs with equal weights based on the reconstructed (or randomized) connectivity graph. We assumed a binary output GrC activation function12,13 but with an activity level of 30% instead of 10% to be more consistent with recent in vivo recordings30,31. To maintain a steady-state activity level in a GrC population with a varying number of dendrites per neuron (as observed in Extended Data Fig. 3c), we tuned the activation threshold of each GrC by simulating 512 random input patterns and getting the average number of spiking inputs that would produce a 30% activity level for this neuron. Because the MF inputs are binary, however, a single threshold will not produce the precise 30% activity level. As an example, in our configuration, for GrCs with 5 dendrites, an activation threshold of 3 produces a 54% activation rate while a threshold of 4 would produce a 21% activation rate. It is then necessary to further divide the GrC population into an x% population of “low” and a 1-x% population of “high” activation level. In the above example, we solved for x% where .21 * x% + .54 * (1 - x%) = .30, then randomly choose x% of GrCs to have an activation threshold of 3 and 1 - x% to have 4. Such random assignment of “low” and “high” activation levels allowed us to overcome limitations in previous work12,13.
Computing dimensionality.
Dimensionality is a measure of the number of independent variables embedded in a population’s activity which we used to quantify network encoding capacity. To compute the dimensionality of the population activity matrix x = (x0, x1, …) describing activity of the MF or GrC population across trials, we used a previously defined equation13.
where 𝜆i are the eigenvalues of the covariance matrix of x. We also computed population correlation12 as:
where N is the number of neurons in the population.
For Fig. 4b,c and Extended Data Fig. 9g, we sought to quantify the dimensionality of the GrC population when given a set of MF input vectors that are different by a specific f_variability factor, where low f_variability means highly similar input vectors and high f_variability means highly different input vectors. To achieve this, we constructed the input MF population activity matrix m as follows. For each simulation trial, a base vector m0 was randomly generated: each element in m0 was drawn independently as a binary number from a Bernoulli distribution with probability p = 0.5. Derived vectors mi (n = 512 vectors) were then generated from m0 by randomly selecting f_variability * len(m0) elements and randomly rerolling their binary values. In this manner, vectors created with f_variability = 1.0 would be entirely uncorrelated to the base input pattern. Each data point (e.g., for each input pattern difference) consisted of the average of 100 such simulation trials and the resulting 95% confidence interval.
For Fig. 4c and Extended Data 6b, variability was restricted and normalized to the 33% MF subpopulation being tested (random, over-, and under-sampled). Within a trial, the variation of MF activity vectors mi was sampled from said fixed MF bouton subpopulations. This was slightly different for Fig. 4b where the variation of MF activity vectors mi was sampled across the entire MF population, even if the effective f_variability was the same. For Extended Data Fig. 9g, to measure dimensionality of a varying GrC population size with a parameter y%, we ran MF→GrC simulation as above, but only performed dimensionality analysis on a random y% subset of the GrCs. The subset is re-randomized for each trial and MF input variability is 100%.
For Extended Data Fig. 9a, we modeled MF inputs as continuous variables representing input spiking frequencies13. Like the binary input model, we also performed GrC activation threshold tuning to produce a precise coding level for the GrC population (i.e., 30%), though we did not need “low” and “high” assignments since inputs are continuous. In this experiment, we sought to quantify the GrC population correlation/stability when given a set of inputs that were the same, but corrupted by an f_noise factor (a higher population correlation value could denote higher noise resilience and learning generalization44). Compared to Fig. 4b, besides using continuous input variables versus discrete variables, the other primary difference is that in this study all inputs were changed (% magnitude based on f_noise), while in the binary study a subset of the input was flipped (% based on f_variability). Specifically, for each simulation trial, we generated a random input vector m0 with a uniform distribution from 0 to 1. Derived input vectors mi were then generated by adding m0 to a random noise vector ni (also using a uniform distribution) multiplied by the noise factor f_noise:
The noise factor f_noise effectively controls the degree of difference between input vectors within the MF activity matrix m. At low f_noise, m is more static, while at high f_noise, m is composed of more independently drawn random input vectors.
Computing learned signal size.
We computed SNR with the signal size defined as the maximum linear distance of GrC representations between two input patterns and noise defined as the variation of the modeled binary activity12,13 of the GrC population. It is thought that the GrC population activity is normalized through inhibition from the Golgi cells8,9, and thus we tuned the GrC population to have the same mean activity across input MF patterns. Assuming equal weights, the presynaptic sum of GrC activations is expected to be the same across patterns. To differentiate patterns, it is theorized that PCs can manipulate their synaptic weights to affect (increase or decrease) the postsynaptic linear sum of specific GrC activation7–9,13,17,20,67. For simplicity, these analyses assumed that GrC inputs were equal in strength. However, we note that some GrC inputs may be more influential than others68 (cf.69,70) and that >25% of unitary connections have multiple synapses (Extended Data Fig. 4c) which would further improve SNR of these inputs.
For Fig. 4d and Extended Data Fig. 9c–d, we modeled the PCs linear/logistic regression models with multiple variables (input GrCs) and binary outputs (PC firing low/high). The models assumed a linear relationship between the input patterns and the outputs, hence the “linear sum.” The model’s predictive power is maximized when the weights/parameters of the model maximize the difference between the input patterns, and so we defined and measured “learned signal size” as the maximum difference of the linear sums.
In Fig. 4d, the objective is to measure SNR of the difference or discriminatory power of the GrC population when encoding MF vectors which can be similar or different based on a variability factor f_variability. Given two GrC activation vectors gj and gk, and a set of weights w, the linear sum difference is:
Given the constraints that weights are positive (GrC→PC synapses are excitatory) and that weights range from 0.0 to 1.0, the max learned difference from gj to gk would be:
Alternatively, the max learned difference from gk to gj is:
MaxLearnedDiffk→j = Σi (gki - gji) if gki > gji |
For simplicity and without loss of generality we simply calculated a combined metric, which equates to the Hamming distance between two GrC activation vectors when GrC activation vectors are binary and weights are between 0.0 and 1.0:
For each simulation trial, we generated a base input MF vector m0 and a set of derived vectors m1, .., n (n = 512) as described previously. We then computed the pairwise Hamming distance of the resulting GrC activation vector g0 against g1, 2, .., n to measure the average and variance of responses. Each data point (e.g., for each input pattern difference) consists of 40 such simulation trials and the resulting 95% confidence interval. For the “selective subsampled” model in Fig. 4d, we chose the 50% subset of GrCs that would produce the greatest Hamming distance across the derived activation vectors m1, .., n to m0.
In Extended Data Fig. 9c–d, the objective was to measure the robustness of the GrCs encoding a number of input patterns when they are corrupted by noise (f_variability). We then scored and ranked the GrCs with a robustness_score to show that SNR is higher and more concentrated in a smaller subpopulation of GrCs in the high MF input sharing models than in the low input sharing model. Given m1..n (n = 8 in this study) random input patterns, we computed the noisy patterns as described for Fig. 4b,c twenty times each, ten used for ranking GrCs (ranking patterns) and then for computing SNR (testing patterns). These input patterns are then used to compute the resulting GrC activation vectors. To estimate the most robust GrC subpopulation, we compute robustness_score by ranking each GrC by its probability to be activated across all ranking patterns - GrCs that are most consistently activated here should also be robust across testing patterns, and PCs that made synapses to these GrCs will also more reliably have a high SNR linear sum with respect to MF input patterns. After ranking, for each subpopulation_threshold, we computed SNR on the subpopulation_threshold of GrCs with the highest robustness_score: signal is the difference between the linear sum of activations across testing patterns and background patterns, and noise is the standard deviation of the activations across background patterns (see “Computing noise” below). Note, we included a pattern_size parameter that controls whether a randomly selected pattern_size subset of inputs is relevant (and kept constant during noise additions) and the rest of the input is irrelevant (and replaced with random inputs during noise additions). Extended Data Fig. 9c–d used an example pattern_size of 0.3 but the general trend was seen in all pattern_size that we tested (data not shown).
Computing noise.
For Fig. 4d and Extended Data Fig. 9c–d, we computed noise as the magnitude of variation of the modeled GrC activity across random background MF input patterns. To calculate the standard deviation of background activity noise, we ran simulations of 512 random MF input patterns and calculated the standard deviation of the sums of GrC outputs.
Prediction accuracy.
To evaluate prediction accuracy in Extended Data Fig. 9f,h, we used gunpowder (https://github.com/funkey/gunpowder), JAX71, Haiku72, Optax73, and Chex73. As above, we trained the output decoder as a linear layer with stochastic gradient descent using the output of the MF-GrC model as the input. The task was a standard binary classification where the network discriminates two classes of MF input patterns, where MF patterns were randomly varied, with different levels of variability, starting from a randomly chosen input vector. A standard sigmoid activation function was used. For comparison, we picked the number of test patterns that resulted in ~90% accuracy performance at 100% variability, which was 800 patterns. Each model was trained for 300 epochs with 5% MF noise augmentation on each batch, with a learning rate of 1 × 10−2. The models are evaluated with different noise than in training. Since accuracy varied from run to run, we ran each model 20 times to produce averages and bootstrapped 95% confidence intervals. To better approximate the performance of a Purkinje cell decoder, we further modified the weight update mechanism to restrict the input weights to be positive with a logistic/sigmoid activation function that centers on a positive value sigmoid_center instead of on zero because GrC->PC synapses are excitatory and the activation of many GrCs are needed to influence the firing rate of PCs. We then defined max_weight, the maximum value of any individual synapse weight to be:
The first and second term normalize max_weight to the total number of GrCs and activation_rate respectively. The third term scales this by the sigmoid_center value, essentially controlling the steepness of the sigmoid function. Lastly, the fourth term controls synapse strength. In these simulations, we set weight_scale to be 4, which means that at activation_rate = 30%, 7.5% (30% ÷ 4) of the inputs would need to be activated at max weight to reach the center of the sigmoid function.
Statistics and reproducibility.
We did not use statistical methods to pre-determine sample sizes. Technical limitations made it only feasible to analyze one mouse in this study. However, connectivity trends were observed for many different neurons, and significance was tested with statistical tests (Wilcoxon rank-sum, Kruskal-Wallis, and permutation tests). Our study did not contain experimental groups, so randomization and blinding do not apply.
Extended Data
Supplementary Material
Acknowledgements
We thank X. Guan, Y. Hu, M. Liu, E. Mayette, M. Narwani, M. Osman, D. Patel, E. Phalen, R. Singh, and K. Yu for neuron reconstruction and manual annotation of ground truth for cell segmentation and synapse predictions; J. Rozowsky for neuron reconstruction, analysis, and the initial finding of redundant MF→GrC connectivity; P. Li and V. Jain for access to and support with Google Armitage/BrainMaps software; Y. Hu and M. Osman for contributing code for segmentation accuracy measurements; M. Narwani, K. Yu and X. Guan for contributing code for the proofreading platform; J. Buhmann and N. Eckstein for discussions and machine learning advice; C. Guo, T. Osorno, S. Rudolph, and L. Witter, for discussion, advice, and help with EM preparations; C. Bolger for EM sample preparation; T. Ayers, R. Smith, and Luxel Corporation for coating GridTape; O. Rhoades for help with illustrations; D. Bock, J. Drugowitsch, A. Handler, C. Ott, R. Wilson, A. Kuan, and members of the Lee lab for comments on the manuscript. This work was supported by the NIH (R21NS085320, RF1MH114047) to W.C.A.L. and (R01MH122570, R35NS097284) to W.G.R., the Bertarelli Program in Translational Neuroscience and Neuroengineering, Stanley and Theodora Feldberg Fund, and the Edward R. and Anne G. Lefler Center. Portions of this research were conducted on the O2 High Performance Compute Cluster at Harvard Medical School partially provided through NIH NCRR (1S10RR028832–01) and a Foundry Award for the HMS Connectomics Core.
Footnotes
Competing interests
W.C.A.L. and D.G.C.H. declare the following competing interest: Harvard University filed a patent application regarding GridTape (WO2017184621A1) on behalf of the inventors including W.C.A.L. and D.G.C.H., and negotiated licensing agreements with interested partners.
Software and Code Availability
- MD-Seg front-end: https://github.com/htem/neuroglancer_mdseg/tree/segway_pr_v2MD-Seg back-end: https://github.com/htem/segway.mdseg
- Segmentation and synapse prediction scripts: https://github.com/htem/segway
- Analysis code: https://github.com/htem/cb2_project_analysis
Please see “Software and code” in the Reporting Summary for more details on other software packages used in this work.
Data Availability
The EM dataset is publicly available at the BossDB (https://bossdb.org/) repository: https://bossdb.org/project/nguyen_thomas2022. Source data for Figs. 2c,e, 3c,d, and 4b,c,d are provided with the paper. Further directions for accessing neuron 3D renderings and derived neuron connectivity graphs are available at https://github.com/htem/cb2_project_analysis.
References
- 1.Wolpert DM, Miall RC & Kawato M Internal models in the cerebellum. Trends Cogn. Sci 2, 338–347 (1998). [DOI] [PubMed] [Google Scholar]
- 2.Ebner TJ & Pasalar S Cerebellum predicts the future motor state. Cerebellum 7, 583–588 (2008). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3.Strick PL, Dum RP & Fiez JA Cerebellum and nonmotor function. Annu. Rev. Neurosci 32, 413–434 (2009). [DOI] [PubMed] [Google Scholar]
- 4.Koziol LF et al. Consensus paper: the cerebellum’s role in movement and cognition. Cerebellum 13, 151–177 (2014). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5.Schmahmann JD Disorders of the cerebellum: ataxia, dysmetria of thought, and the cerebellar cognitive affective syndrome. J. Neuropsychiatry Clin. Neurosci 16, 367–378 (2004). [DOI] [PubMed] [Google Scholar]
- 6.Wagner MJ & Luo L Neocortex–Cerebellum Circuits for Cognitive Processing. Trends Neurosci. 43, 42–54 (2020). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7.Cayco-Gajic NA & Silver RA Re-evaluating Circuit Mechanisms Underlying Pattern Separation. Neuron 101, 584–602 (2019). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Marr D A theory of cerebellar cortex. J. Physiol 202, 437–470 (1969). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9.Albus JS A theory of cerebellar function. Math. Biosci 10, 25–61 (1971). [Google Scholar]
- 10.Babadi B & Sompolinsky H Sparseness and expansion in sensory representations. Neuron 83, 1213–1226 (2014). [DOI] [PubMed] [Google Scholar]
- 11.Billings G, Piasini E, Lőrincz A, Nusser Z & Silver RA Network structure within the cerebellar input layer enables lossless sparse encoding. Neuron 83, 960–974 (2014). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12.Cayco-Gajic NA, Clopath C & Silver RA Sparse synaptic connectivity is required for decorrelation and pattern separation in feedforward networks. Nat. Commun 8, 1116 (2017). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13.Litwin-Kumar A, Harris KD, Axel R, Sompolinsky H & Abbott LF Optimal Degrees of Synaptic Connectivity. Neuron 93, 1153–1164.e7 (2017). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Eccles JC, Ito M & Szentagothai J The cerebellum as a neuronal machine. (Springer-Verlag, 1967). [Google Scholar]
- 15.Palay SL & Chan-Palay V Cerebellar Cortex: Cytology and Organization. vol. Berlin (Springer-Verlag, 1974). [Google Scholar]
- 16.Liu A & Regehr WG Normalization of input patterns in an associative network. J. Neurophysiol 111, 544–551 (2014). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.Walter JT & Khodakhah K The advantages of linear information processing for cerebellar computation. Proc. Natl. Acad. Sci. U. S. A 106, 4471–4476 (2009). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18.Ohmae S & Medina JF Climbing fibers encode a temporal-difference prediction error during cerebellar learning in mice. Nat. Neurosci 18, 1798–1803 (2015). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19.Heffley W et al. Coordinated cerebellar climbing fiber activity signals learned sensorimotor predictions. Nat. Neurosci 21, 1431–1441 (2018). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20.Sanger TD, Yamashita O & Kawato M Expansion coding and computation in the cerebellum: 50 years after the Marr-Albus codon theory. J. Physiol 598, 913–928 (2020). [DOI] [PubMed] [Google Scholar]
- 21.Cao Y, Geddes TA, Yang JYH & Yang P Ensemble deep learning in bioinformatics. Nature Machine Intelligence 2, 500–508 (2020). [Google Scholar]
- 22.Bell CC, Han V & Sawtell NB Cerebellum-like structures and their implications for cerebellar function. Annu. Rev. Neurosci 31, 1–24 (2008). [DOI] [PubMed] [Google Scholar]
- 23.Stevens CF What the fly’s nose tells the fly’s brain. Proc. Natl. Acad. Sci. U. S. A 112, 9460–9465 (2015). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24.Eichler K et al. The complete connectome of a learning and memory centre in an insect brain. Nature 548, 175–182 (2017). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25.LeCun Y, Bengio Y & Hinton G Deep learning. Nature 521, 436–444 (2015). [DOI] [PubMed] [Google Scholar]
- 26.Houston CM et al. Exploring the significance of morphological diversity for cerebellar granule cell excitability. Sci. Rep 7, 46147 (2017). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27.Sultan F Distribution of mossy fibre rosettes in the cerebellum of cat and mice: evidence for a parasagittal organization at the single fibre level. Eur. J. Neurosci 13, 2123–2130 (2001). [DOI] [PubMed] [Google Scholar]
- 28.Gilmer JI & Person AL Morphological Constraints on Cerebellar Granule Cell Combinatorial Diversity. J. Neurosci 37, 12153–12166 (2017). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29.Knogler LD, Markov DA, Dragomir EI, Štih V & Portugues R Sensorimotor Representations in Cerebellar Granule Cells in Larval Zebrafish Are Dense, Spatially Organized, and Non-temporally Patterned. Curr. Biol 27, 1288–1302 (2017). [DOI] [PubMed] [Google Scholar]
- 30.Wagner MJ, Kim TH, Savall J, Schnitzer MJ & Luo L Cerebellar granule cells encode the expectation of reward. Nature 544, 96–100 (2017). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 31.Giovannucci A et al. Cerebellar granule cells acquire a widespread predictive feedback signal during motor learning. Nat. Neurosci 20, 727–734 (2017). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 32.Heck DH, Thach WT & Keating JG On-beam synchrony in the cerebellum as the mechanism for the timing and coordination of movement. Proceedings of the National Academy of Sciences 104, 7658–7663 (2007). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 33.de Solages C et al. High-frequency organization and synchrony of activity in the purkinje cell layer of the cerebellum. Neuron 58, 775–788 (2008). [DOI] [PubMed] [Google Scholar]
- 34.Wise AK, Cerminara NL, Marple-Horvat DE & Apps R Mechanisms of synchronous activity in cerebellar Purkinje cells. J. Physiol 588, 2373–2390 (2010). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 35.Harvey RJ & Napper RMA Quantitatives studies on the mammalian cerebellum. Prog. Neurobiol 36, 437–463 (1991). [DOI] [PubMed] [Google Scholar]
- 36.Napper RM & Harvey RJ Number of parallel fiber synapses on an individual Purkinje cell in the cerebellum of the rat. J. Comp. Neurol 274, 168–177 (1988). [DOI] [PubMed] [Google Scholar]
- 37.Isope P & Barbour B Properties of unitary granule cell-->Purkinje cell synapses in adult rat cerebellar slices. J. Neurosci 22, 9668–9678 (2002). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 38.Lanore F, Cayco-Gajic NA, Gurnani H, Coyle D & Silver RA Cerebellar granule cell axons support high-dimensional representations. Nat. Neurosci 24, 1142–1150 (2021). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 39.Dittman JS, Kreitzer AC & Regehr WG Interplay between facilitation, depression, and residual calcium at three presynaptic terminals. J. Neurosci 20, 1374–1385 (2000). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 40.Huang C-C et al. Convergence of pontine and proprioceptive streams onto multimodal cerebellar granule cells. Elife 2, e00400 (2013). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 41.Chabrol FP, Arenz A, Wiechert MT, Margrie TW & DiGregorio DA Synaptic diversity enables temporal coding of coincident multisensory inputs in single neurons. Nat. Neurosci 18, 718–727 (2015). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 42.Wagner MJ et al. Shared Cortex-Cerebellum Dynamics in the Execution and Learning of a Motor Task. Cell (2019) doi: 10.1016/j.cell.2019.02.019. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 43.Kuan AT et al. Dense neuronal reconstruction through X-ray holographic nano-tomography. Nat. Neurosci 23, 1637–1643 (2020). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 44.Zavitz D, Amematsro EA, Borisyuk A & Caron SJ C. Connectivity patterns that shape olfactory representation in a mushroom body network model. bioRxiv (2021) doi: 10.1101/2021.02.10.430647. [DOI] [Google Scholar]
- 45.Zheng Z et al. Structured sampling of olfactory input by the fly mushroom body. Curr. Biol 2020.04.17.047167 (2022). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 46.Guo C et al. Purkinje Cells Directly Inhibit Granule Cells in Specialized Regions of the Cerebellar Cortex. Neuron 91, 1330–1341 (2016). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 47.Rowan MJM et al. Graded Control of Climbing-Fiber-Mediated Plasticity and Learning by Inhibition in the Cerebellum. Neuron 99, 999–1015.e6 (2018). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 48.Kozareva V et al. A transcriptomic atlas of mouse cerebellar cortex comprehensively defines cell types. Nature 598, 214–219 (2021). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 49.Osorno T et al. Candelabrum cells are molecularly distinct, ubiquitous interneurons of the cerebellar cortex with specialized circuit properties. bioRxiv (2021). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 50.Gao Z et al. A cortico-cerebellar loop for motor planning. Nature 563, 113–116 (2018). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 51.Hua Y, Laserstein P & Helmstaedter M Large-volume en-bloc staining for electron microscopy-based connectomics. Nat. Commun 6, 7923 (2015). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 52.Phelps JS et al. Reconstruction of motor control circuits in adult Drosophila using automated transmission electron microscopy. Cell (2021) doi: 10.1016/j.cell.2020.12.013. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 53.Deerinck TJ, Bushong EA, Thor A & Ellisman MH NCMIR methods for 3D EM: a new protocol for preparation of biological specimens for serial block face scanning electron microscopy. Microscopy 6–8 (2010). [Google Scholar]
- 54.Saalfeld S, Cardona A, Hartenstein V & Tomancak P CATMAID: collaborative annotation toolkit for massive amounts of image data. Bioinformatics 25, 1984–1986 (2009). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 55.Funke J et al. Large Scale Image Segmentation with Structured Loss based Deep Learning for Connectome Reconstruction. IEEE Trans. Pattern Anal. Mach. Intell (2018). [DOI] [PubMed] [Google Scholar]
- 56.Sheridan A et al. Local Shape Descriptors for Neuron Segmentation. bioRxiv 2021.01.18.427039 (2021) doi: 10.1101/2021.01.18.427039. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 57.Falk T et al. U-Net: deep learning for cell counting, detection, and morphometry. Nat. Methods 16, 67–70 (2018). [DOI] [PubMed] [Google Scholar]
- 58.Lee K, Zung J, Li P, Jain V & Seung HS Superhuman accuracy on the SNEMI3D connectomics challenge. arXiv preprint arXiv:1706.00120 (2017). [Google Scholar]
- 59.Berg S et al. ilastik: interactive machine learning for (bio)image analysis. Nat. Methods 16, 1226–1232 (2019). [DOI] [PubMed] [Google Scholar]
- 60.Nguyen T, Malin-Mayor C, Patton W & Funke J Daisy: block-wise task dependencies for luigi. (2022).
- 61.Buhmann J et al. Automatic detection of synaptic partners in a whole-brain Drosophila electron microscopy data set. Nat. Methods 18, 771–774 (2021). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 62.Boergens KM et al. webKnossos: efficient online 3D data annotation for connectomics. Nat. Methods 14, 691–694 (2017). [DOI] [PubMed] [Google Scholar]
- 63.Scheffer LK et al. A connectome and analysis of the adult Drosophila central brain. Elife 9, (2020). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 64.Ester M, Kriegel H-P, Sander J & Xu X A density-based algorithm for discovering clusters in large spatial databases with noise. in Proceedings of the Second International Conference on Knowledge Discovery and Data Mining 226–231 (AAAI Press, 1996). [Google Scholar]
- 65.Hamming RW Error detecting and error correcting codes. The Bell System Technical Journal 29, 147–160 (1950). [Google Scholar]
- 66.Dhar M, Hantman AW & Nishiyama H Developmental pattern and structural factors of dendritic survival in cerebellar granule cells in vivo. Sci. Rep 8, 17561 (2018). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 67.Tyrrell T & Willshaw D Cerebellar cortex: its simulation and the relevance of Marr’s theory. Philos. Trans. R. Soc. Lond. B Biol. Sci 336, 239–257 (1992). [DOI] [PubMed] [Google Scholar]
- 68.Gundappa-Sulur G, De Schutter E & Bower JM Ascending granule cell axon: an important component of cerebellar cortical circuitry. J. Comp. Neurol 408, 580–596 (1999). [DOI] [PubMed] [Google Scholar]
- 69.Huang C-M, Wang L & Huang RH Cerebellar granule cell: ascending axon and parallel fiber. Eur. J. Neurosci 23, 1731–1737 (2006). [DOI] [PubMed] [Google Scholar]
- 70.Walter JT, Dizon M-J & Khodakhah K The functional equivalence of ascending and parallel fiber inputs in cerebellar computation. J. Neurosci 29, 8462–8473 (2009). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 71.Bradbury J et al. JAX: composable transformations of Python+NumPy programs. (2018).
- 72.Hennigan T, Cai T, Norman T & Babuschkin I Haiku: Sonnet for JAX. (2020).
- 73.Babuschkin I et al. The DeepMind JAX Ecosystem. (2020).
- 74.Ito M The molecular organization of cerebellar long-term depression. Nat. Rev. Neurosci 3, 896–902 (2002). [DOI] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Data Availability Statement
The EM dataset is publicly available at the BossDB (https://bossdb.org/) repository: https://bossdb.org/project/nguyen_thomas2022. Source data for Figs. 2c,e, 3c,d, and 4b,c,d are provided with the paper. Further directions for accessing neuron 3D renderings and derived neuron connectivity graphs are available at https://github.com/htem/cb2_project_analysis.