Abstract
The development of in vivo imaging and optogenetic tools makes it possible to control neural circuit activities in an all-optical, closed-loop manner, but such applications are limited by the lack of software for online analysis of neuronal imaging data. We developed an analysis software ORCA (Online Real-time activity and offline Cross-session Analysis), which performs image registration, neuron segmentation, and activity extraction at over 100 frames per second, fast enough to support real-time detection and readout of neural activity. Our active neuron detection algorithm is purely statistical, achieving a much higher speed than previous methods. We demonstrated closed-loop control of neurons that were identified on the fly, without prior recording or image processing. ORCA also includes a cross-session alignment module that efficiently tracks neurons across multiple sessions. In summary, ORCA is a powerful toolbox for fast imaging data analysis and provides a solution for all-optical closed-loop control of neuronal activity.
Keywords: calcium imaging, closed-loop optogenetic stimulation, image registration, neuron segmentation, fast image processing, image alignment
Introduction
Establishing a causal relationship between neural activity and behavior is central to understanding brain function, and this endeavor has been greatly facilitated by optogenetics (Boyden et al., 2005; Deisseroth et al., 2006; Lutz et al., 2008; Zhang et al., 2010; Packer et al., 2012; Yang and Yuste, 2018). Currently, most optogenetic experiments use pre-determined stimulation protocols, without considering the ongoing activity of relevant neurons. However, given that neuronal activities are highly heterogeneous—closely linked to brain states and behaviors—employing real-time activity-dependent stimulation protocols will better reveal the dynamic interactions between neural circuits and behavior (O’Connor et al., 2013; Grosenick et al., 2015; Peters et al., 2017; Prsa et al., 2017; Zhang et al., 2018; Robinson et al., 2020). Such “closed-loop experiments,” in which the input to the system (i.e., optogenetic stimulation) depends on the output (i.e., neural activity), are now within reach owing to optical imaging techniques for monitoring ongoing neural activities (Tian et al., 2009; Chen et al., 2013). However, accurately identifying active neurons in real-time is an open challenge.
In principle, identifying active neurons from raw imaging data requires two essential steps. The first is image registration, to compensate for the shift between image frames, for which several algorithms are available (Thevenaz et al., 1998; Guizar-Sicairos et al., 2008; Dubbs et al., 2016); in this paper, we further accelerated this process by optimization. The second step and the major hurdle is to identify active neurons from the registered movies of ongoing activity. The current methods that feature either dimensionality reduction or deep learning are such that they require pre-processing of imaging datasets spanning a relatively long time for effective neuronal identification (Pnevmatikakis et al., 2016; Pachitariu et al., 2017; Mitani and Komiyama, 2018; Giovannucci et al., 2019; Soltanian-Zadeh et al., 2019; Speen et al., 2019), and are therefore unsuitable for this purpose.
In addition, these imaging analysis pipelines cannot automatically track the same regions of interest (ROIs) across multiple sessions. Imaging the same population of neurons in vivo for an extended period of time is now possible with genetically encoded indicators and the implantation of chronic imaging windows or GRIN lenses (Holtmaat et al., 2009; Driscoll et al., 2017; Peters et al., 2017; Xin et al., 2019), but manually identifying the same neurons from multiple imaging sessions is time-consuming and error-prone. Additionally, some microscopes allow the user to rotate the optical axis. This feature adds flexibility to in vivo imaging, but at the same time exacerbates the difficulty of identifying the same neurons captured with slightly different angles, as rigid transformation alone cannot correct such distortions (Sheintuch et al., 2017; Giovannucci et al., 2019).
We envisioned that using statistical methods based on computing the temporal variance or deviation of the fluorescent signals of each pixel would allow us to achieve online identification of active neurons from streaming images. Pursuing this, we developed an image processing toolbox ORCA (Online Real-time activity and offline Cross-session Analysis), for fast image registration and online active neuron identification. ORCA also contains a cross-session alignment module for automatic tracking of neurons in long-term imaging (Figure 1). We demonstrate that ORCA can perform online identification of active neurons for closed-loop control, by identifying sound-responsive neurons in the mouse auditory cortex and selectively suppressing them with two-photon targeted optogenetic inhibition.
Results
Fast and Accurate Graphics Processing Unit-Based Image Registration
The first module, Image Registration, removes motion artifacts caused by brain pulsation and body movement. To achieve high processing speed, we implemented an algorithm to search for the optimal translation optimized for Graphics Processing Unit (GPU) acceleration (see section “Materials and Methods” for details). We compared the processing speed of ORCA with single-step DFT (Guizar-Sicairos et al., 2008), TurboReg (Thevenaz et al., 1998), and moco (Dubbs et al., 2016), using an example movie of 5,040 frames of 512×512 pixels, containing both high-frequency lateral shifts and gradual drifting in the X–Y plane (Figure 2A). ORCA outperforms the other methods, especially when running on GPU (Figure 2B), with comparable registration accuracy as shown by the magnified z-stack images (Figure 2C, top panels). Examination of the offset of each frame relative to the template image indicated that ORCA captures both small and fast shifts, and large and slow drifts (Figure 2C, bottom panels).
To quantitatively assess the registration accuracy of ORCA, a simulated movie was created by compiling a series of identical images with pseudo-random x and y shifts, termed “true shift values” (Figure 2D). Image registration on this movie was conducted using ORCA and the other three methods mentioned earlier. ORCA achieves similar or better performance, as quantified by the discrepancy between calculated shifts and the ground truth, suggesting that ORCA is a more efficient tool in image registration (Figures 2E,F).
Online Identification and Segmentation of Active Neurons
Most behavioral or neurophysiological experiments run in multiple trials. In these trial-based experiments, a typical trial lasts a few seconds, as does the inter-trial interval (ITI, Figure 3A). With a typical image acquisition rate of 10–30 frames/s, one trial generates a few dozen images. Adjusting optogenetic stimulation parameters based on neuronal activities in the preceding or the current trial requires identifying neurons and extracting their activities from a small set of images, and the computation must be completed within seconds. To satisfy such requirements, we developed two novel statistical algorithms based on computing temporal variance of fluorescent signals, termed algorithm #1 (Figures 3B,C) and algorithm #2 (Figures 3D,E), to identify active neurons within hundreds of milliseconds.
To demonstrate the online identification of active neurons, we performed in vivo two-photon calcium imaging in the mouse auditory cortex (ACx) and used ORCA to process the images. Each imaging trial lasted 4 s, including a 1-s “baseline” period and a 3-s “response” period. At an acquisition rate of 15 frames/s, each trial yielded 60 image frames. A 0.2-s tone was played after the baseline period (Figure 3A). As the mouse was anesthetized during imaging, shifting between frames was negligible during the 4-s imaging period, so we skipped image registration to further accelerate processing.
We used both algorithms of the Mask Segmentation module to process the images. Algorithm #1 first computed cumulative dF/F for each pixel in the response period, and F is defined as the average intensity of each pixel during the baseline period (Figure 3B, “cumulative dF/F,” see section “Materials and Methods” for details). To separate major activities from background fluctuations, the second step was auto-thresholding of cumulative dF/F by Renyi’s Entropy (Sahoo et al., 1997) (Figure 3B, “entropy”). To segment active neurons into ROIs, the third step was additional thresholding based on user-defined ROI size and fluorescence level (Figure 3B, “labeling”). The activity extraction module then extracted the activity trace of each identified ROI. The whole identification process took no more than 0.3 s (Figure 3C), well within the range of inter-trial intervals for sensory and behavioral experiments. The identified neuron mask and the activity traces of 9 out of 15 identified ROIs were shown in Figures 3F,G and can be exported to downstream control systems. The ROIs were sorted by their peak dF/F in descending order.
To further speed up the processing, we designed algorithm #2, which first computed the standard deviation of baseline activities of all pixels from 60 images (Figure 3D, “baseline,” see section “Materials and Methods” for details), and each pixel’s fluorescence level was thresholded by its baseline. Continuous above-threshold signals were exponentially amplified (Figure 3D, “amplification”). To segment active components into ROIs, we performed thresholding based on user-defined ROI size and signal level (Figure 3D, “labeling”), and extracted the activity trace of each identified ROI using the activity extraction module. The whole identification process took only 0.2 s (Figure 3E), even faster than algorithm #1. The identified neuron mask and the activity traces of 9 out of 20 identified ROIs were shown in Figures 3H,I, sorted by their peak dF/F.
To evaluate the identification accuracy of ORCA, we used algorithm #2 to process a publically available two-photon calcium imaging dataset from Svoboda Lab, Janelia Farm,1 with ground truth of labeled neurons also available. Because the dataset was not trial-based, we truncated it into 4-s long imaging windows as artificial “trials” (see section “Materials and Methods” for details). We ran ORCA on three trials and compared the results with ground truth (true labeled neurons with peak dF/F sorted from high to low, Supplementary Figure 1) and manual labeling (active ROIs identified by expert user, Supplementary Figure 2). With the default setting, ORCA achieved good performance on highly active neurons (Supplementary Figures 1C,F,I, 2C,F,I). However, ORCA missed some neurons with lower dF/F, which may be improved by fine-tuning parameters, and using a higher frame rate for imaging (see section “Materials and Methods” for details).
Offline Segmentation of Active Neurons for Imaging Sessions
For trial-based experiments, ORCA also provides a fast offline image analysis solution. An imaging session with hundreds of trials can be analyzed within minutes using the algorithms for online identification, with one extra step to integrate ROIs identified from each trial to form a unified mask for the whole session (Figure 4A). For non-trial-based experiments, we incorporated a published algorithm, HNCcorr (Speen et al., 2019), for ROI segmentation. HNCcorr can also be used for trial-based experiments. We compared the processing speed and identification accuracy of ORCA and HNCcorr in an example trial-based imaging session. In this experiment, ACx neurons were imaged while a series of different pure tones were played. From this session, ORCA and HNCcorr identified 40 and 28 ROIs, respectively, with 17 ROIs identified by both (Figures 4C,D and Supplementary Figures 1, 2). ORCA identified more ROIs than HNCcorr, including some ROIs that showed significant calcium activity (dF/F) but were missed by HNCcorr (Figures 4C,D, ORCA ROI 13 and 15; Supplementary Figure 3). ROIs that were only identified by HNCcorr but were missed by ORCA did not seem to have significant calcium activity (Figures 4C,D, HNCcorr ROI 10 and 11; Supplementary Figure 4). Thus, ORCA is more effective in identifying active ROIs. Furthermore, the processing speed of ORCA is over 1000 times faster than HNCcorr (Figure 4B, 12 s compared to 4 h, for 2,880 image frames).
Cross-Session Alignment for Repeated Imaging
One challenge for long-term imaging is to track the same neurons across multiple imaging sessions obtained over an extended period of time (Figure 5A). Different sessions may have slightly different imaging angles due to imperfect adjustment of the microscope objective’s orientation relative to the plane of the imaging window, which hampers the identification of the same neurons across sessions. Furthermore, manual tracking is time-consuming and prone to human inconsistency. To address these issues, ORCA uses affine transformation, a linear mapping method that deals with unidirectional distortion, to correct for the angle differences between sessions (section “Materials and Methods”). An example of two overlay imaging sessions of the same field of view (FOV) is shown in Figure 5B (left panel). The same neurons from different sessions may overlap completely or partially in the overlay image if they are in the center of the FOV or not overlap if located in the periphery (Figure 5B, center panel). Our algorithm successfully aligned all the cells (Figure 5B, right panel). Masks from individual sessions were merged into one unified mask with a “capture-all” strategy: all ROIs identified in any session were kept; for ROIs identified in multiple sessions with similar (Figure 5C, blue arrow and box) or different shapes (Figure 5C, green arrow and box), the shape in the unified mask is the union of the shapes across all sessions. After indexing all ROIs in all imaging sessions, the activity extraction module extracts and plots the neural responses across different sessions. As an example, we plotted the individual and averaged changes in calcium signals (ÄÄF/F) of one ROI in two imaging sessions responding to pure tones of different frequencies (Figure 5D).
A Demo for All-Optical Closed-Loop Control of Neuronal Activities
To demonstrate ORCA’s capacity for closed-loop photostimulation, we incorporated ORCA into a Thorlabs Bergamo II two-photon microscope with a spatial light modulator (SLM) for online manipulation of tone-responsive neurons in the mouse ACx (Figure 6A). We labeled ACx neurons with genetically encoded calcium indicator and optogenetic silencer by co-injecting AAV2/9-Syn-Cre, AAV2/9-hSyn-FLEX-GCaMP6s, and AAV2/9-hSyn-DIO-hGtACR1-mCherry viruses. A subset of neurons was colabeled with GCaMP6s and GtACR1 (Supplementary Figure 5). To avoid photo-stimulation of GtACR1 while performing two-photon calcium imaging, we chose a wavelength (880 nm) that could discriminate calcium-bound GCaMP6s from calcium-free GCaMP6s and did not excite GtACR1 ((Chen et al., 2013; Govorunova et al., 2015; Mohammad et al., 2017; Mardinly et al., 2018), section “Materials and Methods”).
We delivered a variety of pure tones (0.5 s each) and imaged neuronal calcium responses starting 1 s before the tone onset for a total of 4 s per tone presentation. The ITI was 1 s, common for trial-based sensory and behavioral experiments. We used ORCA (algorithm #1) to identify responsive ROIs and plot calcium responses online, before the onset of the next trial (Figure 6A). By loading identified ROIs into the stimulation software (ThorImage 4.3) before each stimulation trial, we demonstrated successful targeted optogenetic inhibition using SLM (Figures 6B,C and Supplementary Figure 5) in stimulated but not control trials. The identification was proven accurate as in the stimulated trials, only the targeted ROIs were effectively inhibited, leaving the activities of the untargeted ROIs intact (Figure 6C). Thus, we showed that ORCA is well-adapted to commercial imaging and stimulation systems for activity-based closed-loop control.
Discussion
Combining several novel algorithms, we developed ORCA, an imaging analysis toolbox that can process a small set of imaging data quickly for closed-loop neuronal stimulation, and perform cross-session image analysis for large datasets with high speed and accuracy. We validated our toolbox with calcium imaging experiments, but the software is in principle equally suitable for imaging studies using other fluorescent sensors, such as acetylcholine, dopamine, and serotonin sensors (Sun et al., 2018; Feng et al., 2019; Patriarchi et al., 2019; Wan et al., 2021).
To date, most optogenetic experiments are still open-loop, using stimulation parameters that are pre-determined rather than activity-guided. Only recently have neuronal activities been considered in guiding optogenetic manipulations, in the form of large-scale imaging data (Robinson et al., 2020) or real-time activities of pre-identified neurons (Zhang et al., 2018). Nonetheless, an imaging analysis pipeline that identifies active neurons from ongoing imaging experiments, essential for effective manipulation of highly dynamic and heterogeneous neuronal populations, is still lacking. Such a pipeline needs to complete raw image registration, active component identification, and activity extraction within a few seconds. Our GPU-based registration algorithm processes at >200 frames/s (512 × 512 pixels/frame). To our knowledge, only one method written in C++ achieves comparable speed (Mitani and Komiyama, 2018), but it requires an external OpenCV package, whereas our method is MATLAB-based and is simpler to run. For online active component identification, we designed two novel algorithms, with algorithm #1 more sensitive to active neurons with extremely low baseline fluorescence and algorithm #2 having a higher speed. Our method is distinguished from previous methods (Pnevmatikakis et al., 2016; Giovannucci et al., 2019; Soltanian-Zadeh et al., 2019), in that those methods require recording neuronal activity over an extended period and pre-processing before online segmentation, whereas our method can identify active neurons on the fly at >200 frames/s. This feature improves the flexibility of neuron selection, which is especially valuable for experiments involving one-trial learning such as fear conditioning and novelty detection.
Many studies have used long-term in vivo imaging to analyze the dynamics of the same population of neurons in multiple sessions. In such studies, neurons are typically tracked manually, which is error-prone, especially with imaging angle differences. ORCA is the first to use spatial features of the original FOV to align different sessions with possible imaging angle discrepancies. The angle difference up to 5° can be corrected. Notably, angle differences that are too large will result in data loss, especially in the margin, which cannot be fixed by post-processing. With its fast processing speed, the affine alignment function of ORCA can also be used for quality checks and calibration during image acquisition. Users can first take a small number of images, calculate the angle offset from previous sessions, and adjust the microscope accordingly to match previous recordings. As more and more microscopes feature flexible imaging angles, ORCA is a good complement and widely applicable to these systems.
Segmenting neurons from noisy in vivo imaging data is challenging. To achieve high accuracy, segmentation methods often have to sacrifice processing speed and rely on large quantities of imaging data. In all-optical closed-loop experiments, however, data analysis requires fast processing using only small datasets. Using novel statistical algorithms, ORCA bridges the gap and can be flexibly implemented into existing imaging systems to facilitate the application of closed-loop manipulations owing to its modular design.
Materials and Methods
Data Availability
The code used in this paper is available at https://github.com/YangYangLab/ORCA.
Animals
The C57BL/6 mice were purchased from Slac Laboratory Animals (Shanghai, China). Mice were housed and bred in a 12-h light-dark cycle (7 am to 7 pm light) in the animal facility of the ShanghaiTech University. Both male and female mice were used for the experiments. All procedures were approved by the Animal Committee of ShanghaiTech University.
Virus Injection
AAV2/9-Syn-Cre, AAV2/9-hSyn-Flpo-WPRE-pA, AAV2/9-hEFla-fDIO-GCaMP6s, AAV2/9-hSyn-FLEX-GCaMP6s, and AAV2/9-hSyn-DIO-hGtACR1-P2A-mCherry-WPRE-pA were purchased from Taitool Co., Shanghai, China. For virus injection, mice were anesthetized with isofluorane (induction, 4%; maintenance, 1–2%) and positioned onto a stereotaxic frame (Reward Co., Shanghai, China). Body temperature was maintained at 37°C using a heating pad. Viruses were injected using a glass micropipette with a tip diameter of 15–20 μm through a small skull opening (<0.5 mm2) with a micro-injector (Nanoject3, Drummond Scientific Co., Broomall, United States). Stereotaxic coordinates for auditory cortex (ACx): 2.46 mm posterior to the Bregma, 4.5 mm lateral from the midline, and 1.2 mm vertical from the cortical surface. For calcium imaging experiments, we mixed AAV2/9-hSyn-Flpo-WPRE-pA and AAV2/9-hEFla-fDIO-GCaMP6s with the final titer of 1.3 × 1012 and 2.1 × 1012 viral particles per ml, respectively. For closed-loop stimulation experiments, we mixed AAV2/9-Syn-Cre, AAV2/9-hSyn-FLEX-GCaMP6f-WPRE-pA, and AAV2/9-hSyn-DIO-hGtACR1-P2A-mCherry-WPRE-pA with the final titer of 6.4 × 109, 6.9 × 1012, and 1.4 × 1012 viral particles per ml, respectively. We injected a 0.2-μl virus mixture into the auditory cortex for all experiments and waited 3–4 weeks before two-photon imaging experiments.
Cranial Window Implantation
Mice were anesthetized with isoflurane (induction, 4%; maintenance, 1–2%) and positioned onto a stereotaxic frame (Reward Co.). Body temperature was maintained at 37°C using a heating pad. Lidocaine was administered subcutaneously. The muscle covering the auditory cortex was carefully removed with a scalpel. A 2 × 2 mm2 piece of the skull over ACx was removed, exposing the dura. The cranial window was sealed with a custom-made double-layered cover glass. UV-cure glue and dental acrylic were used to cement the cover glass onto the skull. A custom-made stainless steel head plate with a screw hole was embedded into the dental acrylic for head-fixed two-photon imaging.
Two-Photon Calcium Imaging and Optogenetic Stimulation Using the Spatial Light Modulator
Mice were injected with pentobarbital sodium (20 mg/kg) and head-fixed using the implanted head plate. Image series were taken at 15 Hz with a two-photon microscope (Bergamo II, Thorlabs, Newton, NJ, United States) equipped with a 25X/NA 1.05 objective (Olympus, Kyoto, Japan) and a Ti:sapphire laser (DeepSee, Spectra-Physics, Santa Clara, CA, United States) tuned to 880 nm. We chose this wavelength rather than the optimal wavelength for imaging GCaMP6s (930 nm) to avoid exciting GtACR1 during imaging. Mice were imaged when a series of pure tones (4, 12, and 20 kHz) were played. The sound was controlled using a Bpod state machine (Sanworks Co., Rochester, NY, United States) and open-source Bpod software (Sanworks Co.).
A fixed-wavelength laser (1,040 nm, Spectra-Physics, United States) and spatial light modulator (Thorlabs) were used for two-photon targeted optogenetic silencing. A cycle of 40 was used for inhibiting selected ROIs (0.1 s per cycle). The SLM was controlled by ThorImage (Thorlabs) and the control module of ORCA.
Implementation of Online Real-Time Activity and Offline Cross-Session Analysis to Microscope Software
We used ThorImage 4.0 (Thorlabs) for two-photon imaging and SLM stimulation. ORCA was implemented by adding an output interface to ThorImage 4.0 for acquiring two-photon imaging data, and an input interface to the stimulation module of ThorImage 4.0 for sending identified ROIs after image processing, with help from Thorlabs software engineers.
Hardware for Imaging Data Analysis
We used a Dell workstation to perform all computations. The workstation is equipped with two Intel Xeon Gold 5122 CPU, 64 GB DDR4 RAM, and an Nvidia Quadro P6000 GPU, running Linux Debian 10. ORCA uses GPU to accelerate most processing steps, while CPU versions are also provided in the package. We recommend the GPU version for best performance.
Preparation and Evaluation of Image Registration
To compare the speed of different algorithms, we performed registration in their default environments: single-step DFT and ORCA were tested in MATLAB R2020a (MathWorks, Natick, MA, United States), while TurboReg and moco in Fiji (ImageJ 1.48v, Java 1.6.0_24, 64bit). File read/write time was excluded for a better evaluation of the computational efficiency of each algorithm. Timing of TurboReg, single-step DFT, and ORCA started after files were loaded into the memory and stopped as soon as algorithms produced registered results. The timing of moco was evaluated using a screen recording software (Kazam Screencaster) and inspected manually.
To quantitatively compare different algorithms, we generated a simulated movie as follows: An example movie consisting of 1,200 frames was repeatedly registered using moco, TurboReg (accurate mode), and ORCA for 10 times each. This movie was then carefully examined by tracking major features in the images and measuring their movements frame-by-frame, to ensure it reached maximum possible stability. Then, periodical shifts caused by heartbeats were simulated by adding directional movements using the following equation:
where F indicated frames-per-second during acquisition, Mx the maximum shifts in x-direction, and i the current frame count. This kept all positive values in sx and set all negative offsets to zero. An additional random distortion was also introduced using pseudo-random numbers as offsets. Thus, shifts in x-direction, sx, became
where rand(−Mr, Mr) generated a pseudo-random sequence ranging from [−Mr, Mr].
True shift values were generated in both x and y directions in the same manner with varying parameters, and a simulated movie was tested among all four registration methods.
Registration accuracy was quantified as
where t was the true values and e was the estimated shifts generated by each algorithm. The registration accuracy was calculated separately for the x and y directions.
Image Registration Algorithms
Inspired by the moco algorithm (Dubbs et al., 2016), we combined multiple computational ideas to produce a faster implementation. Let a be an image frame of h rows (height) and w columns (width), and ai,j denote the pixel of the ith row, jth column. Let t be the template image (usually the first frame in an image series). Images t and a are standardized by mean-centering before registration:
Let Ms be the user-defined maximum shift value. We set its default value to 1/5 of the frame size. Users are advised to start with the default setting and adjust the parameter from there, as an appropriate choice of Ms is important for the best performance. An Ms too large (e.g., for a 512 × 512 pixel image, Ms = 170) may result in faulty matches, while an Ms too small (e.g., for a 512*512 image, Ms = 1) will cause the registration to fail, because the x–y shift between frames is likely to exceed 1 pixel.
We search for the optimal x, y offsets within the maximum shifts to minimize the difference between a and Tm in their overlapping region, defined as
Substituting the denominator (h−|x|)(w−|y|) with Ax,y, the area of overlap between the template and the target frame, we can rewrite the equation as:
in which Dx,y is the sum of all pixels squared in the overlapping area of the template, and i, j are corresponding indices. Given that x, y ∈ [−ms, ms]∩ℤ, there are finite possibilities of these values, and they can be calculated for any given x, y in the range beforehand.
To efficiently calculate these values, we register template t onto current frame a, so that ∑t2 only need to be calculated once throughout the whole registration process. ∑a2 and ∑a⋅t, are both related to the current frame a. Notice that, once we determined the maximum shift ms, only a central fraction of the original images are then registered to the template. To further accelerate the computation, we crop out the marginal areas of the original image, using only the central part to register against the full template. The cropped image, a′, should be
Where i′ ∈ [Ms,h−Ms], and j′ ∈ [Ms,w−Ms]. Thus we rewrite the equation into:
In this equation, a′ are the same cropped image for any valid x, y ∈ [−ms, ms]∩ℤ and the overlapping area A′ becomes static. The only dynamic part in the equation thus becomes ∑a⋅t. ∑a⋅t is the convolution of a and evaluated at (0,0). Letting be the rotation of t by 180 degrees, we have
Convolution can be efficiently calculated using two-dimensional fast Fourier transform (2D-FFT):
and thus we developed a highly optimized conv2 function (conv2_fftvalid.m in the source code) for our registration, which is much faster than MATLAB’s default implementation using basic addition and multiplication.
∑a2 can also be calculated as
where ones(h−2Ms, w−2Ms)) is a matrix of size (h−2Ms, w−2Ms) with all elements being 1, in the MATLAB notation.
The whole registration process can be further accelerated by downsampling and doing a local search after upsampling. Running on GPU, our code is much faster than most other registration algorithms.
Online Trial-By-trial Identification of Active Neurons
Algorithm #1
We combined several statistical tools for online segmentation of active ROIs on a trial-by-trial basis, each trial consisting of N frames of size (H, W) that constitute an image stack.
The dF/F (here denoted by for pixel (i, j)) is computed by subtracting the baseline intensity () of each pixel in the image stack and divided by :
where si, j(t) represents a pixel (i, j) of the tth image in the stack and tbl is the baseline period (defined by the user). Then, the dF/F of each pixel is summed up to form a two-dimensional time-average image. Values below zero are substituted by zeros.
Here, represents the cumulative dF/F for pixel (i, j) in the 2-D image. To increase the signal-to-noise ratio, we use the standard deviation of as a scaling factor, and multiply by :
The product matrix vi,jundergoes a filter that converts it to a binary image using Renyi Entropy-based auto thresholding (Sahoo et al., 1997) by calling the MIJ plugin in Fiji (Daniel Sage et al., 2012). The binary image is smoothed using a Gaussian filter to combine neighboring pixels. Two additional user-defined thresholds can be applied to determine whether to keep the identified ROIs (area threshold, default set at 16 pixels; intensity threshold, default set to 10). Pixels containing 1 (“positive pixels”) indicate a masked region.
Algorithm #2
The stack is first binarized to zeros and ones with a brightness detection threshold of 3-fold standard deviation over the average of baseline pixels:
where di,j represents the detection threshold, and χ is a step function representing the binarized stack, here we call binary sensitivity index. Fluorescent calcium signals generated by neural activities, rather than random noise, will be continuously above the threshold in the time domain, generating continuous ones in χi,j. On the other hand, random noises may go above and below the detection threshold in a stochastic manner, generating alternating ones and zeros. We, therefore, apply an exponential-like filter on this binary sensitivity index to amplify continuous ones:
Here, α, β are user-defined amplifying coefficients, and by default β = α−1. Continuous ones in the stack are amplified exponentially, while noise signals remained low in the temporal trace. This filter is then applied to the binary stack χt and summed up the results in the temporal trace to get a sensitivity index image L. This image L summarized the degree to which each pixel may contain fluorescent activity beyond noise level. We smoothed L with a gaussian (σ=1) convolution kernel of the cell-sized window and applied a threshold related to fluorescent indicator dynamics:
where F is the dynamic time of the fluorescent indicator used, which tells the algorithm how many continuous frames above the detection threshold should be considered a valid activity, and can be adjusted accordingly. Because the algorithm favors larger numbers of continuous above-threshold frames, it is much more suitable for imaging data acquired with high frame rates. k is a parameter that arbitrarily adjusts threshold T and is set to 0 by default. Connected areas in the final mask (using MATLAB bwconncomp function) after thresholding are used as estimates of areas with neural activity.
Benchmarking Using a Public Dataset
We downloaded a publically available two-photon imaging dataset from NeuroFinder2 (data from Svoboda Lab, Janelia Farm), and truncated 4-s imaging windows as artificial “trials” for online neuronal segmentation. We chose three “trials” for benchmarking, the criteria being: (1) the 1-s “baseline” is relatively flat and (2) there are some calcium activities in the 3-s “response” period. Since the ROIs of all labeled neurons were also available on the website, we extracted the dF/F of all labeled neurons and sorted them by peak dF/F (top 20 shown in Supplementary Figures 1C,F,I for three trials). For manual annotation, the first author of this paper hand-drew ROIs with identifiable fluorescence blind to the ground truth and ORCA identification results. ROIs were sorted by peak dF/F and the top 20 were shown in Supplementary Figures 2C,F,I. False-negative rate can also be reduced by fine-tuning the parameters, such as reducing F, or increasing the imaging frame rate (such as 15 fps or higher).
Offline Manual Segmentation of Components in a Movie
Since automated identification may falsely identify unwanted components or neglect true active neurons, we provide a user-friendly interface for manual labeling of active neurons in a given movie, incorporated in the mask_segmentation module. Our manual labeling program allows users to play the movie for full length or just one trial, and add or erase components.
Cross Session Alignment
For multi-session imaging data of one FOV, we use time-averaged images to represent the imaging sessions and define the time-average image of the first imaging session as the reference image. In three-dimensional space, we define the xy plane as the plane of the reference image, and z-axis perpendicular to xy. We then define the target image to be the time-average of a different imaging session. For sessions without rotation on the z-axis, displacements in x- and y-axis were corrected simply by moving target images in x and y directions for optimum alignment. For sessions with rotation, because the image plane preserves the collinearity of the FOV, displacements can be corrected using affine transformation. We first computed similarity between the reference image and the target image, then applied different transformations to the target image and search for maximum similarity between them. We used MATLAB function imregtform() with “rigid” or “affine” parameters and applied the transformation matrix by the function. We incorporated manual inspection and refinement by the users in this module. After user confirmation of the alignment results, masks with segmented ROIs from different sessions were combined to generate a superset of ROIs. These ROIs were then transformed to the original session based on respective transformation matrices. ROIs in the margins may be cropped out in the cross-session alignment.
Activity Extraction and Additional Thresholding
As ORCA extracts neuronal activity trial-by-trial, the user needs to supply information about the trial structure, in particular, timing parameters. The baseline activity F0 of each ROI is the average of all pixels over a user-defined time window before stimulus onset:
where ai is a pixel in the corresponding ROI and t ∈ [BaselineFrames]. Then, the normalized activity of all ROIs is calculated by subtracting and dividing baseline activities:
By default, an ROI is considered active if its peak activity rises above five standard deviations than the mean baseline activity (Dombeck et al., 2007; O’Connor et al., 2010). Active ROIs are then sorted based on their peak Factivity activities, and plotted or exported to a closed-loop controlling system.
Users can use other criteria for defining active ROIs, by (1) manually choosing a dF/F threshold, and ROIs with peak activity Factivity lower than this threshold will be discarded; or (2) manually setting a user-defined minimum ROI size to exclude smaller ROIs.
Data Availability Statement
The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.
Ethics Statement
The animal study was reviewed and approved by Animal Committee of ShanghaiTech University.
Author Contributions
WS and YY conceived the project. WS, XZ, and XH performed the experiments. WS developed the algorithms. WS and YY wrote the manuscript with contributions from all authors.
Conflict of Interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher’s Note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
Acknowledgments
We thank J. Lu, M. Ho, S.T. Han, and M. Tsakiris for critical reading and editing of the manuscript. We thank the imaging facility of the School of Life Sciences and Technology, ShanghaiTech University, for technical support. We also thank the funding agencies (2018YFC1005004 and NSFC31970960 to YY) for the support of the work.
Footnotes
Supplementary Material
The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fncel.2022.917713/full#supplementary-material
References
- Boyden E. S., Zhang F., Bamberg E., Nagel G., Deisseroth K. (2005). Millisecond-timescale, genetically targeted optical control of neural activity. Nat. Neurosci. 8 1263–1268. 10.1038/nn1525 [DOI] [PubMed] [Google Scholar]
- Chen T. W., Wardill T. J., Sun Y., Pulver S. R., Renninger S. L., Baohan A., et al. (2013). Ultrasensitive fluorescent proteins for imaging neuronal activity. Nature 499 295–300. 10.1038/nature12354 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Daniel Sage D. P., Tinevez J.-Y., Schindelin J. (2012). “MIJ: Making Interoperability Between ImageJ and Matlab Possible,” in Proceedings of the ImageJ User & Developer Conference, Luxembourg. [Google Scholar]
- Deisseroth K., Feng G. P., Majewska A. K., Miesenbock G., Ting A., Schnitzer M. J. (2006). Next-generation optical technologies for illuminating genetically targeted brain circuits. J. Neurosci. 26 10380–10386. 10.1523/JNEUROSCI.3863-06.2006 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Dombeck D. A., Khabbaz A. N., Collman F., Adelman T. L., Tank D. W. (2007). Imaging large-scale neural activity with cellular resolution in awake, mobile mice. Neuron 56 43–57. 10.1016/j.neuron.2007.08.003 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Driscoll L. N., Pettit N. L., Minderer M., Chettih S. N., Harvey C. D. (2017). Dynamic reorganization of neuronal activity patterns in parietal cortex. Cell 170 986–999.e16. 10.1016/j.cell.2017.07.021 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Dubbs A., Guevara J., Yuste R. (2016). moco: Fast Motion Correction for Calcium Imaging. Front. Neuroinform. 10:6. 10.3389/fninf.2016.00006 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Feng J., Zhang C. M., Lischinsky J. E., Jing M., Zhou J. H., Wang H., et al. (2019). A genetically encoded fluorescent sensor for rapid and specific in vivo detection of norepinephrine. Neuron 102 745–761.e8. 10.1016/j.neuron.2019.02.037 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Giovannucci A., Friedrich J., Gunn P., Kalfon J., Brown B. L., Koay S. A. (2019). CaImAn an open source tool for scalable calcium imaging data analysis. eLife 8:e38173. 10.7554/eLife.38173 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Govorunova E. G., Sineshchekov O. A., Janz R., Liu X., Spudich J. L. (2015). Natural light-gated anion channels: a family of microbial rhodopsins for advanced optogenetics. Science 349 647–650. 10.1126/science.aaa7484 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Grosenick L., Marshel J. H., Deisseroth K. (2015). Closed-loop and activity-guided optogenetic control. Neuron 86 106–139. 10.1016/j.neuron.2015.03.034 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Guizar-Sicairos M., Thurman S. T., Fienup J. R. (2008). Efficient subpixel image registration algorithms. Opt. Lett. 33 156–158. 10.1364/OL.33.000156 [DOI] [PubMed] [Google Scholar]
- Holtmaat A., Bonhoeffer T., Chow D. K., Chuckowree J., De Paola V., Hofer S. B., et al. (2009). Long-term, high-resolution imaging in the mouse neocortex through a chronic cranial window. Nat. Protoc. 4 1128–1144. 10.1038/nprot.2009.89 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Lutz C., Otis T. S., DeSars V., Charpak S., DiGregorio D. A., Emiliani V. (2008). Holographic photolysis of caged neurotransmitters. Nat. Methods 5 821–827. 10.1038/nmeth.1241 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Mardinly A. R., Oldenburg I. A., Pegard N. C., Sridharan S., Lyall E. H., Chesnov K., et al. (2018). Precise multimodal optical control of neural ensemble activity. Nat. Neurosci. 21 881–893. 10.1038/s41593-018-0139-8 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Mitani A., Komiyama T. (2018). Real-time processing of two-photon calcium imaging data including lateral motion artifact correction. Front. Neuroinform. 12:98. 10.3389/fninf.2018.00098 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Mohammad F., Stewart J. C., Ott S., Chlebikova K., Chua J. Y., Koh T. W., et al. (2017). Optogenetic inhibition of behavior with anion channelrhodopsins. Nat. Methods 14 271–274. 10.1038/nmeth.4148 [DOI] [PubMed] [Google Scholar]
- O’Connor D. H., Hires S. A., Guo Z. V., Li N., Yu J., Sun Q. Q., et al. (2013). Neural coding during active somatosensation revealed using illusory touch. Nat. Neurosci. 16 958–965. 10.1038/nn.3419 [DOI] [PMC free article] [PubMed] [Google Scholar]
- O’Connor D. H., Peron S. P., Huber D., Svoboda K. (2010). Neural activity in barrel cortex underlying vibrissa-based object localization in mice. Neuron 67 1048–1061. 10.1016/j.neuron.2010.08.026 [DOI] [PubMed] [Google Scholar]
- Pachitariu M., Stringer C., Dipoppa M., Schröder S., Rossi L. F., Dalgleish H., et al. (2017). Suite2p: beyond 10,000 neurons with standard two-photon microscopy. bioRxiv [Preprint]. 10.1101/061507 [DOI] [Google Scholar]
- Packer A. M., Peterka D. S., Hirtz J. J., Prakash R., Deisseroth K., Yuste R. (2012). Two-photon optogenetics of dendritic spines and neural circuits. Nat. Methods 9 1202–U103. 10.1038/nmeth.2249 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Patriarchi T., Cho J. R., Merten K., Marley A., Broussard G. J., Liang R., et al. (2019). Imaging neuromodulators with high spatiotemporal resolution using genetically encoded indicators. Nat. Protoc. 14 3471–3505. 10.1038/s41596-019-0239-2 [DOI] [PubMed] [Google Scholar]
- Peters A. J., Lee J., Hedrick N. G., O’Neil K., Komiyama T. (2017). Reorganization of corticospinal output during motor learning. Nat. Neurosci. 20 1133–1141. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Pnevmatikakis E. A., Soudry D., Gao Y., Machado T. A., Merel J., Pfau D., et al. (2016). Simultaneous denoising, deconvolution, and demixing of calcium imaging data. Neuron 89 285–299. 10.1016/j.neuron.2015.11.037 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Prsa M., Galinanes G. L., Huber D. (2017). Rapid Integration of Artificial Sensory Feedback during Operant Conditioning of Motor Cortex Neurons. Neuron 93 929–939.e6. 10.1016/j.neuron.2017.01.023 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Robinson N. T. M., Descamps L. A. L., Russell L. E., Buchholz M. O., Bicknell B. A., Antonov G. K., et al. (2020). Targeted activation of hippocampal place cells drives memory-guided spatial behavior. Cell 183 2041–2042. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Sahoo P., Wilkins C., Yeager J. (1997). Threshold selection using Renyi’s entropy. Pattern Recognit. 30 71–84. [Google Scholar]
- Sheintuch L., Rubin A., Brande-Eilat N., Geva N., Sadeh N., Pinchasof O., et al. (2017). Tracking the Same Neurons across Multiple Days in Ca2+ Imaging Data. Cell Rep. 21 1102–1115. 10.1016/j.celrep.2017.10.013 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Soltanian-Zadeh S., Sahingur K., Blau S., Gong Y. Y., Farsiu S. (2019). Fast and robust active neuron segmentation in two-photon calcium imaging using spatiotemporal deep learning. Proc. Natl. Acad. Sci. U.S.A. 116 8554–8563. 10.1073/pnas.1812995116 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Speen Q., Asin-Acha R., Chettih S. N., Minderer M., Harvey C., Hochbaum D. S. (2019). HNCcorr: a novel combinatorial approach for cell identification in calcium-imaging movies. Eneuro 6:ENEURO.0304-18.2019. 10.1523/ENEURO.0304-18.2019 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Sun F. M., Zeng J. Z., Jing M., Zhou J. H., Feng J. S., Owen S. F., et al. (2018). A genetically encoded fluorescent sensor enables rapid and specific detection of dopamine in flies, fish, and mice. Cell 174 481–496.e19. 10.1016/j.cell.2018.06.042 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Thevenaz P., Ruttimann U. E., Unser M. (1998). A pyramid approach to subpixel registration based on intensity. IEEE Trans. Image Process. 7 27–41. [DOI] [PubMed] [Google Scholar]
- Tian L., Hires S. A., Mao T., Huber D., Chiappe M. E., Chalasani S. H., et al. (2009). Imaging neural activity in worms, flies and mice with improved GCaMP calcium indicators. Nat. Methods 6 875–113. 10.1038/nmeth.1398 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Wan J., Peng W., Li X., Qian T., Song K., Zeng J., et al. (2021). A genetically encoded sensor for measuring serotonin dynamics. Nat. Neurosci. 24 746–752. 10.1038/s41593-021-00823-7 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Xin Y., Zhong L., Zhang Y., Zhou T. T., Pan J. W., Xu N. L. (2019). Sensory-to-category transformation via dynamic reorganization of ensemble structures in mouse auditory cortex. Neuron 103 909–921.e6. 10.1016/j.neuron.2019.06.004 [DOI] [PubMed] [Google Scholar]
- Yang W. J., Yuste R. (2018). Holographic imaging and photostimulation of neural activity. Curr. Opin. Neurobiol. 50 211–221. 10.1016/j.conb.2018.03.006 [DOI] [PubMed] [Google Scholar]
- Zhang F., Gradinaru V., Adamantidis A. R., Durand R., Airan R. D., de Lecea L., et al. (2010). Optogenetic interrogation of neural circuits: technology for probing mammalian brain structures. Nat. Protoc. 5 439–456. 10.1038/nprot.2009.226 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Zhang Z., Russell L. E., Packer A. M., Gauld O. M., Hausser M. (2018). Closed-loop all-optical interrogation of neural circuits in vivo. Nat Methods 15 1037–1040. 10.1038/s41592-018-0183-z [DOI] [PMC free article] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Data Availability Statement
The code used in this paper is available at https://github.com/YangYangLab/ORCA.
The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.