Abstract
Background
Manual axon histomorphometry (AH) is time- and resource-intensive, which has inspired many attempts at automation. However, there has been little investigation on implementation of automated programs for widespread use. Ideally such a program should be able to perform AH across imaging modalities and nerve states. AxonDeepSeg (ADS) is an open source deep learning program that has previously been validated in electron microscopy. We evaluated the robustness of ADS for peripheral nerve axonal histomorphometry in light micrographs prepared using two different methods.
Methods
Axon histomorphometry using ADS and manual analysis (gold-standard) was performed on light micrographs of naïve or regenerating rat median nerve cross-sections prepared with either toluidine-resin or osmium-paraffin embedding protocols. The parameters of interest included axon count, axon diameter, myelin thickness, and g-ratio.
Results
Manual and automatic ADS axon counts demonstrated good agreement in naïve nerves and moderate agreement on regenerating nerves. There were small but consistent differences in measured axon diameter, myelin thickness and g-ratio; however, absolute differences were small. Both methods appropriately identified differences between naïve and regenerating nerves. ADS was faster than manual axon analysis.
Conclusions
Without any algorithm retraining, ADS was able to appropriately identify critical differences between naïve and regenerating nerves and work with different sample preparation methods of peripheral nerve light micrographs. While there were differences between absolute values between manual and ADS, ADS performed consistently and required much less time. ADS is an accessible and robust tool for AH that can provide consistent analysis across protocols and nerve states.
Introduction
Axon histomorphometry (AH) is the most commonly used outcome measure in nerve research. It involves axon quantification and measurement of axonal micro-structure parameters on nerve cross-sections [1,2]. Manual measurements have traditionally been the gold-standard for AH, but this approach is both time- and resource-intensive and has known limitations related to differing sample preparation protocols and inter-rater reliability [1,2].
Axon histomorphometry is used in studying both the central and peripheral nervous systems, but recent advances in peripheral nerve regeneration have increased its use as an outcome measure. There are now countless techniques and therapeutics being tested to improve peripheral nerve regeneration, including three-dimensional printing of nerve guides, nerve wraps, electrical stimulation, and growth hormone [3–6]. This increased interest on nerve regeneration has highlighted the need for accessible, streamlined, and robust methods of performing AH.
The recognized limitations of manual AH have inspired many tools to achieve some degree of automation [7–9]. A major obstacle to full automation using traditional programming methods is the difficulty of achieving axonal segmentation; that is, the process by which the boundaries of the myelin sheath surrounding individual axons are defined so that each axon can be measured separately [7,10,11]. To illustrate the challenge of axonal segmentation, axons in undamaged nerves reside in close proximity with myelin sheaths tending to abut each other; in comparison, debris is frequently observed in regenerating nerves that can be mislabeled as an axon. These and other nuances need to be carefully defined to improve axonal segmentation, but manual correction is still required due to frequent errors that can affect downstream calculations and substantially reduce measurement accuracy.
Machine learning has the potential to address the barriers to automating AH. Unlike the rigid design and inflexibility of traditional programs, machine learning algorithms iteratively create a set of rules by training on a large dataset [11,12]. Deep learning, also known as computer vision, is a subset of machine learning that can use convolutional neural networks, where each network hierarchically defines specific features of images and does not require structured numerical input data [13,14]. Various programs have leveraged deep learning for nerve analysis, but their generalizability has not been investigated for implementation outside of the research groups in which they were developed [8–11,14–16] and this has limited their widespread use.
AxonDeepSeg (ADS) is a novel deep learning program for fully-automated AH originally trained and validated using both scanning- and transmission-electron micrographs of the brain and spinal cord in mice, macaques, and humans [11]. Given the increased accessibility of light microscopy, and its prevalent use in the field of peripheral nerve research, the purpose of the present study was to validate the performance of ADS “out-of-the-box” in peripheral nerve light micrographs to determine if it can be implemented reliably without retraining. Given the robustness of its deep learning algorithm, we hypothesized that ADS could perform AH on light micrographs of peripheral nerves across different tissue preparation protocols and nerve states. To test this hypothesis, we performed a validation study using adult rat median nerve preparations, comparing ADS to the gold-standard of manual axon quantification and critical parameter measurements.
Materials and methods
Animals
This study was carried out in accordance with National Institutes of Health recommendations in the Guide for the Care and Use of Laboratory Animals. The protocol was approved by the Johns Hopkins University Animal Care and Use Committee (Protocol number: RA18M74). All surgery was performed under isoflurane anesthesia with pre- and post-procedure subcutaneous buprenorphine analgesia. We used four adult male Lewis rats (Envigo, Frederick, MD), 12 to 24 weeks old and weighing 300 to 400 g. Two animals had no intervention prior to median nerve harvest. The other two underwent cut and repair of the proximal aspect of the median nerve at the level of the pectoralis and were sacrificed 16 weeks later. There were no specific exclusion criteria and no animals were excluded from the analysis.
Tissue harvest
The rats were anesthetized with isoflurane prior to exposure of the heart for cardiac perfusion. Each rat was perfused through the left ventricle with 200 mL of phosphate buffered saline (PBS) followed by 200 mL of 4% paraformaldehyde (Sigma Aldrich, St Louis, MO) in PBS. Unilateral median nerves were exposed through a volar approach and dissected free of surrounding tissue under a dissection microscope. To compare embedding protocols, in the two naïve rats, 10 mm samples of each nerve were collected at the mid-humerus level. These were further cut into two 5 mm samples, each of which was allocated to one of two processing protocols: osmium-paraffin or toluidine-resin. For the regenerating samples, the median nerve was harvested at the level of the mid-forearm and nerve was processed in osmium-paraffin only.
Tissue preparation
For this study, we compared two staining and embedding protocols as our lab commonly uses both for our peripheral nerve analysis. Toluidine blue staining and resin embedding has traditionally been used for both light and electron microscopy. This process produces very high quality micrographs, but requires a specialized set up for embedding and sectioning and can be cost prohibitive [17]. Osmium staining with paraffin embedding is a cost-effective solution that has increased accessibility in regards to time and equipment while still providing suitable image quality for analysis using light microscopy [17].
Toluidine blue with resin embedding
One 5 mm sample from each naïve nerve was fixed for 48 hours at 4˚ C in a solution of 2% glutaraldehyde (16216, EMS), 3% paraformalydehyde (15754-S, EMS) and 0.1 M Sorensen’s phosphate buffer pH 7.2 solution. Specimens were post-fixed in 2% osmium tetroxide, dehydrated in ascending alcohol series (starting at 50%) and embedded in Araldite® 502 resin (Polyscience). Semi-thin (1 μm) tissue sections were cut on an Ultracut E microtome (Reichert Inc, Buffalo, NY) and stained with 1% Toluidine blue.
Osmium tetroxide staining only with paraffin embedding
This protocol is based on a previously published staining and embedding procedure with all steps carried out at 4° C [17]. One 5 mm sample from each nerve was fixed in a 4% paraformaldehyde in PBS solution for 2 hours. Samples were then placed in two serial washes of PBS solution with gentle agitation for 30 seconds each prior to being submerged for 2 hours in 2% osmium solution (RT 19172, EMS). Thereafter, the washing step was repeated, and samples were stored in 0.2% glycine (Sigma) in PBS solution until embedding. Samples were processed by graded ethanol dehydration (starting at 30%), cleared with Pro-Par (Anatech Ltd) and infiltrated with paraffin. After processing, samples were embedded in paraffin, then cut with a microtome into 7 μm thick sections.
Imaging
Two representative slices of naïve median nerve (one from each animal) were chosen from each of the two staining/embedding protocols (toluidine-resin, osmium-paraffin), and two representative slices of regenerating nerve (one from each animal) in osmium-paraffin were chosen. Microscopy was performed using a Zeiss Axioplan 2 (Carl Zeiss Microscopy LLC, White Plains, NY) with a 100x oil lens (numerical aperture 1.30) and digital images were captured using a Jenoptik ProgRes C5 camera (Jupiter, FL) mounted to the microscope. At this magnification, a single nerve slice produced 12 to 17 micrograph segments.
Manual analysis
For manual analysis, the 12 to 17 non-overlapping micrograph segments for each nerve slice were analyzed individually by a blinded assessor. These micrograph segments (.tiff files, 2580x1944 pixels, 15 MB) were imported into ImageJ (FIJI Package, version 2.0, NIH, Bethesda, MD) and, per stereological principles, at the center of each micrograph segment we sampled a box measuring either 25x25 μm (625 μm2; naïve nerves) or 40x40 μm (1600 μm2; regenerating nerves) (Fig 1) [18]. This method created non-overlapping sample areas across the entire nerve. Axons touching the top and right borders of the box were included in the measurements while those touching the bottom and left borders were not [1]. The larger sampling area was used for regenerating nerves due to their lower axon density. If the sampled area was at an edge, and axons only covered a portion of the area, a multiplication factor was applied to ensure that the counts were comparable across all samples from the same nerve. Thresholding to create a black and white image was performed manually so that small axons were captured while simultaneously trying to avoid artificially thickening myelin sheaths. For fibers where myelin touched, lines were drawn between them so that they would be individually segmented. Counting was performed manually and then axon diameter and myelin thickness were individually measured using the “minferet” command. G-ratio (axon diameter divided by fiber diameter) was calculated from these values. “Minferet” is the minimum caliper diameter of Feret’s diameter, which is the longest distance between any two points along the selection boundary, it is analogous to the minor axis of an ellipse.
Automated analysis
AxonDeepSeg (ADS) is an open source program created using the Python coding language, and its development has been previously described [11]. In brief, deep learning models were trained from a dataset containing transmission electron micrographs with different acquisition resolutions in order to increase variability and improve generalization. All samples were from the central nervous system (brain and spinal cord) of mice, rats, and humans. The pipeline of ADS development consisted of four steps: data preparation, learning, evaluation, and prediction. Ground truth labeling for segmentation was created using the image processing software GIMP (https://www.gimp.org/) and was cross-checked by at least two researchers. The deep learning architecture is based on the U-net, which combined a contracting path with traditional convolutions and then an expanding path with up-convolutions. This allowed for prediction of 3 classes: axon, myelin, and background. To avoid border effects during prediction, inference is run on a larger sample square but only calculates within a smaller square, then iterates over the entire image by shifting this sample square.
For our study, the same .tiff images used in the manual analysis were imported into ADS (version 2.1, with FSLeyes extension) and the resolution was entered (0.05 micrometer per pixel). The transmission electron microscopy pre-set was chosen because it has dark myelin and lighter axons, similar to light microscopy. Segmentation was performed automatically. Though manual correction is easily performed in the graphic user interface of ADS, to assess the accuracy of the automatic segmentation we refrained from correction beyond removal of segmented connective tissue (Fig 2). ADS then computed axon measurements for the entire micrograph and created a .csv file of each axon, its coordinates, and its respective measurements. To match the stereology sampling used in manual analysis, only axons within the same coordinates as in the manual analysis were included in our results.
Statistical analysis
Data analyses were conducted using R (R Core Team, 3.6.1). Descriptive statistics were calculated for axon count, myelin thickness, and g-ratio. The primary outcome of comparability of axon counts was evaluated using Bland-Altman plots and intraclass correlation (ICC). The Bland-Altman plots show the difference between each value obtained from ADS and manual measurement, plotted against the mean of each measurement pair [19]; 95% of differences are expected to be contained within 1.96 standard deviations of the mean difference (i.e., the “limits of agreement”), otherwise fixed measurement bias must be ruled out. Secondary outcomes of myelin thickness, axon diameter, and g-ratio were compared using two-sample t-tests and Cohen’s d for effect size. A P-value of < 0.05 was considered statistically significant. Cohen’s d effect size magnitudes were defined as: negligible (< 0.2), small (< 0.5), medium (< 0.8), and large (> 0.8).
Results
The average time for ADS to segment a 15 MB tiff file measuring 2580x1944 pixels was 20 seconds (2018 MacBook Pro, 3.1 GHz Intel Core i5, 8 GB 2133 MHz LPDDR3 RAM), with an additional minute for minimal manual correction and 1 second for measurement output. Average time to import the same image into ImageJ, perform manual thresholding, counting and measurement for a single micrograph was 5 minutes. In total there were 27 micrographs of the two toluidine-resin naïve nerves (12 and 15), 27 micrographs of the two osmium-paraffin naïve nerves (13 and 14), and 33 micrographs of the two osmium-paraffin regenerating nerves (16 and 17).
Axon count
Average axon count in the 25x25 μm sample areas of the 27 toluidine-resin micrographs for the two naïve nerves was 12.31 ± 4.12 when counted manually and 11.79 ± 3.8 (mean ± standard deviation) in the ADS output (Fig 3A). Bland-Altman showed limits of agreement ranging from -5 to 5 and the ICC was 0.76 (P < 0.001; Fig 4A).
Axon count in the same sized sample area for the 27 micrographs osmium-paraffin protocol naïve nerves was 13.72 ± 5.19 when counted manually, and was 13.72 ± 5.78 in the ADS output (Fig 3B). Bland-Altman comparison showed good agreement between the counts, with limits of agreement ranging from -6 to 6. ICC was 0.84 (P < 0.001), indicating good to excellent agreement (Fig 4B). In the regenerating nerves, the average axon count per 40x40 μm sample area was 15.96 ± 7.48 when counted manually and 11.09 ± 5.74 in the ADS output (Fig 3C). Bland-Altman limits of agreement were -2 to 12 and the ICC was 0.56 (P <0.001; Fig 4C).
Measurements
In the naïve nerve toluidine-resin protocol samples, there was a small difference in measured axon diameters, with larger measurements in ADS compared to manual (3.67 ± 1.92 vs. 3.06 ± 1.71, t = 4.19, P < 0.001; d = 0.34; Fig 5A). Myelin was thicker in the ADS compared to manual measuring (1.16 ± 0.52 vs. 1.06 ± 0.46, t = 2.63, P = 0.0089; d = 0.21; Fig 6A). Thus g-ratio (axon diameter divided by fiber diameter) was larger in ADS (0.59 ± 0.12 vs. 0.57 ± 0.11, t = -2.22, P = 0.027; d = 0.18; Fig 7A).
In naïve nerve osmium-paraffin protocol samples, there was no difference in measured axon diameter measured by ADS or manually (2.45 ± 1.16 vs. 2.47 ± 1.13, t = -0.29, P = 0.78; Fig 5B). Myelin was measured as thicker in ADS compared to manual measurements (1.54 ± 0.45 vs. 1.35 ± 0.54, t = 4.86, P < 0.001; d = 0.38; Fig 6B), and this difference was consistent across micrographs. In turn, ADS had a smaller g-ratio, defined as axon diameter divided by fiber diameter (0.425 ± 0.10 vs. 0.474 ± 0.10, t = -6.28, P < 0.001; d = 0.49; Fig 7B).
In regenerating nerves (osmium-paraffin protocol), axon diameters were notably smaller than in the naïve nerves in both ADS (regenerating vs. naïve: 1.84 ± 1.00 vs. 2.45 ± 1.16, t = 7.13, P < 0.001, d = 0.56) and manual measurements (1.27 ± 0.87 vs. 2.47 ± 1.13, t = 16.3, P <0.001, d = 1.23). ADS measured axon diameter as larger compared to manual measurements (1.84 ± 1.00 vs. 1.27 ± 0.87, t = 8.46, P < 0.001; d = 0.62; Fig 5C). Unlike in naïve samples, ADS measured myelin thickness as less than manual measuring (0.86 ± 0.30 vs. 1.06 ± 0.32, t = -8.56, P < 0.001; d = -0.60; Fig 6C). This created a significantly larger calculated g-ratio in the ADS data (0.50 ± 0.16 vs. 0.39 ± 0.14, t = 13.8, P < 0.001; d = 1.0; Fig 7C).
Discussion
We aimed to validate the use of the newly developed deep learning program AxonDeepSeg (ADS) for automated axon histomorphometry (AH) and found that even without algorithm retraining, ADS was able to appropriately identify critical differences between naïve and regenerating nerves and work in different sample preparation methods. AH is one of the most important measures in peripheral nerve research, yet it is a time- and resource-intensive task, which limits its accessibility as an outcome measure. While there has been a boom in programs created for AH, none of these have been validated for use outside of the specific protocols for which they were developed and therefore widespread implementation has been poor.
The ADS deep learning algorithm was originally trained and validated using electron micrographs, but given the increased accessibility of light microscopy, we tested its application in light micrographs of peripheral nerves [11]. We found that ADS was able to perform near automatic AH of peripheral nerve light micrographs with results that correlated well to manual analysis but was much faster. Axon quantification in naïve nerves in both sample preparation protocols showed very good agreement with good interclass correlation between the manual and ADS counts. In regenerating nerves, the agreement was moderate, with ADS consistently counting fewer axons. Though ADS allows for manual correction of, we refrained from significant correction apart from removal of obvious connective tissue. More rigorous correction likely would have improved axon counts as it would have further minimized segmentation errors but it would be unlikely to change measurements (i.e. diameter, myelin thickness) as these were consistent within micrographs.
One of the key differences between ADS and manual axon quantification is that ADS counted and measured every axon within a micrograph. With our manual analysis we used a stereologic sampling approach where only axons within a 25x25 μm (naïve) or 40x40 μm (regenerating) box at the center of each micrograph segment was counted, resulting in 12–17 sampled areas for each nerve. This approach is widely used in AH, in order to expedite time-consuming manual measurements (REF). Using a stereologic sampling approach, however, increases the risk of sampling error and bias as a difference in that small sample area can get compounded, leading to a significant difference in the number of total axons. For our comparisons we specifically analyzed only the ADS output for the same coordinates as in our manual analysis; however, ADS can evaluate the entirety of each micrograph rapidly which negates the need for stereologic sampling. It therefore does not need to be limited by local variation in axon density, and can avoid the error and bias propagation inherent in extrapolating from a stereologic approach.
Axon diameter, myelin thickness, and g-ratio are important AH parameters as they differentiate subpopulations of axons and can be used for evaluating changes following denervation and reinnervation [20]. We found that ADS was able to appropriately identify differences in axon diameter, myelin thickness and g-ratio between naïve and regenerating nerves. Though there were small differences between ADS and manual measurements, they were in a consistent direction. With our large dataset, this resulted in statistically significant differences even though the absolute differences were small. The absolute values of these parameters are known to differ between research labs due to differences in staining and imaging protocols [9,10], and the judgment of the individual raters doing the segmentations. Therefore, consistency within studies is most important, and ADS was able to provide reliable measurements with comparable variability as manual analysis.
Conventional machine learning has been applied to axon histomorphometry in the past in the form of random forest and bootstrapping methods, yet such techniques were still not robust enough to handle variations in these data [8,9,21]. More recently, other research groups have started applying deep learning to axon histomorphometry, these all follow the basic tenants of developing a deep learning program but differ in their training datasets and network structure. Naito et al. [10] developed a deep learning algorithm using a database of light micrographs of diseased human sural nerves. Their algorithm was trained using these light micrographs but was not tested outside of their initial database. Janjic et al. [15] developed a segmentation program using electron micrographs of human prefrontal cortex. While both of these programs showed excellent agreement when compared to manual analysis of micrographs from the same dataset, the robustness of the algorithms was not tested.
A known limitation of all machine learning programs is that accuracy and robustness are dependent on the images used to develop the model, what is known as the “ground truth”. This presents a problem for AH since as previously discussed, micrographs can vary depending on nerve state and sample processing. Re-training an algorithm model can provide more accurate results, at the cost of having to re-train a model, which is time consuming, requires expertise, and still lacks generalizability [22].
Besides leveraging deep learning, ADS has a number of advantages over other programs that have been developed for axon histomorphometry [9–11,23]. ADS is open source and freely available for download from GitHub (https://github.com/neuropoly/axondeepseg). It was developed using Python, a commonly used programming language, and remains under continuous active development more than 2 years after its initial release. It has a graphical user interface and does not require advanced programming knowledge to use. Our results demonstrate that the current ADS algorithm is very robust across multiple nerve states and sample processing protocols.
Deep learning and computer vision programs are an important step forward in the analysis of outcomes measures in nerve research. The speed and reliability of these methods promise to be substantially improved over traditional manual measurement. Our work shows that in its current iteration, ADS is an accessible and robust tool for peripheral nerve AH across multiple image preparation modalities without retraining. In the future we hope to further test ADS’s ability to analysis non-myelinated axons for sensory nerve recovery.
Data Availability
All files are available from the Harvard Dataverse database (https://dataverse.harvard.edu/dataset.xhtml?persistentId=doi:10.7910/DVN/H9N9ZU).
Funding Statement
The author(s) received no specific funding for this work.
References
- 1.Kaplan S, Geuna S, Ronchi G, Ulkay MB, Bartheld CS von. Calibration of the stereological estimation of the number of myelinated axons in the rat sciatic nerve: a multicenter study. Journal of Neuroscience Methods. 2010;187: 90–99. doi: 10.1016/j.jneumeth.2010.01.001 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2.Kemp SWP, Cederna PS, Midha R. Comparative outcome measures in peripheral regeneration studies. Experimental neurology. 2017;287: 348–357. doi: 10.1016/j.expneurol.2016.04.011 [DOI] [PubMed] [Google Scholar]
- 3.Petcu EB, Midha R, McColl E, Popa-Wagner A, Chirila TV, Dalton PD. 3D printing strategies for peripheral nerve regeneration. Biofabrication. 2018;10: 032001. doi: 10.1088/1758-5090/aaaf50 [DOI] [PubMed] [Google Scholar]
- 4.Hanwright PJ, Rath JB, Guionneau N von, Slavin B, Pinni S, Zlotolow D, et al. The Effects of a Porcine Extracellular Matrix Nerve Wrap as an Adjunct to Primary Epineurial Repair. J Hand Surg. 2021. doi: 10.1016/j.jhsa.2020.11.023 [DOI] [PubMed] [Google Scholar]
- 5.Zuo KJ, Gordon T, Chan KM, Borschel GH. Electrical stimulation to enhance peripheral nerve regeneration: Update in molecular investigations and clinical translation. Exp Neurol. 2020;332: 113397. doi: 10.1016/j.expneurol.2020.113397 [DOI] [PubMed] [Google Scholar]
- 6.Lopez J, Quan A, Budihardjo J, Xiang S, Wang H, Koshy K, et al. Growth Hormone Improves Nerve Regeneration, Muscle Re-innervation, and Functional Outcomes After Chronic Denervation Injury. Sci Rep-uk. 2019;9: 3117. doi: 10.1038/s41598-019-39738-6 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7.More HL, Chen J, Gibson E, Donelan JM, Beg MF. A semi-automated method for identifying and measuring myelinated nerve fibers in scanning electron microscope images. Journal of Neuroscience Methods. 2011;201: 149–158. doi: 10.1016/j.jneumeth.2011.07.026 [DOI] [PubMed] [Google Scholar]
- 8.Leckenby JI, Chacon MA, Grobbelaar AO, Lichtman JW. Imaging Peripheral Nerve Regeneration: A New Technique for 3D Visualization of Axonal Behavior. The Journal of surgical research. 2019;242: 207–213. doi: 10.1016/j.jss.2019.04.046 [DOI] [PubMed] [Google Scholar]
- 9.Kaynig V, Vazquez-Reina A, Knowles-Barley S, Roberts M, Jones TR, Kasthuri N, et al. Large-scale automatic reconstruction of neuronal processes from electron microscopy images. Medical image analysis. 2015;22: 77–88. doi: 10.1016/j.media.2015.02.001 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Naito T, Nagashima Y, Taira K, Uchio N, Tsuji S, Shimizu J. Identification and segmentation of myelinated nerve fibers in a cross-sectional optical microscopic image using a deep learning model. Journal of Neuroscience Methods. 2017;291: 141–149. doi: 10.1016/j.jneumeth.2017.08.014 [DOI] [PubMed] [Google Scholar]
- 11.Zaimi A, Wabartha M, Herman V, Antonsanti P-L, Perone CS, Cohen-Adad J. AxonDeepSeg: automatic axon and myelin segmentation from microscopy data using convolutional neural networks. Scientific reports. 2018;8: 3816. doi: 10.1038/s41598-018-22181-4 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12.Brent R, Boucheron L. Deep learning to predict microscope images. Nature methods. 2018;15: 868–870. doi: 10.1038/s41592-018-0194-9 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13.Guo Y, Liu Y, Oerlemans A, Lao S, Wu S, Lew MS. Deep learning for visual understanding: A review. Neurocomputing. 2016;187: 27–48. doi: 10.1016/j.neucom.2015.09.116 [DOI] [Google Scholar]
- 14.Januszewski M, Kornfeld J, Li PH, Pope A, Blakely T, Lindsey L, et al. High-precision automated reconstruction of neurons with flood-filling networks. Nature methods. 2018; 1–11. doi: 10.1038/s41592-018-0049-4 [DOI] [PubMed] [Google Scholar]
- 15.Janjic P, Petrovski K, Dolgoski B, Smiley J, Zdravkovski P, Pavlovski G, et al. Measurement-oriented deep-learning workflow for improved segmentation of myelin and axons in high-resolution images of human cerebral white matter. J Neurosci Meth. 2019;326: 108373. doi: 10.1016/j.jneumeth.2019.108373 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16.Moiseev D, Hu B, Li J. Morphometric analysis of peripheral myelinated nerve fibers through deep learning. J Peripher Nerv Syst. 2019;24: 87–93. doi: 10.1111/jns.12293 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.Scipio FD, Raimondo S, Tos P, Geuna S. A simple protocol for paraffin-embedded myelin sheath staining with osmium tetroxide for light microscope observation. Microsc Res Techniq. 2008;71: 497–502. doi: 10.1002/jemt.20577 [DOI] [PubMed] [Google Scholar]
- 18.Schindelin J, Arganda-Carreras I, Frise E, Kaynig V, Longair M, Pietzsch T, et al. Fiji: an open-source platform for biological-image analysis. Nature methods. 2012;9: 676–682. doi: 10.1038/nmeth.2019 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19.Bland JM, Altman D. Statistical methods for assessing agreement between two methods of clinical measurement. Lancet. 1986;327: 307–310. doi: 10.1016/s0140-6736(86)90837-8 [DOI] [PubMed] [Google Scholar]
- 20.Sulaiman OA, Gordon T. Effects of short- and long-term Schwann cell denervation on peripheral nerve regeneration, myelination, and size. Glia. 2000;32: 234–246. doi: [DOI] [PubMed] [Google Scholar]
- 21.Hunter DA, Pan D, Wood MD, Snyder-Warwick AK, Moore AM, Feldman EL, et al. Design-Based Stereology and Binary Image Histomorphometry in Nerve Assessment. J Neurosci Meth. 2020;336: 108635. doi: 10.1016/j.jneumeth.2020.108635 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 22.Scandalea ML di, Perone CS, Boudreau M, Cohen-Adad J. Deep Active Learning for Axon-Myelin Segmentation on Histology Data. Arxiv. 2019. [Google Scholar]
- 23.Ossinger A, Bajic A, Pan S, Andersson B, Ranefall P, Hailer NP, et al. A rapid and accurate method to quantify neurite outgrowth from cell and tissue cultures: Two image analytic approaches using adaptive thresholds or machine learning. J Neurosci Meth. 2019;331: 108522. doi: 10.1016/j.jneumeth.2019.108522 [DOI] [PubMed] [Google Scholar]