Abstract
Human behaviors are extremely sophisticated, relying on the adaptive, plastic and event-driven network of sensory neurons. Such neuronal system analyzes multiple sensory cues efficiently to establish accurate depiction of the environment. Here, we develop a bimodal artificial sensory neuron to implement the sensory fusion processes. Such a bimodal artificial sensory neuron collects optic and pressure information from the photodetector and pressure sensors respectively, transmits the bimodal information through an ionic cable, and integrates them into post-synaptic currents by a synaptic transistor. The sensory neuron can be excited in multiple levels by synchronizing the two sensory cues, which enables the manipulating of skeletal myotubes and a robotic hand. Furthermore, enhanced recognition capability achieved on fused visual/haptic cues is confirmed by simulation of a multi-transparency pattern recognition task. Our biomimetic design has the potential to advance technologies in cyborg and neuromorphic systems by endowing them with supramodal perceptual capabilities.
Subject terms: Electrical and electronic engineering, Sensors and biosensors, Nanosensors
Designing bioinspired perceptual system remains a challenge. Here, the authors report a bimodal artificial sensory neuron, integrating a resistive pressure sensor, a perovskite-based photodetector, a hydrogel-based ionic cable, and a synaptic transistor, to implement the visual-haptic fusion for motion control and patterns recognition.
Introduction
When interacting with the real, dynamic world, biological systems always outshine their electronic counterparts due to their sophisticated sensorimotor skills1–3. Emulating the functionality and/or structuralism of the natural system would intrinsically address unmet needs in current digital systems4–6. Although bioinspired systems using silicon-based circuits and software have realized some complicated and dexterous sensorimotor functions7–9, their efficiency still suffers when data size increases due to centralized and sequential operation. In contrast, the biological systems are essentially run on distributed computing paradigm, whose superior fault tolerance and power efficiency are inherent in the adaptive, plastic, and event-driven network of sensory neurons10,11. Therefore, emulating the biological processes from the level of sensory neuron would fundamentally achieve biological perceptual capabilities.
Artificial sensory neuron development can benefit from an improved understanding of sensory processing in biology. One notable advantage of the biological sensory systems is that they analyze multiple cues, making reactions more reliable than with a unimodal cue. For example, visual and haptic cues are intensively received and perceived in our interactions with the surroundings, and the two cues are closely associated and interpreted in the inferior parietal cortex to provide supramodal spatial ability12–14 (Fig. 1a). The neurons in this cortex area subsume both the macrospace of vision and the microspace encompassed by the hand, to avoid misjudgment from environmental complexity such as variances in the object’s pose9. The behavioral and psychological experiments also indicate a higher precision of appreciating an object when combining the two cues9,15.
Artificial sensory neurons/synapses with either haptic or visual modalities16,17, have been achieved for applications such as pattern recognition18,19 and muscular contraction control20. However, supramodal perceptual capabilities that can increase the reliability and accuracy of these machineries are still absent, possibly due to a lack of platform for mediating multimodal sensory data. Synaptic transistors allow parallel gating of active channels via ions in electrolytes, hence representing a core element for implementing sensory fusion at a neuronal level21.
Here, we develop a bimodal artificial sensory neuron (BASE) based on ionic/electronic hybrid neuromorphic electronics to implement the visual-haptic fusion. This BASE unit consists of four core components: resistive pressure sensor, perovskite-based photodetector, hydrogel-based ionic cable, and a synaptic transistor (Fig. 1b). The photodetector and pressure sensor function as the receptors in the retina and skin, respectively, converting external haptic and visual stimuli into electrical signals. The electrical signals from the two sensors are then transmitted through the ionic cable to the synaptic transistors for integration and conversion into a transient channel current, analogous to the biological excitatory postsynaptic currents (EPSC). Bimodal stimuli in closer succession can induce stronger changes in EPSC, which can be used to determine the extent of synchronization between the two. This, in turn, is used to provide multi-dimensional spatial information thereby controlling a biohybrid neuromuscular junction or a robotic hand, mimicking the process of ‘perception for action’22. We also design and simulate a matrix of BASE as the feature extraction layer of a perceptron for recognition of multi-transparency alphabetic patterns. The results further confirm that multimodal sensory fusion by BASE can increase the recognition rate (ratio of successful recognition of both the letter and its transparency to total trials) even with a reduced data size. Using BASE for constructing a bioinspired perceptual system would have profound implications for neurorobotics, cyborg systems, and autonomous artificial intelligence.
Result
Fabrication and characterization of the BASE
The hybrid neural circuits essentially encompass two artificial sensory channels: the visual and the haptic channel. The visual channel comprises a perovskite-based photodetector, mimicking the photoreceptors in the retina, with a configuration of Zn2SnO4/PEA2MA2Pb3I10/Poly[bis(4-phenyl) (2,4,6-trimethylphenyl) amine] (PTAA)/Au on PET/indium tin oxide (ITO)-coated substrate. 2D perovskite — PEA2MA2Pb3I10 is stable in ambient environment for reliable and repeatable photo-detection23,24 (Supplementary Fig. 1). The haptic channel comprises a pressure sensor that incorporating microstructures in the top carbon nanotube (CNT)-coated poly (dimethylsiloxane) (PDMS) layer18,25 (Supplementary Figs. 2 and 3). When a pressure is loaded on this layer, it forms a resistive pathway with the electrodes on the bottom layer. An increase in pressure increases the contact area and therefore decreases the resistance between the top CNT film and the bottom electrodes. The polyvinyl alcohol (PVA) hydrogel-based ionic cables represent the ionic transmission pathway as the axon in an afferent nerve, which carry information from the two artificial sensory channels to the electrolyte gated synaptic transistors for further integration and processing26. The electrolyte gated synaptic transistors are fabricated as shown in Supplementary Fig. 4. The exponential relaxation of ions in the electrolyte after an applied voltage stimulus on the gate (as the presynaptic terminal) explains the slow decay of channel currents (as the EPSC) based on electrostatic coupling27. Such decay properties have been investigated for mimicking some essential synaptic plasticity — the basic neurological principle underling learning and memory28–31.
These components are wired by the CNT electrodes fabricated through printing-filtration-transferring processes (Supplementary Figs. 5 and 6) on PDMS films, and such ionic/electronic hybrid circuits are illustrated in Fig. 2a. Three input terminals are connected to the voltage supply for photodetector (VV), pressure sensor (VH), and global inhibitory input (VIH), respectively. The energy consumption of the BASE could decrease down to <0.05 μW by introducing the global inhibitory input (Supplementary Fig. 7). Two output terminals (VD and VS) are connected to the drain and source electrodes of the synaptic transistor, and a voltage bias (VDS = 0.5 V) is applied between them for measuring the EPSC. The lowest sheet resistance of 7.0 Ω sq−1 of such CNT electrodes has been obtained with a density of 33 μg cm−2 (Fig. 2b), indicating decent conductivity for applying/collecting electronic signals from each terminal. Moreover, such CNT electrodes exhibit much lower interfacial impedance (f < 1 kHz) with hydrogel than that of gold electrodes (Supplementary Fig. 8), enabling effective electronic–ionic current transduction at the electrode-hydrogel interfaces. Figure 2c shows the impedance characterizations of PVA hydrogel rods bonded to gold and CNT electrodes, respectively, and the latter is almost three orders of magnitude lower than the former. The ionic cable is then fixed on the CNT electrodes by using the instant tough bonding method32, maintaining low interfacial impedance (Supplementary Fig. 9).
The information flow in such circuits starts from triggering specific sensors in response to an external stimulus. For artificial visual receptor, the resistance of the photodetector would decrease with an increase in incident light intensity as shown by the current–voltage (IV) measurement (Fig. 2d). The voltage supply (VV) thus could induce an ion flux through the ionic cable due to the reduced resistance of the sensor. As the other side of the hydrogel is connected to the gate of the synaptic transistor, the accumulation of the ions would electrostatically couple an EPSC through the semiconducting channel of the transistor. The peak EPSCs are influenced by not only the magnitude of the pressure stimulus (Fig. 2e) but also the duration of the light stimulus (Fig. 2f). A similar trend can be found also in the artificial haptic receptor (Fig. 2g–i). Although slight differences between the output characteristics of the two artificial receptors are observed — the pressure sensors show a better linearity and faster responsivity than the photodetectors, the output range of the two kinds of receptors are very similar.
Motion control based on visual-haptic fusion
Next, we have designed and fabricated a biohybrid neuromuscular junction (BNJ) to transmit signals from the BASE and to innervate the skeletal myotubes, mimicking body motion control based on visual-haptic fusion (Fig. 3a). The BNJ consists of interdigital electrodes and cultured skeletal myotubes (Fig. 3b). To reduce the impedance between the gold electrode and the myotubes, a layer of Polypyrrole (PPy) has been coated onto the gold electrode by electroplating. As the conducting polymers can transfer charge both by electronic and ionic mechanisms33,34, thus the impedance magnitude has been greatly reduced from ~1 MΩ to ~3 kΩ at f = 1 Hz (Supplementary Fig. 10). C2C12 myoblast have been seeded after fibronectin functionalization of the BNJ surface. Myoblasts adhere and spread evenly upon overnight incubation in the seeding medium (Fig. 3b). To induce formation of myotubes, the cells were incubated in the differentiation medium for another 5 days. The voltage stimulation can then be applied on the BNJ as shown in Fig. 3c. Particle image velocimetry (PIV) analysis confirms the adhesion between electrodes and cells, and several areas of myotubes can be effectively activated by applying 1.0 V stimulation within 1 min (Fig. 3d). We intentionally choose three regions (region I to III) which are apparently active under electric field. Generally, the myotubes would shrink and show a motion pattern directing from the positive terminal to the ground (Supplementary Fig. 11). Region III is used to illustrate the stimulation intensity dependent contraction property (Fig. 3e), and a clearly positive correlation between voltage intensity and activeness (indicated by mapping of the magnitude of velocity) can be observed. The histogram of x component of velocity has also quantitatively verified such conclusion (Fig. 3f).
To localize a nearby object, we might glance by eye (t = TV) and explore by finger (t = TH). The eye-hand coordination enables us to combine the two cues for spatial inference. Then we grasp the object via the contraction on muscle groups of the hand allows us to grasp the object, once its location has been confirmed. Moreover, human beings usually tolerate certain amount of temporal discrepancy between two signals while still perceiving them as being synchronous35,36. To mimic such synchronous action, the visual stimulus (~4.8 mW cm−2, ~250 ms) and haptic stimulus (~2 kPa, ~350 ms) with different time intervals (∆T = TH–TV) are applied on the visual and haptic sensory channels of the BASE, respectively. The voltages applied on the two channels (VV and VH) are 1.0 and −1.0 V, respectively. Since both the EPSCs triggered by each of the two stimuli exhibit a gradual decay as demonstrated in Fig. 2, they may affect the response following an immediate subsequent stimulus37. Consequently, the EPSC amplitudes (∆EPSC) measured at t = TV show no obvious change compared with ∆EPSCV (the EPSC amplitude triggered by visual stimulation only) when |∆T| > 1 s, while with shorter time intervals, the amplitudes decrease gradually with decreasing |ΔT| (Fig. 3g). Here, we set the relative change of ∆EPSC against ∆EPSCV ((∆EPSC-∆EPSCV)/∆EPSCV, annotated as ∆EPSC%) of 20% as a threshold for determining synchronous (|ΔEPSC%| > 20%) or asynchronous (|ΔEPSC%| ≤ 20%) signals. The ΔEPSC are detected and linearly converted to voltage outputs within the range of 0–1.0 V, which are then sent to the biohybrid neuromuscular junction for myotubes control (Supplementary Fig. 12). Our results show the voltage applied on the BNJ can trigger a perceptible migration when the voltage is >0.4 V (Supplementary Fig. 13). The output voltage is no more than 0.4 V for the asynchronous signals, which cannot trigger the perceptible action of the myotubes (Fig. 3h and Supplementary Movies 1–3). Conversely, the synchronous signals can induce the obvious migration of the myotubes.
The strategy of fusing visual and haptic feedbacks can also enable robotic motion control and is superior to a unimodal one. A tennis ball located at the proximity of a robotic hand can be noted as “YES” (or “NO”) if it could (or not) deliver either the visual or haptic feedback to trigger the hand closing for catching the ball (Fig. 4i, visual feedback for z-axis or haptic feedback for y-axis). The ball can only be caught when it is at the “YES” position for both directions. The unimodal feedback could only be used to differentiate “YES” or “NO” at one-dimension, due to the binary states of each sensor. By contrast, a modified BASE patch is able to obtain the spatial information from the two dimensions (Fig. 4j) for guiding a robotic hand. The spatial information is obtained through an exploration process similar as before: (1) triggering a LED on the ball, providing the visual stimulus; (2) half-closing the hand to touch the ball, providing the haptic stimulus. As shown in Fig. 4k, the BASE can generate three types of responses by the exploration process (ΔT = 100 ms), when the ball is at three typical positions (V = YES, H = NO; V = NO, H = YES; V = YES, H = YES). In this way, the robotic hand can take more appropriate action (open or close) based on the multi-dimension information. Otherwise, the utilization of one-dimension spatial information might lead to wrong decision for the robotic hand (Supplementary Fig. 14, Tables 1 and 2, Scheme 1, and Movies 4–8).
Multi-transparency pattern recognition
Subsequently, we design BASE-based sensory fusion matrices to use the fused visual and haptic cues for recognition tasks (Fig. 4a). Multi-transparency alphabetic patterns (a mixture of PDMS and carbon black, Supplementary Figs. 15 and 16) have been used in this work. The alphabetic patterns are labeled based on the shape and transparency of their main body. To verify the robustness of the recognition, 720 alphabetic patterns with different combinations of random noise pixels have been used for recognition (Supplementary Fig. 17). The sensing data from each matrix are then fed into a perceptron with one hidden layer built by MATLAB for the pattern recognition tasks (Supplementary Fig. 18 and Supplementary Training Samples). In this case, this matrix serves as the feature extraction layer of this artificial neural network.
The bimodal fusion matrix and unimodal matrices (optics or pressure) are designed with 10 × 10 pixels. Each pixel of the bimodal fusion matrix contains one photodetector and one pressure sensor (noted as VH unit). In each line, the ten VH units are connected to synaptic transistors through the ionic cable. As shown in Fig. 4b, the transistor can integrate stimuli from the VH unit inducing a joint EPSC current which is dependent on the intensity of the two stimuli and the weight of the input (w). Our results show that the summation output (measured sum) from one VH unit is slightly lower than the arithmetic summation (expected sum) of the outputs (Fig. 4b and Supplementary Fig. 19). The weight of the input (w(m)) is dependent on the distance between the transistor and the VH unit — a gradual decay with an increasing distance (Fig. 4c and Supplementary Fig. 20). Therefore, the integration effect of the synaptic transistor to multiple inputs could be deemed as the integral of the product of the input intensities and their distance-dependent weights. Here, we define the kernel size (n) of the feature extraction layer as the numbers of VH units in one row connected to one transistor in series through one common ionic cable (i.e., VH unit i to i + n−1 (i = 1, 2, … 10−n + 1). Such configuration thereby could implement the convolution-like operation to the input from each VH unit (annotated by n-VH, Supplementary Fig. 21). The convolutional layer is the core building block of a convolutional neural network (CNN), which serves as the filter and efficiently extracts the essential features of input. As such, the mapping output (EPSC amplitude from each transistor) of the sensory fusion matrices can be calculated by computer using the extracted device parameters from these results (Supplementary Fig. 22 and Note 1).
To obtain the visual and haptic cues of the multi-transparency patterns, we place the patterns on the arrays for ~1.5 s with a light source (~4.8 mW cm−2) turned on at the same time, and thereafter remove the pattern with the light turned off. The normalized sensing data by unimodal matrices and bimodal matrix are shown in Fig. 4d. When mapping the multi-transparency patterns, simply exploiting the optic matrix would lose some essential features, while simply exploiting the pressure matrix could only provide the shape information without the degree of transparency. Only the visual-haptic fusion matrix could extract both the shape and the transparency adequately. Consequently, the recognition rate based on the visual-haptic fusion matrix shows the best outcome as shown in Fig. 4e. Interestingly, although the features of the alphabetic patterns extracted by increasing the kernel size of the matrix would be lost (60% of original data size with n = 5) due to the spatial integration, the recognition rate (~66%) by the fusion mode with n = 5 is still slightly better than that of the unimodal mode (~65% for the optic matrix with n = 1). Such results indicate we can achieve robust recognition by using the VH fusion matrix as the feature extraction layer of an artificial neural network.
Discussion
In summary, inspired by the supramodal sensory fusion in the sensory nervous system, we developed an ionic/electronic hybrid neuromorphic device that combines optic and pressure stimuli to generate a summed EPSC through the synaptic transistor. Such combined current carries the bimodal information in time-dependent and nonlinear manners, which closely resemble the neuronal behaviors. The fused signal was then used to innervate the skeletal myotubes and provide multi-dimensional spatial information for the robotic hand, successfully mimicking the motion control by using bimodal sensory cues at the cellular level. More interestingly, the bimodal sensory data could be implemented for recognition of multi-transparency alphabetic patterns and exhibits superior performance to unimodal sensory data. The multi-transparency patterns could represent and abstract the core factors of some real scenarios where the transparency of the object could be inferred by visual feedback and the shape or weight could be inferred by tactile feedback.
We have summarized recent achievements in artificial sensory neurons as shown in Table 1. Compared with other works, our BASE can fuse the cues of two sensory modalities and use them for manipulation, recognition, and synaptic emulations. In addition, with the successful incorporation of myotubes as the actuator, all components in the current system are performing functions at the cellular or equivalent level, which is of great potential for assembling the bionic sensorimotor system from the bottom up. As distinct from previous biohybrid robots using direct bias or optogenetic stimuli not representative of actual real-world cues38–40, our biohybrid neuromuscular junction can trigger the contraction of myoblasts by visual and/or tactile -like stimulation, hence moving towards biologically relevant sensorimotor systems. Conventional approaches for sensory processing rely on centralized and sequential operations of data, in which the efficiency dramatically decreases with the increase in data size. In contrast, the neuronal-level implementation, incorporating sensing, refining, and processing would eventually achieve the biological advantages of fault tolerance and power efficiency. Although multi-sensory capabilities have been realized in some e-skin systems41–43, their sensory processing has always been completed using conventional digital units. As such, mimicking sensory fusion from the neuronal-level could help to build a highly integrated perceptual system to access massive sensory data for improving current cyborg technologies44,45 and artificial intelligence46–48.
Table 1.
Name | Sensory modality | Actuator & size | Feature action dimension | Recognition pattern | Synaptic emulations |
---|---|---|---|---|---|
Artificial afferent nerve19 | Pressure | Insect leg ~3 cm | ~1 mm | Braille pattern | Short-term potentiation |
NeuTap18 | Pressure | NA | NA | Braille pattern | Short-term potentiation; dendritic integration |
Artificial somatic reflex Arc49 | Pressure | Electrochemical actuator ~2.5 cm | ~1 mm | NA | Neural all-or-none law |
Artificial optic-neural synapse50 | Light | NA | NA | Colour-mixed pattern | Long-term potentiation & depression |
Optoelectronic sensorimotor artificial synapse20 | Light | Polymer actuator ~2 cm | ~1 mm | NA | Short-term potentiation |
Optoelectronic neuromorphic device51 | Light | NA | NA | Illuminance gradation pattern | Visual synaptic functions; environment-adaptable perception |
This work (BASE) | Pressure & light | Myotubes ~1.5 mm | ~5 µm | Multi-transparency pattern | Dendritic integration |
Methods
Fabrication of the photodetectors
The fabrication processes of the photodetectors are briefed as follows. ITO/PET was chemically etched by reaction of zinc powder with HCl (2 M) for the desired pattern before being cut into 2.5 × 2.5 cm2 pieces. The ITO/PET was washed and rinsed with detergent, DI water, and ethanol sequentially. After drying with nitrogen gas, the ITO/PET was secured onto glass substrates and treated in UV-ozone plasma for 15 min. Zn2SnO4 nanoparticle solution was deposited on prepared ITO substrate by spin coating at 3000 rpm for 30 s. The coated substrates were dried at 100 °C for 10 min, followed by annealing at 150 °C for 2 h. Perovskite precursor (1.2 M) and substrates were preheated to 100 °C, before the hot precursor solution was spin coated onto substrates at 4000 rpm for 30 s, followed by annealing at 100 °C for 15 min. PTAA solution was spin coated on cool perovskite films at 3000 rpm for 60 s. Au electrodes (~100 nm) were thermal evaporated on samples with the size of 0.2 cm × 0.2 cm.
Fabrication of pyramidal structured PDMS film
Silicon masters with recessed pyramidal microstructure arrays were fabricated using our previous method by photolithography and wet etching process17. The mixture of PDMS elastomer and crosslinker in 10:1 (w/w) ratio (Sylgard 184, Dow Corning) were spin coated on PDMS at 800 rpm. The elastomer mixture was degassed in vacuum and cured at 90 °C for 1 h. The films were then sectioned by a scalpel and peeled off from the silicon master.
CNT deposition through ultrasonic vibrating method
The CNT were sprayed on the microstructured PDMS film through ultrasonic vibrating method (Supplementary Fig. 2a). In brief, 5 μL CNT solution (1 mg/mL) was diluted in 2 mL deionized water to obtain a homogeneous CNT solution of 2.5 μg/mL through stirring. A commercial ultrasonic humidifier was used to generate microdroplets of the CNT solution through a filter screen with hole size of 60 μm. A time controller was used to trigger/stop the generation process. For fabrication, the PDMS films were placed on a hotplate at 120 °C. One fabrication cycle takes 1 min. The CNT microdroplet were generated through the ultrasonic vibrator and sprayed onto the PDMS film for 5 s. It was then dried out to form a CNT film on the electrodes for another 55 s.
Fabrication of the synaptic transistors
The fabrication of the PVA gated synaptic transistors can be obtained in previous reports17. In brief, the patterned ITO electrodes (drain, source, and gate electrodes) with a thickness of ~60 nm were deposited by radio-frequency magnetron sputtering through a stainless steel shadow mask. Then a ~10 nm ITO channel was deposited using the same method. PVA powder was dissolved in 5% CaCl2 solution at a concentration of 10 wt%, then heated gently using a hot oil bath for complete dissolution without thermal decomposition of the polymer. Then the PVA gate dielectric was cast onto the channel and gate electrode, and dried at 60 °C for 2 h.
CNT patterning process
The patterns were designed using PowerPoint software. The Nylon filter membrane (1 μm) was adhered to A4 paper for printing. The designed patterns thus were printed onto the filter membrane. Then the CNT solution was vacuum-filtered on the filter membrane. CNT could only be concentrated on the white areas of the filter membrane, because the micropores of the membrane at the black areas were penetrated with the printing ink which impedes the passing through of solvent. The membrane was then thoroughly rinsed with DI water to remove the residual sodium dodecyl benzene sulfonate (SDBS) and CNT on the black area. Next, the PDMS (10:1) which was heated in the oven for 35 min (60 °C), was used for transferring the CNT pattern by the hot-press approach. The membrane with the CNT pattern face was put onto the half-cured PDMS, and a glass and a 200-g weight were covered onto the membrane for achieving close contact between the membrane and PDMS. The PDMS and the membrane were then heated in the oven for 4 h (60 °C) before separating of them.
PVA based hydrogel fabrication process
The PVA (Aldrich, MW = 124,000) powder was dissolved in deionized water at a concentration of 10 wt%, then heated gently using a hot oil bath (90 °C) for complete dissolution without thermal decomposition of the polymer. The hot solution was stirred until the polymer was completely dissolved and a clear viscous solution is formed. Then the PVA hydrogel is obtained through three freeze-thaw cycles. To obtain a high ionic conductivity, the PVA hydrogel was immersed in the 1 mol L−1 LiCl for 4 h.
Fabrication of the biohybrid neuromuscular junction
The interdigital Au electrodes with thickness of 70 nm were deposited through thermal evaporation using a shadow mask. Then the Polypyrrole (PPy) was coated onto the Au electrode by electroplating. The PPy was synthesized on the interdigital Au electrodes by anodic oxidation of pyrrole in an electrochemical cell containing 0.42 g pyrrole monomer, p-toluenesulfonic acid (PTSA, 25 mg) dissolved in 30 mL 5% phosphate buffer saline (PBS) at temperature of ~0 °C. Pyrrole was first distilled and kept refrigerated until use. Dissolved oxygen was removed by bubbling the solution with nitrogen for several minutes prior to polymerization. The galvanostatic polymerization was carried out by using Keithley 4200 with the constant current density of 0.8 mA/cm2. An Au plate serves as the counter electrode. The solution was kept under stirring by a magnetic stirrer during the polymerization process. After 3 min of electrochemical polymerization, a dense layer of PPy was deposited on the surface of Au electrode. The obtained electrodes were dipped in DI water for three times at 20 min each. C2C12 myoblasts were seeded onto the electrode substrate in the seeding medium (DMEM with 10% fetal bovine serum) overnight. Subsequently, the cells were induced to form myotubes in the differentiation medium (DMEM supplemented with 2% fetal bovine serum and 1 µg/mL insulin) for 5 days, forming the biohybrid neuromuscular junction.
The recognition simulation
The six basic alphabetic patterns were extended with three transparencies: transparent, translucent, and opaque. Each pattern was added with five noise pixels (40 cases in total) as one multi-transparency pattern for recognition task. The distribution and the transparency of noise pixel were generated by MATLAB. As such, there are 720 multi-transparency patterns in total. The data obtained from the sensory fusion matrix was fed into the two layers of perceptron. There are 10 nodes in the hidden layer and 18 nodes in the output layer corresponding to the 18 labels of patterns. The network was then trained by MATLAB.
Reporting summary
Further information on research design is available in the Nature Research Reporting Summary linked to this article.
Supplementary information
Acknowledgements
The authors thank the financial support from the Agency for Science, Technology and Research (A*STAR) under its AME Programmatic Funding Scheme (project no. A18A1b0045) Cyber-Physiochemical Interfaces (CPI) Programme, the National Research Foundation, Prime Minister’s office, Singapore, under its NRF Investigatorship (NRF-NRFI2017-07), and Singapore Ministry of Education (MOE2017-T2-2-107).
Author contributions
X.C. and C.W. conceived the project and constructed the research frame. These authors contributed equally: C.W. and P.C.; C.W. and P.C. built the biohybrid neuromuscular junction and performed the cell-related measurements. C.W. and X.G. prepared the devices and performed the hardware measurements. C.W. and M.W. built the software package. X.C. and C.W. analyzed the experimental data and simulation results. P.C., Y.L., and Z.L. prepared the hydrogels and CNT electrodes. C.W., L.Y., N.M., Y.L., X.L., and X.C. edited the manuscript. All authors discussed the results and implications and commented on the manuscript at all stages.
Data availability
The data that support the plots within this paper and other finding of this study are available from the corresponding author on reasonable request.
Code availability
Code from this study is available from the corresponding author upon request.
Competing interests
The authors declare no competing interests.
Footnotes
Peer review information Nature Communications thanks the anonymous reviewer(s) for their contribution to the peer review of this work.
Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary information
Supplementary information is available for this paper at 10.1038/s41467-020-18375-y.
References
- 1.Yang J., Liu H., Sun F. & Gao M. in 2015 IEEE Int. Conf. Robot. Biomimetics (ROBIO) (IEEE, 2015).
- 2.Güler P., Bekiroglu Y., Gratal X., Pauwels K. & Kragic D. What’s in the container? Classifying object contents from vision and touch. in Proc.2014 IEEE/RSJ Int. Conf. Intell. Robots Syst. (IEEE, 2014).
- 3.Zidan MA, Strachan JP, Lu WD. The future of electronics based on memristive systems. Nat. Electron. 2018;1:22–29. [Google Scholar]
- 4.Singh S, Okun A, Jackson A. Learning to play Go from scratch. Nature. 2017;550:336–337. doi: 10.1038/550336a. [DOI] [PubMed] [Google Scholar]
- 5.Shen Y, et al. Deep learning with coherent nanophotonic circuits. Nat. Photon. 2017;11:441–446. [Google Scholar]
- 6.Park H-L, et al. Flexible neuromorphic electronics for computing, soft robotics, and neuroprosthetics. Adv. Mater. 2019;32:1903558. doi: 10.1002/adma.201903558. [DOI] [PubMed] [Google Scholar]
- 7.Yue S, Rind FC, Keil MS, Cuadri J, Stafford R. A bio-inspired visual collision detection mechanism for cars: optimisation of a model of a locust neuron to a novel environment. Neurocomputing. 2006;69:1591–1598. [Google Scholar]
- 8.Yue S, Rind FC. Collision detection in complex dynamic scenes using an LGMD-based visual neural network with feature enhancement. IEEE Trans. Neural Netw. 2006;17:705–716. doi: 10.1109/TNN.2006.873286. [DOI] [PubMed] [Google Scholar]
- 9.Toprak S, Navarro-Guerrero N, Wermter S. Evaluating integration strategies for visuo-haptic object recognition. Cognit. Comput. 2018;10:408–425. doi: 10.1007/s12559-017-9536-7. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Abraira VE, Ginty DD. The sensory neurons of touch. Neuron. 2013;79:618–639. doi: 10.1016/j.neuron.2013.07.051. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11.Lumpkin EA, Caterina MJ. Mechanisms of sensory transduction in the skin. Nature. 2007;445:858–865. doi: 10.1038/nature05662. [DOI] [PubMed] [Google Scholar]
- 12.Mishkin M, Ungerleider LG, Macko KA. Object vision and spatial vision: two cortical pathways. Trends Neurosci. 1983;6:414–417. [Google Scholar]
- 13.van Polanen V, Davare M. Interactions between dorsal and ventral streams for controlling skilled grasp. Neuropsychologia. 2015;79:186–191. doi: 10.1016/j.neuropsychologia.2015.07.010. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Freud E, Plaut DC, Behrmann M. ‘What’ is happening in the dorsal visual pathway. Trends Cogn. Sci. 2016;20:773–784. doi: 10.1016/j.tics.2016.08.003. [DOI] [PubMed] [Google Scholar]
- 15.Ernst MO, Banks MS. Humans integrate visual and haptic information in a statistically optimal fashion. Nature. 2002;415:429–433. doi: 10.1038/415429a. [DOI] [PubMed] [Google Scholar]
- 16.Wan C, et al. Artificial sensory memory. Adv. Mater. 2019;32:1902434. doi: 10.1002/adma.201902434. [DOI] [PubMed] [Google Scholar]
- 17.Lee Y, Ahn J-H. Biomimetic tactile sensors based on nanomaterials. ACS Nano. 2020;14:1220–1226. doi: 10.1021/acsnano.0c00363. [DOI] [PubMed] [Google Scholar]
- 18.Wan C, et al. An artificial sensory neuron with tactile perceptual learning. Adv. Mater. 2018;30:1801291. doi: 10.1002/adma.201801291. [DOI] [PubMed] [Google Scholar]
- 19.Kim Y, et al. A bioinspired flexible organic artificial afferent nerve. Science. 2018;360:998–1003. doi: 10.1126/science.aao0098. [DOI] [PubMed] [Google Scholar]
- 20.Lee Y, et al. Stretchable organic optoelectronic sensorimotor synapse. Sci. Adv. 2018;4:eaat7387. doi: 10.1126/sciadv.aat7387. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21.Park H-L, et al. Flexible neuromorphic electronics for computing, soft robotics, and neuroprosthetics. Adv. Mater. 2020;32:1903558. doi: 10.1002/adma.201903558. [DOI] [PubMed] [Google Scholar]
- 22.Ravinder S. & Dahiya M. V. in Robotic Tactile Sensing 14–15 (Springer, 2013).
- 23.Huang X, et al. Two-dimensional additive diethylammonium iodide promoting crystal growth for efficient and stable perovskite solar cells. RSC Adv. 2019;9:7984–7991. doi: 10.1039/c9ra01186b. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24.Yang HG, Zeng HC. Preparation of hollow anatase TiO2 nanospheres via ostwald ripening. J. Phys. Chem. B. 2004;108:3492–3495. doi: 10.1021/jp0377782. [DOI] [PubMed] [Google Scholar]
- 25.Zhu B, et al. Skin-inspired haptic memory arrays with an electrically reconfigurable architecture. Adv. Mater. 2016;28:1559–1566. doi: 10.1002/adma.201504754. [DOI] [PubMed] [Google Scholar]
- 26.Yang C, Suo Z. Hydrogel ionotronics. Nat. Rev. Mater. 2018;3:125–142. [Google Scholar]
- 27.Wan C, Zhu L, Wan X, Shi Y, Wan Q. Organic/inorganic hybrid synaptic transistors gated by proton conducting methylcellulose films. Appl. Phys. Lett. 2016;108:043508. [Google Scholar]
- 28.Wan C, et al. Flexible metal oxide/graphene oxide hybrid neuromorphic transistors on flexible conducting graphene substrates. Adv. Mater. 2016;28:5878–5885. doi: 10.1002/adma.201600820. [DOI] [PubMed] [Google Scholar]
- 29.John RA, et al. Synergistic gating of electro-iono-photoactive 2D chalcogenide neuristors: coexistence of hebbian and homeostatic synaptic metaplasticity. Adv. Mater. 2018;30:1800220. doi: 10.1002/adma.201800220. [DOI] [PubMed] [Google Scholar]
- 30.Zhu J, et al. Ion gated synaptic transistors based on 2D van der waals crystals with tunable diffusive dynamics. Adv. Mater. 2018;30:1800195. doi: 10.1002/adma.201800195. [DOI] [PubMed] [Google Scholar]
- 31.Moon J, et al. Temporal data classification and forecasting using a memristor-based reservoir computing system. Nat. Electron. 2019;2:480–487. [Google Scholar]
- 32.Wirthl D, et al. Instant tough bonding of hydrogels for soft machines and electronics. Sci. Adv. 2017;3:e1700053. doi: 10.1126/sciadv.1700053. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 33.Qi D, et al. Highly stretchable, compliant, polymeric microelectrode arrays for in vivo electrophysiological interfacing. Adv. Mater. 2017;29:1702800. doi: 10.1002/adma.201702800. [DOI] [PubMed] [Google Scholar]
- 34.Liu Y, et al. Soft conductive micropillar electrode arrays for biologically relevant electrophysiological recording. Proc. Natl Acad. Sci. USA. 2018;115:11718–11723. doi: 10.1073/pnas.1810827115. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 35.Vogels IM. Detection of temporal delays in visual-haptic interfaces. Hum. Factors. 2004;46:118–134. doi: 10.1518/hfes.46.1.118.30394. [DOI] [PubMed] [Google Scholar]
- 36.Lewkowicz DJ. Perception of auditory–visual temporal synchrony in human infants. J. Exp. Psychol. Hum. Percept. Perform. 1996;22:1094–1106. doi: 10.1037//0096-1523.22.5.1094. [DOI] [PubMed] [Google Scholar]
- 37.Kim K, Chen C-L, Truong Q, Shen AM, Chen Y. A carbon nanotube synapse with dynamic logic and learning. Adv. Mater. 2013;25:1693–1698. doi: 10.1002/adma.201203116. [DOI] [PubMed] [Google Scholar]
- 38.Nawroth JC, et al. A tissue-engineered jellyfish with biomimetic propulsion. Nat. Biotechnol. 2012;30:792–797. doi: 10.1038/nbt.2269. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 39.Park S-J, et al. Phototactic guidance of a tissue-engineered soft-robotic ray. Science. 2016;353:158–162. doi: 10.1126/science.aaf4292. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 40.Shin SR, et al. Electrically driven microengineered bioinspired soft robots. Adv. Mater. 2018;30:1704189. doi: 10.1002/adma.201704189. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 41.Hua Q, et al. Skin-inspired highly stretchable and conformable matrix networks for multifunctional sensing. Nat. Commun. 2018;9:244. doi: 10.1038/s41467-017-02685-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 42.Lee WW, et al. A neuro-inspired artificial peripheral nervous system for scalable electronic skins. Sci. Robot. 2019;4:eaax2198. doi: 10.1126/scirobotics.aax2198. [DOI] [PubMed] [Google Scholar]
- 43.Kim MK, et al. Soft-packaged sensory glove system for human-like natural interaction and control of prosthetic hands. NPG Asia Mater. 2019;11:43. [Google Scholar]
- 44.Lee G-H, et al. Multifunctional materials for implantable and wearable photonic healthcare devices. Nat. Rev. Mater. 2020;5:149–165. doi: 10.1038/s41578-019-0167-3. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 45.Arab Hassani F, Jin H, Yokota T, Someya T, Thakor NV. Soft sensors for a sensing-actuation system with high bladder voiding efficiency. Sci. Adv. 2020;6:eaba0412. doi: 10.1126/sciadv.aba0412. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 46.Kim H, Hwang S, Park J, Park B-G. Silicon synaptic transistor for hardware-based spiking neural network and neuromorphic system. Nanotechnology. 2017;28:405202. doi: 10.1088/1361-6528/aa86f8. [DOI] [PubMed] [Google Scholar]
- 47.Cai F, et al. A fully integrated reprogrammable memristor–CMOS system for efficient multiply–accumulate operations. Nat. Electron. 2019;2:290–299. [Google Scholar]
- 48.Cai P, et al. Locally coupled electromechanical interfaces based on cytoadhesion-inspired hybrids to identify muscular excitation-contraction signatures. Nat. Commun. 2020;11:2183. doi: 10.1038/s41467-020-15990-7. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 49.He K, et al. An artificial somatic reflex arc. Adv. Mater. 2020;32:1905399. doi: 10.1002/adma.201905399. [DOI] [PubMed] [Google Scholar]
- 50.Seo S, et al. Artificial optic-neural synapse for colored and color-mixed pattern recognition. Nat. Commun. 2018;9:5106. doi: 10.1038/s41467-018-07572-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 51.Kwon SM, et al. Environment-adaptable artificial visual perception behaviors using a light-adjustable optoelectronic neuromorphic device array. Adv. Mater. 2019;31:1906433. doi: 10.1002/adma.201906433. [DOI] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Data Availability Statement
The data that support the plots within this paper and other finding of this study are available from the corresponding author on reasonable request.
Code from this study is available from the corresponding author upon request.