Significance
In the last century, robots have been revolutionizing our lives, augmenting human actions with greater precision and repeatability. Unfortunately, most robotic systems can only operate in controlled environments. While increasing the complexity of the centralized controller is an instinctive direction to enable robots that are capable of autonomously adapting to their environment, there are ample examples in nature where adaptivity emerges from simpler decentralized processes. Here we perform experiments and simulations on a modular and scalable robotic platform in which each unit is stochastically updating its own behavior to explore requirements needed for a decentralized learning strategy capable of achieving locomotion in a continuously changing environment or when undergoing damage.
Keywords: emergent behavior, reinforced learning, modular robot, dynamic environment
Abstract
One of the main challenges in robotics is the development of systems that can adapt to their environment and achieve autonomous behavior. Current approaches typically aim to achieve this by increasing the complexity of the centralized controller by, e.g., direct modeling of their behavior, or implementing machine learning. In contrast, we simplify the controller using a decentralized and modular approach, with the aim of finding specific requirements needed for a robust and scalable learning strategy in robots. To achieve this, we conducted experiments and simulations on a specific robotic platform assembled from identical autonomous units that continuously sense their environment and react to it. By letting each unit adapt its behavior independently using a basic Monte Carlo scheme, the assembled system is able to learn and maintain optimal behavior in a dynamic environment as long as its memory is representative of the current environment, even when incurring damage. We show that the physical connection between the units is enough to achieve learning, and no additional communication or centralized information is required. As a result, such a distributed learning approach can be easily scaled to larger assemblies, blurring the boundaries between materials and robots, paving the way for a new class of modular “robotic matter” that can autonomously learn to thrive in dynamic or unfamiliar situations, for example, encountered by soft robots or self-assembled (micro)robots in various environments spanning from the medical realm to space explorations.
Traditional robots typically exhibit well-defined motions and are often controlled by a centralized controller that uses sensor data to provide feedback to its actuators (1, 2). While these systems are capable of operating in well-known and controlled environments, adapting their behavior to various unforeseen circumstances remains one of the main challenges in robotics (3). Toward this goal, reinforced learning strategies can be used to deal with more complex tasks that cannot directly be programmed in the behavior of the robot. By sensing their environment, robots can build models of themselves (4) or optimize human-made models with machine learning strategies to improve their behavior (1, 2, 4–8). Apart from a clear division between training and task execution, these systems rely on a centralized architecture that—with an increasing number of active components—demands advanced models and higher computational power.
A recent key idea to simplify the computational task, while allowing for adaptability to the environment, is to make soft robots. These are robots fabricated with materials with stiffness resembling human tissue (3, 9–12), that exhibit so-called embodied intelligence (13–15) and adjust their shape when subjected to forces resulting from interactions with their environment. While this physical adaptability enhances the robustness of the robot’s behavior, allowing it to, e.g., operate in more complex environments, it is not guaranteed that they operate optimally, or even operate at all. For example, a soft robot that is able to walk on various surfaces still requires a different actuation sequence to crawl through a narrow opening (16). At the moment, and because of the unpredictable nature of the robot’s behavior in varying environments, this sequence is externally (and manually) adjusted. A natural question to ask is if such higher levels of decision-making and learning can also be embodied in robots, similar to, e.g., the decentralized neural networks of squids (17, 18).
In fact, decentralized approaches are already applied in swarm robotics, where an increasing number of active components can cooperatively achieve complex behavior when sensing or communication capabilities are incorporated into the individual units (19–22). In such a robotic system, behavior emerges from local interactions, rather than being centrally controlled. Inspired by social animals such as ants, bees, or birds, researchers have been able to achieve emerging properties out of simple unit-to-unit interactions, such as adaptation to mechanical stimuli (23), self-assembly (22, 24), construction (25), and locomotion (20, 26). Interestingly, such behavior does not have to be programmed and could also be learned using the same decentralized architecture (27, 28).
In this work, we develop and perform experiments on modular robots that are connected to form a single body, in order to explore similar decentralized strategies. Using this platform, our aim is to simplify the learning strategy, while focusing on the fundamental challenge to ensure robust and optimal behavior under varying external conditions and damage. We do so by having each element continuously learn and adapt to changing circumstances. To that end, we develop a fully decentralized and universal reinforced learning strategy that can be easily implemented, is scalable, and allows the robotic system to autonomously achieve and maintain specific emergent behavior.
Individual Unit Behavior
Such an approach holds potential for a wide variety of robotic systems, and we illustrate this here with a robot made of several identical units connected by soft actuators that periodically extend and contract along one direction. Each robotic unit operates independently and consists of a microcontroller, optical motion sensor, pump, pneumatic actuator, and battery (Fig. 1A and SI Appendix). The actuator is activated by cyclically turning the pump on and off at a ratio of with a total cycle duration of . By continuously venting the air in the actuator through a needle, the actuator extends and contracts (Fig. 1A). Note that the units are not capable of locomotion by themselves and require being physically attached to other units to move, which we refer to as the emergent behavior that our assembled robot exhibits.
Importantly, each unit aims to move as quickly as possible in the same predefined direction. To achieve this, it can adapt its behavior by varying the phase of actuation . As a first trial for a potential learning algorithm that is capable of learning emergent behavior, we implement a Monte Carlo scheme (29) in each unit. At the beginning of every learning step, the unit perturbs its phase according to , in which is a random number drawn from a uniform distribution on the interval and is the step size indicating the maximum change in phase allowed per learning step. Then, after performing cycles the unit determines its average velocity via the motion sensor and compares it with the velocity stored in memory. The unit will update the phases and velocities in its memory with probability , in which is a “temperature” that can be used to tune the acceptance probability when . We refer to this algorithm as the “Thermal algorithm” (see pseudocode in SI Appendix, Fig. S1).
Results of Two Units
In order to test if the system can optimize its movement in a predefined direction, we first focus on the most basic assembly of two active units (Fig. 1B). Note that we also need one dummy unit without an actuator at the end of the assembly to make sure all of the other actuators are connected on both sides. We place the units on a circular track and initialize the active units with a random actuation phase (Fig. 1C) and let the system run for learning iterations. At the beginning of each learning step the units update their phases independently, without exchanging any information. The pushing and pulling forces applied by the pneumatic actuators are the only interactions between the units (Movie S1). The results in Fig. 1 D and E show that the assembled robot learns to move forward and reaches its fastest speed after just (i.e., 20 learning iterations). This grants each unit a forward motion of (see SI Appendix for specifics on the velocity calculation).
For the remaining the units try to continuously increase their speed, resulting in random perturbations around the optimal behavior. We then repeat this learning experiment 56 times with the same settings, but with different random initial phases. We additionally model the robot behavior with a simple mass-spring system that includes static friction (SI Appendix) and perform 112 learning experiments. The distribution of the average robot velocities during learning is shown in Fig. 1F. We find that, while the average speed after learning is distributed around for both simulations and experiments, in the latter case we also find an additional optimum for . The distribution stabilizes at these two values after learning steps.
To determine the underlying cause for this bimodal velocity profile shown in experiments we map the behavior that this assembled robot can exhibit. Although each unit updates its phase individually, the overall motion of the assembled robot depends only on the phase difference between the two units. Fig. 1G shows a scan of the potential velocities the robot can exhibit, obtained in experiments by setting and varying between 0 and 1. From these results we find that the optimal speed is achieved at a phase difference of , at which the robot reaches a velocity of . At the optimal behavior, the system seems to undergo peristaltic behavior, where the phase difference is approximately equal to the actuation period (SI Appendix, Fig. S2). The same phase difference of is observed in our first learning experiment (Fig. 1D). Similarly, for the robot moves in the opposite direction with a reversed actuation pattern with a velocity of , while at we find a local maximum in experiments with lower speed of . These three cases are demonstrated in Fig. 1H, where we show both the starting and the end position after . It is important to note that while we expected symmetric behavior between forward and backward motion, as also demonstrated by the mass-spring model (Fig. 1G), in experiments we observe asymmetric motion. We believe that this asymmetry arises from a reduction of the friction due to the vibration of the pumps, which only occurs in the active units. The placement of the dummy unit therefore influences the behavior the robot can achieve. Also note that when applying the learning strategy in the simulations (Fig. 1 F and G) we find that the system only stabilizes around the maximum at , since no significant local optima are observed.
Results of Three Units
While the results for a robot assembled from two active units seem promising, the system’s average score function is relatively simple (Fig. 1G). Therefore, we next increase the complexity by performing experiments on a robot assembled from three active units and one dummy unit (Movie S1). In Fig. 2A we show the distribution of the average velocities during learning, of 112 simulations and 56 experiments. We observe both in experiments and simulations an initial improvement of during the first few learning iterations, similar to the assembled robot containing two active units. Interestingly, while the simulations stabilize at a with a mean positive value and an acceptance rate of (Fig. 2B), for experiments we observe a decrease of the mean after learning steps. This is accompanied by a decrease in the average acceptance rate of new phases, which approaches zero soon after the learning starts.
To understand the qualitative difference between simulations and experiments, and specifically the decrease in in experiments, we measure for all potential phase combinations in both simulations (Fig. 2C) and experiments (Fig. 2D). As expected, the results show a diverse range of possibilities, with multiple maxima. While our relatively simple model seems to capture the behavior of the assembled robot, there are two main differences in experiments. First, in experiments we observe significant noise, which could, e.g., be the result of how the measurements are taken (SI Appendix). Second, in experiments the performance of the robot seems to depend on where it is located on the circular track, which could be a result of variations in track width and the resulting friction (SI Appendix and SI Appendix, Fig. S3). This is supported when considering the standard deviation of the observed measured over 20 experiments, which exhibits nonnegligible variations (Fig. 2E and SI Appendix, Fig. S4A). For example, when running an experiment with and (bottom circle in Fig. 2 D and E), we find that for some parts of the track this assembled robot is capable of achieving , while the robot comes to a complete stop when reaching different parts of the track (Fig. 2F and Movie S2). In contrast, other regions in the phase space appear to be robust against track variations; see, e.g., the phase combination and (top circle in Fig. 2 D and E), for which the robot displacement is at a maximum (Fig. 2F and Movie S2).
A More Robust Learning Strategy
To understand the effect that noise and changes in environment can have on our proposed decentralized learning algorithm and the reason that the algorithm stops accepting new phases, we need to consider how the memory is updated. At the moment, each unit only updates its that is stored in memory when achieving a higher , with the exception of a probability to accept lower depending on the specified temperature . If in the previous step the noise caused an overestimation of the that was stored in memory, the probability to accept realistic becomes smaller. Similarly, if changes in the environment cause a decrease in , it will become less likely that the value of that is stored in memory is updated. This will influence the overall acceptance probability (Fig. 2B) and reduce the ability to adapt to variations and noise as also demonstrated by including noise in the simulations (SI Appendix, Fig. S5 A and B).
Therefore, while the Thermal algorithm performs well in a noiseless and static environment, in order to improve the adaptability of our system in real dynamic experimental conditions we need to change how memory is stored in each active unit. So far, we stored the last accepted phase and corresponding displacement in memory. Instead, we propose an improved algorithm in which we update the memory according to the newly obtained velocity , while still only updating the phase to in case of acceptance. We refer to the this new strategy as the “Flaky algorithm” (see pseudocode in SI Appendix, Fig. S1). As such, with this new formulation each unit is constantly aware of the landscape and does not rely on past information that may not be reliable anymore (Movie S1).
To determine the performance of this updated algorithm, we repeated the learning study and gathered data from 112 simulations and 56 experiments. The results in Fig. 2G show that with the Flaky algorithm the active units are now capable of adapting to variations in their environment, achieving velocities of . Interestingly, the units now accept of all new candidates, independent of the learning step (Fig. 2H). As a result, the robot is able to adapt its behavior to noise and variations in the environment, which significantly improves its overall robustness. The difference in adaptivity between the Thermal and Flaky algorithms can also be illustrated by simulations in which we implement a sudden reduction in friction during learning (SI Appendix, Fig. S6). Moreover, simulations indicate that the acceptance rate of the Flaky algorithm is not affected by the noise level, in contrast to the Thermal algorithm (SI Appendix, Fig. S5C).
While the Flaky algorithm seems more suitable for dynamic environments than the Thermal algorithm, we can still observe certain conditions for which the Flaky algorithm is starting to have problems. For example, when increasing the friction in this system as shown in SI Appendix, Fig. S3 E and F we observe potential behavior that changes more abruptly with a change in actuation phase. As such, the derivative of the score with respect to the phases is close to zero everywhere, except for a few boundaries between domains. In this case the memory in the Flaky algorithm can be reset almost instantaneously, after which it will accept any other phase combination. As such, the Flaky algorithm requires some notion of smoothness, with variations occurring over domains that are larger than the step size.
Adaptivity to Damage
With the adoption of the Flaky algorithm, we next explore if and how the assembled robot adapts to suddenly applied damage. In order to show this, we consider a learning experiment with three active units and one dummy unit, in which we intentionally damage the robot after learning steps by removing the venting needle in the second unit (Fig. 3 A and B and Movie S3). This dramatically alters the possible behaviors the robot can exhibit, as indicated by the contour plots of the score before and after damage (Fig. 3 C and D and SI Appendix, Fig. S4). Once damaged, the robot must find a new locomotion pattern. It should be noted that although the second actuator cannot inflate after damage we still observe the presence of a global optimum at , for which unit two and three are actuating simultaneously. While one would expect no influence of the damaged unit’s phase on the overall score (SI Appendix, Fig. S7), we find that by turning the pump on and off unit two seems to control its friction properties and as a result still plays a role in the optimal behavior. Importantly, when directly comparing the experiment and simulation results of the Thermal and the Flaky algorithms as shown in Movie S3, Fig. 3 E and F, and SI Appendix, Fig. S7, we find that the Thermal algorithm is not able to adapt to the damage and gets stuck at a certain phase, while the Flaky algorithm is able to overcome damage by completely adapting its behavior.
Scalability of the Learning Behavior
Finally, to verify if our learning strategy is scalable and can be applied to assembled robots consisting of more than three active units, we performed 112 simulations and 56 experiments on a robot consisting of seven active units and one dummy unit (Fig. 4A and Movie S4). In Fig. 4B we show the distribution of average velocity as a function of the learning step. While the average displacement is lower than before, the initial learning occurs within a comparable number of learning steps. Note that during learning the assembled robot occasionally does reach higher speeds up to , however it is not able to maintain these speeds, likely due to the random guesses made by each unit during normal operation. To better understand how the system’s behavior is affected by its size, we performed 112 simulations for sizes up to 20 active units. To determine the equilibrium velocity as shown in Fig. 4C, we fitted an exponential function to the average learning behavior (SI Appendix, Fig. S8A). We find that the equilibrium velocity decreases for larger number of units but appears to asymptotically reach a value close to .
Here, it should be noted that scalability could refer to two things, which are the behavior that the robotic platform can exhibit and the ability to learn “good” behavior. In the results shown in Fig. 4C we are investigating both factors for scalability simultaneously. One could argue that the velocity at the typical optimal peristaltic behavior that our system is aiming to achieve should not depend on the number of units, such that the system is apparently not able to find the optimal behavior for larger system size. This might be explained by the fact that our learning algorithm is exploring the behavior in a probabilistic way, such that equilibrium relates to some extent to the standard deviation of the system’s potential behavior (SI Appendix, Fig. S9). However, further studies should be performed to support this statement.
Interestingly, we do find that an increase in the number of units does not affect the rate of learning, as indicated by Fig. 4D, and that on average the units only need 35 learning steps to reach speeds within 80% of the equilibrium velocity . Furthermore, we can increase the equilibrium velocity at the cost of learning rate by decreasing the steps size used to update the phases (Fig. 4 C and D), which underpins the assumption that better behavior is attainable but depends on the stochastic nature of the algorithm.
Conclusion
In this work, we implemented a basic decentralized reinforced learning algorithm in a modular robot to improve and adapt its emergent global behavior. We demonstrated that such a decentralized learning approach is an effective way to adapt the robot’s behavior to a changing environment or damage. One of the most important findings of this work is the requirement on how previous scores (measured by the sensors) should be stored in each unit, where we find that the memory should be representative of the current environment in order for the algorithm to be adaptive. We implemented this in the Flaky algorithm by always keeping the last measured displacement, in contrast to the more classic Thermal algorithm that keeps the last accepted displacement, and therefore can only overcome sudden changes in conditions according to a specific temperature parameter.
Interestingly, our proposed system was able to learn peristaltic locomotion gaits (20, 30, 31) without any prior knowledge, making this decentralized learning strategy suitable for the exploration of environments whose governing physical conditions are unknown. Given that each unit in the robot is identical, no electrical connections are needed, and there is no centralized controller; the robot can be reconfigured into any shape and maintain its capability to learn, for example after shuffling the units (Movie S5). As a result, each module represents a unit cell capable of learning, such that the learning capability becomes essentially a material property, e.g., when cutting the robot in two, both parts would maintain the ability to learn. We therefore refer to this kind of system as “robotic matter” (32–34).
Although we performed one-dimensional experiments to simplify the potential behaviors, the robot does not need to be confined to a track (Movie S5). Importantly, we believe that our approach is applicable to a large range of robotic systems with different architectures, actuators, and sensors, including (soft) robots with many degrees of freedom that are assembled in two-dimensional patterns (20), or that fully deform in three dimensions (16). Furthermore, the simplicity of our learning strategy and the fact that it can be coded using only a few lines makes it suitable for (microsized) robots that have limited computational power and dedicated electronic circuits (35–37) and could potentially even be realized through other means than an electronic circuit such as microfluidics (9, 38), and even material behavior (39). The simplicity and robustness of our proposed robotic matter is an important step toward robotic systems that can autonomously learn to thrive in a broad range of applications at different length scales, spanning from in vivo healthcare to disaster relief and space exploration.
Materials and Methods
Details on the materials and methods are provided in SI Appendix.
Supplementary Material
Acknowledgments
This work is part of the Dutch Research Council (NWO) and was performed at the research institute of AMOLF. This project is part of the research program Innovational Research Incentives Scheme Veni from NWO with project number 15868 (NWO) and has received funding from the European Union’s Horizon 2020 research and innovation program under grant agreement 767195.
Footnotes
The authors declare no competing interest.
This article is a PNAS Direct Submission. C.D. is a guest editor invited by the Editorial Board.
This article contains supporting information online at https://www.pnas.org/lookup/suppl/doi:10.1073/pnas.2017015118/-/DCSupplemental.
Data Availability
The experimental and numerical data that support the findings of this study and all computer algorithms necessary to run the models have been deposited at Figshare (10.6084/m9.figshare.13626110).
References
- 1.Hwangbo J., et al. , Learning agile and dynamic motor skills for legged robots. Sci. Robot. 4, eaau5872 (2019). [DOI] [PubMed] [Google Scholar]
- 2.Kwiatkowski R., Lipson H., Task-agnostic self-modeling machines. Sci. Robot. 4, eaau9354 (2019). [DOI] [PubMed] [Google Scholar]
- 3.Rus D., Tolley M. T., Design, fabrication and control of soft robots. Nature 521, 467–475 (2015). [DOI] [PubMed] [Google Scholar]
- 4.Cully A., Clune J., Tarapore D., Mouret J.-B., Robots that can adapt like animals. Nature 521, 503–507 (2015). [DOI] [PubMed] [Google Scholar]
- 5.Haarnoja T., et al., Learning to walk via deep reinforcement learning. arXiv [Preprint] (2018). https://arxiv.org/abs/1812.11103v3 (Accessed 1 May 2020).
- 6.Ficuciello F., Migliozzi A., Laudante G., Falco P., Siciliano B., Vision-based grasp learning of an anthropomorphic hand-arm system in a synergy-based control framework. Sci. Robot. 4, eaao4900 (2019). [DOI] [PubMed] [Google Scholar]
- 7.Mahler J., et al. , Learning ambidextrous robot grasping policies. Sci. Robot. 4, eaau4984 (2019). [DOI] [PubMed] [Google Scholar]
- 8.Fazeli N., et al. , See, feel, act: Hierarchical learning for complex manipulation skills with multisensory fusion. Sci. Robot. 4, eaav3123 (2019). [DOI] [PubMed] [Google Scholar]
- 9.Wehner M., et al. , An integrated design and fabrication strategy for entirely soft, autonomous robots. Nature 536, 451–455 (2016). [DOI] [PubMed] [Google Scholar]
- 10.Majidi C., Soft-matter engineering for soft robotics. Adv. Mater. Tech. 4, 1800477 (2019). [Google Scholar]
- 11.Tolley M. T., et al. , A resilient, untethered soft robot. Soft Robot. 1, 213–223 (2014). [Google Scholar]
- 12.Chen T., Bilal O. R., Shea K., Daraio C., Harnessing bistability for directional propulsion of soft, untethered robots. Proc. Natl. Acad. Sci. U.S.A. 115, 5698–5702 (2018). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13.Howard D., et al. , Evolving embodied intelligence from materials to machines. Nat. Mach. Intel. 1, 12–19 (2019). [Google Scholar]
- 14.Laschi C., Cianchetti M., Soft robotics: New perspectives for robot bodyware and control. Front. Bioeng. Biotechnol. 2, (2014). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.Mengüç Y., Correll N., Kramer R., Paik J., Will robots be bodies with brains or brains with bodies? Sci. Robot. 2, eaar4527 (2017). [DOI] [PubMed] [Google Scholar]
- 16.Shepherd R. F., et al. , Multigait soft robot. Proc. Natl. Acad. Sci. U.S.A. 108, 20400–20403 (2011). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.Williamson R., Chrachri A., Cephalopod neural networks. Neurosignals 13, 87–98 (2004). [DOI] [PubMed] [Google Scholar]
- 18.Sünderhauf N., et al. , The limits and potentials of deep learning for robotics. Int. J. Robot Res. 37, 405–420 (2018). [Google Scholar]
- 19.Slavkov I., et al. , Morphogenesis in robot swarms. Sci. Robot. 3, eaau9178 (2018). [DOI] [PubMed] [Google Scholar]
- 20.Li S., et al. , Particle robotics based on statistical mechanics of loosely coupled components. Nature 567, 361–365 (2019). [DOI] [PubMed] [Google Scholar]
- 21.Brambilla M., Ferrante E., Birattari M., Dorigo M., Swarm robotics: A review from the swarm engineering perspective. Swarm Intel. 7, 1–41 (2013). [Google Scholar]
- 22.Rubenstein M., Cornejo A., Nagpal R., Programmable self-assembly in a thousand-robot swarm. Science 345, 795–799 (2014). [DOI] [PubMed] [Google Scholar]
- 23.Peleg O., Peters J. M., Salcedo M. K., Mahadevan L., Collective mechanical adaptation of honeybee swarms. Nat. Phys. 14, 1193 (2018). [Google Scholar]
- 24.Romanishin J. W., Gilpin K., Claici S., Rus D., “3D M-blocks: Self-reconfiguring robots capable of locomotion via pivoting in three dimensions” in 2015 IEEE International Conference on Robotics and Automation (ICRA) (IEEE, 2015), pp. 1925–1932. [Google Scholar]
- 25.Augugliaro F., et al. , The flight assembled architecture installation: Cooperative construction with flying machines. IEEE Contr. Syst. Mag. 34, 46–64 (2014). [Google Scholar]
- 26.Shimizu M., Ishiguro A., Kawakatsu T., “Slimebot: A modular robot that exploits emergent phenomena” in Proceedings of the 2005 IEEE International Conference on Robotics and Automation (IEEE, 2005), pp. 2982–2987. [Google Scholar]
- 27.Leottau D. L., Ruiz-del Solar J., Babuška R., Decentralized reinforcement learning of robot behaviors. Artif. Intell. 256, 130–159 (2018). [Google Scholar]
- 28.Christensen D. J., Schultz U. P., Stoy K., A distributed and morphology-independent strategy for adaptive locomotion in self-reconfigurable modular robots. Robot. Autonom. Syst. 61, 1021–1035 (2013). [Google Scholar]
- 29.Metropolis N., Rosenbluth A. W., Rosenbluth M. N., Teller A. H., Teller E., Equation of state calculations by fast computing machines. J. Chem. Phys. 21, 1087–1092 (1953). [Google Scholar]
- 30.Seok S., Onal C. D., Cho K.-J., Wood R. J., Rus D., Kim S., Meshworm: A peristaltic soft robot with antagonistic nickel titanium coil actuators. IEEE ASME Trans. Mechatron. 18, 1485–1497 (2012). [Google Scholar]
- 31.Boxerbaum A. S., Shaw K. M., Chiel H. J., Quinn R. D., Continuous wave peristaltic motion in a robot. Int. J. Robot Res. 31, 302–318 (2012). [Google Scholar]
- 32.Brandenbourger M., Locsin X., Lerner E., Coulais C., Non-reciprocal robotic metamaterials. Nat. Commun. 10, 1–8 (2019). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 33.McEvoy M. A., Correll N., Materials that couple sensing, actuation, computation, and communication. Science 347, 1261689 (2015). [DOI] [PubMed] [Google Scholar]
- 34.Kotikian A., et al. , Untethered soft robotic matter with passive control of shape morphing and propulsion. Sci. Robot. 4, 7044 (2019). [DOI] [PubMed] [Google Scholar]
- 35.Miskin M. Z., et al. , Graphene-based bimorphs for micron-sized, autonomous origami machines. Proc. Natl. Acad. Sci. U.S.A. 115, 466–470 (2018). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 36.Ceylan H., Giltinan J., Kozielski K., Metin S., Mobile microrobots for bioengineering applications. Lab Chip 17, 1705–1724 (2017). [DOI] [PubMed] [Google Scholar]
- 37.Medina-Sánchez M., Magdanz V., Guix M., Fomin V. M., Schmidt O. G., Swimming microrobots: Soft, reconfigurable, and smart. Adv. Funct. Mater. 28, 1707228 (2018). [Google Scholar]
- 38.Rothemund P., et al. , A soft, bistable valve for autonomous control of soft actuators. Sci. Robot. 3, eaar7986 (2018). [DOI] [PubMed] [Google Scholar]
- 39.Zhang H., Zeng H., Priimagi A., Ikkala O., Programmable responsive hydrogels inspired by classical conditioning algorithm. Nat. Commun. 10, 3267 (2019). [DOI] [PMC free article] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Data Availability Statement
The experimental and numerical data that support the findings of this study and all computer algorithms necessary to run the models have been deposited at Figshare (10.6084/m9.figshare.13626110).