Skip to main content
Proceedings of the National Academy of Sciences of the United States of America logoLink to Proceedings of the National Academy of Sciences of the United States of America
. 2021 May 10;118(21):e2017015118. doi: 10.1073/pnas.2017015118

Continuous learning of emergent behavior in robotic matter

Giorgio Oliveri a, Lucas C van Laake a, Cesare Carissimo a, Clara Miette a, Johannes T B Overvelde a,1
PMCID: PMC8166149  PMID: 33972408

Significance

In the last century, robots have been revolutionizing our lives, augmenting human actions with greater precision and repeatability. Unfortunately, most robotic systems can only operate in controlled environments. While increasing the complexity of the centralized controller is an instinctive direction to enable robots that are capable of autonomously adapting to their environment, there are ample examples in nature where adaptivity emerges from simpler decentralized processes. Here we perform experiments and simulations on a modular and scalable robotic platform in which each unit is stochastically updating its own behavior to explore requirements needed for a decentralized learning strategy capable of achieving locomotion in a continuously changing environment or when undergoing damage.

Keywords: emergent behavior, reinforced learning, modular robot, dynamic environment

Abstract

One of the main challenges in robotics is the development of systems that can adapt to their environment and achieve autonomous behavior. Current approaches typically aim to achieve this by increasing the complexity of the centralized controller by, e.g., direct modeling of their behavior, or implementing machine learning. In contrast, we simplify the controller using a decentralized and modular approach, with the aim of finding specific requirements needed for a robust and scalable learning strategy in robots. To achieve this, we conducted experiments and simulations on a specific robotic platform assembled from identical autonomous units that continuously sense their environment and react to it. By letting each unit adapt its behavior independently using a basic Monte Carlo scheme, the assembled system is able to learn and maintain optimal behavior in a dynamic environment as long as its memory is representative of the current environment, even when incurring damage. We show that the physical connection between the units is enough to achieve learning, and no additional communication or centralized information is required. As a result, such a distributed learning approach can be easily scaled to larger assemblies, blurring the boundaries between materials and robots, paving the way for a new class of modular “robotic matter” that can autonomously learn to thrive in dynamic or unfamiliar situations, for example, encountered by soft robots or self-assembled (micro)robots in various environments spanning from the medical realm to space explorations.


Traditional robots typically exhibit well-defined motions and are often controlled by a centralized controller that uses sensor data to provide feedback to its actuators (1, 2). While these systems are capable of operating in well-known and controlled environments, adapting their behavior to various unforeseen circumstances remains one of the main challenges in robotics (3). Toward this goal, reinforced learning strategies can be used to deal with more complex tasks that cannot directly be programmed in the behavior of the robot. By sensing their environment, robots can build models of themselves (4) or optimize human-made models with machine learning strategies to improve their behavior (1, 2, 48). Apart from a clear division between training and task execution, these systems rely on a centralized architecture that—with an increasing number of active components—demands advanced models and higher computational power.

A recent key idea to simplify the computational task, while allowing for adaptability to the environment, is to make soft robots. These are robots fabricated with materials with stiffness resembling human tissue (3, 912), that exhibit so-called embodied intelligence (1315) and adjust their shape when subjected to forces resulting from interactions with their environment. While this physical adaptability enhances the robustness of the robot’s behavior, allowing it to, e.g., operate in more complex environments, it is not guaranteed that they operate optimally, or even operate at all. For example, a soft robot that is able to walk on various surfaces still requires a different actuation sequence to crawl through a narrow opening (16). At the moment, and because of the unpredictable nature of the robot’s behavior in varying environments, this sequence is externally (and manually) adjusted. A natural question to ask is if such higher levels of decision-making and learning can also be embodied in robots, similar to, e.g., the decentralized neural networks of squids (17, 18).

In fact, decentralized approaches are already applied in swarm robotics, where an increasing number of active components can cooperatively achieve complex behavior when sensing or communication capabilities are incorporated into the individual units (1922). In such a robotic system, behavior emerges from local interactions, rather than being centrally controlled. Inspired by social animals such as ants, bees, or birds, researchers have been able to achieve emerging properties out of simple unit-to-unit interactions, such as adaptation to mechanical stimuli (23), self-assembly (22, 24), construction (25), and locomotion (20, 26). Interestingly, such behavior does not have to be programmed and could also be learned using the same decentralized architecture (27, 28).

In this work, we develop and perform experiments on modular robots that are connected to form a single body, in order to explore similar decentralized strategies. Using this platform, our aim is to simplify the learning strategy, while focusing on the fundamental challenge to ensure robust and optimal behavior under varying external conditions and damage. We do so by having each element continuously learn and adapt to changing circumstances. To that end, we develop a fully decentralized and universal reinforced learning strategy that can be easily implemented, is scalable, and allows the robotic system to autonomously achieve and maintain specific emergent behavior.

Individual Unit Behavior

Such an approach holds potential for a wide variety of robotic systems, and we illustrate this here with a robot made of several identical units connected by soft actuators that periodically extend and contract along one direction. Each robotic unit operates independently and consists of a microcontroller, optical motion sensor, pump, pneumatic actuator, and battery (Fig. 1A and SI Appendix). The actuator is activated by cyclically turning the pump on and off at a ratio of α=ton/tcycle=0.4 with a total cycle duration of tcycle=2s. By continuously venting the air in the actuator through a needle, the actuator extends and contracts (Fig. 1A). Note that the units are not capable of locomotion by themselves and require being physically attached to other units to move, which we refer to as the emergent behavior that our assembled robot exhibits.

Fig. 1.

Fig. 1.

Robotic unit design and initial learning experiments for an assembled robot consisting of two active units. (A) Design of a single robotic unit in its relaxed and extended state. (B) Robot assembled from two active units and one dummy unit. (C) Each active unit can vary its phase ϕitcycle, which affects the actuation timing as shown by the cyclic pressure in the soft actuators. (D and E) Evolution of phases ϕi and measured velocity Ui in the active units as a function of time for a single learning experiment. (F) Distribution of average robot velocity Ū as a function of the learning step for 112 learning simulations and 56 learning experiments. (G) Simulated and experimental velocity as function of phase difference Δϕ=ϕ2ϕ1, exemplified by three experiments (H).

Importantly, each unit aims to move as quickly as possible in the same predefined direction. To achieve this, it can adapt its behavior by varying the phase of actuation ϕitcycle. As a first trial for a potential learning algorithm that is capable of learning emergent behavior, we implement a Monte Carlo scheme (29) in each unit. At the beginning of every learning step, the unit perturbs its phase according to ϕi=ϕi+ϵΔs, in which ϵ is a random number drawn from a uniform distribution on the interval [1,1] and Δs=0.1 is the step size indicating the maximum change in phase allowed per learning step. Then, after performing nact=2 cycles the unit determines its average velocity Ui via the motion sensor and compares it with the velocity Ui stored in memory. The unit will update the phases and velocities in its memory with probability e(UiUi)/T, in which T is a “temperature” that can be used to tune the acceptance probability when UiUi<0. We refer to this algorithm as the “Thermal algorithm” (see pseudocode in SI Appendix, Fig. S1).

Results of Two Units

In order to test if the system can optimize its movement in a predefined direction, we first focus on the most basic assembly of two active units (Fig. 1B). Note that we also need one dummy unit without an actuator at the end of the assembly to make sure all of the other actuators are connected on both sides. We place the units on a circular track and initialize the active units with a random actuation phase ϕi (Fig. 1C) and let the system run for nlearn=100 learning iterations. At the beginning of each learning step the units update their phases independently, without exchanging any information. The pushing and pulling forces applied by the pneumatic actuators are the only interactions between the units (Movie S1). The results in Fig. 1 D and E show that the assembled robot learns to move forward and reaches its fastest speed after just 80s (i.e., 20 learning iterations). This grants each unit a forward motion of 4.5mm/cycle (see SI Appendix for specifics on the velocity calculation).

For the remaining 320s the units try to continuously increase their speed, resulting in random perturbations around the optimal behavior. We then repeat this learning experiment 56 times with the same settings, but with different random initial phases. We additionally model the robot behavior with a simple mass-spring system that includes static friction (SI Appendix) and perform 112 learning experiments. The distribution of the average robot velocities Ū=iUi during learning is shown in Fig. 1F. We find that, while the average speed after learning is distributed around Ū4.5mm/cycle for both simulations and experiments, in the latter case we also find an additional optimum for Ū1mm/cycle. The distribution stabilizes at these two values after 20 learning steps.

To determine the underlying cause for this bimodal velocity profile shown in experiments we map the behavior that this assembled robot can exhibit. Although each unit updates its phase individually, the overall motion of the assembled robot depends only on the phase difference Δϕ=ϕ2ϕ1 between the two units. Fig. 1G shows a scan of the potential velocities Ū the robot can exhibit, obtained in experiments by setting ϕ1=0 and varying ϕ2 between 0 and 1. From these results we find that the optimal speed is achieved at a phase difference of Δϕ0.4, at which the robot reaches a velocity of 4.5mm/cycle. At the optimal behavior, the system seems to undergo peristaltic behavior, where the phase difference is approximately equal to the actuation period Δϕα (SI Appendix, Fig. S2). The same phase difference of Δϕ0.4 is observed in our first learning experiment (Fig. 1D). Similarly, for Δϕ=0.6 the robot moves in the opposite direction with a reversed actuation pattern with a velocity of 3mm/cycle, while at Δϕ0.9 we find a local maximum in experiments with lower speed of 1mm/cycle. These three cases are demonstrated in Fig. 1H, where we show both the starting and the end position after 60s. It is important to note that while we expected symmetric behavior between forward and backward motion, as also demonstrated by the mass-spring model (Fig. 1G), in experiments we observe asymmetric motion. We believe that this asymmetry arises from a reduction of the friction due to the vibration of the pumps, which only occurs in the active units. The placement of the dummy unit therefore influences the behavior the robot can achieve. Also note that when applying the learning strategy in the simulations (Fig. 1 F and G) we find that the system only stabilizes around the maximum at Δϕ0.4, since no significant local optima are observed.

Results of Three Units

While the results for a robot assembled from two active units seem promising, the system’s average score function Ū is relatively simple (Fig. 1G). Therefore, we next increase the complexity by performing experiments on a robot assembled from three active units and one dummy unit (Movie S1). In Fig. 2A we show the distribution of the average velocities Ū during learning, of 112 simulations and 56 experiments. We observe both in experiments and simulations an initial improvement of Ū during the first few 50 learning iterations, similar to the assembled robot containing two active units. Interestingly, while the simulations stabilize at a Ū with a mean positive value and an acceptance rate of 0.25 (Fig. 2B), for experiments we observe a decrease of the mean Ū after 200 learning steps. This is accompanied by a decrease in the average acceptance rate of new phases, which approaches zero soon after the learning starts.

Fig. 2.

Fig. 2.

Adaptability to variations in the environment of an assembled robot consisting of three active units. (A and B) Distribution of average velocities Ū and average acceptance rate p¯ for the Thermal algorithm, as a function of the learning step, for 112 simulations and 56 experiments. (C and D) Average velocity Ū obtained in simulations and experiments as a function of all of the possible combinations of phases ϕ3ϕ1 and ϕ2ϕ1 (with ϕ1=0). (E) Standard deviation of the average velocity Ū observed over 20 experimental runs. (F) Two different experiments with fixed phases, to highlight the effect of the track on the robot’s behavior. (G and H) Distribution of average velocities Ū and average acceptance rate p¯ for the Flaky algorithm as a function of the learning step, for 112 simulations and 56 experiments.

To understand the qualitative difference between simulations and experiments, and specifically the decrease in Ū in experiments, we measure Ū for all potential phase combinations in both simulations (Fig. 2C) and experiments (Fig. 2D). As expected, the results show a diverse range of possibilities, with multiple maxima. While our relatively simple model seems to capture the behavior of the assembled robot, there are two main differences in experiments. First, in experiments we observe significant noise, which could, e.g., be the result of how the measurements are taken (SI Appendix). Second, in experiments the performance of the robot seems to depend on where it is located on the circular track, which could be a result of variations in track width and the resulting friction (SI Appendix and SI Appendix, Fig. S3). This is supported when considering the standard deviation of the observed Ū measured over 20 experiments, which exhibits nonnegligible variations (Fig. 2E and SI Appendix, Fig. S4A). For example, when running an experiment with ϕ2ϕ1=0.4 and ϕ3ϕ1=0.2 (bottom circle in Fig. 2 D and E), we find that for some parts of the track this assembled robot is capable of achieving Ū2mm/cycle, while the robot comes to a complete stop when reaching different parts of the track (Fig. 2F and Movie S2). In contrast, other regions in the phase space appear to be robust against track variations; see, e.g., the phase combination ϕ2ϕ10.4 and ϕ3ϕ10.8 (top circle in Fig. 2 D and E), for which the robot displacement Ū4.5mm/cycle is at a maximum (Fig. 2F and Movie S2).

A More Robust Learning Strategy

To understand the effect that noise and changes in environment can have on our proposed decentralized learning algorithm and the reason that the algorithm stops accepting new phases, we need to consider how the memory is updated. At the moment, each unit only updates its Ui that is stored in memory when achieving a higher Ui, with the exception of a probability to accept lower Ui depending on the specified temperature T. If in the previous step the noise caused an overestimation of the Ui that was stored in memory, the probability to accept realistic Ui becomes smaller. Similarly, if changes in the environment cause a decrease in Ui, it will become less likely that the value of Ui that is stored in memory is updated. This will influence the overall acceptance probability (Fig. 2B) and reduce the ability to adapt to variations and noise as also demonstrated by including noise in the simulations (SI Appendix, Fig. S5 A and B).

Therefore, while the Thermal algorithm performs well in a noiseless and static environment, in order to improve the adaptability of our system in real dynamic experimental conditions we need to change how memory is stored in each active unit. So far, we stored the last accepted phase ϕi and corresponding displacement Ui in memory. Instead, we propose an improved algorithm in which we update the memory according to the newly obtained velocity Ui, while still only updating the phase to ϕi in case of acceptance. We refer to the this new strategy as the “Flaky algorithm” (see pseudocode in SI Appendix, Fig. S1). As such, with this new formulation each unit is constantly aware of the landscape and does not rely on past information that may not be reliable anymore (Movie S1).

To determine the performance of this updated algorithm, we repeated the learning study and gathered data from 112 simulations and 56 experiments. The results in Fig. 2G show that with the Flaky algorithm the active units are now capable of adapting to variations in their environment, achieving velocities of Ū4mm/cycle. Interestingly, the units now accept 60% of all new candidates, independent of the learning step (Fig. 2H). As a result, the robot is able to adapt its behavior to noise and variations in the environment, which significantly improves its overall robustness. The difference in adaptivity between the Thermal and Flaky algorithms can also be illustrated by simulations in which we implement a sudden reduction in friction during learning (SI Appendix, Fig. S6). Moreover, simulations indicate that the acceptance rate of the Flaky algorithm is not affected by the noise level, in contrast to the Thermal algorithm (SI Appendix, Fig. S5C).

While the Flaky algorithm seems more suitable for dynamic environments than the Thermal algorithm, we can still observe certain conditions for which the Flaky algorithm is starting to have problems. For example, when increasing the friction in this system as shown in SI Appendix, Fig. S3 E and F we observe potential behavior that changes more abruptly with a change in actuation phase. As such, the derivative of the score with respect to the phases is close to zero everywhere, except for a few boundaries between domains. In this case the memory in the Flaky algorithm can be reset almost instantaneously, after which it will accept any other phase combination. As such, the Flaky algorithm requires some notion of smoothness, with variations occurring over domains that are larger than the step size.

Adaptivity to Damage

With the adoption of the Flaky algorithm, we next explore if and how the assembled robot adapts to suddenly applied damage. In order to show this, we consider a learning experiment with three active units and one dummy unit, in which we intentionally damage the robot after 300 learning steps by removing the venting needle in the second unit (Fig. 3 A and B and Movie S3). This dramatically alters the possible behaviors the robot can exhibit, as indicated by the contour plots of the score before and after damage (Fig. 3 C and D and SI Appendix, Fig. S4). Once damaged, the robot must find a new locomotion pattern. It should be noted that although the second actuator cannot inflate after damage we still observe the presence of a global optimum at ϕ2ϕ1ϕ3ϕ1, for which unit two and three are actuating simultaneously. While one would expect no influence of the damaged unit’s phase on the overall score (SI Appendix, Fig. S7), we find that by turning the pump on and off unit two seems to control its friction properties and as a result still plays a role in the optimal behavior. Importantly, when directly comparing the experiment and simulation results of the Thermal and the Flaky algorithms as shown in Movie S3, Fig. 3 E and F, and SI Appendix, Fig. S7, we find that the Thermal algorithm is not able to adapt to the damage and gets stuck at a certain phase, while the Flaky algorithm is able to overcome damage by completely adapting its behavior.

Fig. 3.

Fig. 3.

Adaptability of the assembled robot to damage using the Thermal and Flaky algorithms. (A and B) An assembled robot consisting of three active units is damaged by removing one of the needles. (C and D) Average velocity Ū obtained in experiments for an intact and damaged robot, as a function of the possible range of combinations of phases ϕ3ϕ1 and ϕ2ϕ1 (with ϕ1=0). The white and black dots represent the tried phases in two learning experiments in which the robot is damaged after 300 learning steps, using the (E) Thermal and (F) Flaky algorithms, respectively.

Scalability of the Learning Behavior

Finally, to verify if our learning strategy is scalable and can be applied to assembled robots consisting of more than three active units, we performed 112 simulations and 56 experiments on a robot consisting of seven active units and one dummy unit (Fig. 4A and Movie S4). In Fig. 4B we show the distribution of average velocity Ū as a function of the learning step. While the average displacement Ū2.5mm/cycle is lower than before, the initial learning occurs within a comparable number of learning steps. Note that during learning the assembled robot occasionally does reach higher speeds up to Ū4mm/cycle, however it is not able to maintain these speeds, likely due to the random guesses made by each unit during normal operation. To better understand how the system’s behavior is affected by its size, we performed 112 simulations for sizes up to 20 active units. To determine the equilibrium velocity Ūeq as shown in Fig. 4C, we fitted an exponential function to the average learning behavior (SI Appendix, Fig. S8A). We find that the equilibrium velocity decreases for larger number of units but appears to asymptotically reach a value close to Ū1.7mm/cycle.

Fig. 4.

Fig. 4.

Scalability of the learning behavior for an assembled robot ranging from 2 to 20 active units. (A) Initial and final position after 300 learning steps (1,200 s) of an assembled robot consisting of seven active units. (B) Distribution of average velocities Ū as a function of the learning step, for 112 simulations and 56 experiments. (C) Relation between the number of active units and the equilibrium velocity Ūeq, obtained using simulations, for Δs=0.1 and Δs=0.1/4. (D) Average number of learning steps, and standard deviation, needed to reach equilibrium (i.e., 0.8Ūeq) as a function of the number of active units. Here we only consider the simulation which reach the 0.8Ūeq threshold within the assigned number of learning steps. While for Δs=0.1 all of the simulations reach the threshold, for Δs=0.1/4 we find that a percentage of simulations do not reach 0.8Ūeq (SI Appendix, Fig. S8B).

Here, it should be noted that scalability could refer to two things, which are the behavior that the robotic platform can exhibit and the ability to learn “good” behavior. In the results shown in Fig. 4C we are investigating both factors for scalability simultaneously. One could argue that the velocity at the typical optimal peristaltic behavior that our system is aiming to achieve should not depend on the number of units, such that the system is apparently not able to find the optimal behavior for larger system size. This might be explained by the fact that our learning algorithm is exploring the behavior in a probabilistic way, such that equilibrium relates to some extent to the standard deviation of the system’s potential behavior (SI Appendix, Fig. S9). However, further studies should be performed to support this statement.

Interestingly, we do find that an increase in the number of units does not affect the rate of learning, as indicated by Fig. 4D, and that on average the units only need 35 learning steps to reach speeds within 80% of the equilibrium velocity Ūeq. Furthermore, we can increase the equilibrium velocity Ūeq at the cost of learning rate by decreasing the steps size used to update the phases (Fig. 4 C and D), which underpins the assumption that better behavior is attainable but depends on the stochastic nature of the algorithm.

Conclusion

In this work, we implemented a basic decentralized reinforced learning algorithm in a modular robot to improve and adapt its emergent global behavior. We demonstrated that such a decentralized learning approach is an effective way to adapt the robot’s behavior to a changing environment or damage. One of the most important findings of this work is the requirement on how previous scores (measured by the sensors) should be stored in each unit, where we find that the memory should be representative of the current environment in order for the algorithm to be adaptive. We implemented this in the Flaky algorithm by always keeping the last measured displacement, in contrast to the more classic Thermal algorithm that keeps the last accepted displacement, and therefore can only overcome sudden changes in conditions according to a specific temperature parameter.

Interestingly, our proposed system was able to learn peristaltic locomotion gaits (20, 30, 31) without any prior knowledge, making this decentralized learning strategy suitable for the exploration of environments whose governing physical conditions are unknown. Given that each unit in the robot is identical, no electrical connections are needed, and there is no centralized controller; the robot can be reconfigured into any shape and maintain its capability to learn, for example after shuffling the units (Movie S5). As a result, each module represents a unit cell capable of learning, such that the learning capability becomes essentially a material property, e.g., when cutting the robot in two, both parts would maintain the ability to learn. We therefore refer to this kind of system as “robotic matter” (3234).

Although we performed one-dimensional experiments to simplify the potential behaviors, the robot does not need to be confined to a track (Movie S5). Importantly, we believe that our approach is applicable to a large range of robotic systems with different architectures, actuators, and sensors, including (soft) robots with many degrees of freedom that are assembled in two-dimensional patterns (20), or that fully deform in three dimensions (16). Furthermore, the simplicity of our learning strategy and the fact that it can be coded using only a few lines makes it suitable for (microsized) robots that have limited computational power and dedicated electronic circuits (3537) and could potentially even be realized through other means than an electronic circuit such as microfluidics (9, 38), and even material behavior (39). The simplicity and robustness of our proposed robotic matter is an important step toward robotic systems that can autonomously learn to thrive in a broad range of applications at different length scales, spanning from in vivo healthcare to disaster relief and space exploration.

Materials and Methods

Details on the materials and methods are provided in SI Appendix.

Supplementary Material

Supplementary File
pnas.2017015118.sapp.pdf (10.3MB, pdf)
Supplementary File
Download video file (52.9MB, mp4)
Supplementary File
Download video file (31.2MB, mp4)
Supplementary File
Download video file (59MB, mp4)
Supplementary File
Download video file (20.7MB, mp4)
Supplementary File
Download video file (35.3MB, mp4)

Acknowledgments

This work is part of the Dutch Research Council (NWO) and was performed at the research institute of AMOLF. This project is part of the research program Innovational Research Incentives Scheme Veni from NWO with project number 15868 (NWO) and has received funding from the European Union’s Horizon 2020 research and innovation program under grant agreement 767195.

Footnotes

The authors declare no competing interest.

This article is a PNAS Direct Submission. C.D. is a guest editor invited by the Editorial Board.

This article contains supporting information online at https://www.pnas.org/lookup/suppl/doi:10.1073/pnas.2017015118/-/DCSupplemental.

Data Availability

The experimental and numerical data that support the findings of this study and all computer algorithms necessary to run the models have been deposited at Figshare (10.6084/m9.figshare.13626110).

References

  • 1.Hwangbo J., et al. , Learning agile and dynamic motor skills for legged robots. Sci. Robot. 4, eaau5872 (2019). [DOI] [PubMed] [Google Scholar]
  • 2.Kwiatkowski R., Lipson H., Task-agnostic self-modeling machines. Sci. Robot. 4, eaau9354 (2019). [DOI] [PubMed] [Google Scholar]
  • 3.Rus D., Tolley M. T., Design, fabrication and control of soft robots. Nature 521, 467–475 (2015). [DOI] [PubMed] [Google Scholar]
  • 4.Cully A., Clune J., Tarapore D., Mouret J.-B., Robots that can adapt like animals. Nature 521, 503–507 (2015). [DOI] [PubMed] [Google Scholar]
  • 5.Haarnoja T., et al., Learning to walk via deep reinforcement learning. arXiv [Preprint] (2018). https://arxiv.org/abs/1812.11103v3 (Accessed 1 May 2020).
  • 6.Ficuciello F., Migliozzi A., Laudante G., Falco P., Siciliano B., Vision-based grasp learning of an anthropomorphic hand-arm system in a synergy-based control framework. Sci. Robot. 4, eaao4900 (2019). [DOI] [PubMed] [Google Scholar]
  • 7.Mahler J., et al. , Learning ambidextrous robot grasping policies. Sci. Robot. 4, eaau4984 (2019). [DOI] [PubMed] [Google Scholar]
  • 8.Fazeli N., et al. , See, feel, act: Hierarchical learning for complex manipulation skills with multisensory fusion. Sci. Robot. 4, eaav3123 (2019). [DOI] [PubMed] [Google Scholar]
  • 9.Wehner M., et al. , An integrated design and fabrication strategy for entirely soft, autonomous robots. Nature 536, 451–455 (2016). [DOI] [PubMed] [Google Scholar]
  • 10.Majidi C., Soft-matter engineering for soft robotics. Adv. Mater. Tech. 4, 1800477 (2019). [Google Scholar]
  • 11.Tolley M. T., et al. , A resilient, untethered soft robot. Soft Robot. 1, 213–223 (2014). [Google Scholar]
  • 12.Chen T., Bilal O. R., Shea K., Daraio C., Harnessing bistability for directional propulsion of soft, untethered robots. Proc. Natl. Acad. Sci. U.S.A. 115, 5698–5702 (2018). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Howard D., et al. , Evolving embodied intelligence from materials to machines. Nat. Mach. Intel. 1, 12–19 (2019). [Google Scholar]
  • 14.Laschi C., Cianchetti M., Soft robotics: New perspectives for robot bodyware and control. Front. Bioeng. Biotechnol. 2, (2014). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Mengüç Y., Correll N., Kramer R., Paik J., Will robots be bodies with brains or brains with bodies? Sci. Robot. 2, eaar4527 (2017). [DOI] [PubMed] [Google Scholar]
  • 16.Shepherd R. F., et al. , Multigait soft robot. Proc. Natl. Acad. Sci. U.S.A. 108, 20400–20403 (2011). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Williamson R., Chrachri A., Cephalopod neural networks. Neurosignals 13, 87–98 (2004). [DOI] [PubMed] [Google Scholar]
  • 18.Sünderhauf N., et al. , The limits and potentials of deep learning for robotics. Int. J. Robot Res. 37, 405–420 (2018). [Google Scholar]
  • 19.Slavkov I., et al. , Morphogenesis in robot swarms. Sci. Robot. 3, eaau9178 (2018). [DOI] [PubMed] [Google Scholar]
  • 20.Li S., et al. , Particle robotics based on statistical mechanics of loosely coupled components. Nature 567, 361–365 (2019). [DOI] [PubMed] [Google Scholar]
  • 21.Brambilla M., Ferrante E., Birattari M., Dorigo M., Swarm robotics: A review from the swarm engineering perspective. Swarm Intel. 7, 1–41 (2013). [Google Scholar]
  • 22.Rubenstein M., Cornejo A., Nagpal R., Programmable self-assembly in a thousand-robot swarm. Science 345, 795–799 (2014). [DOI] [PubMed] [Google Scholar]
  • 23.Peleg O., Peters J. M., Salcedo M. K., Mahadevan L., Collective mechanical adaptation of honeybee swarms. Nat. Phys. 14, 1193 (2018). [Google Scholar]
  • 24.Romanishin J. W., Gilpin K., Claici S., Rus D., “3D M-blocks: Self-reconfiguring robots capable of locomotion via pivoting in three dimensions” in 2015 IEEE International Conference on Robotics and Automation (ICRA) (IEEE, 2015), pp. 1925–1932. [Google Scholar]
  • 25.Augugliaro F., et al. , The flight assembled architecture installation: Cooperative construction with flying machines. IEEE Contr. Syst. Mag. 34, 46–64 (2014). [Google Scholar]
  • 26.Shimizu M., Ishiguro A., Kawakatsu T., “Slimebot: A modular robot that exploits emergent phenomena” in Proceedings of the 2005 IEEE International Conference on Robotics and Automation (IEEE, 2005), pp. 2982–2987. [Google Scholar]
  • 27.Leottau D. L., Ruiz-del Solar J., Babuška R., Decentralized reinforcement learning of robot behaviors. Artif. Intell. 256, 130–159 (2018). [Google Scholar]
  • 28.Christensen D. J., Schultz U. P., Stoy K., A distributed and morphology-independent strategy for adaptive locomotion in self-reconfigurable modular robots. Robot. Autonom. Syst. 61, 1021–1035 (2013). [Google Scholar]
  • 29.Metropolis N., Rosenbluth A. W., Rosenbluth M. N., Teller A. H., Teller E., Equation of state calculations by fast computing machines. J. Chem. Phys. 21, 1087–1092 (1953). [Google Scholar]
  • 30.Seok S., Onal C. D., Cho K.-J., Wood R. J., Rus D., Kim S., Meshworm: A peristaltic soft robot with antagonistic nickel titanium coil actuators. IEEE ASME Trans. Mechatron. 18, 1485–1497 (2012). [Google Scholar]
  • 31.Boxerbaum A. S., Shaw K. M., Chiel H. J., Quinn R. D., Continuous wave peristaltic motion in a robot. Int. J. Robot Res. 31, 302–318 (2012). [Google Scholar]
  • 32.Brandenbourger M., Locsin X., Lerner E., Coulais C., Non-reciprocal robotic metamaterials. Nat. Commun. 10, 1–8 (2019). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33.McEvoy M. A., Correll N., Materials that couple sensing, actuation, computation, and communication. Science 347, 1261689 (2015). [DOI] [PubMed] [Google Scholar]
  • 34.Kotikian A., et al. , Untethered soft robotic matter with passive control of shape morphing and propulsion. Sci. Robot. 4, 7044 (2019). [DOI] [PubMed] [Google Scholar]
  • 35.Miskin M. Z., et al. , Graphene-based bimorphs for micron-sized, autonomous origami machines. Proc. Natl. Acad. Sci. U.S.A. 115, 466–470 (2018). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 36.Ceylan H., Giltinan J., Kozielski K., Metin S., Mobile microrobots for bioengineering applications. Lab Chip 17, 1705–1724 (2017). [DOI] [PubMed] [Google Scholar]
  • 37.Medina-Sánchez M., Magdanz V., Guix M., Fomin V. M., Schmidt O. G., Swimming microrobots: Soft, reconfigurable, and smart. Adv. Funct. Mater. 28, 1707228 (2018). [Google Scholar]
  • 38.Rothemund P., et al. , A soft, bistable valve for autonomous control of soft actuators. Sci. Robot. 3, eaar7986 (2018). [DOI] [PubMed] [Google Scholar]
  • 39.Zhang H., Zeng H., Priimagi A., Ikkala O., Programmable responsive hydrogels inspired by classical conditioning algorithm. Nat. Commun. 10, 3267 (2019). [DOI] [PMC free article] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Supplementary File
pnas.2017015118.sapp.pdf (10.3MB, pdf)
Supplementary File
Download video file (52.9MB, mp4)
Supplementary File
Download video file (31.2MB, mp4)
Supplementary File
Download video file (59MB, mp4)
Supplementary File
Download video file (20.7MB, mp4)
Supplementary File
Download video file (35.3MB, mp4)

Data Availability Statement

The experimental and numerical data that support the findings of this study and all computer algorithms necessary to run the models have been deposited at Figshare (10.6084/m9.figshare.13626110).


Articles from Proceedings of the National Academy of Sciences of the United States of America are provided here courtesy of National Academy of Sciences

RESOURCES