These self-sufficient robots can have ‘babies’ and colonize distant planets
Robots rising
Robots have come a long way since our first clumsy forays into artificial movement many decades ago. Today, companies such as Boston Dynamics produceultra-efficient robotsthat load trucks, build pallets, and move boxes around factories, undertaking tasks you might think only humans could perform.
Despite these advances, designing robots to work in unknown or inhospitable environments – like exoplanets or deep ocean trenches – still poses a considerable challenge for scientists and engineers. Out in the cosmos, what shape and size should the ideal robot be? Should it crawl or walk? What tools will it need to manipulate its environment – and how will it survive extremes of pressure, temperature, and chemical corrosion?
[Read:How much does it cost to buy, own, and run an EV? It’s not as much as you think]
An impossible brainteaser for humans, nature has already solved this problem. Darwinian evolution has resulted in millions of species that are perfectly adapted to their environment. Although biological evolution takes millions of years,artificial evolution– modeling evolutionary processes inside a computer – can take place in hours, or even minutes. Computer scientists have been harnessing its power for decades, resulting ingas nozzles to satellite antennasthat are ideally suited to their function, for instance.
But current artificial evolution of moving, physical objects still requires a great deal of human oversight, requiring a tight feedback loop between robot and human. If artificial evolution is to design a useful robot for exoplanetary exploration, we’ll need to remove the human from the loop. In essence, evolved robot designs must manufacture, assemble and test themselves autonomously – untethered from human oversight.
Unnatural selection
Any evolved robots will need to be capable of sensing their environment and have diverse means of moving – for example using wheels, jointed legs, or even mixtures of the two. And to address the inevitable reality gap that occurs when transferring a design from software to hardware, it is also desirable for at least some evolution to take place in hardware – within an ecosystem of robots that evolve in real-time and real space.
The Autonomous Robot Evolution (ARE) project addresses exactly this, bringing together scientists and engineers from four universities in an ambitious four-year project to develop this radical new technology.
As depicted above, robots will be “born” through the use of 3D manufacturing. We use a new kind of hybrid hardware-software evolutionary architecture for design. That means that every physical robot has a digital clone. Physical robots are performance-tested in real-world environments, while their digital clones enter a software program, where they undergo rapid simulated evolution. This hybrid system introduces a novel type of evolution: new generations can be produced from a union of the most successful traits from a virtual “mother” and a physical “father.”
As well as being rendered in our simulator, “child” robots produced via our hybrid evolution are also 3D-printed and introduced into a real-world, creche-like environment. The most successful individuals within this physical training center make their “genetic code” available for reproduction and for the improvement of future generations, while less “fit” robots can simply be hoisted away and recycled into new ones as part of an ongoing evolutionary cycle.
Two years into the project, significant advances have been made. From a scientific perspective, we have designed newartificial evolutionary algorithmsthat have produced a diverse set of robots that drive or crawl and can learn to navigate through complex mazes. These algorithms evolve both thebody-plan, and brain of the robot.
Thebraincontains a controller that determines how the robot moves, interpreting sensory information from the environment and translating this into motor controls. Once the robot is built, alearning algorithmquickly refines the child brain to account for any potential mismatch between its new body and its inherited brain.
From an engineering perspective, we have designed the “RoboFab” to fullyautomate manufacturing. This robotic arm attaches wires, sensors, and other “organs” chosen by evolution to the robot’s 3D-printed chassis. We designed these components to facilitate swift assembly, giving the RoboFab access to a bigtoolboxof robot limbs and organs.
Waste disposal
The first major use case we plan to address is deploying this technology to design robots to undertake clean-up of legacy waste in a nuclear reactor – like that seen in the TV miniseries Chernobyl. Using humans for this task is both dangerous and expensive, and necessaryrobotic solutionsremain to be developed.
Looking forward, the long-term vision is to develop the technology sufficiently to enable the evolution of entireautonomous robotic ecosystemsthat live and work for long periods in challenging and dynamic environments without the need for direct human oversight.
In this radical new paradigm, robots are conceived and born, rather than designed and manufactured. Such robots will fundamentally change the concept of machines, showcasing a new breed that can change their form and behavior over time – just like us.
This article byEmma Hart, Chair in Natural Computation,Edinburgh Napier Universityis republished fromThe Conversationunder a Creative Commons license. Read theoriginal article.
Story byThe Conversation
An independent news and commentary website produced by academics and journalists.An independent news and commentary website produced by academics and journalists.
Get the TNW newsletter
Get the most important tech news in your inbox each week.