Reaching Out to Robots

An electronic skin could help robots and prosthetic devices, such as this 3D-printed model, mimic the sense of touch Credit: American Chemical Society

Human skin contains sensitive nerve cells that detect pressure, temperature and other sensations that allow tactile interactions with the environment. To help robots and prosthetic devices attain these abilities, scientists are trying to develop electronic skins. Now researchers report a new method in ACS Applied Materials & Interfaces that creates an ultrathin, stretchable electronic skin, which could be used for a variety of human-machine interactions.

Electronic skin could be used for many applications, including prosthetic devices, wearable health monitors, robotics and virtual reality. A major challenge is transferring ultrathin electrical circuits onto complex 3D surfaces and then having the electronics be bendable and stretchable enough to allow movement. Some scientists have developed flexible “electronic tattoos” for this purpose, but their production is typically slow, expensive and requires clean-room fabrication methods such as photolithography. Mahmoud Tavakoli, Carmel Majidi and colleagues wanted to develop a fast, simple and inexpensive method for producing thin-film circuits with integrated microelectronics.

In the new approach, the researchers patterned a circuit template onto a sheet of transfer tattoo paper with an ordinary desktop laser printer. They then coated the template with silver paste, which adhered only to the printed toner ink. On top of the silver paste, the team deposited a gallium–indium liquid metal alloy that increased the electrical conductivity and flexibility of the circuit. Finally, they added external electronics, such as microchips, with a conductive “glue” made of vertically aligned magnetic particles embedded in a polyvinyl alcohol gel. The researchers transferred the electronic tattoo to various objects and demonstrated several applications of the new method, such as controlling a robot prosthetic arm, monitoring human skeletal muscle activity and incorporating proximity sensors into a 3D model of a hand.

This article has been republished from materials provided by ACS. Note: material may have been edited for length and content. For further information, please contact the cited source.

RELATED ARTICLES

Can artificial intelligence (AI) help us understand how the brain understands language? Can neuroscience help us understand why AI and neural networks are effective at predicting human perception? New research suggests both are possible.

y conjuring the spell “Lumos!” wizards in the mythical world of Harry Potter could light up the tip of their magic wands and illuminate their surroundings. So, too, does LumosVar, a computer program that “lights up” cancer-causing genetic Var-ients, or mutations, illuminating how physicians might best treat their patients.