Tag Archives: skin

#439110 Robotic Exoskeletons Could One Day Walk ...

Engineers, using artificial intelligence and wearable cameras, now aim to help robotic exoskeletons walk by themselves.

Increasingly, researchers around the world are developing lower-body exoskeletons to help people walk. These are essentially walking robots users can strap to their legs to help them move.

One problem with such exoskeletons: They often depend on manual controls to switch from one mode of locomotion to another, such as from sitting to standing, or standing to walking, or walking on the ground to walking up or down stairs. Relying on joysticks or smartphone apps every time you want to switch the way you want to move can prove awkward and mentally taxing, says Brokoslaw Laschowski, a robotics researcher at the University of Waterloo in Canada.

Scientists are working on automated ways to help exoskeletons recognize when to switch locomotion modes — for instance, using sensors attached to legs that can detect bioelectric signals sent from your brain to your muscles telling them to move. However, this approach comes with a number of challenges, such as how how skin conductivity can change as a person’s skin gets sweatier or dries off.

Now several research groups are experimenting with a new approach: fitting exoskeleton users with wearable cameras to provide the machines with vision data that will let them operate autonomously. Artificial intelligence (AI) software can analyze this data to recognize stairs, doors, and other features of the surrounding environment and calculate how best to respond.

Laschowski leads the ExoNet project, the first open-source database of high-resolution wearable camera images of human locomotion scenarios. It holds more than 5.6 million images of indoor and outdoor real-world walking environments. The team used this data to train deep-learning algorithms; their convolutional neural networks can already automatically recognize different walking environments with 73 percent accuracy “despite the large variance in different surfaces and objects sensed by the wearable camera,” Laschowski notes.

According to Laschowski, a potential limitation of their work their reliance on conventional 2-D images, whereas depth cameras could also capture potentially useful distance data. He and his collaborators ultimately chose not to rely on depth cameras for a number of reasons, including the fact that the accuracy of depth measurements typically degrades in outdoor lighting and with increasing distance, he says.

In similar work, researchers in North Carolina had volunteers with cameras either mounted on their eyeglasses or strapped onto their knees walk through a variety of indoor and outdoor settings to capture the kind of image data exoskeletons might use to see the world around them. The aim? “To automate motion,” says Edgar Lobaton an electrical engineering researcher at North Carolina State University. He says they are focusing on how AI software might reduce uncertainty due to factors such as motion blur or overexposed images “to ensure safe operation. We want to ensure that we can really rely on the vision and AI portion before integrating it into the hardware.”

In the future, Laschowski and his colleagues will focus on improving the accuracy of their environmental analysis software with low computational and memory storage requirements, which are important for onboard, real-time operations on robotic exoskeletons. Lobaton and his team also seek to account for uncertainty introduced into their visual systems by movements .

Ultimately, the ExoNet researchers want to explore how AI software can transmit commands to exoskeletons so they can perform tasks such as climbing stairs or avoiding obstacles based on a system’s analysis of a user's current movements and the upcoming terrain. With autonomous cars as inspiration, they are seeking to develop autonomous exoskeletons that can handle the walking task without human input, Laschowski says.

However, Laschowski adds, “User safety is of the utmost importance, especially considering that we're working with individuals with mobility impairments,” resulting perhaps from advanced age or physical disabilities.
“The exoskeleton user will always have the ability to override the system should the classification algorithm or controller make a wrong decision.” Continue reading

Posted in Human Robots

#439077 How Scientists Grew Human Muscles in Pig ...

The little pigs bouncing around the lab looked exceedingly normal. Yet their adorable exterior hid a remarkable secret: each piglet carried two different sets of genes. For now, both sets came from their own species. But one day, one of those sets may be human.

The piglets are chimeras—creatures with intermingled sets of genes, as if multiple entities were seamlessly mashed together. Named after the Greek lion-goat-serpent monsters, chimeras may hold the key to an endless supply of human organs and tissues for transplant. The crux is growing these human parts in another animal—one close enough in size and function to our own.

Last week, a team from the University of Minnesota unveiled two mind-bending chimeras. One was joyous little piglets, each propelled by muscles grown from a different pig. Another was pig embryos, transplanted into surrogate pigs, that developed human muscles for more than 20 days.

The study, led by Drs. Mary and Daniel Garry at the University of Minnesota, had a therapeutic point: engineering a brilliant way to replace muscle loss, especially for the muscles around our skeletons that allow us to move and navigate the world. Trauma and injury, such as from firearm wounds or car crashes, can damage muscle tissue beyond the point of repair. Unfortunately, muscles are also stubborn in that donor tissue from cadavers doesn’t usually “take” at the injury site. For now, there are no effective treatments for severe muscle death, called volumetric muscle loss.

The new human-pig hybrids are designed to tackle this problem. Muscle wasting aside, the study also points to a clever “hack” that increases the amount of human tissue inside a growing pig embryo.

If further improved, the technology could “provide an unlimited supply of organs for transplantation,” said Dr. Mary Garry to Inverse. What’s more, because the human tissue can be sourced from patients themselves, the risk of rejection by the immune system is relatively low—even when grown inside a pig.

“The shortage of organs for heart transplantation, vascular grafting, and skeletal muscle is staggering,” said Garry. Human-animal chimeras could have a “seismic impact” that transforms organ transplantation and helps solve the organ shortage crisis.

That is, if society accepts the idea of a semi-humanoid pig.

Wait…But How?
The new study took a page from previous chimera recipes.

The main ingredients and steps go like this: first, you need an embryo that lacks the ability to develop a tissue or organ. This leaves an “empty slot” of sorts that you can fill with another set of genes—pig, human, or even monkey.

Second, you need to fine-tune the recipe so that the embryos “take” the new genes, incorporating them into their bodies as if they were their own. Third, the new genes activate to instruct the growing embryo to make the necessary tissue or organs without harming the overall animal. Finally, the foreign genes need to stay put, without cells migrating to another body part—say, the brain.

Not exactly straightforward, eh? The piglets are technological wonders that mix cutting-edge gene editing with cloning technologies.

The team went for two chimeras: one with two sets of pig genes, the other with a pig and human mix. Both started with a pig embryo that can’t make its own skeletal muscles (those are the muscles surrounding your bones). Using CRISPR, the gene-editing Swiss Army Knife, they snipped out three genes that are absolutely necessary for those muscles to develop. Like hitting a bullseye with three arrows simultaneously, it’s already a technological feat.

Here’s the really clever part: the muscles around your bones have a slightly different genetic makeup than the ones that line your blood vessels or the ones that pump your heart. While the resulting pig embryos had severe muscle deformities as they developed, their hearts beat as normal. This means the gene editing cut only impacted skeletal muscles.

Then came step two: replacing the missing genes. Using a microneedle, the team injected a fertilized and slightly developed pig egg—called a blastomere—into the embryo. If left on its natural course, a blastomere eventually develops into another embryo. This step “smashes” the two sets of genes together, with the newcomer filling the muscle void. The hybrid embryo was then placed into a surrogate, and roughly four months later, chimeric piglets were born.

Equipped with foreign DNA, the little guys nevertheless seemed totally normal, nosing around the lab and running everywhere without obvious clumsy stumbles. Under the microscope, their “xenomorph” muscles were indistinguishable from run-of-the-mill average muscle tissue—no signs of damage or inflammation, and as stretchy and tough as muscles usually are. What’s more, the foreign DNA seemed to have only developed into muscles, even though they were prevalent across the body. Extensive fishing experiments found no trace of the injected set of genes inside blood vessels or the brain.

A Better Human-Pig Hybrid
Confident in their recipe, the team next repeated the experiment with human cells, with a twist. Instead of using controversial human embryonic stem cells, which are obtained from aborted fetuses, they relied on induced pluripotent stem cells (iPSCs). These are skin cells that have been reverted back into a stem cell state.

Unlike previous attempts at making human chimeras, the team then scoured the genetic landscape of how pig and human embryos develop to find any genetic “brakes” that could derail the process. One gene, TP53, stood out, which was then promptly eliminated with CRISPR.

This approach provides a way for future studies to similarly increase the efficiency of interspecies chimeras, the team said.

The human-pig embryos were then carefully grown inside surrogate pigs for less than a month, and extensively analyzed. By day 20, the hybrids had already grown detectable human skeletal muscle. Similar to the pig-pig chimeras, the team didn’t detect any signs that the human genes had sprouted cells that would eventually become neurons or other non-muscle cells.

For now, human-animal chimeras are not allowed to grow to term, in part to stem the theoretical possibility of engineering humanoid hybrid animals (shudder). However, a sentient human-pig chimera is something that the team specifically addressed. Through multiple experiments, they found no trace of human genes in the embryos’ brain stem cells 20 and 27 days into development. Similarly, human donor genes were absent in cells that would become the hybrid embryos’ reproductive cells.

Despite bioethical quandaries and legal restrictions, human-animal chimeras have taken off, both as a source of insight into human brain development and a well of personalized organs and tissues for transplant. In 2019, Japan lifted its ban on developing human brain cells inside animal embryos, as well as the term limit—to global controversy. There’s also the question of animal welfare, given that hybrid clones will essentially become involuntary organ donors.

As the debates rage on, scientists are nevertheless pushing the limits of human-animal chimeras, while treading as carefully as possible.

“Our data…support the feasibility of the generation of these interspecies chimeras, which will serve as a model for translational research or, one day, as a source for xenotransplantation,” the team said.

Image Credit: Christopher Carson on Unsplash Continue reading

Posted in Human Robots

#439062 Xenobots 2.0: These Living Robots ...

The line between animals and machines was already getting blurry after a team of scientists and roboticists unveiled the first living robots last year. Now the same team has released version 2.0 of their so-called xenobots, and they’re faster, stronger, and more capable than ever.

In January 2020, researchers from Tufts University and the University of Vermont laid out a method for building tiny biological machines out of the eggs of the African claw frog Xenopus laevis. Dubbed xenobots after their animal forebear, they could move independently, push objects, and even team up to create swarms.

Remarkably, building them involved no genetic engineering. Instead, the team used an evolutionary algorithm running on a supercomputer to test out thousands of potential designs made up of different configurations of cells.

Once they’d found some promising candidates that could solve the tasks they were interested in, they used microsurgical tools to build real-world versions out of living cells. The most promising design was built by splicing heart muscle cells (which could contract to propel the xenobots), and skin cells (which provided a rigid support).

Impressive as that might sound, having to build each individual xenobot by hand is obviously tedious. But now the team has devised a new approach that works from the bottom up by getting the xenobots to self-assemble their bodies from single cells. Not only is the approach more scalable, the new xenobots are faster, live longer, and even have a rudimentary memory.

In a paper in Science Robotics, the researchers describe how they took stem cells from frog embryos and allowed them to grow into clumps of several thousand cells called spheroids. After a few days, the stem cells had turned into skin cells covered in small hair-like projections called cilia, which wriggle back and forth.

Normally, these structures are used to spread mucus around on the frog’s skin. But when divorced from their normal context they took on a function more similar to that seen in microorganisms, which use cilia to move about by acting like tiny paddles.

“We are witnessing the remarkable plasticity of cellular collectives, which build a rudimentary new ‘body’ that is quite distinct from their default—in this case, a frog—despite having a completely normal genome,” corresponding author Michael Levin from Tufts University said in a press release.

“We see that cells can re-purpose their genetically encoded hardware, like cilia, for new functions such as locomotion. It is amazing that cells can spontaneously take on new roles and create new body plans and behaviors without long periods of evolutionary selection for those features,” he said.

Not only were the new xenobots faster and longer-lived, they were also much better at tasks like working together as a swarm to gather piles of iron oxide particles. And while the form and function of the xenobots was achieved without any genetic engineering, in an extra experiment the team injected them with RNA that caused them to produce a fluorescent protein that changes color when exposed to a particular color of light.

This allowed the xenobots to record whether they had come into contact with a specific light source while traveling about. The researchers say this is a proof of principle that the xenobots can be imbued with a molecular memory, and future work could allow them to record multiple stimuli and potentially even react to them.

What exactly these xenobots could eventually be used for is still speculative, but they have features that make them a promising alternative to non-organic alternatives. For a start, robots made of stem cells are completely biodegradable and also have their own power source in the form of “yolk platelets” found in all amphibian embryos. They are also able to self-heal in as little as five minutes if cut, and can take advantage of cells’ ability to process all kinds of chemicals.

That suggests they could have applications in everything from therapeutics to environmental engineering. But the researchers also hope to use them to better understand the processes that allow individual cells to combine and work together to create a larger organism, and how these processes might be harnessed and guided for regenerative medicine.

As these animal-machine hybrids advance, they are sure to raise ethical concerns and question marks over the potential risks. But it looks like the future of robotics could be a lot more wet and squishy than we imagined.

Image Credit: Doug Blackiston/Tufts University Continue reading

Posted in Human Robots

#438807 Visible Touch: How Cameras Can Help ...

The dawn of the robot revolution is already here, and it is not the dystopian nightmare we imagined. Instead, it comes in the form of social robots: Autonomous robots in homes and schools, offices and public spaces, able to interact with humans and other robots in a socially acceptable, human-perceptible way to resolve tasks related to core human needs.

To design social robots that “understand” humans, robotics scientists are delving into the psychology of human communication. Researchers from Cornell University posit that embedding the sense of touch in social robots could teach them to detect physical interactions and gestures. They describe a way of doing so by relying not on touch but on vision.

A USB camera inside the robot captures shadows of hand gestures on the robot’s surface and classifies them with machine-learning software. They call this method ShadowSense, which they define as a modality between vision and touch, bringing “the high resolution and low cost of vision-sensing to the close-up sensory experience of touch.”

Touch-sensing in social or interactive robots is usually achieved with force sensors or capacitive sensors, says study co-author Guy Hoffman of the Sibley School of Mechanical and Aerospace Engineering at Cornell University. The drawback to his group’s approach has been that, even to achieve coarse spatial resolution, many sensors are needed in a small area.

However, working with non-rigid, inflatable robots, Hoffman and his co-researchers installed a consumer-grade USB camera to which they attached a fisheye lens for a wider field of vision.

“Given that the robot is already hollow, and has a soft and translucent skin, we could do touch interaction by looking at the shadows created by people touching the robot,” says Hoffman. They used deep neural networks to interpret the shadows. “And we were able to do it with very high accuracy,” he says. The robot was able to interpret six different gestures, including one- or two-handed touch, pointing, hugging and punching, with an accuracy of 87.5 to 96 percent, depending on the lighting.

This is not the first time that computer vision has been used for tactile sensing, though the scale and application of ShadowSense is unique. “Photography has been used for touch mainly in robotic grasping,” says Hoffman. By contrast, Hoffman and collaborators wanted to develop a sense that could be “felt” across the whole of the device.

The potential applications for ShadowSense include mobile robot guidance using touch, and interactive screens on soft robots. A third concerns privacy, especially in home-based social robots. “We have another paper currently under review that looks specifically at the ability to detect gestures that are further away [from the robot’s skin],” says Hoffman. This way, users would be able to cover their robot’s camera with a translucent material and still allow it to interpret actions and gestures from shadows. Thus, even though it’s prevented from capturing a high-resolution image of the user or their surrounding environment, using the right kind of training datasets, the robot can continue to monitor some kinds of non-tactile activities.

In its current iteration, Hoffman says, ShadowSense doesn’t do well in low-light conditions. Environmental noise, or shadows from surrounding objects, also interfere with image classification. Relying on one camera also means a single point of failure. “I think if this were to become a commercial product, we would probably [have to] work a little bit better on image detection,” says Hoffman.

As it was, the researchers used transfer learning—reusing a pre-trained deep-learning model in a new problem—for image analysis. “One of the problems with multi-layered neural networks is that you need a lot of training data to make accurate predictions,” says Hoffman. “Obviously, we don’t have millions of examples of people touching a hollow, inflatable robot. But we can use pre-trained networks trained on general images, which we have billions of, and we only retrain the last layers of the network using our own dataset.” Continue reading

Posted in Human Robots

#438790 Novel soft tactile sensor with ...

A joint research team co-led by City University of Hong Kong (CityU) has developed a new soft tactile sensor with skin-comparable characteristics. A robotic gripper with the sensor mounted at the fingertip could accomplish challenging tasks such as stably grasping fragile objects and threading a needle. Their research provided new insight into tactile sensor design and could contribute to various applications in the robotics field, such as smart prosthetics and human-robot interaction. Continue reading

Posted in Human Robots