Tag Archives: object
#435664 Swarm Robots Mimic Ant Jaws to Flip and ...
Small robots are appealing because they’re simple, cheap, and it’s easy to make a lot of them. Unfortunately, being simple and cheap means that each robot individually can’t do a whole lot. To make up for this, you can do what insects do—leverage that simplicity and low-cost to just make a huge swarm of simple robots, and together, they can cooperate to carry out relatively complex tasks.
Using insects as an example does set a bit of an unfair expectation for the poor robots, since insects are (let’s be honest) generally smarter and much more versatile than a robot on their scale could ever hope to be. Most robots with insect-like capabilities (like DASH and its family) are really too big and complex to be turned into swarms, because to make a vast amount of small robots, things like motors aren’t going to work because they’re too expensive.
The question, then, is to how to make a swarm of inexpensive small robots with insect-like mobility that don’t need motors to get around, and Jamie Paik’s Reconfigurable Robotics Lab at EPFL has an answer, inspired by trap-jaw ants.
Let’s talk about trap-jaw ants for just a second, because they’re insane. You can read this 2006 paper about them if you’re particularly interested in insane ants (and who isn’t!), but if you just want to hear the insane bit, it’s that trap-jaw ants can fire themselves into the air by biting the ground (!). In just 0.06 millisecond, their half-millimeter long mandibles can close at a top speed of 64 meters per second, which works out to an acceleration of about 100,000 g’s. Biting the ground causes the ant’s head to snap back with a force of 300 times the body weight of the ant itself, which launches the ant upwards. The ants can fly 8 centimeters vertically, and up to 15 cm horizontally—this is a lot, for an ant that’s just a few millimeters long.
Trap-jaw ants can fire themselves into the air by biting the ground, causing the ant’s head to snap back with a force of 300 times the body weight of the ant itself
EPFL’s robots, called Tribots, look nothing at all like trap-jaw ants, which personally I am fine with. They’re about 5 cm tall, weighing 10 grams each, and can be built on a flat sheet, and then folded into a tripod shape, origami-style. Or maybe it’s kirigami, because there’s some cutting involved. The Tribots are fully autonomous, meaning they have onboard power and control, including proximity sensors that allow them to detect objects and avoid them.
Photo: Marc Delachaux/EPFL
EPFL researchers Zhenishbek Zhakypov and Jamie Paik.
Avoiding objects is where the trap-jaw ants come in. Using two different shape-memory actuators (a spring and a latch, similar to how the ant’s jaw works), the Tribots can move around using a bunch of different techniques that can adapt to the terrain that they’re on, including:
Vertical jumping for height
Horizontal jumping for distance
Somersault jumping to clear obstacles
Walking on textured terrain with short hops (called “flic-flac” walking)
Crawling on flat surfaces
Here’s the robot in action:
Tribot’s maximum vertical jump is 14 cm (2.5 times its height), and horizontally it can jump about 23 cm (almost 4 times its length). Tribot is actually quite efficient in these movements, with a cost of transport much lower than similarly-sized robots, on par with insects themselves.
Working together, small groups of Tribots can complete tasks that a single robot couldn’t do alone. One example is pushing a heavy object a set distance. It turns out that you need five Tribots for this task—a leader robot, two worker robots, a monitor robot to measure the distance that the object has been pushed, and then a messenger robot to relay communications around the obstacle.
Image: EPFL
Five Tribots collaborate to move an object to a desired position, using coordination between a leader, two workers, a monitor, and a messenger robot. The leader orders the two worker robots to push the object while the monitor measures the relative position of the object. As the object blocks the two-way link between the leader and the monitor, the messenger maintains the communication link.
The researchers acknowledge that the current version of the hardware is limited in pretty much every way (mobility, sensing, and computation), but it does a reasonable job of demonstrating what’s possible with the concept. The plan going forward is to automate fabrication in order to “enable on-demand, ’push-button-manufactured’” robots.
“Designing minimal and scalable insect-inspired multi-locomotion millirobots,” by Zhenishbek Zhakypov, Kazuaki Mori, Koh Hosoda, and Jamie Paik from EPFL and Osaka University, is published in the current issue of Nature.
[ RRL ] via [ EPFL ] Continue reading
#435619 Video Friday: Watch This Robot Dog ...
Video Friday is your weekly selection of awesome robotics videos, collected by your Automaton bloggers. We’ll also be posting a weekly calendar of upcoming robotics events for the next few months; here’s what we have so far (send us your events!):
IEEE Africon 2019 – September 25-27, 2019 – Accra, Ghana
RoboBusiness 2019 – October 1-3, 2019 – Santa Clara, CA, USA
ISRR 2019 – October 6-10, 2019 – Hanoi, Vietnam
Ro-Man 2019 – October 14-18, 2019 – New Delhi, India
Humanoids 2019 – October 15-17, 2019 – Toronto, Canada
ARSO 2019 – October 31-1, 2019 – Beijing, China
ROSCon 2019 – October 31-1, 2019 – Macau
IROS 2019 – November 4-8, 2019 – Macau
Let us know if you have suggestions for next week, and enjoy today’s videos.
Team PLUTO (University of Pennsylvania, Ghost Robotics, and Exyn Technologies) put together this video giving us a robot’s-eye-view (or whatever they happen to be using for eyes) of the DARPA Subterranean Challenge tunnel circuits.
[ PLUTO ]
Zhifeng Huang has been improving his jet-stepping humanoid robot, which features new hardware and the ability to take larger and more complex steps.
This video reported the last progress of an ongoing project utilizing ducted-fan propulsion system to improve humanoid robot’s ability in stepping over large ditches. The landing point of the robot’s swing foot can be not only forward but also side direction. With keeping quasi-static balance, the robot was able to step over a ditch with 450mm in width (up to 97% of the robot’s leg’s length) in 3D stepping.
[ Paper ]
Thanks Zhifeng!
These underacuated hands from Matei Ciocarlie’s lab at Columbia are magically able to reconfigure themselves to grasp different object types with just one or two motors.
[ Paper ] via [ ROAM Lab ]
This is one reason we should pursue not “autonomous cars” but “fully autonomous cars” that never require humans to take over. We can’t be trusted.
During our early days as the Google self-driving car project, we invited some employees to test our vehicles on their commutes and weekend trips. What we were testing at the time was similar to the highway driver assist features that are now available on cars today, where the car takes over the boring parts of the driving, but if something outside its ability occurs, the driver has to take over immediately.
What we saw was that our testers put too much trust in that technology. They were doing things like texting, applying makeup, and even falling asleep that made it clear they would not be ready to take over driving if the vehicle asked them to. This is why we believe that nothing short of full autonomy will do.
[ Waymo ]
Buddy is a DIY and fetchingly minimalist social robot (of sorts) that will be coming to Kickstarter this month.
We have created a new arduino kit. His name is Buddy. He is a DIY social robot to serve as a replacement for Jibo, Cozmo, or any of the other bots that are no longer available. Fully 3D printed and supported he adds much more to our series of Arduino STEM robotics kits.
Buddy is able to look around and map his surroundings and react to changes within them. He can be surprised and he will always have a unique reaction to changes. The kit can be built very easily in less than an hour. It is even robust enough to take the abuse that kids can give it in a classroom.
[ Littlebots ]
The android Mindar, based on the Buddhist deity of mercy, preaches sermons at Kodaiji temple in Kyoto, and its human colleagues predict that with artificial intelligence it could one day acquire unlimited wisdom. Developed at a cost of almost $1 million (¥106 million) in a joint project between the Zen temple and robotics professor Hiroshi Ishiguro, the robot teaches about compassion and the dangers of desire, anger and ego.
[ Japan Times ]
I’m not sure whether it’s the sound or what, but this thing scares me for some reason.
[ BIRL ]
This gripper uses magnets as a sort of adjustable spring for dynamic stiffness control, which seems pretty clever.
[ Buffalo ]
What a package of medicine sees while being flown by drone from a hospital to a remote clinic in the Dominican Republic. The drone flew 11 km horizontally and 800 meters vertically, and I can’t even imagine what it would take to make that drive.
[ WeRobotics ]
My first ride in a fully autonomous car was at Stanford in 2009. I vividly remember getting in the back seat of a descendant of Junior, and watching the steering wheel turn by itself as the car executed a perfect parking maneuver. Ten years later, it’s still fun to watch other people have that experience.
[ Waymo ]
Flirtey, the pioneer of the commercial drone delivery industry, has unveiled the much-anticipated first video of its next-generation delivery drone, the Flirtey Eagle. The aircraft designer and manufacturer also unveiled the Flirtey Portal, a sophisticated take off and landing platform that enables scalable store-to-door operations; and an autonomous software platform that enables drones to deliver safely to homes.
[ Flirtey ]
EPFL scientists are developing new approaches for improved control of robotic hands – in particular for amputees – that combines individual finger control and automation for improved grasping and manipulation. This interdisciplinary proof-of-concept between neuroengineering and robotics was successfully tested on three amputees and seven healthy subjects.
[ EPFL ]
This video is a few years old, but we’ll take any excuse to watch the majestic sage-grouse be majestic in all their majesticness.
[ UC Davis ]
I like the idea of a game of soccer (or, football to you weirdos in the rest of the world) where the ball has a mind of its own.
[ Sphero ]
Looks like the whole delivery glider idea is really taking off! Or, you know, not taking off.
Weird that they didn’t show the landing, because it sure looked like it was going to plow into the side of the hill at full speed.
[ Yates ] via [ sUAS News ]
This video is from a 2018 paper, but it’s not like we ever get tired of seeing quadrupeds do stuff, right?
[ MIT ]
Founder and Head of Product, Ian Bernstein, and Head of Engineering, Morgan Bell, have been involved in the Misty project for years and they have learned a thing or two about building robots. Hear how and why Misty evolved into a robot development platform, learn what some of the earliest prototypes did (and why they didn’t work for what we envision), and take a deep dive into the technology decisions that form the Misty II platform.
[ Misty Robotics ]
Lex Fridman interviews Vijay Kumar on the Artifiical Intelligence Podcast.
[ AI Podcast ]
This week’s CMU RI Seminar is from Ross Knepper at Cornell, on Formalizing Teamwork in Human-Robot Interaction.
Robots out in the world today work for people but not with people. Before robots can work closely with ordinary people as part of a human-robot team in a home or office setting, robots need the ability to acquire a new mix of functional and social skills. Working with people requires a shared understanding of the task, capabilities, intentions, and background knowledge. For robots to act jointly as part of a team with people, they must engage in collaborative planning, which involves forming a consensus through an exchange of information about goals, capabilities, and partial plans. Often, much of this information is conveyed through implicit communication. In this talk, I formalize components of teamwork involving collaboration, communication, and representation. I illustrate how these concepts interact in the application of social navigation, which I argue is a first-class example of teamwork. In this setting, participants must avoid collision by legibly conveying intended passing sides via nonverbal cues like path shape. A topological representation using the braid groups enables the robot to reason about a small enumerable set of passing outcomes. I show how implicit communication of topological group plans achieves rapid covergence to a group consensus, and how a robot in the group can deliberately influence the ultimate outcome to maximize joint performance, yielding pedestrian comfort with the robot.
[ CMU RI ]
In this week’s episode of Robots in Depth, Per speaks with Julien Bourgeois about Claytronics, a project from Carnegie Mellon and Intel to develop “programmable matter.”
Julien started out as a computer scientist. He was always interested in robotics privately but then had the opportunity to get into micro robots when his lab was merged into the FEMTO-ST Institute. He later worked with Seth Copen Goldstein at Carnegie Mellon on the Claytronics project.
Julien shows an enlarged mock-up of the small robots that make up programmable matter, catoms, and speaks about how they are designed. Currently he is working on a unit that is one centimeter in diameter and he shows us the very small CPU that goes into that model.
[ Robots in Depth ] Continue reading