Tag Archives: nasa

#437224 This Week’s Awesome Tech Stories From ...

VIRTUAL REALITY
How Holographic Tech Is Shrinking VR Displays to the Size of Sunglasses
Kyle Orland | Ars Technica
“…researchers at Facebook Reality Labs are using holographic film to create a prototype VR display that looks less like ski goggles and more like lightweight sunglasses. With a total thickness less than 9mm—and without significant compromises on field of view or resolution—these displays could one day make today’s bulky VR headset designs completely obsolete.”

TRANSPORTATION
Stock Surge Makes Tesla the World’s Most Valuable Automaker
Timothy B. Lee | Ars Technica
“It’s a remarkable milestone for a company that sells far fewer cars than its leading rivals. …But Wall Street is apparently very optimistic about Tesla’s prospects for future growth and profits. Many experts expect a global shift to battery electric vehicles over the next decade or two, and Tesla is leading that revolution.”

FUTURE OF FOOD
These Plant-Based Steaks Come Out of a 3D Printer
Adele Peters | Fast Company
“The startup, launched by cofounders who met while developing digital printers at HP, created custom 3D printers that aim to replicate meat by printing layers of what they call ‘alt-muscle,’ ‘alt-fat,’ and ‘alt-blood,’ forming a complex 3D model.”

AUTOMATION
The US Air Force Is Turning Old F-16s Into AI-Powered Fighters
Amit Katwala | Wired UK
“Maverick’s days are numbered. The long-awaited sequel to Top Gun is due to hit cinemas in December, but the virtuoso fighter pilots at its heart could soon be a thing of the past. The trustworthy wingman will soon be replaced by artificial intelligence, built into a drone, or an existing fighter jet with no one in the cockpit.”

ROBOTICS
NASA Wants to Build a Steam-Powered Hopping Robot to Explore Icy Worlds
Georgina Torbet | Digital Trends
“A bouncing, ball-like robot that’s powered by steam sounds like something out of a steampunk fantasy, but it could be the ideal way to explore some of the distant, icy environments of our solar system. …This round robot would be the size of a soccer ball, with instruments held in the center of a metal cage, and it would use steam-powered thrusters to make jumps from one area of terrain to the next.”

FUTURE
Could Teleporting Ever Work?
Daniel Kolitz | Gizmodo
“Have the major airlines spent decades suppressing teleportation research? Have a number of renowned scientists in the field of teleportation studies disappeared under mysterious circumstances? Is there a cork board at the FBI linking Delta Airlines, shady foreign security firms, and dozens of murdered research professors? …No. None of that is the case. Which begs the question: why doesn’t teleportation exist yet?”

ENERGY
Nuclear ‘Power Balls’ Could Make Meltdowns a Thing of the Past
Daniel Oberhaus | Wired
“Not only will these reactors be smaller and more efficient than current nuclear power plants, but their designers claim they’ll be virtually meltdown-proof. Their secret? Millions of submillimeter-size grains of uranium individually wrapped in protective shells. It’s called triso fuel, and it’s like a radioactive gobstopper.”

TECHNOLOGY
A Plan to Redesign the Internet Could Make Apps That No One Controls
Will Douglas Heaven | MIT Techology Review
“[John Perry] Barlow’s ‘home of Mind’ is ruled today by the likes of Google, Facebook, Amazon, Alibaba, Tencent, and Baidu—a small handful of the biggest companies on earth. Yet listening to the mix of computer scientists and tech investors speak at an online event on June 30 hosted by the Dfinity Foundation…it is clear that a desire for revolution is brewing.”

IMPACT
To Save the World, the UN Is Turning It Into a Computer Simulation
Will Bedingfield | Wired
“The UN has now announced its new secret recipe to achieve [its 17 sustainable development goals or SDGs]: a computer simulation called Policy Priority Inference (PPI). …PPI is a budgeting software—it simulates a government and its bureaucrats as they allocate money on projects that might move a country closer to an SDG.”

Image credit: Benjamin Suter / Unsplash Continue reading

Posted in Human Robots

#436507 The Weird, the Wacky, the Just Plain ...

As you know if you’ve ever been to, heard of, or read about the annual Consumer Electronics Show in Vegas, there’s no shortage of tech in any form: gadgets, gizmos, and concepts abound. You probably couldn’t see them all in a month even if you spent all day every day trying.

Given the sheer scale of the show, the number of exhibitors, and the inherent subjectivity of bestowing superlatives, it’s hard to pick out the coolest tech from CES. But I’m going to do it anyway; in no particular order, here are some of the products and concepts that I personally found most intriguing at this year’s event.

e-Novia’s Haptic Gloves
Italian startup e-Novia’s Weart glove uses a ‘sensing core’ to record tactile sensations and an ‘actuation core’ to reproduce those sensations onto the wearer’s skin. Haptic gloves will bring touch to VR and AR experiences, making them that much more life-like. The tech could also be applied to digitization of materials and in gaming and entertainment.

e-Novia’s modular haptic glove
I expected a full glove, but in fact there were two rings that attached to my fingers. Weart co-founder Giovanni Spagnoletti explained that they’re taking a modular approach, so as to better tailor the technology to different experiences. He then walked me through a virtual reality experience that was a sort of simulated science experiment: I had to lift a glass beaker, place it on a stove, pour in an ingredient, open a safe to access some dry ice, add that, and so on. As I went through the steps, I felt the beaker heat up and cool off at the expected times, and felt the liquid moving inside, as well as the pressure of my fingertips against the numbered buttons on the safe.

A virtual (but tactile) science experiment
There was a slight delay between my taking an action and feeling the corresponding tactile sensation, but on the whole, the haptic glove definitely made the experience more realistic—and more fun. Slightly less fun but definitely more significant, Spagnoletti told me Weart is working with a medical group to bring tactile sensations to VR training for surgeons.

Sarcos Robotics’ Exoskeleton
That tire may as well be a feather
Sarcos Robotics unveiled its Guardian XO full-body exoskeleton, which it says can safely lift up to 200 pounds across an extended work session. What’s cool about this particular exoskeleton is that it’s not just a prototype; the company announced a partnership with Delta airlines, which will be trialing the technology for aircraft maintenance, engine repair, and luggage handling. In a demo, I watched a petite female volunteer strap into the exoskeleton and easily lift a 50-pound weight with one hand, and a Sarcos employee lift and attach a heavy component of a propeller; she explained that the strength-augmenting function of the exoskeleton can easily be switched on or off—and the wearer’s hands released—to facilitate multi-step tasks.

Hyundai’s Flying Taxi
Where to?
Hyundai and Uber partnered to unveil an air taxi concept. With a 49-foot wingspan, 4 lift rotors, and 4 tilt rotors, the aircraft would be manned by a pilot and could carry 4 passengers at speeds up to 180 miles per hour. The companies say you’ll be able to ride across your city in one of these by 2030—we’ll see if the regulatory environment, public opinion, and other factors outside of technological capability let that happen.

Mercedes’ Avatar Concept Car
Welcome to the future
As evident from its name, Mercedes’ sweet new Vision AVTR concept car was inspired by the movie Avatar; director James Cameron helped design it. The all-electric car has no steering wheel, transparent doors, seats made of vegan leather, and 33 reptilian-scale-like flaps on the back; its design is meant to connect the driver with both the car and the surrounding environment in a natural, seamless way.

Next-generation scrolling
Offered the chance to ‘drive’ the car, I jumped on it. Placing my hand on the center console started the engine, and within seconds it had synced to my heartbeat, which reverberated through the car. The whole dashboard, from driver door to passenger door, is one big LED display. It showed a virtual landscape I could select by holding up my hand: as I moved my hand from left to right, different images were projected onto my open palm. Closing my hand on an image selected it, and suddenly it looked like I was in the middle of a lush green mountain range. Applying slight forward pressure on the center console made the car advance in the virtual landscape; it was essentially like playing a really cool video game.

Mercedes is aiming to have a carbon-neutral production fleet by 2039, and to reduce the amount of energy it uses during production by 40 percent by 2030. It’s unclear when—or whether—the man-machine-nature connecting features of the Vision AVTR will start showing up in production, but I for one will be on the lookout.

Waverly Labs’ In-Ear Translator
Waverly Labs unveiled its Ambassador translator earlier this year and has it on display at the show. It’s worn on the ear and uses a far-field microphone array with speech recognition to translate real-time conversations in 20 different languages. Besides in-ear audio, translations can also appear as text on an app or be broadcast live in a conference environment.

It’s kind of like a giant talking earring
I stopped by the booth and tested out the translator with Waverly senior software engineer Georgiy Konovalov. We each hooked on an earpiece, and first, he spoke to me in Russian. After a delay of a couple seconds, I heard his words in—slightly robotic, but fully comprehensible—English. Then we switched: I spoke to him in Spanish, my words popped up on his phone screen in Cyrillic, and he translated them back to English for me out loud.

On the whole, the demo was pretty cool. If you’ve ever been lost in a foreign country whose language you don’t speak, imagine how handy a gadget like this would come in. Let’s just hope that once they’re more widespread, these products don’t end up discouraging people from learning languages.

Not to be outdone, Google also announced updates to its Translate product, which is being deployed at information desks in JFK airport’s international terminal, in sports stadiums in Qatar, and by some large hotel chains.

Stratuscent’s Digital Nose
AI is making steady progress towards achieving human-like vision and hearing—but there’s been less work done on mimicking our sense of smell (maybe because it’s less useful in everyday applications). Stratuscent’s digital nose, which it says is based on NASA patents, uses chemical receptors and AI to identify both simple chemicals and complex scents. The company is aiming to create the world’s first comprehensive database of everyday scents, which it says it will use to make “intelligent decisions” for customers. What kind of decisions remains to be seen—and smelled.

Banner Image Credit: The Mercedes Vision AVTR concept car. Photo by Vanessa Bates Ramirez Continue reading

Posted in Human Robots

#436207 This Week’s Awesome Tech Stories From ...

COMPUTING
A Giant Superfast AI Chip Is Being Used to Find Better Cancer Drugs
Karen Hao | MIT Technology Review
“Thus far, Cerebras’s computer has checked all the boxes. Thanks to its chip size—it is larger than an iPad and has 1.2 trillion transistors for making calculations—it isn’t necessary to hook multiple smaller processors together, which can slow down model training. In testing, it has already shrunk the training time of models from weeks to hours.”

MEDICINE
Humans Put Into Suspended Animation for First Time
Ian Sample | The Guardian
“The process involves rapidly cooling the brain to less than 10C by replacing the patient’s blood with ice-cold saline solution. Typically the solution is pumped directly into the aorta, the main artery that carries blood away from the heart to the rest of the body.”

DRONES
This Transforming Drone Can Be Fired Straight Out of a Cannon
James Vincent | The Verge
“Drones are incredibly useful machines in the air, but getting them up and flying can be tricky, especially in crowded, windy, or emergency scenarios when speed is a factor. But a group of researchers from Caltech university and NASA’s Jet Propulsion Laboratory have come up with an elegant and oh-so-fun solution: fire the damn thing out of a cannon.”

ROBOTICS
Alphabet’s Dream of an ‘Everyday Robot’ Is Just Out of Reach
Tom Simonite | Wired
“Sorting trash was chosen as a convenient challenge to test the project’s approach to creating more capable robots. It’s using artificial intelligence software developed in collaboration with Google to make robots that learn complex tasks through on-the-job experience. The hope is to make robots less reliant on human coding for their skills, and capable of adapting quickly to complex new tasks and environments.”

ENVIRONMENT
The Electric Car Revolution May Take a Lot Longer Than Expected
James Temple | MIT Technology Review
“A new report from the MIT Energy Initiative warns that EVs may never reach the same sticker price so long as they rely on lithium-ion batteries, the energy storage technology that powers most of today’s consumer electronics. In fact, it’s likely to take another decade just to eliminate the difference in the lifetime costs between the vehicle categories, which factors in the higher fuel and maintenance expenses of standard cars and trucks.”

SPACE
How Two Intruders From Interstellar Space Are Upending Astronomy
Alexandra Witze | Nature
“From the tallest peak in Hawaii to a high plateau in the Andes, some of the biggest telescopes on Earth will point towards a faint smudge of light over the next few weeks. …What they’re looking for is a rare visitor that is about to make its closest approach to the Sun. After that, they have just months to grab as much information as they can from the object before it disappears forever into the blackness of space.”

Image Credit: Simone Hutsch / Unsplash Continue reading

Posted in Human Robots

#436186 Video Friday: Invasion of the Mini ...

Video Friday is your weekly selection of awesome robotics videos, collected by your Automaton bloggers. We’ll also be posting a weekly calendar of upcoming robotics events for the next few months; here's what we have so far (send us your events!):

DARPA SubT Urban Circuit – February 18-27, 2020 – Olympia, Wash., USA
Let us know if you have suggestions for next week, and enjoy today’s videos.

There will be a Mini-Cheetah Workshop (sponsored by Naver Labs) a year from now at IROS 2020 in Las Vegas. Mini-Cheetahs for everyone!

That’s just a rendering, of course, but this isn’t:

[ MCW ]

I was like 95 percent sure that the Urban Circuit of the DARPA SubT Challenge was going to be in something very subway station-y. Oops!

In the Subterranean (SubT) Challenge, teams deploy autonomous ground and aerial systems to attempt to map, identify, and report artifacts along competition courses in underground environments. The artifacts represent items a first responder or service member may encounter in unknown underground sites. This video provides a preview of the Urban Circuit event location. The Urban Circuit is scheduled for February 18-27, 2020, at Satsop Business Park west of Olympia, Washington.

[ SubT ]

Researchers at SEAS and the Wyss Institute for Biologically Inspired Engineering have developed a resilient RoboBee powered by soft artificial muscles that can crash into walls, fall onto the floor, and collide with other RoboBees without being damaged. It is the first microrobot powered by soft actuators to achieve controlled flight.

To solve the problem of power density, the researchers built upon the electrically-driven soft actuators developed in the lab of David Clarke, the Extended Tarr Family Professor of Materials. These soft actuators are made using dielectric elastomers, soft materials with good insulating properties, that deform when an electric field is applied. By improving the electrode conductivity, the researchers were able to operate the actuator at 500 Hertz, on par with the rigid actuators used previously in similar robots.

Next, the researchers aim to increase the efficiency of the soft-powered robot, which still lags far behind more traditional flying robots.

[ Harvard ]

We present a system for fast and robust handovers with a robot character, together with a user study investigating the effect of robot speed and reaction time on perceived interaction quality. The system can match and exceed human speeds and confirms that users prefer human-level timing.

In a 3×3 user study, we vary the speed of the robot and add variable sensorimotor delays. We evaluate the social perception of the robot using the Robot Social Attribute Scale (RoSAS). Inclusion of a small delay, mimicking the delay of the human sensorimotor system, leads to an improvement in perceived qualities over both no delay and long delay conditions. Specifically, with no delay the robot is perceived as more discomforting and with a long delay, it is perceived as less warm.

[ Disney Research ]

When cars are autonomous, they’re not going to be able to pump themselves full of gas. Or, more likely, electrons. Kuka has the solution.

[ Kuka ]

This looks like fun, right?

[ Robocoaster ]

NASA is leading the way in the use of On-orbit Servicing, Assembly, and Manufacturing to enable large, persistent, upgradable, and maintainable spacecraft. This video was developed by the Advanced Concepts Lab (ACL) at NASA Langley Research Center.

[ NASA ]

The noisiest workshop by far at Humanoids last month (by far) was Musical Interactions With Humanoids, the end result of which was this:

[ Workshop ]

IROS is an IEEE event, and in furthering the IEEE mission to benefit humanity through technological innovation, IROS is doing a great job. But don’t take it from us – we are joined by IEEE President-Elect Professor Toshio Fukuda to find out a bit more about the impact events like IROS can have, as well as examine some of the issues around intelligent robotics and systems – from privacy to transparency of the systems at play.

[ IROS ]

Speaking of IROS, we hope you’ve been enjoying our coverage. We have already featured Harvard’s strange sea-urchin-inspired robot and a Japanese quadruped that can climb vertical ladders, with more stories to come over the next several weeks.

In the mean time, enjoy these 10 videos from the conference (as usual, we’re including the title, authors, and abstract for each—if you’d like more details about any of these projects, let us know and we’ll find out more for you).

“A Passive Closing, Tendon Driven, Adaptive Robot Hand for Ultra-Fast, Aerial Grasping and Perching,” by Andrew McLaren, Zak Fitzgerald, Geng Gao, and Minas Liarokapis from the University of Auckland, New Zealand.

Current grasping methods for aerial vehicles are slow, inaccurate and they cannot adapt to any target object. Thus, they do not allow for on-the-fly, ultra-fast grasping. In this paper, we present a passive closing, adaptive robot hand design that offers ultra-fast, aerial grasping for a wide range of everyday objects. We investigate alternative uses of structural compliance for the development of simple, adaptive robot grippers and hands and we propose an appropriate quick release mechanism that facilitates an instantaneous grasping execution. The quick release mechanism is triggered by a simple distance sensor. The proposed hand utilizes only two actuators to control multiple degrees of freedom over three fingers and it retains the superior grasping capabilities of adaptive grasping mechanisms, even under significant object pose or other environmental uncertainties. The hand achieves a grasping time of 96 ms, a maximum grasping force of 56 N and it is able to secure objects of various shapes at high speeds. The proposed hand can serve as the end-effector of grasping capable Unmanned Aerial Vehicle (UAV) platforms and it can offer perching capabilities, facilitating autonomous docking.

“Unstructured Terrain Navigation and Topographic Mapping With a Low-Cost Mobile Cuboid Robot,” by Andrew S. Morgan, Robert L. Baines, Hayley McClintock, and Brian Scassellati from Yale University, USA.

Current robotic terrain mapping techniques require expensive sensor suites to construct an environmental representation. In this work, we present a cube-shaped robot that can roll through unstructured terrain and construct a detailed topographic map of the surface that it traverses in real time with low computational and monetary expense. Our approach devolves many of the complexities of locomotion and mapping to passive mechanical features. Namely, rolling movement is achieved by sequentially inflating latex bladders that are located on four sides of the robot to destabilize and tip it. Sensing is achieved via arrays of fine plastic pins that passively conform to the geometry of underlying terrain, retracting into the cube. We developed a topography by shade algorithm to process images of the displaced pins to reconstruct terrain contours and elevation. We experimentally validated the efficacy of the proposed robot through object mapping and terrain locomotion tasks.

“Toward a Ballbot for Physically Leading People: A Human-Centered Approach,” by Zhongyu Li and Ralph Hollis from Carnegie Mellon University, USA.

This work presents a new human-centered method for indoor service robots to provide people with physical assistance and active guidance while traveling through congested and narrow spaces. As most previous work is robot-centered, this paper develops an end-to-end framework which includes a feedback path of the measured human positions. The framework combines a planning algorithm and a human-robot interaction module to guide the led person to a specified planned position. The approach is deployed on a person-size dynamically stable mobile robot, the CMU ballbot. Trials were conducted where the ballbot physically led a blindfolded person to safely navigate in a cluttered environment.

“Achievement of Online Agile Manipulation Task for Aerial Transformable Multilink Robot,” by Fan Shi, Moju Zhao, Tomoki Anzai, Keita Ito, Xiangyu Chen, Kei Okada, and Masayuki Inaba from the University of Tokyo, Japan.

Transformable aerial robots are favorable in aerial manipulation tasks for their flexible ability to change configuration during the flight. By assuming robot keeping in the mild motion, the previous researches sacrifice aerial agility to simplify the complex non-linear system into a single rigid body with a linear controller. In this paper, we present a framework towards agile swing motion for the transformable multi-links aerial robot. We introduce a computational-efficient non-linear model predictive controller and joints motion primitive frame-work to achieve agile transforming motions and validate with a novel robot named HYRURS-X. Finally, we implement our framework under a table tennis task to validate the online and agile performance.

“Small-Scale Compliant Dual Arm With Tail for Winged Aerial Robots,” by Alejandro Suarez, Manuel Perez, Guillermo Heredia, and Anibal Ollero from the University of Seville, Spain.

Winged aerial robots represent an evolution of aerial manipulation robots, replacing the multirotor vehicles by fixed or flapping wing platforms. The development of this morphology is motivated in terms of efficiency, endurance and safety in some inspection operations where multirotor platforms may not be suitable. This paper presents a first prototype of compliant dual arm as preliminary step towards the realization of a winged aerial robot capable of perching and manipulating with the wings folded. The dual arm provides 6 DOF (degrees of freedom) for end effector positioning in a human-like kinematic configuration, with a reach of 25 cm (half-scale w.r.t. the human arm), and 0.2 kg weight. The prototype is built with micro metal gear motors, measuring the joint angles and the deflection with small potentiometers. The paper covers the design, electronics, modeling and control of the arms. Experimental results in test-bench validate the developed prototype and its functionalities, including joint position and torque control, bimanual grasping, the dynamic equilibrium with the tail, and the generation of 3D maps with laser sensors attached at the arms.

“A Novel Small-Scale Turtle-inspired Amphibious Spherical Robot,” by Huiming Xing, Shuxiang Guo, Liwei Shi, Xihuan Hou, Yu Liu, Huikang Liu, Yao Hu, Debin Xia, and Zan Li from Beijing Institute of Technology, China.

This paper describes a novel small-scale turtle-inspired Amphibious Spherical Robot (ASRobot) to accomplish exploration tasks in the restricted environment, such as amphibious areas and narrow underwater cave. A Legged, Multi-Vectored Water-Jet Composite Propulsion Mechanism (LMVWCPM) is designed with four legs, one of which contains three connecting rod parts, one water-jet thruster and three joints driven by digital servos. Using this mechanism, the robot is able to walk like amphibious turtles on various terrains and swim flexibly in submarine environment. A simplified kinematic model is established to analyze crawling gaits. With simulation of the crawling gait, the driving torques of different joints contributed to the choice of servos and the size of links of legs. Then we also modeled the robot in water and proposed several underwater locomotion. In order to assess the performance of the proposed robot, a series of experiments were carried out in the lab pool and on flat ground using the prototype robot. Experiments results verified the effectiveness of LMVWCPM and the amphibious control approaches.

“Advanced Autonomy on a Low-Cost Educational Drone Platform,” by Luke Eller, Theo Guerin, Baichuan Huang, Garrett Warren, Sophie Yang, Josh Roy, and Stefanie Tellex from Brown University, USA.

PiDrone is a quadrotor platform created to accompany an introductory robotics course. Students build an autonomous flying robot from scratch and learn to program it through assignments and projects. Existing educational robots do not have significant autonomous capabilities, such as high-level planning and mapping. We present a hardware and software framework for an autonomous aerial robot, in which all software for autonomy can run onboard the drone, implemented in Python. We present an Unscented Kalman Filter (UKF) for accurate state estimation. Next, we present an implementation of Monte Carlo (MC) Localization and Fast-SLAM for Simultaneous Localization and Mapping (SLAM). The performance of UKF, localization, and SLAM is tested and compared to ground truth, provided by a motion-capture system. Our evaluation demonstrates that our autonomous educational framework runs quickly and accurately on a Raspberry Pi in Python, making it ideal for use in educational settings.

“FlightGoggles: Photorealistic Sensor Simulation for Perception-driven Robotics using Photogrammetry and Virtual Reality,” by Winter Guerra, Ezra Tal, Varun Murali, Gilhyun Ryou and Sertac Karaman from the Massachusetts Institute of Technology, USA.

FlightGoggles is a photorealistic sensor simulator for perception-driven robotic vehicles. The key contributions of FlightGoggles are twofold. First, FlightGoggles provides photorealistic exteroceptive sensor simulation using graphics assets generated with photogrammetry. Second, it provides the ability to combine (i) synthetic exteroceptive measurements generated in silico in real time and (ii) vehicle dynamics and proprioceptive measurements generated in motio by vehicle(s) in flight in a motion-capture facility. FlightGoggles is capable of simulating a virtual-reality environment around autonomous vehicle(s) in flight. While a vehicle is in flight in the FlightGoggles virtual reality environment, exteroceptive sensors are rendered synthetically in real time while all complex dynamics are generated organically through natural interactions of the vehicle. The FlightGoggles framework allows for researchers to accelerate development by circumventing the need to estimate complex and hard-to-model interactions such as aerodynamics, motor mechanics, battery electrochemistry, and behavior of other agents. The ability to perform vehicle-in-the-loop experiments with photorealistic exteroceptive sensor simulation facilitates novel research directions involving, e.g., fast and agile autonomous flight in obstacle-rich environments, safe human interaction, and flexible sensor selection. FlightGoggles has been utilized as the main test for selecting nine teams that will advance in the AlphaPilot autonomous drone racing challenge. We survey approaches and results from the top AlphaPilot teams, which may be of independent interest. FlightGoggles is distributed as open-source software along with the photorealistic graphics assets for several simulation environments, under the MIT license at http://flightgoggles.mit.edu.

“An Autonomous Quadrotor System for Robust High-Speed Flight Through Cluttered Environments Without GPS,” by Marc Rigter, Benjamin Morrell, Robert G. Reid, Gene B. Merewether, Theodore Tzanetos, Vinay Rajur, KC Wong, and Larry H. Matthies from University of Sydney, Australia; NASA Jet Propulsion Laboratory, California Institute of Technology, USA; and Georgia Institute of Technology, USA.

Robust autonomous flight without GPS is key to many emerging drone applications, such as delivery, search and rescue, and warehouse inspection. These and other appli- cations require accurate trajectory tracking through cluttered static environments, where GPS can be unreliable, while high- speed, agile, flight can increase efficiency. We describe the hardware and software of a quadrotor system that meets these requirements with onboard processing: a custom 300 mm wide quadrotor that uses two wide-field-of-view cameras for visual- inertial motion tracking and relocalization to a prior map. Collision-free trajectories are planned offline and tracked online with a custom tracking controller. This controller includes compensation for drag and variability in propeller performance, enabling accurate trajectory tracking, even at high speeds where aerodynamic effects are significant. We describe a system identification approach that identifies quadrotor-specific parameters via maximum likelihood estimation from flight data. Results from flight experiments are presented, which 1) validate the system identification method, 2) show that our controller with aerodynamic compensation reduces tracking error by more than 50% in both horizontal flights at up to 8.5 m/s and vertical flights at up to 3.1 m/s compared to the state-of-the-art, and 3) demonstrate our system tracking complex, aggressive, trajectories.

“Morphing Structure for Changing Hydrodynamic Characteristics of a Soft Underwater Walking Robot,” by Michael Ishida, Dylan Drotman, Benjamin Shih, Mark Hermes, Mitul Luhar, and Michael T. Tolley from the University of California, San Diego (UCSD) and University of Southern California, USA.

Existing platforms for underwater exploration and inspection are often limited to traversing open water and must expend large amounts of energy to maintain a position in flow for long periods of time. Many benthic animals overcome these limitations using legged locomotion and have different hydrodynamic profiles dictated by different body morphologies. This work presents an underwater legged robot with soft legs and a soft inflatable morphing body that can change shape to influence its hydrodynamic characteristics. Flow over the morphing body separates behind the trailing edge of the inflated shape, so whether the protrusion is at the front, center, or back of the robot influences the amount of drag and lift. When the legged robot (2.87 N underwater weight) needs to remain stationary in flow, an asymmetrically inflated body resists sliding by reducing lift on the body by 40% (from 0.52 N to 0.31 N) at the highest flow rate tested while only increasing drag by 5.5% (from 1.75 N to 1.85 N). When the legged robot needs to walk with flow, a large inflated body is pushed along by the flow, causing the robot to walk 16% faster than it would with an uninflated body. The body shape significantly affects the ability of the robot to walk against flow as it is able to walk against 0.09 m/s flow with the uninflated body, but is pushed backwards with a large inflated body. We demonstrate that the robot can detect changes in flow velocity with a commercial force sensor and respond by morphing into a hydrodynamically preferable shape. Continue reading

Posted in Human Robots

#436165 Video Friday: DJI’s Mavic Mini Is ...

Video Friday is your weekly selection of awesome robotics videos, collected by your Automaton bloggers. We’ll also be posting a weekly calendar of upcoming robotics events for the next few months; here’s what we have so far (send us your events!):

IROS 2019 – November 4-8, 2019 – Macau
Let us know if you have suggestions for next week, and enjoy today’s videos.

DJI’s new Mavic Mini looks like a pretty great drone for US $400 ($500 for a combo with more accessories): It’s tiny, flies for 30 minutes, and will do what you need as far as pictures and video (although not a whole lot more).

DJI seems to have put a bunch of effort into making the drone 249 grams, 1 gram under what’s required for FAA registration. That means you save $5 and a few minutes of your time, but that does not mean you don’t have to follow the FAA’s rules and regulations governing drone use.

[ DJI ]

Don’t panic, but Clearpath and HEBI Robotics have armed the Jackal:

After locking eyes across a crowded room at ICRA 2019, Clearpath Robotics and HEBI Robotics basked in that warm and fuzzy feeling that comes with starting a new and exciting relationship. Over a conference hall coffee, they learned that the two companies have many overlapping interests. The most compelling was the realization that customers across a variety of industries are hunting for an elusive true love of their own – a robust but compact robotic platform combined with a long reach manipulator for remote inspection tasks.

After ICRA concluded, Arron Griffiths, Application Engineer at Clearpath, and Matthew Tesch, Software Engineer at HEBI, kept in touch and decided there had been enough magic in the air to warrant further exploration. A couple of months later, Matthew arrived at Clearpath to formally introduce the HEBI’s X-Series Arm to Clearpath’s Jackal UGV. It was love.

[ Clearpath ]

Thanks Dave!

I’m really not a fan of the people-carrying drones, but heavy lift cargo drones seem like a more okay idea.

Volocopter, the pioneer in Urban Air Mobility, presented the demonstrator of its VoloDrone. This marks Volocopters expansion into the logistics, agriculture, infrastructure and public services industry. The VoloDrone is an unmanned, fully electric, heavy-lift utility drone capable of carrying a payload of 200 kg (440 lbs) up to 40 km (25 miles). With a standardized payload attachment, VoloDrone can serve a great variety of purposes from transporting boxes, to liquids, to equipment and beyond. It can be remotely piloted or flown in automated mode on pre-set routes.

[ Volocopter ]

JAY is a mobile service robot that projects a display on the floor and plays sound with its speaker. By playing sounds and videos, it provides visual and audio entertainment in various places such as exhibition halls, airports, hotels, department stores and more.

[ Rainbow Robotics ]

The DARPA Subterranean Challenge Virtual Tunnel Circuit concluded this week—it was the same idea as the physical challenge that took place in August, just with a lot less IRL dirt.

The awards ceremony and team presentations are in this next video, and we’ll have more on this once we get back from IROS.

[ DARPA SubT ]

NASA is sending a mobile robot to the south pole of the Moon to get a close-up view of the location and concentration of water ice in the region and for the first time ever, actually sample the water ice at the same pole where the first woman and next man will land in 2024 under the Artemis program.

About the size of a golf cart, the Volatiles Investigating Polar Exploration Rover, or VIPER, will roam several miles, using its four science instruments — including a 1-meter drill — to sample various soil environments. Planned for delivery in December 2022, VIPER will collect about 100 days of data that will be used to inform development of the first global water resource maps of the Moon.

[ NASA ]

Happy Halloween from HEBI Robotics!

[ HEBI ]

Happy Halloween from Soft Robotics!

[ Soft Robotics ]

Halloween must be really, really confusing for autonomous cars.

[ Waymo ]

Once a year at Halloween, hardworking JPL engineers put their skills to the test in a highly competitive pumpkin carving contest. The result: A pumpkin gently landed on the Moon, its retrorockets smoldering, while across the room a Nemo-inspired pumpkin explored the sub-surface ocean of Jupiter moon Europa. Suffice to say that when the scientists and engineers at NASA’s Jet Propulsion Laboratory compete in a pumpkin-carving contest, the solar system’s the limit. Take a look at some of the masterpieces from 2019.

Now in its ninth year, the contest gives teams only one hour to carve and decorate their pumpkin though they can prepare non-pumpkin materials – like backgrounds, sound effects and motorized parts – ahead of time.

[ JPL ]

The online autonomous navigation and semantic mapping experiment presented [below] is conducted with the Cassie Blue bipedal robot at the University of Michigan. The sensors attached to the robot include an IMU, a 32-beam LiDAR and an RGB-D camera. The whole online process runs in real-time on a Jetson Xavier and a laptop with an i7 processor.

[ BPL ]

Misty II is now available to anyone who wants one, and she’s on sale for a mere $2900.

[ Misty ]

We leveraged LIDAR-based slam, in conjunction with our specialized relative localization sensor UVDAR to perform a de-centralized, communication-free swarm flight without the units knowing their absolute locations. The swarming and obstacle avoidance control is based on a modified Boids-like algorithm, while the whole swarm is controlled by directing a selected leader unit.

[ MRS ]

The MallARD robot is an autonomous surface vehicle (ASV), designed for the monitoring and inspection of wet storage facilities for example spent fuel pools or wet silos. The MallARD is holonomic, uses a LiDAR for localisation and features a robust trajectory tracking controller.

The University of Manchester’s researcher Dr Keir Groves designed and built the autonomous surface vehicle (ASV) for the challenge which came in the top three of the second round in Nov 2017. The MallARD went on to compete in a final 3rd round where it was deployed in a spent fuel pond at a nuclear power plant in Finland by the IAEA, along with two other entries. The MallARD came second overall, in November 2018.

[ RNE ]

Thanks Jennifer!

I sometimes get the sense that in the robotic grasping and manipulation world, suction cups are kinda seen as cheating at times. But, their nature allows you to do some pretty interesting things.

More clever octopus footage please.

[ CMU ]

A Personal, At-Home Teacher For Playful Learning: From academic topics to child-friendly news bulletins, fun facts and more, Miko 2 is packed with relevant and freshly updated content specially designed by educationists and child-specialists. Your little one won’t even realize they’re learning.

As we point out pretty much every time we post a video like this, keep in mind that you’re seeing a heavily edited version of a hypothetical best case scenario for how this robot can function. And things like “creating a relationship that they can then learn how to form with their peers” is almost certainly overselling things. But at $300 (shipping included), this may be a decent robot as long as your expectations are appropriately calibrated.

[ Miko ]

ICRA 2018 plenary talk by Rodney Brooks: “Robots and People: the Research Challenge.”

[ IEEE RAS ]

ICRA-X 2018 talk by Ron Arkin: “Lethal Autonomous Robots and the Plight of the Noncombatant.”

[ IEEE RAS ]

On the most recent episode of the AI Podcast, Lex Fridman interviews Garry Kasparov.

[ AI Podcast ] Continue reading

Posted in Human Robots