Tag Archives: good

#430761 How Robots Are Getting Better at Making ...

The multiverse of science fiction is populated by robots that are indistinguishable from humans. They are usually smarter, faster, and stronger than us. They seem capable of doing any job imaginable, from piloting a starship and battling alien invaders to taking out the trash and cooking a gourmet meal.
The reality, of course, is far from fantasy. Aside from industrial settings, robots have yet to meet The Jetsons. The robots the public are exposed to seem little more than over-sized plastic toys, pre-programmed to perform a set of tasks without the ability to interact meaningfully with their environment or their creators.
To paraphrase PayPal co-founder and tech entrepreneur Peter Thiel, we wanted cool robots, instead we got 140 characters and Flippy the burger bot. But scientists are making progress to empower robots with the ability to see and respond to their surroundings just like humans.
Some of the latest developments in that arena were presented this month at the annual Robotics: Science and Systems Conference in Cambridge, Massachusetts. The papers drilled down into topics that ranged from how to make robots more conversational and help them understand language ambiguities to helping them see and navigate through complex spaces.
Improved Vision
Ben Burchfiel, a graduate student at Duke University, and his thesis advisor George Konidaris, an assistant professor of computer science at Brown University, developed an algorithm to enable machines to see the world more like humans.
In the paper, Burchfiel and Konidaris demonstrate how they can teach robots to identify and possibly manipulate three-dimensional objects even when they might be obscured or sitting in unfamiliar positions, such as a teapot that has been tipped over.
The researchers trained their algorithm by feeding it 3D scans of about 4,000 common household items such as beds, chairs, tables, and even toilets. They then tested its ability to identify about 900 new 3D objects just from a bird’s eye view. The algorithm made the right guess 75 percent of the time versus a success rate of about 50 percent for other computer vision techniques.
In an email interview with Singularity Hub, Burchfiel notes his research is not the first to train machines on 3D object classification. How their approach differs is that they confine the space in which the robot learns to classify the objects.
“Imagine the space of all possible objects,” Burchfiel explains. “That is to say, imagine you had tiny Legos, and I told you [that] you could stick them together any way you wanted, just build me an object. You have a huge number of objects you could make!”
The infinite possibilities could result in an object no human or machine might recognize.
To address that problem, the researchers had their algorithm find a more restricted space that would host the objects it wants to classify. “By working in this restricted space—mathematically we call it a subspace—we greatly simplify our task of classification. It is the finding of this space that sets us apart from previous approaches.”
Following Directions
Meanwhile, a pair of undergraduate students at Brown University figured out a way to teach robots to understand directions better, even at varying degrees of abstraction.
The research, led by Dilip Arumugam and Siddharth Karamcheti, addressed how to train a robot to understand nuances of natural language and then follow instructions correctly and efficiently.
“The problem is that commands can have different levels of abstraction, and that can cause a robot to plan its actions inefficiently or fail to complete the task at all,” says Arumugam in a press release.
In this project, the young researchers crowdsourced instructions for moving a virtual robot through an online domain. The space consisted of several rooms and a chair, which the robot was told to manipulate from one place to another. The volunteers gave various commands to the robot, ranging from general (“take the chair to the blue room”) to step-by-step instructions.
The researchers then used the database of spoken instructions to teach their system to understand the kinds of words used in different levels of language. The machine learned to not only follow instructions but to recognize the level of abstraction. That was key to kickstart its problem-solving abilities to tackle the job in the most appropriate way.
The research eventually moved from virtual pixels to a real place, using a Roomba-like robot that was able to respond to instructions within one second 90 percent of the time. Conversely, when unable to identify the specificity of the task, it took the robot 20 or more seconds to plan a task about 50 percent of the time.
One application of this new machine-learning technique referenced in the paper is a robot worker in a warehouse setting, but there are many fields that could benefit from a more versatile machine capable of moving seamlessly between small-scale operations and generalized tasks.
“Other areas that could possibly benefit from such a system include things from autonomous vehicles… to assistive robotics, all the way to medical robotics,” says Karamcheti, responding to a question by email from Singularity Hub.
More to Come
These achievements are yet another step toward creating robots that see, listen, and act more like humans. But don’t expect Disney to build a real-life Westworld next to Toon Town anytime soon.
“I think we’re a long way off from human-level communication,” Karamcheti says. “There are so many problems preventing our learning models from getting to that point, from seemingly simple questions like how to deal with words never seen before, to harder, more complicated questions like how to resolve the ambiguities inherent in language, including idiomatic or metaphorical speech.”
Even relatively verbose chatbots can run out of things to say, Karamcheti notes, as the conversation becomes more complex.
The same goes for human vision, according to Burchfiel.
While deep learning techniques have dramatically improved pattern matching—Google can find just about any picture of a cat—there’s more to human eyesight than, well, meets the eye.
“There are two big areas where I think perception has a long way to go: inductive bias and formal reasoning,” Burchfiel says.
The former is essentially all of the contextual knowledge people use to help them reason, he explains. Burchfiel uses the example of a puddle in the street. People are conditioned or biased to assume it’s a puddle of water rather than a patch of glass, for instance.
“This sort of bias is why we see faces in clouds; we have strong inductive bias helping us identify faces,” he says. “While it sounds simple at first, it powers much of what we do. Humans have a very intuitive understanding of what they expect to see, [and] it makes perception much easier.”
Formal reasoning is equally important. A machine can use deep learning, in Burchfiel’s example, to figure out the direction any river flows once it understands that water runs downhill. But it’s not yet capable of applying the sort of human reasoning that would allow us to transfer that knowledge to an alien setting, such as figuring out how water moves through a plumbing system on Mars.
“Much work was done in decades past on this sort of formal reasoning… but we have yet to figure out how to merge it with standard machine-learning methods to create a seamless system that is useful in the actual physical world.”
Robots still have a lot to learn about being human, which should make us feel good that we’re still by far the most complex machines on the planet.
Image Credit: Alex Knight via Unsplash Continue reading

Posted in Human Robots

#430734 Why XPRIZE Is Asking Writers to Take Us ...

In a world of accelerating change, educating the public about the implications of technological advancements is extremely important. We can continue to write informative articles and speculate about the kind of future that lies ahead. Or instead, we can take readers on an immersive journey by using science fiction to paint vivid images of the future for society.
The XPRIZE Foundation recently announced a science fiction storytelling competition. In recent years, the organization has backed and launched a range of competitions to propel innovation in science and technology. These have been aimed at a variety of challenges, such as transforming the lives of low-literacy adults, tackling climate change, and creating water from thin air.
Their sci-fi writing competition asks participants to envision a groundbreaking future for humanity. The initiative, in partnership with Japanese airline ANA, features 22 sci-fi stories from noteworthy authors that are now live on the website. Each of these stories is from the perspective of a different passenger on a plane that travels 20 years into the future through a wormhole. Contestants will compete to tell the story of the passenger in Seat 14C.
In addition to the competition, XPRIZE has brought together a science fiction advisory council to work with the organization and imagine what the future will look like. According to Peter Diamandis, founder and executive chairman, “As the future becomes harder and harder to predict, we look forward to engaging some of the world’s most visionary storytellers to help us imagine what’s just beyond the horizon and chart a path toward a future of abundance.”
The Importance of Science Fiction
Why is an organization like XPRIZE placing just as much importance on fiction as it does on reality? As Isaac Asimov has pointed out, “Modern science fiction is the only form of literature that consistently considers the nature of the changes that face us.” While the rest of the world reports on a new invention, sci-fi authors examine how these advancements affect the human condition.
True science fiction is distinguished from pure fantasy in that everything that happens is within the bounds of the physical laws of the universe. We’ve already seen how sci-fi can inspire generations and shape the future. 3D printers, wearable technology, and smartphones were first seen in Star Trek. Targeted advertising and air touch technology was first seen in Philip K. Dick’s 1958 story “The Minority Report.” Tanning beds, robot vacuums, and flatscreen TVs were seen in The Jetsons. The internet and a world of global instant communication was predicted by Arthur C. Clarke in his work long before it became reality.
Sci-fi shows like Black Mirror or Star Trek aren’t just entertainment. They allow us to imagine and explore the influence of technology on humanity. For instance, how will artificial intelligence impact human relationships? How will social media affect privacy? What if we encounter alien life? Good sci-fi stories take us on journeys that force us to think critically about the societal impacts of technological advancements.
As sci-fi author Yaasha Moriah points out, the genre is universal because “it tackles hard questions about human nature, morality, and the evolution of society, all through the narrative of speculation about the future. If we continue to do A, will it necessarily lead to problems B and C? What implicit lessons are being taught when we insist on a particular policy? When we elevate the importance of one thing over another—say, security over privacy—what could be the potential benefits and dangers of that mentality? That’s why science fiction has such an enduring appeal. We want to explore deep questions, without being preached at. We want to see the principles in action, and observe their results.”
An Extension of STEAM Education
At its core, this genre is a harmonious symbiosis between two distinct disciplines: science and literature. It is an extension of STEAM education, an educational approach that combines science, technology, engineering, the arts, and mathematics. Story-telling with science fiction allows us to use the arts in order to educate and engage the public about scientific advancements and its implications.
According to the National Science Foundation, research on art-based learning of STEM, including the use of narrative writing, works “beyond expectation.” It has been shown to have a powerful impact on creative thinking, collaborative behavior and application skills.
What does it feel like to travel through a wormhole? What are some ethical challenges of AI? How could we terraform Mars? For decades, science fiction writers and producers have answered these questions through the art of storytelling.
What better way to engage more people with science and technology than through sparking their imaginations? The method makes academic subject areas many traditionally perceived as boring or dry far more inspiring and engaging.
A Form of Time Travel
XPRIZE’s competition theme of traveling 20 years into the future through a wormhole is an appropriate beacon for the genre. In many ways, sci-fi is a precautionary form of time travel. Before we put a certain technology, scientific invention, or policy to use, we can envision and explore what our world would be like if we were to do so.
Sci-fi lets us explore different scenarios for the future of humanity before deciding which ones are more desirable. Some of these scenarios may be radically beyond our comfort zone. Yet when we’re faced with the seemingly impossible, we must remind ourselves that if something is within the domain of the physical laws of the universe, then it’s absolutely possible.
Stock Media provided by NASA_images / Pond5 Continue reading

Posted in Human Robots

#430656 Code @ TACC robotics camp delivers on ...

On a hot and breezy June day in Austin, parents, friends, brothers and sisters navigated through main campus at The University of Texas at Austin and helped carry luggage for the new arrivals to their dorm rooms. Thirty-four high school students from mostly low-income Title I schools in Central Texas, some from as far away as Houston, said good-bye to their families. Continue reading

Posted in Human Robots

#430556 Forget Flying Cars, the Future Is ...

Flying car concepts have been around nearly as long as their earthbound cousins, but no one has yet made them a commercial success. MIT engineers think we’ve been coming at the problem from the wrong direction; rather than putting wings on cars, we should be helping drones to drive.
The team from the university’s Computer Science and Artificial Intelligence Laboratory (CSAIL) added wheels to a fleet of eight mini-quadcopters and tested driving and flying them around a tiny toy town made out of cardboard and fabric.
Adding the ability to drive reduced the distance the drone could fly by 14 percent compared to a wheel-less version. But while driving was slower, the drone could travel 150 percent further than when flying. The result is a vehicle that combines the speed and mobility of flying with the energy-efficiency of driving.

CSAIL director Daniela Rus told MIT News their work suggested that when looking to create flying cars, it might make more sense to build on years of research into drones rather than trying to simply “put wings on cars.”
Historically, flying car concepts have looked like someone took apart a Cessna light aircraft and a family sedan, mixed all the parts up, and bolted them back together again. Not everyone has abandoned this approach—two of the most developed flying car designs from Terrafugia and AeroMobil are cars with folding wings that need an airstrip to take off.
But flying car concepts are looking increasingly drone-like these days, with multiple small rotors, electric propulsion and vertical take-off abilities. Take the eHang 184 autonomous aerial vehicle being developed in China, the Kitty Hawk all-electric aircraft backed by Google founder Larry Page, which is little more than a quadcopter with a seat, the AirQuadOne designed by UK consortium Neva Aerospace, or Lilium Aviation’s Jet.
The attraction is obvious. Electric-powered drones are more compact, maneuverable, and environmentally friendly, making them suitable for urban environments.
Most of these vehicles are not quite the same as those proposed by the MIT engineers, as they’re pure flying machines. But a recent Airbus concept builds on the same principle that the future of urban mobility is vehicles that can both fly and drive. Its Pop.Up design is a two-passenger pod that can either be clipped to a set of wheels or hang under a quadcopter.
Importantly, they envisage their creation being autonomous in both flight and driving modes. And they’re not the only ones who think the future of flying cars is driverless. Uber has committed to developing a network of autonomous air taxis within a decade. This spring, Dubai announced it would launch a pilotless passenger drone service using the Ehang 184 as early as next month (July).
While integrating fully-fledged autonomous flying cars into urban environments will be far more complex, the study by Rus and her colleagues provides a good starting point for the kind of 3D route-planning and collision avoidance capabilities this would require.
The team developed multi-robot path planning algorithms that were able to control all eight drones as they flew and drove around their mock up city, while also making sure they didn’t crash into each other and avoided no-fly zones.
“This work provides an algorithmic solution for large-scale, mixed-mode transportation and shows its applicability to real-world problems,” Jingjin Yu, a computer science professor at Rutgers University who was not involved in the research, told MIT News.
This vision of a driverless future for flying cars might be a bit of a disappointment for those who’d envisaged themselves one day piloting their own hover car just like George Jetson. But autonomy and Uber-like ride-hailing business models are likely to be attractive, as they offer potential solutions to three of the biggest hurdles drone-like passenger vehicles face.
Firstly, it makes the vehicles accessible to anyone by removing the need to learn how to safely pilot an aircraft. Secondly, battery life still limits most electric vehicles to flight times measured in minutes. For personal vehicles this could be frustrating, but if you’re just hopping in a driverless air taxi for a five minute trip across town it’s unlikely to become apparent to you.
Operators of the service simply need to make sure they have a big enough fleet to ensure a charged vehicle is never too far away, or they’ll need a way to swap out batteries easily, such as the one suggested by the makers of the Volocopter electric helicopter.
Finally, there has already been significant progress in developing technology and regulations needed to integrate autonomous drones into our airspace that future driverless flying cars can most likely piggyback off of.
Safety requirements will inevitably be more stringent, but adding more predictable and controllable autonomous drones to the skies is likely to be more attractive to regulators than trying to license and police thousands of new amateur pilots.
Image Credit: Lilium Continue reading

Posted in Human Robots

#430550 This Week’s Awesome Stories From ...

DRONES
MIT Is Building Autonomous Drones That Can Both Drive and FlyApril Glaser | Recode“The drones, which were built at MIT’s Computer Science and Artificial Intelligence Laboratory, also include route-planning software that can help calculate when the flying robot switches from air to ground in order to optimize its battery life.”
SPACE
SpaceX Is Making Commercial Space Launches Look Like Child’s PlayJamie Condliffe | MIT Technology Review“Late Friday, SpaceX launched a satellite into orbit from Florida using one of its refurbished Falcon 9 rockets. Then on Sunday, for good measure, it lofted 10 smaller satellites using a new version of the same rocket, which it launched from California. The feat is a sign that the private space company seems more likely than ever to turn its vision of competitively priced, rapid-turnaround rocket launches into reality.”
CYBERSECURITY
A New Ransomware Attack Is Infecting Airlines, Banks, and Utilities Across EuropeRussell Brandom | The Verge“The origins of the attack are still unclear, but the involvement of Ukraine’s electric utilities is likely to cast suspicion on Russia. Ukraine’s power grid was hit by a persistent and sophisticated attack in December 2015, which many attributed to Russia. The attack ultimately left 230,000 residents without power for as long as six hours.”
SILICON VALLEY NEWS
Mark Zuckerberg’s Probably Nonexistent 2020 Presidential Campaign, ExplainedTimothy B. Lee | VOX“After all, the kind of outreach Zuckerberg would do in a presidential campaign isn’t that different from the kind of outreach he’d do if he were simply trying to understand Facebook users better and build public goodwill for his massive social media site.”
AUTONOMOUS CARS
Riding in a Robocar That Sees Around CornersPhilip E. Ross | IEEE Spectrum“It takes 20 to 30 minutes to fit a car with the necessary hardware: a GPS sensor and a wireless transceiver. Here in the MCity compound, at least, the GPS system uses a repeater to enhance its accuracy down to centimeter level—good enough to locate a car precisely and to allow other cars to figure out its trajectory and measure its speed.”
Image Credit: SpaceX / Flickr Continue reading

Posted in Human Robots