Tag Archives: could

#438925 Nanophotonics Could Be the ‘Dark ...

The race to build the first practical quantum computers looks like a two-horse contest between machines built from superconducting qubits and those that use trapped ions. But new research suggests a third contender—machines based on optical technology—could sneak up on the inside.

The most advanced quantum computers today are the ones built by Google and IBM, which rely on superconducting circuits to generate the qubits that form the basis of quantum calculations. They are now able to string together tens of qubits, and while controversial, Google claims its machines have achieved quantum supremacy—the ability to carry out a computation beyond normal computers.

Recently this approach has been challenged by a wave of companies looking to use trapped ion qubits, which are more stable and less error-prone than superconducting ones. While these devices are less developed, engineering giant Honeywell has already released a machine with 10 qubits, which it says is more powerful than a machine made of a greater number of superconducting qubits.

But despite this progress, both of these approaches have some major drawbacks. They require specialized fabrication methods, incredibly precise control mechanisms, and they need to be cooled to close to absolute zero to protect the qubits from any outside interference.

That’s why researchers at Canadian quantum computing hardware and software startup Xanadu are backing an alternative quantum computing approach based on optics, which was long discounted as impractical. In a paper published last week in Nature, they unveiled the first fully programmable and scalable optical chip that can run quantum algorithms. Not only does the system run at room temperature, but the company says it could scale to millions of qubits.

The idea isn’t exactly new. As Chris Lee notes in Ars Technica, people have been experimenting with optical approaches to quantum computing for decades, because encoding information in photons’ quantum states and manipulating those states is relatively easy. The biggest problem was that optical circuits were very large and not readily programmable, which meant you had to build a new computer for every new problem you wanted to solve.

That started to change thanks to the growing maturity of photonic integrated circuits. While early experiments with optical computing involved complex table-top arrangements of lasers, lenses, and detectors, today it’s possible to buy silicon chips not dissimilar to electronic ones that feature hundreds of tiny optical components.

In recent years, the reliability and performance of these devices has improved dramatically, and they’re now regularly used by the telecommunications industry. Some companies believe they could be the future of artificial intelligence too.

This allowed the Xanadu researchers to design a silicon chip that implements a complex optical network made up of beam splitters, waveguides, and devices called interferometers that cause light sources to interact with each other.

The chip can generate and manipulate up to eight qubits, but unlike conventional qubits, which can simultaneously be in two states, these qubits can be in any configuration of three states, which means they can carry more information.

Once the light has travelled through the network, it is then fed out to cutting-edge photon-counting detectors that provide the result. This is one of the potential limitations of the system, because currently these detectors need to be cryogenically cooled, although the rest of the chip does not.

But most importantly, the chip is easily re-programmable, which allows it to tackle a variety of problems. The computation can be controlled by adjusting the settings of these interferometers, but the researchers have also developed a software platform that hides the physical complexity from users and allows them to program it using fairly conventional code.

The company announced that its chips were available on the cloud in September of 2020, but the Nature paper is the first peer-reviewed test of their system. The researchers verified that the computations being done were genuinely quantum mechanical in nature, but they also implemented two more practical algorithms: one for simulating molecules and the other for judging how similar two graphs are, which has applications in a variety of pattern recognition problems.

In an accompanying opinion piece, Ulrik Andersen from the Technical University of Denmark says the quality of the qubits needs to be improved considerably and photon losses reduced if the technology is ever to scale to practical problems. But, he says, this breakthrough suggests optical approaches “could turn out to be the dark horse of quantum computing.”

Image Credit: Shahadat Rahman on Unsplash Continue reading

Posted in Human Robots

#438886 This Week’s Awesome Tech Stories From ...

ARTIFICIAL INTELLIGENCE
This Chip for AI Works Using Light, Not Electrons
Will Knight | Wired
“As demand for artificial intelligence grows, so does hunger for the computer power needed to keep AI running. Lightmatter, a startup born at MIT, is betting that AI’s voracious hunger will spawn demand for a fundamentally different kind of computer chip—one that uses light to perform key calculations. ‘Either we invent new kinds of computers to continue,’ says Lightmatter CEO Nick Harris, ‘or AI slows down.’i”

BIOTECH
With This CAD for Genomes, You Can Design New Organisms
Eliza Strickland | IEEE Spectrum
“Imagine being able to design a new organism as easily as you can design a new integrated circuit. That’s the ultimate vision behind the computer-aided design (CAD) program being developed by the GP-write consortium. ‘We’re taking the same things we’d do for design automation in electronics, and applying them to biology,’ says Doug Densmore, an associate professor of electrical and computer engineering at Boston University.”

BIOLOGY
Hey, So These Sea Slugs Decapitate Themselves and Grow New Bodies
Matt Simon | Wired
“That’s right: It pulled a Deadpool. Just a few hours after its self-decapitation, the head began dragging itself around to feed. After a day, the neck wound had closed. After a week, it started to regenerate a heart. In less than a month, the whole body had grown back, and the disembodied slug was embodied once more.”

INTERNET
Move Over, Deep Nostalgia, This AI App Can Make Kim Jong-un Sing ‘I Will Survive’
Helen Sullivan | The Guardian
“If you’ve ever wanted to know what it might be like to see Kim Jong-un let loose at karaoke, your wish has been granted, thanks to an app that lets users turn photographs of anyone—or anything remotely resembling a face—into uncanny AI-powered videos of them lip syncing famous songs.”

ENERGY
GM Unveils Plans for Lithium-Metal Batteries That Could Boost EV Range
Steve Dent | Engadget
“GM has released more details about its next-generation Ultium batteries, including plans for lithium-metal (Li-metal) technology to boost performance and energy density. The automaker announced that it has signed an agreement to work with SolidEnergy Systems (SES), an MIT spinoff developing prototype Li-metal batteries with nearly double the capacity of current lithium-ion cells.”

TECHNOLOGY
Xi’s Gambit: China Plans for a World Without American Technology
Paul Mozur and Steven Lee Myers | The New York Times
“China is freeing up tens of billions of dollars for its tech industry to borrow. It is cataloging the sectors where the United States or others could cut off access to crucial technologies. And when its leaders released their most important economic plans last week, they laid out their ambitions to become an innovation superpower beholden to none.”

SCIENCE
Imaginary Numbers May Be Essential for Describing Reality
Charlie Wood | Wired
“…physicists may have just shown for the first time that imaginary numbers are, in a sense, real. A group of quantum theorists designed an experiment whose outcome depends on whether nature has an imaginary side. Provided that quantum mechanics is correct—an assumption few would quibble with—the team’s argument essentially guarantees that complex numbers are an unavoidable part of our description of the physical universe.”

PHILOSOPHY
What Is Life? Its Vast Diversity Defies Easy Definition
Carl Zimmer | Quanta
“i‘It is commonly said,’ the scientists Frances Westall and André Brack wrote in 2018, ‘that there are as many definitions of life as there are people trying to define it.’ …As an observer of science and of scientists, I find this behavior strange. It is as if astronomers kept coming up with new ways to define stars. …With scientists adrift in an ocean of definitions, philosophers rowed out to offer lifelines.”

Image Credit: Kir Simakov / Unsplash Continue reading

Posted in Human Robots

#438809 This Week’s Awesome Tech Stories From ...

ARTIFICIAL INTELLIGENCE
Facebook’s New AI Teaches Itself to See With Less Human Help
Will Knight | Wired
“Peer inside an AI algorithm and you’ll find something constructed using data that was curated and labeled by an army of human workers. Now, Facebook has shown how some AI algorithms can learn to do useful work with far less human help. The company built an algorithm that learned to recognize objects in images with little help from labels.”

CULTURE
New AI ‘Deep Nostalgia’ Brings Old Photos, Including Very Old Ones, to Life
Kim Lyons | The Verge
“The Deep Nostalgia service, offered by online genealogy company MyHeritage, uses AI licensed from D-ID to create the effect that a still photo is moving. It’s kinda like the iOS Live Photos feature, which adds a few seconds of video to help smartphone photographers find the best shot. But Deep Nostalgia can take photos from any camera and bring them to ‘life.’i”

COMPUTING
Could ‘Topological Materials’ Be a New Medium For Ultra-Fast Electronics?
Charles Q. Choi | IEEE Spectrum
“Potential future transistors that can exceed Moore’s law may rely on exotic materials called ‘topological matter’ in which electricity flows across surfaces only, with virtually no dissipation of energy. And now new findings suggest these special topological materials might one day find use in high-speed, low-power electronics and in quantum computers.”

ENERGY
A Chinese Province Could Ban Bitcoin Mining to Cut Down Energy Use
Dharna Noor | Gizmodo
“Since energy prices in Inner Mongolia are particularly low, many bitcoin miners have set up shop there specifically. The region is the third-largest mining site in China. Because the grid is heavily coal-powered, however, that’s led to skyrocketing emissions, putting it in conflict with President Xi Jinping’s promise last September to have China reach peak carbon emissions by 2030 at the latest and achieve carbon neutrality before 2060.”

VIRTUAL REALITY
Mesh Is Microsoft’s Vision for Sending Your Hologram Back to the Office
Sam Rutherford | Gizmodo
“With Mesh, Microsoft is hoping to create a virtual environment capable of sharing data, 3D models, avatars, and more—basically, the company wants to upgrade the traditional remote-working experience with the power of AR and VR. In the future, Microsoft is planning for something it’s calling ‘holoportation,’ which will allow Mesh devices to create photorealistic digital avatars of your body that can appear in virtual spaces anywhere in the world—assuming you’ve been invited, of course.”

SPACE
Rocket Lab Could Be SpaceX’s Biggest Rival
Neel V. Patel | MIT Technology Review
“At 40 meters tall and able to carry 20 times the weight that Electron can, [the new] Neutron [rocket] is being touted by Rocket Lab as its entry into markets for large satellite and mega-constellation launches, as well as future robotics missions to the moon and Mars. Even more tantalizing, Rocket Lab says Neutron will be designed for human spaceflight as well.”

SCIENCE
Can Alien Smog Lead Us to Extraterrestrial Civilizations?
Meghan Herbst | Wired
“Kopparapu is at the forefront of an emerging field in astronomy that is aiming to identify technosignatures, or technological markers we can search for in the cosmos. No longer conceptually limited to radio signals, astronomers are looking for ways we could identify planets or other spacefaring objects by looking for things like atmospheric gases, lasers, and even hypothetical sun-encircling structures called Dyson spheres.”

DIGITAL CURRENCIES
China Charges Ahead With a National Digital Currency
Nathaniel Popper and Cao Li | The New York Times
“China has charged ahead with a bold effort to remake the way that government-backed money works, rolling out its own digital currency with different qualities than cash or digital deposits. The country’s central bank, which began testing eCNY last year in four cities, recently expanded those trials to bigger cities such as Beijing and Shanghai, according to government presentations.”

Image Credit: Leon Seibert / Unsplash Continue reading

Posted in Human Robots

#438807 Visible Touch: How Cameras Can Help ...

The dawn of the robot revolution is already here, and it is not the dystopian nightmare we imagined. Instead, it comes in the form of social robots: Autonomous robots in homes and schools, offices and public spaces, able to interact with humans and other robots in a socially acceptable, human-perceptible way to resolve tasks related to core human needs.

To design social robots that “understand” humans, robotics scientists are delving into the psychology of human communication. Researchers from Cornell University posit that embedding the sense of touch in social robots could teach them to detect physical interactions and gestures. They describe a way of doing so by relying not on touch but on vision.

A USB camera inside the robot captures shadows of hand gestures on the robot’s surface and classifies them with machine-learning software. They call this method ShadowSense, which they define as a modality between vision and touch, bringing “the high resolution and low cost of vision-sensing to the close-up sensory experience of touch.”

Touch-sensing in social or interactive robots is usually achieved with force sensors or capacitive sensors, says study co-author Guy Hoffman of the Sibley School of Mechanical and Aerospace Engineering at Cornell University. The drawback to his group’s approach has been that, even to achieve coarse spatial resolution, many sensors are needed in a small area.

However, working with non-rigid, inflatable robots, Hoffman and his co-researchers installed a consumer-grade USB camera to which they attached a fisheye lens for a wider field of vision.

“Given that the robot is already hollow, and has a soft and translucent skin, we could do touch interaction by looking at the shadows created by people touching the robot,” says Hoffman. They used deep neural networks to interpret the shadows. “And we were able to do it with very high accuracy,” he says. The robot was able to interpret six different gestures, including one- or two-handed touch, pointing, hugging and punching, with an accuracy of 87.5 to 96 percent, depending on the lighting.

This is not the first time that computer vision has been used for tactile sensing, though the scale and application of ShadowSense is unique. “Photography has been used for touch mainly in robotic grasping,” says Hoffman. By contrast, Hoffman and collaborators wanted to develop a sense that could be “felt” across the whole of the device.

The potential applications for ShadowSense include mobile robot guidance using touch, and interactive screens on soft robots. A third concerns privacy, especially in home-based social robots. “We have another paper currently under review that looks specifically at the ability to detect gestures that are further away [from the robot’s skin],” says Hoffman. This way, users would be able to cover their robot’s camera with a translucent material and still allow it to interpret actions and gestures from shadows. Thus, even though it’s prevented from capturing a high-resolution image of the user or their surrounding environment, using the right kind of training datasets, the robot can continue to monitor some kinds of non-tactile activities.

In its current iteration, Hoffman says, ShadowSense doesn’t do well in low-light conditions. Environmental noise, or shadows from surrounding objects, also interfere with image classification. Relying on one camera also means a single point of failure. “I think if this were to become a commercial product, we would probably [have to] work a little bit better on image detection,” says Hoffman.

As it was, the researchers used transfer learning—reusing a pre-trained deep-learning model in a new problem—for image analysis. “One of the problems with multi-layered neural networks is that you need a lot of training data to make accurate predictions,” says Hoffman. “Obviously, we don’t have millions of examples of people touching a hollow, inflatable robot. But we can use pre-trained networks trained on general images, which we have billions of, and we only retrain the last layers of the network using our own dataset.” Continue reading

Posted in Human Robots

#438801 This AI Thrashes the Hardest Atari Games ...

Learning from rewards seems like the simplest thing. I make coffee, I sip coffee, I’m happy. My brain registers “brewing coffee” as an action that leads to a reward.

That’s the guiding insight behind deep reinforcement learning, a family of algorithms that famously smashed most of Atari’s gaming catalog and triumphed over humans in strategy games like Go. Here, an AI “agent” explores the game, trying out different actions and registering ones that let it win.

Except it’s not that simple. “Brewing coffee” isn’t one action; it’s a series of actions spanning several minutes, where you’re only rewarded at the very end. By just tasting the final product, how do you learn to fine-tune grind coarseness, water to coffee ratio, brewing temperature, and a gazillion other factors that result in the reward—tasty, perk-me-up coffee?

That’s the problem with “sparse rewards,” which are ironically very abundant in our messy, complex world. We don’t immediately get feedback from our actions—no video-game-style dings or points for just grinding coffee beans—yet somehow we’re able to learn and perform an entire sequence of arm and hand movements while half-asleep.

This week, researchers from UberAI and OpenAI teamed up to bestow this talent on AI.

The trick is to encourage AI agents to “return” to a previous step, one that’s promising for a winning solution. The agent then keeps a record of that state, reloads it, and branches out again to intentionally explore other solutions that may have been left behind on the first go-around. Video gamers are likely familiar with this idea: live, die, reload a saved point, try something else, repeat for a perfect run-through.

The new family of algorithms, appropriately dubbed “Go-Explore,” smashed notoriously difficult Atari games like Montezuma’s Revenge that were previously unsolvable by its AI predecessors, while trouncing human performance along the way.

It’s not just games and digital fun. In a computer simulation of a robotic arm, the team found that installing Go-Explore as its “brain” allowed it to solve a challenging series of actions when given very sparse rewards. Because the overarching idea is so simple, the authors say, it can be adapted and expanded to other real-world problems, such as drug design or language learning.

Growing Pains
How do you reward an algorithm?

Rewards are very hard to craft, the authors say. Take the problem of asking a robot to go to a fridge. A sparse reward will only give the robot “happy points” if it reaches its destination, which is similar to asking a baby, with no concept of space and danger, to crawl through a potential minefield of toys and other obstacles towards a fridge.

“In practice, reinforcement learning works very well, if you have very rich feedback, if you can tell, ‘hey, this move is good, that move is bad, this move is good, that move is bad,’” said study author Joost Huinzinga. However, in situations that offer very little feedback, “rewards can intentionally lead to a dead end. Randomly exploring the space just doesn’t cut it.”

The other extreme is providing denser rewards. In the same robot-to-fridge example, you could frequently reward the bot as it goes along its journey, essentially helping “map out” the exact recipe to success. But that’s troubling as well. Over-holding an AI’s hand could result in an extremely rigid robot that ignores new additions to its path—a pet, for example—leading to dangerous situations. It’s a deceptive AI solution that seems effective in a simple environment, but crashes in the real world.

What we need are AI agents that can tackle both problems, the team said.

Intelligent Exploration
The key is to return to the past.

For AI, motivation usually comes from “exploring new or unusual situations,” said Huizinga. It’s efficient, but comes with significant downsides. For one, the AI agent could prematurely stop going back to promising areas because it thinks it had already found a good solution. For another, it could simply forget a previous decision point because of the mechanics of how it probes the next step in a problem.

For a complex task, the end result is an AI that randomly stumbles around towards a solution while ignoring potentially better ones.

“Detaching from a place that was previously visited after collecting a reward doesn’t work in difficult games, because you might leave out important clues,” Huinzinga explained.

Go-Explore solves these problems with a simple principle: first return, then explore. In essence, the algorithm saves different approaches it previously tried and loads promising save points—once more likely to lead to victory—to explore further.

Digging a bit deeper, the AI stores screen caps from a game. It then analyzes saved points and groups images that look alike as a potential promising “save point” to return to. Rinse and repeat. The AI tries to maximize its final score in the game, and updates its save points when it achieves a new record score. Because Atari doesn’t usually allow people to revisit any random point, the team used an emulator, which is a kind of software that mimics the Atari system but with custom abilities such as saving and reloading at any time.

The trick worked like magic. When pitted against 55 Atari games in the OpenAI gym, now commonly used to benchmark reinforcement learning algorithms, Go-Explore knocked out state-of-the-art AI competitors over 85 percent of the time.

It also crushed games previously unbeatable by AI. Montezuma’s Revenge, for example, requires you to move Pedro, the blocky protagonist, through a labyrinth of underground temples while evading obstacles such as traps and enemies and gathering jewels. One bad jump could derail the path to the next level. It’s a perfect example of sparse rewards: you need a series of good actions to get to the reward—advancing onward.

Go-Explore didn’t just beat all levels of the game, a first for AI. It also scored higher than any previous record for reinforcement learning algorithms at lower levels while toppling the human world record.

Outside a gaming environment, Go-Explore was also able to boost the performance of a simulated robot arm. While it’s easy for humans to follow high-level guidance like “put the cup on this shelf in a cupboard,” robots often need explicit training—from grasping the cup to recognizing a cupboard, moving towards it while avoiding obstacles, and learning motions to not smash the cup when putting it down.

Here, similar to the real world, the digital robot arm was only rewarded when it placed the cup onto the correct shelf, out of four possible shelves. When pitted against another algorithm, Go-Explore quickly figured out the movements needed to place the cup, while its competitor struggled with even reliably picking the cup up.

Combining Forces
By itself, the “first return, then explore” idea behind Go-Explore is already powerful. The team thinks it can do even better.

One idea is to change the mechanics of save points. Rather than reloading saved states through the emulator, it’s possible to train a neural network to do the same, without needing to relaunch a saved state. It’s a potential way to make the AI even smarter, the team said, because it can “learn” to overcome one obstacle once, instead of solving the same problem again and again. The downside? It’s much more computationally intensive.

Another idea is to combine Go-Explore with an alternative form of learning, called “imitation learning.” Here, an AI observes human behavior and mimics it through a series of actions. Combined with Go-Explore, said study author Adrien Ecoffet, this could make more robust robots capable of handling all the complexity and messiness in the real world.

To the team, the implications go far beyond Go-Explore. The concept of “first return, then explore” seems to be especially powerful, suggesting “it may be a fundamental feature of learning in general.” The team said, “Harnessing these insights…may be essential…to create generally intelligent agents.”

Image Credit: Adrien Ecoffet, Joost Huizinga, Joel Lehman, Kenneth O. Stanley, and Jeff Clune Continue reading

Posted in Human Robots