Tag Archives: looking

#432311 Everyone Is Talking About AI—But Do ...

In 2017, artificial intelligence attracted $12 billion of VC investment. We are only beginning to discover the usefulness of AI applications. Amazon recently unveiled a brick-and-mortar grocery store that has successfully supplanted cashiers and checkout lines with computer vision, sensors, and deep learning. Between the investment, the press coverage, and the dramatic innovation, “AI” has become a hot buzzword. But does it even exist yet?

At the World Economic Forum Dr. Kai-Fu Lee, a Taiwanese venture capitalist and the founding president of Google China, remarked, “I think it’s tempting for every entrepreneur to package his or her company as an AI company, and it’s tempting for every VC to want to say ‘I’m an AI investor.’” He then observed that some of these AI bubbles could burst by the end of 2018, referring specifically to “the startups that made up a story that isn’t fulfillable, and fooled VCs into investing because they don’t know better.”

However, Dr. Lee firmly believes AI will continue to progress and will take many jobs away from workers. So, what is the difference between legitimate AI, with all of its pros and cons, and a made-up story?

If you parse through just a few stories that are allegedly about AI, you’ll quickly discover significant variation in how people define it, with a blurred line between emulated intelligence and machine learning applications.

I spoke to experts in the field of AI to try to find consensus, but the very question opens up more questions. For instance, when is it important to be accurate to a term’s original definition, and when does that commitment to accuracy amount to the splitting of hairs? It isn’t obvious, and hype is oftentimes the enemy of nuance. Additionally, there is now a vested interest in that hype—$12 billion, to be precise.

This conversation is also relevant because world-renowned thought leaders have been publicly debating the dangers posed by AI. Facebook CEO Mark Zuckerberg suggested that naysayers who attempt to “drum up these doomsday scenarios” are being negative and irresponsible. On Twitter, business magnate and OpenAI co-founder Elon Musk countered that Zuckerberg’s understanding of the subject is limited. In February, Elon Musk engaged again in a similar exchange with Harvard professor Steven Pinker. Musk tweeted that Pinker doesn’t understand the difference between functional/narrow AI and general AI.

Given the fears surrounding this technology, it’s important for the public to clearly understand the distinctions between different levels of AI so that they can realistically assess the potential threats and benefits.

As Smart As a Human?
Erik Cambria, an expert in the field of natural language processing, told me, “Nobody is doing AI today and everybody is saying that they do AI because it’s a cool and sexy buzzword. It was the same with ‘big data’ a few years ago.”

Cambria mentioned that AI, as a term, originally referenced the emulation of human intelligence. “And there is nothing today that is even barely as intelligent as the most stupid human being on Earth. So, in a strict sense, no one is doing AI yet, for the simple fact that we don’t know how the human brain works,” he said.

He added that the term “AI” is often used in reference to powerful tools for data classification. These tools are impressive, but they’re on a totally different spectrum than human cognition. Additionally, Cambria has noticed people claiming that neural networks are part of the new wave of AI. This is bizarre to him because that technology already existed fifty years ago.

However, technologists no longer need to perform the feature extraction by themselves. They also have access to greater computing power. All of these advancements are welcomed, but it is perhaps dishonest to suggest that machines have emulated the intricacies of our cognitive processes.

“Companies are just looking at tricks to create a behavior that looks like intelligence but that is not real intelligence, it’s just a mirror of intelligence. These are expert systems that are maybe very good in a specific domain, but very stupid in other domains,” he said.

This mimicry of intelligence has inspired the public imagination. Domain-specific systems have delivered value in a wide range of industries. But those benefits have not lifted the cloud of confusion.

Assisted, Augmented, or Autonomous
When it comes to matters of scientific integrity, the issue of accurate definitions isn’t a peripheral matter. In a 1974 commencement address at the California Institute of Technology, Richard Feynman famously said, “The first principle is that you must not fool yourself—and you are the easiest person to fool.” In that same speech, Feynman also said, “You should not fool the layman when you’re talking as a scientist.” He opined that scientists should bend over backwards to show how they could be wrong. “If you’re representing yourself as a scientist, then you should explain to the layman what you’re doing—and if they don’t want to support you under those circumstances, then that’s their decision.”

In the case of AI, this might mean that professional scientists have an obligation to clearly state that they are developing extremely powerful, controversial, profitable, and even dangerous tools, which do not constitute intelligence in any familiar or comprehensive sense.

The term “AI” may have become overhyped and confused, but there are already some efforts underway to provide clarity. A recent PwC report drew a distinction between “assisted intelligence,” “augmented intelligence,” and “autonomous intelligence.” Assisted intelligence is demonstrated by the GPS navigation programs prevalent in cars today. Augmented intelligence “enables people and organizations to do things they couldn’t otherwise do.” And autonomous intelligence “establishes machines that act on their own,” such as autonomous vehicles.

Roman Yampolskiy is an AI safety researcher who wrote the book “Artificial Superintelligence: A Futuristic Approach.” I asked him whether the broad and differing meanings might present difficulties for legislators attempting to regulate AI.

Yampolskiy explained, “Intelligence (artificial or natural) comes on a continuum and so do potential problems with such technology. We typically refer to AI which one day will have the full spectrum of human capabilities as artificial general intelligence (AGI) to avoid some confusion. Beyond that point it becomes superintelligence. What we have today and what is frequently used in business is narrow AI. Regulating anything is hard, technology is no exception. The problem is not with terminology but with complexity of such systems even at the current level.”

When asked if people should fear AI systems, Dr. Yampolskiy commented, “Since capability comes on a continuum, so do problems associated with each level of capability.” He mentioned that accidents are already reported with AI-enabled products, and as the technology advances further, the impact could spread beyond privacy concerns or technological unemployment. These concerns about the real-world effects of AI will likely take precedence over dictionary-minded quibbles. However, the issue is also about honesty versus deception.

Is This Buzzword All Buzzed Out?
Finally, I directed my questions towards a company that is actively marketing an “AI Virtual Assistant.” Carl Landers, the CMO at Conversica, acknowledged that there are a multitude of explanations for what AI is and isn’t.

He said, “My definition of AI is technology innovation that helps solve a business problem. I’m really not interested in talking about the theoretical ‘can we get machines to think like humans?’ It’s a nice conversation, but I’m trying to solve a practical business problem.”

I asked him if AI is a buzzword that inspires publicity and attracts clients. According to Landers, this was certainly true three years ago, but those effects have already started to wane. Many companies now claim to have AI in their products, so it’s less of a differentiator. However, there is still a specific intention behind the word. Landers hopes to convey that previously impossible things are now possible. “There’s something new here that you haven’t seen before, that you haven’t heard of before,” he said.

According to Brian Decker, founder of Encom Lab, machine learning algorithms only work to satisfy their preexisting programming, not out of an interior drive for better understanding. Therefore, he views AI as an entirely semantic argument.

Decker stated, “A marketing exec will claim a photodiode controlled porch light has AI because it ‘knows when it is dark outside,’ while a good hardware engineer will point out that not one bit in a register in the entire history of computing has ever changed unless directed to do so according to the logic of preexisting programming.”

Although it’s important for everyone to be on the same page regarding specifics and underlying meaning, AI-powered products are already powering past these debates by creating immediate value for humans. And ultimately, humans care more about value than they do about semantic distinctions. In an interview with Quartz, Kai-Fu Lee revealed that algorithmic trading systems have already given him an 8X return over his private banking investments. “I don’t trade with humans anymore,” he said.

Image Credit: vrender / Shutterstock.com Continue reading

Posted in Human Robots

#432051 What Roboticists Are Learning From Early ...

You might not have heard of Hanson Robotics, but if you’re reading this, you’ve probably seen their work. They were the company behind Sophia, the lifelike humanoid avatar that’s made dozens of high-profile media appearances. Before that, they were the company behind that strange-looking robot that seemed a bit like Asimo with Albert Einstein’s head—or maybe you saw BINA48, who was interviewed for the New York Times in 2010 and featured in Jon Ronson’s books. For the sci-fi aficionados amongst you, they even made a replica of legendary author Philip K. Dick, best remembered for having books with titles like Do Androids Dream of Electric Sheep? turned into films with titles like Blade Runner.

Hanson Robotics, in other words, with their proprietary brand of life-like humanoid robots, have been playing the same game for a while. Sometimes it can be a frustrating game to watch. Anyone who gives the robot the slightest bit of thought will realize that this is essentially a chat-bot, with all the limitations this implies. Indeed, even in that New York Times interview with BINA48, author Amy Harmon describes it as a frustrating experience—with “rare (but invariably thrilling) moments of coherence.” This sensation will be familiar to anyone who’s conversed with a chatbot that has a few clever responses.

The glossy surface belies the lack of real intelligence underneath; it seems, at first glance, like a much more advanced machine than it is. Peeling back that surface layer—at least for a Hanson robot—means you’re peeling back Frubber. This proprietary substance—short for “Flesh Rubber,” which is slightly nightmarish—is surprisingly complicated. Up to thirty motors are required just to control the face; they manipulate liquid cells in order to make the skin soft, malleable, and capable of a range of different emotional expressions.

A quick combinatorial glance at the 30+ motors suggests that there are millions of possible combinations; researchers identify 62 that they consider “human-like” in Sophia, although not everyone agrees with this assessment. Arguably, the technical expertise that went into reconstructing the range of human facial expressions far exceeds the more simplistic chat engine the robots use, although it’s the second one that allows it to inflate the punters’ expectations with a few pre-programmed questions in an interview.

Hanson Robotics’ belief is that, ultimately, a lot of how humans will eventually relate to robots is going to depend on their faces and voices, as well as on what they’re saying. “The perception of identity is so intimately bound up with the perception of the human form,” says David Hanson, company founder.

Yet anyone attempting to design a robot that won’t terrify people has to contend with the uncanny valley—that strange blend of concern and revulsion people react with when things appear to be creepily human. Between cartoonish humanoids and genuine humans lies what has often been a no-go zone in robotic aesthetics.

The uncanny valley concept originated with roboticist Masahiro Mori, who argued that roboticists should avoid trying to replicate humans exactly. Since anything that wasn’t perfect, but merely very good, would elicit an eerie feeling in humans, shirking the challenge entirely was the only way to avoid the uncanny valley. It’s probably a task made more difficult by endless streams of articles about AI taking over the world that inexplicably conflate AI with killer humanoid Terminators—which aren’t particularly likely to exist (although maybe it’s best not to push robots around too much).

The idea behind this realm of psychological horror is fairly simple, cognitively speaking.

We know how to categorize things that are unambiguously human or non-human. This is true even if they’re designed to interact with us. Consider the popularity of Aibo, Jibo, or even some robots that don’t try to resemble humans. Something that resembles a human, but isn’t quite right, is bound to evoke a fear response in the same way slightly distorted music or slightly rearranged furniture in your home will. The creature simply doesn’t fit.

You may well reject the idea of the uncanny valley entirely. David Hanson, naturally, is not a fan. In the paper Upending the Uncanny Valley, he argues that great art forms have often resembled humans, but the ultimate goal for humanoid roboticists is probably to create robots we can relate to as something closer to humans than works of art.

Meanwhile, Hanson and other scientists produce competing experiments to either demonstrate that the uncanny valley is overhyped, or to confirm it exists and probe its edges.

The classic experiment involves gradually morphing a cartoon face into a human face, via some robotic-seeming intermediaries—yet it’s in movement that the real horror of the almost-human often lies. Hanson has argued that incorporating cartoonish features may help—and, sometimes, that the uncanny valley is a generational thing which will melt away when new generations grow used to the quirks of robots. Although Hanson might dispute the severity of this effect, it’s clearly what he’s trying to avoid with each new iteration.

Hiroshi Ishiguro is the latest of the roboticists to have dived headlong into the valley.

Building on the work of pioneers like Hanson, those who study human-robot interaction are pushing at the boundaries of robotics—but also of social science. It’s usually difficult to simulate what you don’t understand, and there’s still an awful lot we don’t understand about how we interpret the constant streams of non-verbal information that flow when you interact with people in the flesh.

Ishiguro took this imitation of human forms to extreme levels. Not only did he monitor and log the physical movements people made on videotapes, but some of his robots are based on replicas of people; the Repliee series began with a ‘replicant’ of his daughter. This involved making a rubber replica—a silicone cast—of her entire body. Future experiments were focused on creating Geminoid, a replica of Ishiguro himself.

As Ishiguro aged, he realized that it would be more effective to resemble his replica through cosmetic surgery rather than by continually creating new casts of his face, each with more lines than the last. “I decided not to get old anymore,” Ishiguro said.

We love to throw around abstract concepts and ideas: humans being replaced by machines, cared for by machines, getting intimate with machines, or even merging themselves with machines. You can take an idea like that, hold it in your hand, and examine it—dispassionately, if not without interest. But there’s a gulf between thinking about it and living in a world where human-robot interaction is not a field of academic research, but a day-to-day reality.

As the scientists studying human-robot interaction develop their robots, their replicas, and their experiments, they are making some of the first forays into that world. We might all be living there someday. Understanding ourselves—decrypting the origins of empathy and love—may be the greatest challenge to face. That is, if you want to avoid the valley.

Image Credit: Anton Gvozdikov / Shutterstock.com Continue reading

Posted in Human Robots

#432027 We Read This 800-Page Report on the ...

The longevity field is bustling but still fragmented, and the “silver tsunami” is coming.

That is the takeaway of The Science of Longevity, the behemoth first volume of a four-part series offering a bird’s-eye view of the longevity industry in 2017. The report, a joint production of the Biogerontology Research Foundation, Deep Knowledge Life Science, Aging Analytics Agency, and Longevity.International, synthesizes the growing array of academic and industry ventures related to aging, healthspan, and everything in between.

This is huge, not only in scale but also in ambition. The report, totally worth a read here, will be followed by four additional volumes in 2018, covering topics ranging from the business side of longevity ventures to financial systems to potential tensions between life extension and religion.

And that’s just the first step. The team hopes to publish updated versions of the report annually, giving scientists, investors, and regulatory agencies an easy way to keep their finger on the longevity pulse.

“In 2018, ‘aging’ remains an unnamed adversary in an undeclared war. For all intents and purposes it is mere abstraction in the eyes of regulatory authorities worldwide,” the authors write.

That needs to change.

People often arrive at the field of aging from disparate areas with wildly diverse opinions and strengths. The report compiles these individual efforts at cracking aging into a systematic resource—a “periodic table” for longevity that clearly lays out emerging trends and promising interventions.

The ultimate goal? A global framework serving as a road map to guide the burgeoning industry. With such a framework in hand, academics and industry alike are finally poised to petition the kind of large-scale investments and regulatory changes needed to tackle aging with a unified front.

Infographic depicting many of the key research hubs and non-profits within the field of geroscience.
Image Credit: Longevity.International
The Aging Globe
The global population is rapidly aging. And our medical and social systems aren’t ready to handle this oncoming “silver tsunami.”

Take the medical field. Many age-related diseases such as Alzheimer’s lack effective treatment options. Others, including high blood pressure, stroke, lung or heart problems, require continuous medication and monitoring, placing enormous strain on medical resources.

What’s more, because disease risk rises exponentially with age, medical care for the elderly becomes a game of whack-a-mole: curing any individual disease such as cancer only increases healthy lifespan by two to three years before another one hits.

That’s why in recent years there’s been increasing support for turning the focus to the root of the problem: aging. Rather than tackling individual diseases, geroscience aims to add healthy years to our lifespan—extending “healthspan,” so to speak.

Despite this relative consensus, the field still faces a roadblock. The US FDA does not yet recognize aging as a bona fide disease. Without such a designation, scientists are banned from testing potential interventions for aging in clinical trials (that said, many have used alternate measures such as age-related biomarkers or Alzheimer’s symptoms as a proxy).

Luckily, the FDA’s stance is set to change. The promising anti-aging drug metformin, for example, is already in clinical trials, examining its effect on a variety of age-related symptoms and diseases. This report, and others to follow, may help push progress along.

“It is critical for investors, policymakers, scientists, NGOs, and influential entities to prioritize the amelioration of the geriatric world scenario and recognize aging as a critical matter of global economic security,” the authors say.

Biomedical Gerontology
The causes of aging are complex, stubborn, and not all clear.

But the report lays out two main streams of intervention with already promising results.

The first is to understand the root causes of aging and stop them before damage accumulates. It’s like meddling with cogs and other inner workings of a clock to slow it down, the authors say.

The report lays out several treatments to keep an eye on.

Geroprotective drugs is a big one. Often repurposed from drugs already on the market, these traditional small molecule drugs target a wide variety of metabolic pathways that play a role in aging. Think anti-oxidants, anti-inflammatory, and drugs that mimic caloric restriction, a proven way to extend healthspan in animal models.

More exciting are the emerging technologies. One is nanotechnology. Nanoparticles of carbon, “bucky-balls,” for example, have already been shown to fight viral infections and dangerous ion particles, as well as stimulate the immune system and extend lifespan in mice (though others question the validity of the results).

Blood is another promising, if surprising, fountain of youth: recent studies found that molecules in the blood of the young rejuvenate the heart, brain, and muscles of aged rodents, though many of these findings have yet to be replicated.

Rejuvenation Biotechnology
The second approach is repair and maintenance.

Rather than meddling with inner clockwork, here we force back the hands of a clock to set it back. The main example? Stem cell therapy.

This type of approach would especially benefit the brain, which harbors small, scattered numbers of stem cells that deplete with age. For neurodegenerative diseases like Alzheimer’s, in which neurons progressively die off, stem cell therapy could in theory replace those lost cells and mend those broken circuits.

Once a blue-sky idea, the discovery of induced pluripotent stem cells (iPSCs), where scientists can turn skin and other mature cells back into a stem-like state, hugely propelled the field into near reality. But to date, stem cells haven’t been widely adopted in clinics.

It’s “a toolkit of highly innovative, highly invasive technologies with clinical trials still a great many years off,” the authors say.

But there is a silver lining. The boom in 3D tissue printing offers an alternative approach to stem cells in replacing aging organs. Recent investment from the Methuselah Foundation and other institutions suggests interest remains high despite still being a ways from mainstream use.

A Disruptive Future
“We are finally beginning to see an industry emerge from mankind’s attempts to make sense of the biological chaos,” the authors conclude.

Looking through the trends, they identified several technologies rapidly gaining steam.

One is artificial intelligence, which is already used to bolster drug discovery. Machine learning may also help identify new longevity genes or bring personalized medicine to the clinic based on a patient’s records or biomarkers.

Another is senolytics, a class of drugs that kill off “zombie cells.” Over 10 prospective candidates are already in the pipeline, with some expected to enter the market in less than a decade, the authors say.

Finally, there’s the big gun—gene therapy. The treatment, unlike others mentioned, can directly target the root of any pathology. With a snip (or a swap), genetic tools can turn off damaging genes or switch on ones that promote a youthful profile. It is the most preventative technology at our disposal.

There have already been some success stories in animal models. Using gene therapy, rodents given a boost in telomerase activity, which lengthens the protective caps of DNA strands, live healthier for longer.

“Although it is the prospect farthest from widespread implementation, it may ultimately prove the most influential,” the authors say.

Ultimately, can we stop the silver tsunami before it strikes?

Perhaps not, the authors say. But we do have defenses: the technologies outlined in the report, though still immature, could one day stop the oncoming tidal wave in its tracks.

Now we just have to bring them out of the lab and into the real world. To push the transition along, the team launched Longevity.International, an online meeting ground that unites various stakeholders in the industry.

By providing scientists, entrepreneurs, investors, and policy-makers a platform for learning and discussion, the authors say, we may finally generate enough drive to implement our defenses against aging. The war has begun.

Read the report in full here, and watch out for others coming soon here. The second part of the report profiles 650 (!!!) longevity-focused research hubs, non-profits, scientists, conferences, and literature. It’s an enormously helpful resource—totally worth keeping it in your back pocket for future reference.

Image Credit: Worraket / Shutterstock.com Continue reading

Posted in Human Robots

#432009 How Swarm Intelligence Is Making Simple ...

As a group, simple creatures following simple rules can display a surprising amount of complexity, efficiency, and even creativity. Known as swarm intelligence, this trait is found throughout nature, but researchers have recently begun using it to transform various fields such as robotics, data mining, medicine, and blockchains.

Ants, for example, can only perform a limited range of functions, but an ant colony can build bridges, create superhighways of food and information, wage war, and enslave other ant species—all of which are beyond the comprehension of any single ant. Likewise, schools of fish, flocks of birds, beehives, and other species exhibit behavior indicative of planning by a higher intelligence that doesn’t actually exist.

It happens by a process called stigmergy. Simply put, a small change by a group member causes other members to behave differently, leading to a new pattern of behavior.

When an ant finds a food source, it marks the path with pheromones. This attracts other ants to that path, leads them to the food source, and prompts them to mark the same path with more pheromones. Over time, the most efficient route will become the superhighway, as the faster and easier a path is, the more ants will reach the food and the more pheromones will be on the path. Thus, it looks as if a more intelligent being chose the best path, but it emerged from the tiny, simple changes made by individuals.

So what does this mean for humans? Well, a lot. In the past few decades, researchers have developed numerous algorithms and metaheuristics, such as ant colony optimization and particle swarm optimization, and they are rapidly being adopted.

Swarm Robotics
A swarm of robots would work on the same principles as an ant colony: each member has a simple set of rules to follow, leading to self-organization and self-sufficiency.

For example, researchers at Georgia Robotics and InTelligent Systems (GRITS) created a small swarm of simple robots that can spell and play piano. The robots cannot communicate, but based solely on the position of surrounding robots, they are able to use their specially-created algorithm to determine the optimal path to complete their task.

This is also immensely useful for drone swarms.

Last February, Ehang, an aviation company out of China, created a swarm of a thousand drones that not only lit the sky with colorful, intricate displays, but demonstrated the ability to improvise and troubleshoot errors entirely autonomously.

Further, just recently, the University of Cambridge and Koc University unveiled their idea for what they call the Energy Neutral Internet of Drones. Amazingly, this drone swarm would take initiative to share information or energy with other drones that did not receive a communication or are running low on energy.

Militaries all of the world are utilizing this as well.

Last year, the US Department of Defense announced it had successfully tested a swarm of miniature drones that could carry out complex missions cheaper and more efficiently. They claimed, “The micro-drones demonstrated advanced swarm behaviors such as collective decision-making, adaptive formation flying, and self-healing.”

Some experts estimate at least 30 nations are actively developing drone swarms—and even submersible drones—for military missions, including intelligence gathering, missile defense, precision missile strikes, and enhanced communication.

NASA also plans on deploying swarms of tiny spacecraft for space exploration, and the medical community is looking into using swarms of nanobots for precision delivery of drugs, microsurgery, targeting toxins, and biological sensors.

What If Humans Are the Ants?
The strength of any blockchain comes from the size and diversity of the community supporting it. Cryptocurrencies like Bitcoin, Ethereum, and Litecoin are driven by the people using, investing in, and, most importantly, mining them so their blockchains can function. Without an active community, or swarm, their blockchains wither away.

When viewed from a great height, a blockchain performs eerily like an ant colony in that it will naturally find the most efficient way to move vast amounts of information.

Miners compete with each other to perform the complex calculations necessary to add another block, for which the winner is rewarded with the blockchain’s native currency and agreed-upon fees. Of course, the miner with the more powerful computers is more likely to win the reward, thereby empowering the winner’s ability to mine and receive even more rewards. Over time, fewer and fewer miners are going to exist, as the winners are able to more efficiently shoulder more of the workload, in much the same way that ants build superhighways.

Further, a company called Unanimous AI has developed algorithms that allow humans to collectively make predictions. So far, the AI algorithms and their human participants have made some astoundingly accurate predictions, such as the first four winning horses of the Kentucky Derby, the Oscar winners, the Stanley Cup winners, and others. The more people involved in the swarm, the greater their predictive power will be.

To be clear, this is not a prediction based on group consensus. Rather, the swarm of humans uses software to input their opinions in real time, thus making micro-changes to the rest of the swarm and the inputs of other members.

Studies show that swarm intelligence consistently outperforms individuals and crowds working without the algorithms. While this is only the tip of the iceberg, some have suggested swarm intelligence can revolutionize how doctors diagnose a patient or how products are marketed to consumers. It might even be an essential step in truly creating AI.

While swarm intelligence is an essential part of many species’ success, it’s only a matter of time before humans harness its effectiveness as well.

Image Credit: Nature Bird Photography / Shutterstock.com Continue reading

Posted in Human Robots

#431958 The Next Generation of Cameras Might See ...

You might be really pleased with the camera technology in your latest smartphone, which can recognize your face and take slow-mo video in ultra-high definition. But these technological feats are just the start of a larger revolution that is underway.

The latest camera research is shifting away from increasing the number of mega-pixels towards fusing camera data with computational processing. By that, we don’t mean the Photoshop style of processing where effects and filters are added to a picture, but rather a radical new approach where the incoming data may not actually look like at an image at all. It only becomes an image after a series of computational steps that often involve complex mathematics and modeling how light travels through the scene or the camera.

This additional layer of computational processing magically frees us from the chains of conventional imaging techniques. One day we may not even need cameras in the conventional sense any more. Instead we will use light detectors that only a few years ago we would never have considered any use for imaging. And they will be able to do incredible things, like see through fog, inside the human body and even behind walls.

Single Pixel Cameras
One extreme example is the single pixel camera, which relies on a beautifully simple principle. Typical cameras use lots of pixels (tiny sensor elements) to capture a scene that is likely illuminated by a single light source. But you can also do things the other way around, capturing information from many light sources with a single pixel.

To do this you need a controlled light source, for example a simple data projector that illuminates the scene one spot at a time or with a series of different patterns. For each illumination spot or pattern, you then measure the amount of light reflected and add everything together to create the final image.

Clearly the disadvantage of taking a photo in this is way is that you have to send out lots of illumination spots or patterns in order to produce one image (which would take just one snapshot with a regular camera). But this form of imaging would allow you to create otherwise impossible cameras, for example that work at wavelengths of light beyond the visible spectrum, where good detectors cannot be made into cameras.

These cameras could be used to take photos through fog or thick falling snow. Or they could mimic the eyes of some animals and automatically increase an image’s resolution (the amount of detail it captures) depending on what’s in the scene.

It is even possible to capture images from light particles that have never even interacted with the object we want to photograph. This would take advantage of the idea of “quantum entanglement,” that two particles can be connected in a way that means whatever happens to one happens to the other, even if they are a long distance apart. This has intriguing possibilities for looking at objects whose properties might change when lit up, such as the eye. For example, does a retina look the same when in darkness as in light?

Multi-Sensor Imaging
Single-pixel imaging is just one of the simplest innovations in upcoming camera technology and relies, on the face of it, on the traditional concept of what forms a picture. But we are currently witnessing a surge of interest for systems that use lots of information but traditional techniques only collect a small part of it.

This is where we could use multi-sensor approaches that involve many different detectors pointed at the same scene. The Hubble telescope was a pioneering example of this, producing pictures made from combinations of many different images taken at different wavelengths. But now you can buy commercial versions of this kind of technology, such as the Lytro camera that collects information about light intensity and direction on the same sensor, to produce images that can be refocused after the image has been taken.

The next generation camera will probably look something like the Light L16 camera, which features ground-breaking technology based on more than ten different sensors. Their data are combined using a computer to provide a 50 MB, re-focusable and re-zoomable, professional-quality image. The camera itself looks like a very exciting Picasso interpretation of a crazy cell-phone camera.

Yet these are just the first steps towards a new generation of cameras that will change the way in which we think of and take images. Researchers are also working hard on the problem of seeing through fog, seeing behind walls, and even imaging deep inside the human body and brain.

All of these techniques rely on combining images with models that explain how light travels through through or around different substances.

Another interesting approach that is gaining ground relies on artificial intelligence to “learn” to recognize objects from the data. These techniques are inspired by learning processes in the human brain and are likely to play a major role in future imaging systems.

Single photon and quantum imaging technologies are also maturing to the point that they can take pictures with incredibly low light levels and videos with incredibly fast speeds reaching a trillion frames per second. This is enough to even capture images of light itself traveling across as scene.

Some of these applications might require a little time to fully develop, but we now know that the underlying physics should allow us to solve these and other problems through a clever combination of new technology and computational ingenuity.

This article was originally published on The Conversation. Read the original article.

Image Credit: Sylvia Adams / Shutterstock.com Continue reading

Posted in Human Robots