Tag Archives: Would
#433386 What We Have to Gain From Making ...
The borders between the real world and the digital world keep crumbling, and the latter’s importance in both our personal and professional lives keeps growing. Some describe the melding of virtual and real worlds as part of the fourth industrial revolution. Said revolution’s full impact on us as individuals, our companies, communities, and societies is still unknown.
Greg Cross, chief business officer of New Zealand-based AI company Soul Machines, thinks one inescapable consequence of these crumbling borders is people spending more and more time interacting with technology. In a presentation at Singularity University’s Global Summit in San Francisco last month, Cross unveiled Soul Machines’ latest work and shared his views on the current state of human-like AI and where the technology may go in the near future.
Humanizing Technology Interaction
Cross started by introducing Rachel, one of Soul Machines’ “emotionally responsive digital humans.” The company has built 15 different digital humans of various sexes, groups, and ethnicities. Rachel, along with her “sisters” and “brothers,” has a virtual nervous system based on neural networks and biological models of different paths in the human brain. The system is controlled by virtual neurotransmitters and hormones akin to dopamine, serotonin, and oxytocin, which influence learning and behavior.
As a result, each digital human can have its own unique set of “feelings” and responses to interactions. People interact with them via visual and audio sensors, and the machines respond in real time.
“Over the last 20 or 30 years, the way we think about machines and the way we interact with machines has changed,” Cross said. “We’ve always had this view that they should actually be more human-like.”
The realism of the digital humans’ graphic representations comes thanks to the work of Soul Machines’ other co-founder, Dr. Mark Sager, who has won two Academy Awards for his work on some computer-generated movies, including James Cameron’s Avatar.
Cross pointed out, for example, that rather than being unrealistically flawless and clear, Rachel’s skin has blemishes and sun spots, just like real human skin would.
The Next Human-Machine Frontier
When people interact with each other face to face, emotional and intellectual engagement both heavily influence the interaction. What would it look like for machines to bring those same emotional and intellectual capacities to our interactions with them, and how would this type of interaction affect the way we use, relate to, and feel about AI?
Cross and his colleagues believe that humanizing artificial intelligence will make the technology more useful to humanity, and prompt people to use AI in more beneficial ways.
“What we think is a very important view as we move forward is that these machines can be more helpful to us. They can be more useful to us. They can be more interesting to us if they’re actually more like us,” Cross said.
It is an approach that seems to resonate with companies and organizations. For example, in the UK, where NatWest Bank is testing out Cora as a digital employee to help answer customer queries. In Germany, Daimler Financial Group plans to employ Sarah as something “similar to a personal concierge” for its customers. According to Cross, Daimler is looking at other ways it could deploy digital humans across the organization, from building digital service people, digital sales people, and maybe in the future, digital chauffeurs.
Soul Machines’ latest creation is Will, a digital teacher that can interact with children through a desktop, tablet, or mobile device and help them learn about renewable energy. Cross sees other social uses for digital humans, including potentially serving as doctors to rural communities.
Our Digital Friends—and Twins
Soul Machines is not alone in its quest to humanize technology. It is a direction many technology companies, including the likes of Amazon, also seem to be pursuing. Amazon is working on building a home robot that, according to Bloomberg, “could be a sort of mobile Alexa.”
Finding a more human form for technology seems like a particularly pervasive pursuit in Japan. Not just when it comes to its many, many robots, but also virtual assistants like Gatebox.
The Japanese approach was perhaps best summed up by famous android researcher Dr. Hiroshi Ishiguro, who I interviewed last year: “The human brain is set up to recognize and interact with humans. So, it makes sense to focus on developing the body for the AI mind, as well as the AI. I believe that the final goal for both Japanese and other companies and scientists is to create human-like interaction.”
During Cross’s presentation, Rob Nail, CEO and associate founder of Singularity University, joined him on the stage, extending an invitation to Rachel to be SU’s first fully digital faculty member. Rachel accepted, and though she’s the only digital faculty right now, she predicted this won’t be the case for long.
“In 10 years, all of you will have digital versions of yourself, just like me, to take on specific tasks and make your life a whole lot easier,” she said. “This is great news for me. I’ll have millions of digital friends.”
Image Credit: Soul Machines Continue reading
#433301 ‘Happiness Tech’ Is On the Rise. Is ...
We often get so fixated on technological progress that we forget it’s merely one component of the entirety of human progress. Technological advancement does not necessarily correlate with increases in human mental well-being.
While cleaner energy, access to education, and higher employment rates can improve quality of life, they do not guarantee happiness and inner peace. Amid what appears to be an increasing abundance of resources and ongoing human progress, we are experiencing a mental health epidemic, with high anxiety and depression rates. This is especially true in the developed world, where we have access to luxuries our ancestors couldn’t even dream of—all the world’s information contained in a device we hold in the palm of our hands, for example.
But as you may have realized through your own experience, technology can make us feel worse instead of better. Social media can become a tool for comparison and a source of debilitating status anxiety. Increased access to goods and services, along with the rise of consumerism, can lead people to choose “stuff” over true sources of meaning and get trapped in a hedonistic treadmill of materialism. Tools like artificial intelligence and big data could lead to violation of our privacy and autonomy. The digital world can take us away from the beauty of the present moment.
Understanding Happiness
How we use technology can significantly impact our happiness. In this context, “happiness” refers to a general sense of well-being, gratitude, and inner peace. Even with such a simple definition, it is a state of mind many people will admit they lack.
Eastern philosophies have told us for thousands of years that the problem of human suffering begins with our thoughts and perceptions of the circumstances we are in, as opposed to beginning with the circumstances themselves. As Derren Brown brilliantly points out in Happy: Why More or Less Everything Is Absolutely Fine, “The problem with the modern conception of happiness is that it is seen as some kind of commodity. There is this fantasy that simply by believing in yourself and setting goals you can have anything. But that simply isn’t how life works. The ancients had a much better view of it. They offered an approach of not trying to control things you can’t control, and of lessening your desires and your expectations so you achieve a harmony between what you desire and what you have.”
A core part of feeling more happy is about re-wiring our minds to adjust our expectations, exercise gratitude, escape negative narratives, and live in the present moment.
But can technology help us do that?
Applications for Mental Well-Being
Many doers are asking themselves how they can leverage digital tools to contribute to human happiness.
Meditation and mindfulness are examples of practices we can use to escape the often overwhelming burden of our thoughts and ground our minds into the present. They have become increasingly democratized with the rise of meditation mobile apps, such as Headspace, Gaia, and Calm, that allow millions of people globally to use their phones to learn from experts at a very low cost.
These companies have also partnered with hospitals, airlines, athletic teams, and others that could benefit from increased access to mindfulness and meditation. The popularity of these apps continues to rise as more people recognize their necessity. The combination of mass technology and ancient wisdom is one that can lead to a transformation of the collective consciousness.
Sometimes merely reflecting on the sources of joy in our lives and practicing gratitude can contribute to better well-being. Apps such as Happier encourage users to reflect upon and share pleasant everyday moments in their daily lives. Such exercises are based on the understanding that being happy is a “skill” one can build though practice and through scientifically-proven activities, such as writing down a nice thought and sharing your positivity with the world. Many other tools such as Track Your Happiness and Happstr allow users to track their happiness, which often serves as a valuable source of data to researchers.
There is also a growing body of knowledge that tells us we can achieve happiness by helping others. This “helper’s high” is a result of our brains producing endorphins after having a positive impact on the lives of others. In many shapes and forms, technology has made it easier now more than ever to help other people no matter where they are located. From charitable donations to the rise of social impact organizations, there is an abundance of projects that leverage technology to positively impact individual lives. Platforms like GoVolunteer connect nonprofits with individuals from a variety of skill sets who are looking to gift their abilities to those in need. Kiva allows for fundraising loans that can change lives. These are just a handful of examples of a much wider positive paradigm shift.
The Future of Technology for Well-Being
There is no denying that increasingly powerful and immersive technology can be used to better or worsen the human condition. Today’s leaders will not only have to focus on their ability to use technology to solve a problem or generate greater revenue; they will have to ask themselves if their tech solutions are beneficial or detrimental to human well-being. They will also have to remember that more powerful technology does not always translate to happier users. It is also crucial that future generations be equipped with the values required to use increasingly powerful tools responsibly and ethically.
In the Education 2030 report, the Millennium Project envisions a world wherein portable intelligent devices combined with integrated systems for lifelong learning contribute to better well-being. In this vision, “continuous evaluation of individual learning processes designed to prevent people from growing unstable and/or becoming mentally ill, along with programs aimed at eliminating prejudice and hate, could bring about a more beautiful, loving world.”
There is exciting potential for technology to be leveraged to contribute to human happiness at a massive scale. Yet, technology shouldn’t consume every aspect of our lives, since a life worth living is often about balance. Sometimes, even if just for a few moments, what would make us feel happier is we disconnected from technology to begin with.
Image Credit: 13_Phunkod / Shutterstock.com Continue reading
#433288 The New AI Tech Turning Heads in Video ...
A new technique using artificial intelligence to manipulate video content gives new meaning to the expression “talking head.”
An international team of researchers showcased the latest advancement in synthesizing facial expressions—including mouth, eyes, eyebrows, and even head position—in video at this month’s 2018 SIGGRAPH, a conference on innovations in computer graphics, animation, virtual reality, and other forms of digital wizardry.
The project is called Deep Video Portraits. It relies on a type of AI called generative adversarial networks (GANs) to modify a “target” actor based on the facial and head movement of a “source” actor. As the name implies, GANs pit two opposing neural networks against one another to create a realistic talking head, right down to the sneer or raised eyebrow.
In this case, the adversaries are actually working together: One neural network generates content, while the other rejects or approves each effort. The back-and-forth interplay between the two eventually produces a realistic result that can easily fool the human eye, including reproducing a static scene behind the head as it bobs back and forth.
The researchers say the technique can be used by the film industry for a variety of purposes, from editing facial expressions of actors for matching dubbed voices to repositioning an actor’s head in post-production. AI can not only produce highly realistic results, but much quicker ones compared to the manual processes used today, according to the researchers. You can read the full paper of their work here.
“Deep Video Portraits shows how such a visual effect could be created with less effort in the future,” said Christian Richardt, from the University of Bath’s motion capture research center CAMERA, in a press release. “With our approach, even the positioning of an actor’s head and their facial expression could be easily edited to change camera angles or subtly change the framing of a scene to tell the story better.”
AI Tech Different Than So-Called “Deepfakes”
The work is far from the first to employ AI to manipulate video and audio. At last year’s SIGGRAPH conference, researchers from the University of Washington showcased their work using algorithms that inserted audio recordings from a person in one instance into a separate video of the same person in a different context.
In this case, they “faked” a video using a speech from former President Barack Obama addressing a mass shooting incident during his presidency. The AI-doctored video injects the audio into an unrelated video of the president while also blending the facial and mouth movements, creating a pretty credible job of lip synching.
A previous paper by many of the same scientists on the Deep Video Portraits project detailed how they were first able to manipulate a video in real time of a talking head (in this case, actor and former California governor Arnold Schwarzenegger). The Face2Face system pulled off this bit of digital trickery using a depth-sensing camera that tracked the facial expressions of an Asian female source actor.
A less sophisticated method of swapping faces using a machine learning software dubbed FakeApp emerged earlier this year. Predictably, the tech—requiring numerous photos of the source actor in order to train the neural network—was used for more juvenile pursuits, such as injecting a person’s face onto a porn star.
The application gave rise to the term “deepfakes,” which is now used somewhat ubiquitously to describe all such instances of AI-manipulated video—much to the chagrin of some of the researchers involved in more legitimate uses.
Fighting AI-Created Video Forgeries
However, the researchers are keenly aware that their work—intended for benign uses such as in the film industry or even to correct gaze and head positions for more natural interactions through video teleconferencing—could be used for nefarious purposes. Fake news is the most obvious concern.
“With ever-improving video editing technology, we must also start being more critical about the video content we consume every day, especially if there is no proof of origin,” said Michael Zollhöfer, a visiting assistant professor at Stanford University and member of the Deep Video Portraits team, in the press release.
Toward that end, the research team is training the same adversarial neural networks to spot video forgeries. They also strongly recommend that developers clearly watermark videos that are edited through AI or otherwise, and denote clearly what part and element of the scene was modified.
To catch less ethical users, the US Department of Defense, through the Defense Advanced Research Projects Agency (DARPA), is supporting a program called Media Forensics. This latest DARPA challenge enlists researchers to develop technologies to automatically assess the integrity of an image or video, as part of an end-to-end media forensics platform.
The DARPA official in charge of the program, Matthew Turek, did tell MIT Technology Review that so far the program has “discovered subtle cues in current GAN-manipulated images and videos that allow us to detect the presence of alterations.” In one reported example, researchers have targeted eyes, which rarely blink in the case of “deepfakes” like those created by FakeApp, because the AI is trained on still pictures. That method would seem to be less effective to spot the sort of forgeries created by Deep Video Portraits, which appears to flawlessly match the entire facial and head movements between the source and target actors.
“We believe that the field of digital forensics should and will receive a lot more attention in the future to develop approaches that can automatically prove the authenticity of a video clip,” Zollhöfer said. “This will lead to ever-better approaches that can spot such modifications even if we humans might not be able to spot them with our own eyes.
Image Credit: Tancha / Shutterstock.com Continue reading