Tag Archives: religion

#437216 New Report: Tech Could Fuel an Age of ...

With rapid technological progress running headlong into dramatic climate change and widening inequality, most experts agree the coming decade will be tumultuous. But a new report predicts it could actually make or break civilization as we know it.

The idea that humanity is facing a major shake-up this century is not new. The Fourth Industrial Revolution being brought about by technologies like AI, gene editing, robotics, and 3D printing is predicted to cause dramatic social, political, and economic upheaval in the coming decades.

But according to think tank RethinkX, thinking about the coming transition as just another industrial revolution is too simplistic. In a report released last week called Rethinking Humanity, the authors argue that we are about to see a reordering of our relationship with the world as fundamental as when hunter-gatherers came together to build the first civilizations.

At the core of their argument is the fact that since the first large human settlements appeared 10,000 years ago, civilization has been built on the back of our ability to extract resources from nature, be they food, energy, or materials. This led to a competitive landscape where the governing logic is grow or die, which has driven all civilizations to date.

That could be about to change thanks to emerging technologies that will fundamentally disrupt the five foundational sectors underpinning society: information, energy, food, transportation, and materials. They predict that across all five, costs will fall by 10 times or more, while production processes will become 10 times more efficient and will use 90 percent fewer natural resources with 10 to 100 times less waste.

They say that this transformation has already happened in information, where the internet has dramatically reduced barriers to communication and knowledge. They predict the combination of cheap solar and grid storage will soon see energy costs drop as low as one cent per kilowatt hour, and they envisage widespread adoption of autonomous electric vehicles and the replacement of car ownership with ride-sharing.

The authors laid out their vision for the future of food in another report last year, where they predicted that traditional agriculture would soon be replaced by industrial-scale brewing of single-celled organisms genetically modified to produce all the nutrients we need. In a similar vein, they believe the same processes combined with additive manufacturing and “nanotechnologies” will allow us to build all the materials required for the modern world from the molecule up rather than extracting scarce natural resources.

They believe this could allow us to shift from a system of production based on extraction to one built on creation, as limitless renewable energy makes it possible to build everything we need from scratch and barriers to movement and information disappear. As a result, a lifestyle worthy of the “American Dream” could be available to anyone for as little as $250/month by 2030.

This will require a fundamental reimagining of our societies, though. All great civilizations have eventually hit fundamental limits on their growth and we are no different, as demonstrated by our growing impact on the environment and the increasing concentration of wealth. Historically this stage of development has lead to a doubling down on old tactics in search of short-term gains, but this invariably leads to the collapse of the civilization.

The authors argue that we’re in a unique position. Because of the technological disruption detailed above, we have the ability to break through the limits on our growth. But only if we change what the authors call our “Organizing System.” They describe this as “the prevailing models of thought, belief systems, myths, values, abstractions, and conceptual frameworks that help explain how the world works and our relationship to it.”

They say that the current hierarchical, centralized system based on nation-states is unfit for the new system of production that is emerging. The cracks are already starting to appear, with problems like disinformation campaigns, fake news, and growing polarization demonstrating how ill-suited our institutions are for dealing with the distributed nature of today’s information systems. And as this same disruption comes to the other foundational sectors the shockwaves could lead to the collapse of civilization as we know it.

Their solution is a conscious shift towards a new way of organizing the world. As emerging technology allows communities to become self-sufficient, flows of physical resources will be replaced by flows of information, and we will require a decentralized but highly networked Organizing System.

The report includes detailed recommendations on how to usher this in. Examples include giving individuals control and ownership of data rights; developing new models for community ownership of energy, information, and transportation networks; and allowing states and cities far greater autonomy on policies like immigration, taxation, education, and public expenditure.

How easy it will be to get people on board with such a shift is another matter. The authors say it may require us to re-examine the foundations of our society, like representative democracy, capitalism, and nation-states. While they acknowledge that these ideas are deeply entrenched, they appear to believe we can reason our way around them.

That seems optimistic. Cultural and societal change can be glacial, and efforts to impose it top-down through reason and logic are rarely successful. The report seems to brush over many of the messy realities of humanity, such as the huge sway that tradition and religion hold over the vast majority of people.

It also doesn’t deal with the uneven distribution of the technology that is supposed to catapult us into this new age. And while the predicted revolutions in transportation, energy, and information do seem inevitable, the idea that in the next decade or two we’ll be able to produce any material we desire using cheap and abundant stock materials seems like a stretch.

Despite the techno-utopianism though, many of the ideas in the report hold promise for building societies that are better adapted for the disruptive new age we are about to enter.

Image Credit: Futuristic Society/flickr Continue reading

Posted in Human Robots

#435462 Where Death Ends and Cyborgs Begin, With ...

Transhumanism is a growing movement but also one of the most controversial. Though there are many varying offshoots within the movement, the general core idea is the same: evolve and enhance human beings by integrating biology with technology.

We recently sat down with one of the most influential and vocal transhumanists, author and futurist Zoltan Istvan, on the latest episode of Singularity University Radio’s podcast series: The Feedback Loop, to discuss his ideas on technological implants, religion, transhumanism, and death.

Although Zoltan’s origin story is rooted deeply in his time as a reporter for National Geographic, much of his rise to prominence has been a result of his contributions to a variety of media outlets, including an appearance on the Joe Rogan podcast. Additionally, many of you may know him from his novel, The Transhumanist Wager, and his 2016 presidential campaign, where he drove around the United States in a bus that had been remodeled into the shape of a coffin.

Although Zoltan had no illusions about actually winning the presidency, he had hoped that the “immortality bus” and his campaign might help inject more science, technology, and longevity research into the political discourse, or at the very least spark a more serious conversation around the future of our species.

Only time will tell if his efforts paid off, but in the meantime, you can hear Zoltan discuss religion, transhumanism, implants, the existential motivation of death, and the need for new governmental policies in Episode 7 of The Feedback Loop. To listen in each week you can find us on your favorite podcasting platforms, such as Spotify, Apple, or Google, and you can find links to other podcasting platforms and Singularity Hub’s text-to-speech articles here. You can also find our past episodes with other thought leaders like Douglas Rushkoff and Annaka Harris below.

Image Credit: VAlex / Shutterstock.com Continue reading

Posted in Human Robots

#433717 Could an artificial intelligence be ...

Humans aren't the only people in society – at least according to the law. In the U.S., corporations have been given rights of free speech and religion. Some natural features also have person-like rights. But both of those required changes to the legal system. A new argument has laid a path for artificial intelligence systems to be recognized as people too – without any legislation, court rulings or other revisions to existing law. Continue reading

Posted in Human Robots

#433689 The Rise of Dataism: A Threat to Freedom ...

What would happen if we made all of our data public—everything from wearables monitoring our biometrics, all the way to smartphones monitoring our location, our social media activity, and even our internet search history?

Would such insights into our lives simply provide companies and politicians with greater power to invade our privacy and manipulate us by using our psychological profiles against us?

A burgeoning new philosophy called dataism doesn’t think so.

In fact, this trending ideology believes that liberating the flow of data is the supreme value of the universe, and that it could be the key to unleashing the greatest scientific revolution in the history of humanity.

What Is Dataism?
First mentioned by David Brooks in his 2013 New York Times article “The Philosophy of Data,” dataism is an ethical system that has been most heavily explored and popularized by renowned historian, Yuval Noah Harari.

In his 2016 book Homo Deus, Harari described dataism as a new form of religion that celebrates the growing importance of big data.

Its core belief centers around the idea that the universe gives greater value and support to systems, individuals, and societies that contribute most heavily and efficiently to data processing. In an interview with Wired, Harari stated, “Humans were special and important because up until now they were the most sophisticated data processing system in the universe, but this is no longer the case.”

Now, big data and machine learning are proving themselves more sophisticated, and dataists believe we should hand over as much information and power to these algorithms as possible, allowing the free flow of data to unlock innovation and progress unlike anything we’ve ever seen before.

Pros: Progress and Personal Growth
When you let data run freely, it’s bound to be mixed and matched in new ways that inevitably spark progress. And as we enter the exponential future where every person is constantly connected and sharing their data, the potential for such collaborative epiphanies becomes even greater.

We can already see important increases in quality of life thanks to companies like Google. With Google Maps on your phone, your position is constantly updating on their servers. This information, combined with everyone else on the planet using a phone with Google Maps, allows your phone to inform you of traffic conditions. Based on the speed and location of nearby phones, Google can reroute you to less congested areas or help you avoid accidents. And since you trust that these algorithms have more data than you, you gladly hand over your power to them, following your GPS’s directions rather than your own.

We can do the same sort of thing with our bodies.

Imagine, for instance, a world where each person has biosensors in their bloodstreams—a not unlikely or distant possibility when considering diabetic people already wear insulin pumps that constantly monitor their blood sugar levels. And let’s assume this data was freely shared to the world.

Now imagine a virus like Zika or the Bird Flu breaks out. Thanks to this technology, the odd change in biodata coming from a particular region flags an artificial intelligence that feeds data to the CDC (Center for Disease Control and Prevention). Recognizing that a pandemic could be possible, AIs begin 3D printing vaccines on-demand, predicting the number of people who may be afflicted. When our personal AIs tell us the locations of the spreading epidemic and to take the vaccine it just delivered by drone to our homes, are we likely to follow its instructions? Almost certainly—and if so, it’s likely millions, if not billions, of lives will have been saved.

But to quickly create such vaccines, we’ll also need to liberate research.

Currently, universities and companies seeking to benefit humankind with medical solutions have to pay extensively to organize clinical trials and to find people who match their needs. But if all our biodata was freely aggregated, perhaps they could simply say “monitor all people living with cancer” to an AI, and thanks to the constant stream of data coming in from the world’s population, a machine learning program may easily be able to detect a pattern and create a cure.

As always in research, the more sample data you have, the higher the chance that such patterns will emerge. If data is flowing freely, then anyone in the world can suddenly decide they have a hunch they want to explore, and without having to spend months and months of time and money hunting down the data, they can simply test their hypothesis.

Whether garage tinkerers, at-home scientists, or PhD students—an abundance of free data allows for science to progress unhindered, each person able to operate without being slowed by lack of data. And any progress they make is immediately liberated, becoming free data shared with anyone else that may find a use for it.

Any individual with a curious passion would have the entire world’s data at their fingertips, empowering every one of us to become an expert in any subject that inspires us. Expertise we can then share back into the data stream—a positive feedback loop spearheading progress for the entirety of humanity’s knowledge.

Such exponential gains represent a dataism utopia.

Unfortunately, our current incentives and economy also show us the tragic failures of this model.

As Harari has pointed out, the rise of datism means that “humanism is now facing an existential challenge and the idea of ‘free will’ is under threat.”

Cons: Manipulation and Extortion
In 2017, The Economist declared that data was the most valuable resource on the planet—even more valuable than oil.

Perhaps this is because data is ‘priceless’: it represents understanding, and understanding represents control. And so, in the world of advertising and politics, having data on your consumers and voters gives you an incredible advantage.

This was evidenced by the Cambridge Analytica scandal, in which it’s believed that Donald Trump and the architects of Brexit leveraged users’ Facebook data to create psychological profiles that enabled them to manipulate the masses.

How powerful are these psychological models?

A team who built a model similar to that used by Cambridge Analytica said their model could understand someone as well as a coworker with access to only 10 Facebook likes. With 70 likes they could know them as well as a friend might, 150 likes to match their parents’ understanding, and at 300 likes they could even come to know someone better than their lovers. With more likes, they could even come to know someone better than that person knows themselves.

Proceeding With Caution
In a capitalist democracy, do we want businesses and politicians to know us better than we know ourselves?

In spite of the remarkable benefits that may result for our species by freely giving away our information, do we run the risk of that data being used to exploit and manipulate the masses towards a future without free will, where our daily lives are puppeteered by those who own our data?

It’s extremely possible.

And it’s for this reason that one of the most important conversations we’ll have as a species centers around data ownership: do we just give ownership of the data back to the users, allowing them to choose who to sell or freely give their data to? Or will that simply deter the entrepreneurial drive and cause all of the free services we use today, like Google Search and Facebook, to begin charging inaccessible prices? How much are we willing to pay for our freedom? And how much do we actually care?

If recent history has taught us anything, it’s that humans are willing to give up more privacy than they like to think. Fifteen years ago, it would have been crazy to suggest we’d all allow ourselves to be tracked by our cars, phones, and daily check-ins to our favorite neighborhood locations; but now most of us see it as a worthwhile trade for optimized commutes and dating. As we continue navigating that fine line between exploitation and innovation into a more technological future, what other trade-offs might we be willing to make?

Image Credit: graphicINmotion / Shutterstock.com Continue reading

Posted in Human Robots

#432303 What If the AI Revolution Is Neither ...

Why does everyone assume that the AI revolution will either lead to a fiery apocalypse or a glorious utopia, and not something in between? Of course, part of this is down to the fact that you get more attention by saying “The end is nigh!” or “Utopia is coming!”

But part of it is down to how humans think about change, especially unprecedented change. Millenarianism doesn’t have anything to do with being a “millennial,” being born in the 90s and remembering Buffy the Vampire Slayer. It is a way of thinking about the future that involves a deeply ingrained sense of destiny. A definition might be: “Millenarianism is the expectation that the world as it is will be destroyed and replaced with a perfect world, that a redeemer will come to cast down the evil and raise up the righteous.”

Millenarian beliefs, then, intimately link together the ideas of destruction and creation. They involve the idea of a huge, apocalyptic, seismic shift that will destroy the fabric of the old world and create something entirely new. Similar belief systems exist in many of the world’s major religions, and also the unspoken religion of some atheists and agnostics, which is a belief in technology.

Look at some futurist beliefs around the technological Singularity. In Ray Kurzweil’s vision, the Singularity is the establishment of paradise. Everyone is rendered immortal by biotechnology that can cure our ills; our brains can be uploaded to the cloud; inequality and suffering wash away under the wave of these technologies. The “destruction of the world” is replaced by a Silicon Valley buzzword favorite: disruption. And, as with many millenarian beliefs, your mileage varies on whether this destruction paves the way for a new utopia—or simply ends the world.

There are good reasons to be skeptical and interrogative towards this way of thinking. The most compelling reason is probably that millenarian beliefs seem to be a default mode of how humans think about change; just look at how many variants of this belief have cropped up all over the world.

These beliefs are present in aspects of Christian theology, although they only really became mainstream in their modern form in the 19th and 20th centuries. Ideas like the Tribulations—many years of hardship and suffering—before the Rapture, when the righteous will be raised up and the evil punished. After this destruction, the world will be made anew, or humans will ascend to paradise.

Despite being dogmatically atheist, Marxism has many of the same beliefs. It is all about a deterministic view of history that builds to a crescendo. In the same way as Rapture-believers look for signs that prophecies are beginning to be fulfilled, so Marxists look for evidence that we’re in the late stages of capitalism. They believe that, inevitably, society will degrade and degenerate to a breaking point—just as some millenarian Christians do.

In Marxism, this is when the exploitation of the working class by the rich becomes unsustainable, and the working class bands together and overthrows the oppressors. The “tribulation” is replaced by a “revolution.” Sometimes revolutionary figures, like Lenin, or Marx himself, are heralded as messiahs who accelerate the onset of the Millennium; and their rhetoric involves utterly smashing the old system such that a new world can be built. Of course, there is judgment, when the righteous workers take what’s theirs and the evil bourgeoisie are destroyed.

Even Norse mythology has an element of this, as James Hughes points out in his essay in Nick Bostrom’s book Global Catastrophic Risks. Ragnarok involves men and gods being defeated in a final, apocalyptic battle—but because that was a little bleak, they add in the idea that a new earth will arise where the survivors will live in harmony.

Judgement day is a cultural trope, too. Take the ancient Egyptians and their beliefs around the afterlife; the Lord of the underworld, Osiris, weighs the mortal’s heart against a feather. “Should the heart of the deceased prove to be heavy with wrongdoing, it would be eaten by a demon, and the hope of an afterlife vanished.”

Perhaps in the Singularity, something similar goes on. As our technology and hence our power improve, a final reckoning approaches: our hearts, as humans, will be weighed against a feather. If they prove too heavy with wrongdoing—with misguided stupidity, with arrogance and hubris, with evil—then we will fail the test, and we will destroy ourselves. But if we pass, and emerge from the Singularity and all of its threats and promises unscathed, then we will have paradise. And, like the other belief systems, there’s no room for non-believers; all of society is going to be radically altered, whether you want it to be or not, whether it benefits you or leaves you behind. A technological rapture.

It almost seems like every major development provokes this response. Nuclear weapons did, too. Either this would prove the final straw and we’d destroy ourselves, or the nuclear energy could be harnessed to build a better world. People talked at the dawn of the nuclear age about electricity that was “too cheap to meter.” The scientists who worked on the bomb often thought that with such destructive power in human hands, we’d be forced to cooperate and work together as a species.

When we see the same response over and over again to different circumstances, cropping up in different areas, whether it’s science, religion, or politics, we need to consider human biases. We like millenarian beliefs; and so when the idea of artificial intelligence outstripping human intelligence emerges, these beliefs spring up around it.

We don’t love facts. We don’t love information. We aren’t as rational as we’d like to think. We are creatures of narrative. Physicists observe the world and we weave our observations into narrative theories, stories about little billiard balls whizzing around and hitting each other, or space and time that bend and curve and expand. Historians try to make sense of an endless stream of events. We rely on stories: stories that make sense of the past, justify the present, and prepare us for the future.

And as stories go, the millenarian narrative is a brilliant and compelling one. It can lead you towards social change, as in the case of the Communists, or the Buddhist uprisings in China. It can justify your present-day suffering, if you’re in the tribulation. It gives you hope that your life is important and has meaning. It gives you a sense that things are evolving in a specific direction, according to rules—not just randomly sprawling outwards in a chaotic way. It promises that the righteous will be saved and the wrongdoers will be punished, even if there is suffering along the way. And, ultimately, a lot of the time, the millenarian narrative promises paradise.

We need to be wary of the millenarian narrative when we’re considering technological developments and the Singularity and existential risks in general. Maybe this time is different, but we’ve cried wolf many times before. There is a more likely, less appealing story. Something along the lines of: there are many possibilities, none of them are inevitable, and lots of the outcomes are less extreme than you might think—or they might take far longer than you think to arrive. On the surface, it’s not satisfying. It’s so much easier to think of things as either signaling the end of the world or the dawn of a utopia—or possibly both at once. It’s a narrative we can get behind, a good story, and maybe, a nice dream.

But dig a little below the surface, and you’ll find that the millenarian beliefs aren’t always the most promising ones, because they remove human agency from the equation. If you think that, say, the malicious use of algorithms, or the control of superintelligent AI, are serious and urgent problems that are worth solving, you can’t be wedded to a belief system that insists utopia or dystopia are inevitable. You have to believe in the shades of grey—and in your own ability to influence where we might end up. As we move into an uncertain technological future, we need to be aware of the power—and the limitations—of dreams.

Image Credit: Photobank gallery / Shutterstock.com

We are a participant in the Amazon Services LLC Associates Program, an affiliate advertising program designed to provide a means for us to earn fees by linking to Amazon.com and affiliated sites. Continue reading

Posted in Human Robots