Never Mind The Robots. Beware The ‘Errant’ Humans

has humanity reached its limits to manage accelerating change?

Giovanni Rene Rodriquez

Originally published in
Forbes

Superheroes, messiahs, and martyrs

The stories we tell ourselves is the subject of most of my writing. And when those stories are of a religious-like variety (fear, fantasy, or faith-based), I get particularly interested. In preparing for this piece, I spoke with a former colleague, Peter Hirshberg – who among other roles headed enterprise marketing for Apple reporting to Steve Jobs – about the “messianic” tone that many leaders were striking (in Hirshberg’ss words, a by-product of the “Apocalypse Industrial Complex”). Let’s start with Andreessen himself, who early in his manifesto proclaimed in a Jesus-like voice:

“[T]echnology is the glory of human ambition and achievement, the spearhead of progress, and the realization of our potential. For hundreds of years, we properly glorified this – until recently. I am here to bring the good news. (emphasis added).”

No knock on Andreessen, some of whose contributions – including the first commercial web browser – undeniably have transformed society for the greater good. And it’s hard to know whether he was being playful or facetious. Unfortunately, his argument is further complicated by the fact that he credits people like Tommaso Marinetti – who as Vice noted was “one of the architects of Italian fascism” — as a “patron saint” of techno optimism. It’s not a good look for a Silicon Valley VC. Even giving him the benefit of the doubt, it’s not becoming of someone in his position, validating the ongoing story of the “the hubris of the tech bros.”

In the meantime, says Hirshberg, OpenAI “was stacked with messianic people who believed they were saving the world,” and odd behavior. There was talk about how chief scientist Ilya Sutskever – who originally voted to oust Altman, but quickly changed his mind – of making strange spiritual claims. In a November 19 article for the Atlantic, Karen Hao and Charlie Wertzel reported on an all-hands meeting where Sustkever led a chant invoking the much feared yet greatly anticipated arrival of AGI, artificial general intelligence that can perform any task a human can:

Anticipating the arrival of this all-powerful technology, Sutskever began to behave like a spiritual leader, three employees who worked with him told us. His constant, enthusiastic refrain was ‘feel the AGI,’ a reference to the idea that the company was on the cusp of its ultimate goal. At OpenAI’s 2022 holiday party, held at the California Academy of Sciences, Sutskever led employees in a chant: ‘Feel the AGI! Feel the AGI!’ The phrase itself was popular enough that OpenAI employees created a special ‘Feel the AGI’ reaction emoji in Slack.”

The spiritual and superhero superlatives kept coming ... sometimes from the media. Upon Altman’s return to OpenAI, Wired wrote about his “second coming.” In film fantasy land, one journalist began casting the inevitable biopic (with Nicholas Cage as Elon Musk). But let’s put this in context: 2023 was a year when all sorts of CEO craziness and silliness happened. This was the year when two tech titans taunted one another to take part in a cage match. On New Year’s Eve, the usually sober Anderson Cooper was still talking about it, opining that Musk was not in shape to battle Zuckerberg. It was also the year that former chair and CEO of CNN Walter Isaacson published his biography of Musk, which ended with the question, “Do the audaciousness and hubris that drive him to attempt epic feats excuse his bad behavior, his callousness, his recklessness, the times he’s an a__hole?” Two months later, after advertisers on X (formerly known on Twitter) began to flee because Musk seemed to have endorsed anti-semitic hate speech the platform, Musk sat on the stage of an industry event and told advertisers to “go f_ck yourself.”

If there were a film about 2023, it could be titled, “CEOs Gone Wild.” And along with the messiahs came a martyr for the cause of the so-called techno-pessimists. Helen Toner, the former OpenAI board member who helped lead the charge to boot Altman, got a harsh spotlight in a New York Times article about the chaos within the OpenAI board:

“Hours after Mr. Altman was ousted, OpenAI executives confronted the remaining board members during a video call, according to three people who were on the call. During the call, Jason Kwon, OpenAI’s chief strategy officer, said the board was endangering the future of the company by pushing out Mr. Altman. This, he said, violated the members’ responsibilities. Ms. Toner disagreed. The board’s mission was to ensure that the company creates artificial intelligence that ‘benefits all of humanity’, and if the company was destroyed, she said, that could be consistent with its mission.”

The very idea that a massively financed for-profit could be the servant to a non-profit entity must have been shocking to investors including Khosla and Microsoft which owns nearly half the for-profit entity. But the elevation of Toner in early analysis of the board’s actions marked her for martyrdom for a lost cause. Long gone now, her silence only underscores her martyr identity.

Apocalypse now ... and then

Is the moral of the story is that “tech geniuses make bad governance decisions?” Perhaps. A glance at the Rube Goldberg-inspired governance chart for OpenAI might persuade a lot of folks of that. The failure of the board to act in the interests of investors may have been the result of replacing Musk and Hoffman when they left the board, respectively, in 2009 and 2013 (leaving a gap in entrepreneurial leadership and perspective). Or is it more, as Fortune’s Maria Aspan recently argued, that “when nonprofits and for-profits clash, the one with the money usually wins.”

But a broader view of what happened at OpenAI is that humans may be reaching their limits to understand and manage accelerating change, and that’s why we’re behaving irrationally. Ray Kurzweil, the famed futurist, has written about our inability to grasp exponential change, the rate at which many modern technologies develop. This is a problem that extends beyond corporate leaders drunk with their newly found powers. It’s increasingly a problem for the people they’re supposed to serve.

According to a December 2022 survey by Pew Research, nearly four out of ten American adults believe we are living in “the end times.” When I first learned this, I recalled from my college studies how every 1,000 years, sectors of society – often the poorest and least educated – are possessed by the fear that the end is nigh. Scholars have labeled this phenomenon as millenarianism. We experienced it not long ago, in a secular way, with the Y2K scare. But historically, we can trace apocalyptic thinking to differing phenomena, between millenia, like the Native American Ghost Dance of 1890, in which practitioners believed the dead would reunite with the living, the Mayan Calendar cult, which foretold the world would end in 2012, Judaism, Islam, Mormonism, and Christianity itself which, in the New Testament, prophesied a period of violence and chaos before the second coming of Jesus. As the Pew survey shows, religious apocalyptic thinking is alive and well today. But it’s secular apocalyptic thinking that should concern us when thinking about AI. Why: the number of people is far larger.

Recently, a new phenomenon has come to my attention. It goes by the name of “Apocalyptic Anxiety.” It's become the focus of countless articles by health and wellness professionals who have united around the notion that the recent growth of apocalyptic thinking can be explained by real-world causes for anxiety. This loosely fits with what psychologists call “anticipatory anxiety.” The impact is two-fold. First there’s the physical harm that anticipatory anxiety can bring: emotional numbness, muscle tension and pain, trouble sleeping. Then there’s the long-term impact on the human psyche. An article by Kari Paul in The Guardian opines that “experts say the fallout from powerful AI will be less a nuclear bomb and more a creeping deterioration of society.” Similarly, in a contributed article bylined by Nir Eisikovits, professor of philosophy at UMass Boston, the real existential threat of AI is in the “philosophical rather than apocalyptic sense. AI in its current form can alter the way people view themselves. It can degrade abilities and experiences that people consider essential to being human.”

There’s another legitimate concern. Can the fear of AI distract us from real and present dangers like climate change and the rise of military conflict in the Middle East? In a June 2023 article for Wired entitled, “Humans Aren’t Mentally Ready for an AI-Saturated ‘Post-Truth World’,” Thor Benson interviewed philosopher Larry Rosen who says “he worries that AI will make people more reliant on technology. Humans like things to be as simple and easy as possible, to avoid stress, he says, so people might start automating every aspect of their life that they can.” He quoted Rosen:

“I get concerned about the fact that we just blindly believe the GPS. We don’t question it. Are we just going to blindly believe in AI? … As it is, we’re overwhelmed. We’re so overwhelmed that we can’t make ourselves do a simple task and see it to completion. Anxiety is just going to ratchet up as we’re faced with this unknown thing in our world.”

So, one fear – a rational fear – is that AI may further lull humanity into a state of techno-dependence, robbing humans of one of the principal attributes of being human: agency. It conjures the memory of the 2008 animated film Wall-E, where in the future everyone is overfed and complacent as cattle. But there’s another kind of the distraction, caused apocalypse alarmists: the distraction from real-world applications for AI, said Peter Hirshberg, speaking of the “apocalypse industrial complex”:

“You can build a big audience with fear, and that can lead to academic and foundation funding, a mechanism that works particularly well in a field that gets lots of attention and is built on 100 years frightening myths about robots and machine intelligence. This tends to support a lot of the theoretical fears of AI at the expense the very practical and near term matters of economic impact, understanding how AI will transform enterprises.”

Processing change: the other side of apocalyptic thinking

Then there’s the argument that distraction can be a good thing. Over the weekend, I saw that In a holiday issue of Scientific American, Daisy Yuhas wrote about the “comforts of the apocalypse.” Yuhas quoted Minnesota neuroscientist Shmuel Lissek who believes that “at its heart, the concept of doomsday evokes an innate and ancient bias in most mammals.” She continued:

“Lissek suspects that some apocalyptic believers find the idea that the end is nigh to be validating. Individuals with a history of traumatic experiences, for example, may be fatalistic. For these people, finding a group of like-minded fatalists is reassuring. There may also be comfort in being able to attribute doom to some larger cosmic order—such as an ancient Mayan prophecy. This kind of mythology removes any sense of individual responsibility.”

So, yes, it’s important to consider the other side of apocalyptic thinking: the therapeutic value. Lissek is not the first to suggest this – for many years, social scientists have written about the satisfaction doomers derive from processing a tale of good and evil. It might also explain the success of dramas like The Walking Dead franchise, and the film Don’t Look Up. But digging a little deeper, we can better appreciate the upside to apocalyptic thinking by understanding that “end is nigh” tales aren’t just about destruction. They’re also about reversal and renewal. The apocalyptic view in the dark ages gave birth to imagining a renaissance. The apocalyptic view in the New Testament mapped the coming of the day of divine reckoning. It pays to understand the original meaning of the word apocalypse from the Greek at the time of the writing of the New Testament: “to uncover, reveal, lay bare, or disclose.”

And therein lies the secular opportunity for this world – to uncover, reveal, and act on the good we can do when we understand the real opportunities and threats of AI. But are we up to the task? It’s often been quoted that “when a man knows he is to be hanged in a fortnight, it concentrates his mind wonderfully.” But what if a man had five years, ten years, thirty years? Would his thinking be more rational? Already cooler heads are prevailing. In the business world, leaders are forging straight ahead, but guided under the rubric of “responsible AI.” Yet for the larger societal narrative, leadership has yet to emerge. How do we deal with the inevitable? Does the inevitable need to be told in either a utopian or dystopian narrative?” Kurzweil, who has warned about our inability to comprehend accelerating change, is one of the world’s most ardent apologists for the “technological singularity,” with a more empirical, though controversial, vision for the time when human intelligence is enhanced or overtaken by AI. If this apocalypse were to come true, would it be the end of the world, or a new beginning, a new renaissance, following the new dark age we are living in?

Let’s remind ourselves that to be human means to have agency. That should help us find the right course. Until then, beware the errant humans and the unintentional harm they can bring.