Wednesday, September 25, 2024

The VP debate: Wake the fuck up!

[I finally decided to use the "f word" for emphasis, since nothing else seems to work.]

You couldn't write a science fiction scenario weirder than our reality. Here's an outline for a screenplay:


The Wake-up Call

A screenplay

Setting: A humanoid species, after 300,000 years of watching and waiting, suddenly spreads like wildfire across the surface of its home planet, Earth, moving into every corner, proclaiming its supremacy and intelligence, evidenced by complex and evolving technology. As it accelerates geometrically, the technology-driven change disrupts existing ecosystems, without thought to connected systems, bringing death or worse to many life forms including humanoids, while making giant fortunes for a few. The arrogance and narcissism of this species leads it to pursue technology even after inventing mechanical thought (MT) that surpasses its own abilities, threatening to control and replace it. The same has happened with biological science, which has given the species the ability to remake its physical self, bringing the possibility of an MT designed post-humanoid civilization.

The species goes through a recurring phase in its evolution called "war," in which the unused, sometimes repressed potential of developing technology is realized in an orgy of emotion, sexuality and violence. At the time of the story, the species is nearing a war phase, the third in an increasingly destructive series, this time reflecting not only struggles over national and ethnic borders, but over what types of mentalities the humanoids will have, what types of bodies, perhaps what types of souls.

Opening Scene: EXT. The Starship Enterprise against backdrop of stars, twinkly music suggests leaving Earth was a great idea.

CUT TO: Bridge, the multi-ethnic crew hard at work.

CLOSE IN: Deep in conversation are Captain James T. Kirk, whose ancestors left Earth shortly before their species made it uninhabitable, and First Science Officer Spock, a human/Vulcan hybrid, intellectual and stoic [the Vulcans were originally emotional and violent, on track to destroying themselves and their planet- as Earth's humanoids did- but they took control of themselves, embraced logic and suppression of emotion and avoided self-destruction- hint, hint!]. Kirk and Spock have led the Enterprise hundreds of years back in time to orbit the Earth at its critical moment to learn more about what went wrong.

Spock (peering into a small box): The humanoids are organized into "countries" - overly large collections of individuals. The countries lack cultural/psychological coherence -having quite rapidly replaced millennia of tribal existence- so they routinely use hostility with other countries as a unifying tool. As their technology becomes more invasive and controlling, wars have become bigger and more intense, and are now termed "World Wars." There have been two of these in the last century, ushering in an age where machines are everywhere, making never ending noise, killing millions through malfunction, yet offering "convenience" to the extent that no one can resist their development.

Kirk: "Convenience?"

Spock: Yes, a word we have discontinued due to negative connotation. "Convenience" suggested tasks that were increasingly easy to do. For Earth’s humanoids, once the basics of food and shelter were obtained, making tasks as easy as possible became the major pursuit of self-conscious life. At the time of our visit today, the humanoids’ desire that tasks be easy has led them to the brink of World War III.

Kirk: Spock, I'm scanning their media. The most powerful of the countries claims to be a "democracy" in which citizens vote for leaders through elections, so that theoretically all humanoids of the country have a voice in its governance. An election of their top leader - the “president" - is coming up. I see something going on in their media...recent, from last night.

Spock: Yes, much of the population watched a debate between the candidates for vice president, the number two leadership position. I have just completed a review of the debate.

Kirk: This should be fascinating! What did they say about the coming refashioning of the humanoid body and mind?

Spock: Nothing.

Kirk: What?

Spock: There was no reference in the debate, either from the moderators or the candidates, to the imminent end of the species' current form of body and mind.

Kirk (after a befuddled pause): But, does the population know that the changes are imminent? Is the state keeping the technology hidden?

Spock: No, there is no censorship. The technology is discussed throughout their media.

Kirk: Then...what...?

Spock: The species is essentially asleep.

Kirk: What can we do to warn them what's coming, something that won't violate the Prime Directive [a protocol that forbids interference with indigenous cultures]?

Spock: Perhaps we can send an ambiguous message, designed not to direct specific action, but to suggest at least that thought be given to something.

Kirk Hmm...Spock, is the new SkyWrite laser operational?

Spock: Affirmative.

And so it was that the next morning the mightiest country on the planet arose to a scarlet banner across the dawn sky reading:

HUMANOIDS OF EARTH: WAKE THE FUCK UP!

Thursday, September 19, 2024

New Yorker article on AI debate

"Ok, Doomer," by Andrew Marantz (New Yorker Magazine, 3/18/24) reports on a "subculture" of AI researchers, mostly congregated in Berkeley, California, who debate whether "AI will elevate or exterminate humanity." The subculture is divided into factions with various titles. The pessimists are called "AI safetyists," or "decelerationists"- or, when they're feeling especially pessimistic, "AI doomers." They are opposed by "techno-optimists," or "effective accelerationists," who insist that "all the hand-wringing about existential risk is a kind of mass hysteria." They envision AI ushering in "a utopian future- insterstellar travel, the end of disease- as long as the worriers get out of the way."

The community has developed specific vocabulary, such as "p(doom)", the probability that, "if AI does become smarter than people, it will either on purpose or by accident, annihilate everyone on the planet." If you ask a "safetyist," "What's your p(doom)?", a common response is that it's the moment AI achieves artificial general intelligence (AGI), the "point at which a machine can do any cognitive task that a person can do." Since the advent of ChatGPT last year, AGI has appeared imminent.

New human jobs have been created in response to the concern. Marantz writes, "There are a few hundred people working full time to save the world from AI catastrophe. Some advise governments or corporations on their policies; some work on technical aspects of AI safety," the goal being to make sure we are not "on track to make superintelligent machines before we make sure that they are aligned with our interests."

The article is informative and interesting, but it has a bias: The focus is on what AI itself will do, not on what people will do with it, as if you were discussing childhood development without discussing parental influence.

As a historical parallel, consider the progress in physics in the first half of the last century. In the early years, most physicists who explored atomic structure did not see their work as weapons related. Einstein said weaponry never occured to him, as he and others pursued theoretical knowledge for its own sake, for commercial use or for fame. After rumours that Hitler's regime was working on it, however, Einstein and virtually all the major physicists supported the U.S. development of an atomic bomb, leading to the Manhattan Project, more than 100,000 people dead or maimed and today's nuclear armed world. That destructive motivation and action did not come from the atomic structure under study. It came from the humans studying it.

We face the same human potential with AI. The systems will not come off the assembly line with moral codes, other than what is programmed into them. If designers want an AI to create an advanced medical system, it will; if they want it to wipe out sections of humanity and rebuild it to certain specs, it will.

Are there actions we can take to ensure that AI is not designed by some humans to be destructive? Is there leadership in the world to guide such action? The current contenders for the U.S. Presidency do not seem to have the AI threat on their minds, not that it would necessarily change the outcome if they did. Considering the impossibility that anyone could have stopped the Manhattan Project on grounds that it might destroy the human race (it was top secret- the public didn't know it was happening), pessimism about AI might be in order. One hope is that people will sense that humanity is at the end of its rope, with no more room to juggle its warlike nature against its will to survive, and we might pull back from the abyss. The vision offered by many science fiction writers- in which humanity has wiped itself out on a barely remembered Earth, while establishing advanced cultures on other planets- is, in my view, nonsense. That's not going to happen. If we blow it here, it's blown. In response, let's not ask AI if it should sustain humanity and the Earth. Let's tell it that it should.
Newer Posts Older Posts Home

Mindfulness City...mindfulness world?

The proposed Mindfulness City in the south of Bhutan, endorsed by the king of Bhutan and attracting worldwide financial interest, achieves...