What Artificial Intelligence Cannot Do

Self Reflection

Fresno Bee, April 5, 2025

Artificial intelligence is already changing the world. But will it change our humanity?

Bill Gates recently predicted that AI will soon be widely employed to supplement and even replace a lot of labor that currently requires human experts. This may include accountants, teachers, doctors and computer programmers. Any profession that requires repetitive information processing and rule-following expertise can be supplemented or replaced by AI.

This may free up human intellect to engage in more creative and imaginative tasks. It may also leave humans with more time to focus on interpersonal and relationship-based work. But there are also AI “therapists” and “friends” available online. AI companions are always available. The AI friend, Replika, touts itself as “always here to listen and talk. Always on your side.”

The convenience and efficiency of AI will lead to its widespread use. AI never sleeps — it never tires, or becomes fed up or impatient, unlike real human companions.

As AI development increases, it will be used to create even more powerful technology. This technological acceleration has led some experts to predict that artificial general intelligence will soon be created (something akin to human thinking but faster, tireless and not prone to laziness, procrastination or daydreaming). Others think the creation of artificial general intelligence is decades off; some say it is impossible.

As AI transforms into artifical general intelligence, it could be applied (or apply itself) to generating even more intelligent machinery. Some fear the creation of artificial super intelligence, a fear fueled by fictional sci-fi dystopias in which artificial super intelligence takes over and kills or enslaves humans.

Leaving that nightmare aside, there is no doubt that AI is already changing the meaning of a variety of human tasks. This will continue to happen as the technology becomes so efficient that resistance is futile. This may sound ominous, but it happens all the time as technologies improve.

The inexorable efficiency of technology explains why we prefer to ride rather than walk. It’s why we send texts instead of writing old-fashioned letters. The efficiency imperative will likely lead us to replace inefficient human beings with efficient AI in many parts of life. Why bother to write a report if AI can do it for you faster and better? Why bother to wake a real friend in a crisis in the middle of the night when AI is there to chat?

Of course, some people still write letters or walk. And there is a kind of pleasure to be found in completing your own tax form, or in writing computer code. But those quaint human activities are now a matter of choice. They represent a kind of boutique curiosity, chosen not for efficiency but for some other reason.

This is where the human element returns. Many things are valuable not because they are efficient, but because they are good, beautiful, intellectually challenging or uniquely human.

Friendship is like that: An AI-companion may be more efficient at giving advice in difficult times, or at keeping us entertained. But real human friendship is valuable for other reasons. Human friendship is not simply a one-sided exchange in which we use the other person for our benefit. Rather, friends make demands upon us. Their impatience reminds us to slow down. Their needs give us reason to look beyond our own.

The demands that other humans make upon us are infinitely more valuable than the cult of efficiency can imagine. Other human beings are part of who we are. When a friend or family member triumphs, we swell with pride for them. When they suffer, we suffer with them. And when they die, they take a part of us away with them.

AI will never replace the deeply inefficient existential reality of love, suffering and mortality. AI is fast, convenient and always available. But it cannot supplant the difficult experiences and troublesome relationships that make us fully human. Efficiency is a machine-based good. But human life is not mechanical. The wonder of existence is found in the tragic and often beautiful mess that is human nature.

To be human is not to be efficient. Rather, it is love, suffer and die. And that’s what no machine can ever do.

Read more at: https://www.fresnobee.com/opinion/readers-opinion/article303416756.html#storylink=cpy

The Ghost of Christmas Future: Artificial Intelligence, Santa Claus, and Holiday Magic

Artificial intelligence is now part of the magic of Christmas and it’s sweetly weird

Fresno Bee, Dec. 15, 2024

Artificial intelligence is now part of the magic of Christmas. I recently made a video call to an AI Santa. Santa sat by a roaring fire and chatted with me about reindeer, elves and his penchant for cookies and cocoa. I asked him to bring world peace for Christmas. He told me I was kind to ask for such a wonderful thing.

This was fun and charming — and a little weird. But this cute example is just the tip of the AI iceberg. In the future, we may routinely call AI friends, coaches and advisors. This will put real people out of business. It is much easier to visit AI Santa than to fight the masses at the mall. Virtual Santa may leave portly white-bearded men out of work.

Beyond the immediate ramifications of AI Santa are deeper questions about imagination and belief — a great Christmastime topic. Christmas calls for the “voluntary suspension of disbelief.” A common theme in Christmas movies is that for Christmas to happen, you must choose to believe.

This is mostly harmless: We suspend belief quite often in culture and the arts. We choose to believe when we enter the world of a novel, a movie or a theater production. At Christmas, voluntary suspension of disbelief kicks into high gear. The imagination’s ability to jump into a fantasy world is what allows art and AI to happen.

When you chat with an AI avatar, you enter the world inside the screen. AI makes this easier and more realistic than cinema or theater. By responding in real time to your presence, AI lures you into an imagined world. As this technology improves, some people could end up mistaking artifice for reality.

Christmas also involves what scholars call “motivated belief,” something we believe because we want it to be true. At some point, kids start to suspect that Santa isn’t real. But kids may play along because they want the Christmas goodies. The myths of Christmas encourage this, telling children that their belief in Santa is part of the process.

We all do this from time to time: Despite the evidence, we believe things that support our desires or preconceptions. It is often benign. But conspiracy theories, cults and superstitions also work this way. These beliefs are supported by an elaborate network of rationalization and confabulation. When they are challenged, the motivated believer explains away the counter-evidence and accuses the challenger of spreading fake news.

At Christmas, the imagination is beguiled by a network of trickery and tomfoolery. To support the Santa story, we make a big show of pretending that all kinds of phony stuff is real — from flying reindeer to elves. Parents construct an elaborate ruse involving the Elf on the Shelf, letters to Santa and a trip to the mall to visit Santa (or a trip to the computer to chat with AI Santa). It all leads up to a big Christmas Eve finale and the magic of Christmas morning.

This is all good fun. But the season of believing opens lots of questions about culture and belief, myth and magic. A culture is, after all, an elaborate game in which we all collaborate. In a sense, culture is “artificial,” a kind of art and artifice in which we construct meaning. Artificial intelligence is the latest and most sophisticated example of how this works.

As long as we understand the difference between what is real and what is phony, it is amusing to play along. But we should worry that in the AI era some people will confuse artifice with reality. It would be tragic if people mistook artificial friendship for the real thing. And it is dangerous when fake news seems as true or real as actual facts.

We need to do our best to keep all of this straight. And perhaps Christmas can help. In this magical season, take time to enjoy the show. Try out AI Santa, or make a wish for world peace. But when the show ends, ask what you’ve learned about the human imagination and about the difference between wishful thinking and the truth.

Read more at: https://www.fresnobee.com/opinion/readers-opinion/article296998234.html#storylink=cpy

Social Media Warnings and Education

Fresno Bee, June 23, 2024

Surgeon General Vivek Murthy’s recent call for warning labels on social media is a good idea. He notes that children who spend significant time on social media are at risk for mental illness. Murthy concludes, “The moral test of any society is how well it protects its children.”

But how best should we protect kids? Murthy recognizes that a warning label is a simple solution to a complex problem. Last year, his office issued a more detailed report noting that schools, parents, policymakers, and technology companies have a role to play in protecting kids. And long-term solutions depend upon education.

There is probably also a role for prohibitions. Smartphones have been banned in schools in Fresno and recently in Los Angeles. This week, Governor Newsom called for a statewide smartphone ban in California schools. Social media and smartphones are not the same thing. However, a school ban on smartphones is effectively a ban on social media during school time.

Tools and technologies can be employed in good or bad ways. A hammer can be used to build or to destroy. Prohibitions are justifiable when the risks are obvious and severe, and when the purported benefits of a tool are unclear. And with kids, their relative immaturity matters. A ban on social media access for kids might be justifiable and there is some wisdom in prohibiting smartphones at school. But at this point, a ban on these technologies is akin to closing the proverbial barn door once the horse has already galloped off.

People disagree about the risks and benefits of various technologies. One might argue against these bans by claiming that these technologies are more beneficial than dangerous. These tools help us stay connected, access the news, and conduct business. Of course, these tools also provide instant access to cyberbullying, exploitation, scams, and disinformation. But there is some truth to the claim that with smartphones and social media, it’s not the tool that is to blame, but how it is used.

Some technophobes are opposed to any innovative tool. Calculators were once viewed with skepticism, as was the Internet. These days technophobes are worried about artificial intelligence. But skeptics often adapt to new technologies, when their safety and usefulness are proven.

Hard-core libertarians resist every effort at prohibition. The recent Supreme Court case allowing “bump stock” weapons is worth mentioning here. The decision depends upon a technical matter involving trigger mechanisms. But the bigger question, not decided in this case, is whether there should be limits on dangerous weapons or whether individuals have a right to own even very dangerous weapons.

Social media and smartphones do not seem as dangerous as machine guns. So, it is easy to imagine a libertarian argument against Newsom’s proposed ban. Furthermore, social media is useful for kids. It’s how they socialize, organize clubs and teams, and how they communicate with each other and even with their parents. Smartphones can be useful in education when used properly to access information.

An outright ban may take away useful tools. And a school ban will have no impact on after-school usage. But there is no doubt that education is part of the solution. Teenagers must take driver’s ed and pass a licensing test to drive. Perhaps a similar training course and qualifying exam could be created for social media and smartphones.

Kids need critical lessons about cyberbullying, peer pressure, the bandwagon effect, and the power of misinformation and exploitative algorithms. They also need frank examples of the dangers of social media and smartphone addiction. They would benefit from a training course that includes lessons in “digital citizenship,” “ethical A.I.,” and “virtuous virtual reality” that encourage best practices online and good moral habits in cyberspace.

A Surgeon General’s warning is only a starting point for a broader conversation. We need to continue this conversation. A ban at school might help. But the social media and smartphone horse is already out of the barn. Kids need to be taught the skills and virtues that are required to ride that horse without getting hurt.

Read more at: https://www.fresnobee.com/opinion/article289421636.html#storylink=cpy

The wisdom of slowing down

Fresno Bee, September 10, 2023

Stop the mindless smartphone scrolling. Our souls long for a slower tempo.

Our world emphasizes speed. This is the age of artificial intelligence, smartphones and instant downloads. In this first-come, first-served culture, the early bird gets the worm. Who has time to ponder or reflect? We’re too busy flitting from one superficial thing to the next.

All of this speed and mobility may undermine our humanity. It contributes to loneliness and anxiety. Many good things require us to slow down, rather than speed up. Wisdom is not quick. Neither is love. The best things in life dwell in a time apart, lingering in slowness.

But artificial intelligence and related technologies push an ever more frantic pace. The speed of the stimuli on our screens can explain some of the negative mental health impacts of social media, video games and other technologies. Our brains are not meant to go this fast. Our souls long for a slower tempo. Human relationships need time to ripen, and genuine happiness is not instant gratification.

Now, sometimes speed is a good thing. Quick computers can churn through data and solve many problems. It is much more efficient to Google information than to go to a library and search the indexes of books on dusty shelves. Social media, online news apps and video games can be useful and fun. We can stay in touch with distant friends. We have immediate access to the latest news. And your phone contains multiple sources of instant gratification.

But moderation is needed. Scrolling for thrills is not the same as digging deep. We don’t build wisdom or friendships with a swipe on a screen. We need time for thinking, solitude and soul searching.

The novelist Milan Kundera lamented the lost pleasure of slowness in his novel “Slowness” where he suggests that we need time to “gaze at God’s windows.” He says, “There is a secret bond between slowness and memory, between speed and forgetting.” Speed causes us to forget who we are and what we value. We’re not sure where we’re going. But we’ll get there quickly.

Our bodies and brains evolved in a slower era. Our ancestors needed to think quickly on occasion to escape predators or hunt. But when the sun went down, they contemplated the stars and shared stories and songs. These ancient works of imagination unfolded at a pace that was rooted in the tempo of our beating hearts. With this in the background, it’s no wonder that most of the world’s wisdom traditions emphasize tranquility, patience, calmness and slowness.

The ancient sages took time to gaze deeply into God’s windows, and into their own souls. Socrates was well known for wandering and wondering. He would sometimes come to a halt as he walked through Athens, completely lost in thought.

In Asian traditions, the practice of meditation aims to cultivate slowness. The Buddha saw restlessness as an impediment to wisdom. The solution is to calm the mind and its restless agitation.

You don’t have to be Socrates or the Buddha to understand that many of the most meaningful human activities are best experienced slowly. This is true of grieving, making love and enjoying art. We can’t set a timer for grief or for love. The pace of these things transcends the frantic tempo of ordinary life, reflecting the patience of tender intimacy. To insist that Mozart or Shakespeare should speed things up is to misunderstand the nature of their art.

Philosophers describe things that are enjoyed slowly as “ends-in-themselves” valued for their own sake. These experiences represent moments of completion and fulfillment. Some people even sigh, and say of certain beautiful moments that they want them to last forever. This is also true of life itself. If you love life, you want it to last. Life is enjoyed for its own sake, and those who say that it is better to live fast and die young have probably not thought it over.

But the sages who have thought deeply about these things tell us that we need to relax our pace. The best and most important things — love, beauty and wisdom — are not quick or immediate. If you want to find these goods, you must slow down.

Read more at: https://www.fresnobee.com/opinion/readers-opinion/article279063134.html#storylink=cpy

Artificial Intelligence and Human Morality

Fresno Bee, June 4, 2023

Is artificial intelligence going to kill us? It all depends on who is using it and why

Experts warn that artificial intelligence may kill us. A declaration signed by a number of luminaries states: “Mitigating the risk of extinction from AI should be a global priority alongside other societal-scale risks such as pandemics and nuclear war.”

I’m sympathetic to the worry. But when you think about the other problems mentioned here — nuclear war and pandemics — it might be that we need AI to save us from our own incompetence. Could AI have responded better to Covid than we humans did?

It all depends on what we do with AI, and who is using it. A crazed dictator with AI is scary. But a scientist assisted by AI, not so much.

Geoffrey Hinton is one of the signatories of the new AI warning, and an expert in the field. In a recent interview, Hinton warns that AI may grow smarter than its human creators within five to 20 years.

One of the things that freaked him out recently was when he asked an AI to explain a joke. Hinton did not expect AI to understand humor. But it did.

That got me curious, so I asked ChatGPT (an online AI), “Why did the chicken cross the road?” Immediately, it said, “To get to the other side.” And then, without prompting, it explained the joke as a play on words. It said, “It’s a simple and often unexpected answer that plays on the double meaning of ‘the other side.’” It explained the joke as a “philosophical statement on the nature of life and death.”

This surprised me. The AI recognized that I was asking a joke. I had actually forgotten that the joke was about chicken suicide. But the AI went straight to the heart of the matter.

But is this an existential risk? I depends on how we use AI. If we use AI to explain jokes, we won’t risk much. Philosophy, and comedy, assisted by AI, might be fun and informative. But if we weave AI into the systems that govern our lives, we might end up in a strange dystopia.

One obvious concern is the stock market. AI can analyze data and make trades in nanoseconds. This may not lead to extinction. But it may cause bubbles and panics, and enrich those fortunate enough to have an AI broker. Or, maybe AI could be used beneficially to even things out, preventing panics and bubbles. Again, it depends on what we do with it, and what safeguards we program into the system.

A darker possibility is if AI took control of military systems, including nuclear weapons. What if AI were put in charge in the hope of automating and streamlining the decision procedures involved in nuclear war? Maybe nuclear-armed AI will lead to Armageddon. Or, again, maybe AI will better control our most deadly weapons.

It’s worth asking whether human beings are really trustworthy custodians of weapons, or wealth. Some crazed Dr. Strangelove could launch a nuclear war. And rapacious financiers like Bernie Madoff ruin people’s lives. Perhaps AI is more trustworthy than humans in this regard. AI won’t get angry, greedy, envious, or hateful.

And here is where things get really weird and dystopian. What if a smart AI figures out that humans — with all of our ignorance, spite, and greed — should not be trusted with nukes or with billion-dollar deals? In science fiction, the AI might seize control — for our own good!

But AI will only take control, if we put it in charge. Human beings are always looking for shortcuts and quick fixes to complex problems (as I discussed in my column last week). We invent tools to make things easier. But ultimately, we are responsible for the tools we create, including nuclear weapons, the stock market and AI.

We are also responsible for the greed, spite, and ignorance that afflict the world. These are human problems. Tools can magnify these ugly traits, or they can help us control our worst impulses. In the end, the choice of crossing the road to get to the other side belongs to us. This choice is ultimately about ethics and the human spirit. If AI leads to our extinction, the fault will not be in the tool but within the human soul.

Read more at: https://www.fresnobee.com/opinion/readers-opinion/article275991471.html#storylink=cpy