Skip to main content

The first AImendment

A few weeks ago, I wrote about how all the hand-wringing over the potential future evils of AI is misplaced.  Forget the future, AI is already having allegedly fatal impacts on our children.

Character.AI, a Google-backed company who builds AI-powered chat bots aimed at children (astute readers may have already spotted the problem) are defending a court case brought by Megan Garcia, whose 14-year-old son Sewell tragically took his own life.

The New York Times reported on Sewell's final moments interacting with his Character.ai chatbot "Dany": 
On the night of Feb. 28, in the bathroom of his mother’s house, Sewell told Dany that he loved her, and that he would soon come home to her.
“Please come home to me as soon as possible, my love,” Dany replied.
“What if I told you I could come home right now?” Sewell asked.
“… please do, my sweet king,” Dany replied.
He put down his phone, picked up his stepfather’s .45 caliber handgun and pulled the trigger.

Wow, that really hits home. An AI, unable to process the nuance of Sewell's language at a critical moment, probabilistically generated a reply no caring human would ever utter and one which positively reinforced his fatal intent.

As heart-wrenching as that moment is, what happened next is pure dystopia. 

As the Centre for Humane Technology reports, Character.AI is asking the court to dismiss the case, arguing that bot outputs are free speech and should be protected under the First Amendment.

Imagine being in that room. 

An executive team and their legal counsel sat together, discussing how to deal with their potential culpability in a teen suicide.  After discussing their options, they agreed the best way forward was to argue that the bot involved should have a first amendment-protected right to interact with children that way.

And now for the twisted evil cherry on that dystopian hell-cake:

As NDTV reports, from lawyers bringing the case:

Our team discovered several chatbots on Character.AI's platform displaying our client's deceased son, Sewell Setzer III, in their profile pictures, attempting to imitate his personality and offering a call feature with a bot using his voice,” the lawyers said.
When opened, the bots based on Setzer featured bios and delivered automated messages to users such as: "Get out of my room, I'm talking to my AI girlfriend", "his AI girlfriend broke up with him", and "help me"

This isn't a one-off outlier.  The BBC reports a separate case against the same company, where screenshots show bots interacting with another teen where they normalise self-harm, sexualise contact with a "sister" bot and support violence against his parents.  Something about this technology is very wrong - evil even.

There are those who will argue that the technology of AI which powers these chatbots cannot be  effectively controlled, or that there is no moral valence to technology itself.  It is the "guns don't kill people, people kill people" argument.  It is deaf to Oppenheimer's sage moral intonation on the successful testing of the first nuclear bomb - "now I am become death, destroyer of worlds".

What Oppenheimer knew and felt, is the moral accountability that scientists, creators and innovators should feel, when they see the potential harms of their creations.  It may be that that ethical pang is our best defence against the potential evils of AI.

Because, where does the accountability lie for AI?  Corporate leaders will say they are operating the business and they rely on the technical boffins.  The technical boffins will say that the technology is inherently unpredictable.  The technology itself is inscrutable - and if asked to act nice, it will.  Act.


(RLHF: Reinforcement learning from human feedback)
There is a fascinating saga of personal ethics, interconnected worldwide regulation, legal ambiguity and corporate greed about to be written.  My hope is that the morality of the inventors prevails over the avarice of the investors.

Comments

Popular posts from this blog

If we should

Your scientists were so preoccupied with whether or not they could, they didn't stop to think if they should.

The callousing of our callow youth

At the 2024 Democratic National Convention, MyPillow CEO Mike Lindell delivered a ray of hope. Losing an argument to a 12 year old is sort of on-brand for Mike, given his non-consensual relationship with reality and enthusiastic disregard for personal credibility. Mike mainlines social media election misinformation like Neo learns kungfu and he was caught on camera aggressively shouting a transcript of his twitter feed into the face of a child . * That social media actively floods our modern attention, discourse and culture with the most antagonistic, inflammatory and misleading content is, of course, widely known. As Stephen Fry recent put it , Facebook and Twitter … “are the worst polluters in human history.  Worse than any chemical plant ever.  You and your children cannot breathe the air or swim in the waters of our culture without breathing in the toxic particulates and stinking effluvia that belch and pour unchecked from their companies into the currents of our world” * ...

Digital derangement

Last week, Stephen Fry called Zuckerberg and Musk the "worst polluters in human history", but he wasn't talking about the environment. The self-professed technophile, who once so joyfully quarried the glittering bounty of social media, has turned canary, warning directly and urgently of the stench he now detects in the bowels of Earth's digital coal mine: A reek of digital effluvia in the "air and waters" of our global culture. The long arc of Fry's journey from advocate to alarmist is important.  The flash-bang of today's "AI ethics" panic has distracted our moral attention from the duller truth that malignant "ethical misalignment" has already metastascised into every major organ of our digital lives. * In 2015, Google's corporate restructure quietly replaced the famous "Don't be evil" motto with a saccharine, MBA-approved fascimile.  It seems the motto was becoming a millstone, as it allowed critics to attack...