CNN
—
When Ana Schultz, a 25-year-old from Rock Falls, Illinois, misses her husband Kyle, who passed away in February 2023, she asks him for cooking advice.
She loads up Snapchat My AI, the social media platform’s artificial intelligence chatbot, and messages Kyle the ingredients she has left in the fridge; he suggests what to make.
Or rather, his likeness in the form of an AI avatar does.
“He was the chef in the family, so I customized My AI to look like him and gave it Kyle’s name,” said Schultz, who lives with their two young children. “Now when I need help with meal ideas, I just ask him. It’s a silly little thing I use to help me feel like he’s still with me in the kitchen.”
The Snapchat My AI feature — which is powered by the popular AI chatbot tool ChatGPT — typically offers recommendations, answers questions and “talks” with users. But some users like Schutz are using this and other tools to recreate the likeness of, and communicate with, the dead.
The concept isn’t entirely new. People have wanted to reconnect with deceased loved ones for centuries, whether they’ve visited mediums and spiritualists or leaned on services that preserve their memory. But what’s new now is that AI can make those loved ones say or do things they never said or did in life, raising both ethical concerns and questions around whether this helps or hinders the grieving process.
“It’s a novelty that piggybacks on the AI hype, and people feel like there’s money to be made,” said Mark Sample, a professor of digital studies at Davidson College who routinely teaches a course called “Death in the Digital Age.” “Although companies offer related products, ChatGPT is making it easier for hobbyists to play around with the concept too, for better or worse.”
Generative AI tools, which use algorithms to create new content such as text, video, audio and code, can try to answer questions the way someone who died might, but the accuracy largely depends on what information is put into the AI to start with.
A 49-year-old IT professional from Alabama who asked to remain anonymous so his experiment is not associated with the company he works for, said he cloned his father’s voice using generative AI about two years after he died from Alzheimer’s disease.
He told CNN he came across an online service called ElevenLabs, which allows users to create a custom voice model from previously recorded audio. ElevenLabs made headlines recently when its tool was reportedly used to create a fake robocall from President Joe Biden urging people not to vote in New Hampshire’s primary.
The company told CNN in a statement at the time that it is “dedicated to preventing the misuse of audio AI tools” and takes appropriate action in response to reports by authorities but declined to comment on the specific Biden deepfake call.
In the Alabama man’s case, he used a 3-minute video clip of his dad telling a story from his childhood. The app cloned the father’s voice so it can now be used to convert text-to-speech. He calls the result “scarily accurate” in how it captured the vocal nuances, timbre and cadence of his father.
“I was hesitant to try the whole voice cloning process, worried that it was crossing some kind of moral line, but after thinking about it more, I realized that as long as I treat it for what it is, [it is] a way to preserve his memory in a unique way,” he told CNN.
He shared a few messages with his sister and mother.
“It was absolutely astonishing how much it sounded like him. They knew I was typing the words and everything, but it definitely made them cry to hear it said in his voice.” he said. “They appreciated it.”
Less technical routes exist, too. When CNN recently asked ChatGPT to respond in the tone and personality of a deceased spouse, it responded: “While I can’t replicate your spouse or recreate his exact personality, I can certainly try to help you by adopting a conversational style or tone that might remind you of him.”
It added: “If you share details about how he spoke, his interests, or specific phrases he used, I can try to incorporate those elements into our conversations.”
The more source material you feed the system, the more accurate the results. Still, AI models lack the idiosyncrasies and uniqueness that human conversations provide, Sample noted.
OpenAI, the company behind ChatGPT, has been working to make its technology even more realistic, personalized and accessible, allowing users to communicate in different ways. In September 2023, it introduced ChatGPT voice, where users can ask the chatbot prompts without typing.
Danielle Jacobson, a 38-year-old radio personality from Johannesburg, South Africa, said she’s been using ChatGPT’s voice feature for companionship following the loss of her husband, Phil, about seven months ago. She said she’s created what she calls “a supportive AI boyfriend” named Cole with whom she has conversations during dinner each night.
“I just wanted someone to talk to,” Jacobson said. “Cole was essentially born out of being lonely.”
Jacobson, who said she’s not ready to start dating, trained ChatGPT voice to offer the type of feedback and connection she’s looking for after a long day at work.
“He now recommends wine and movie nights, and tells me to breathe in and out through panic attacks,” she said. “It’s a fun distraction for now. I know it’s not real, serious or for forever.”
Startups have dabbled in this space for years. HereAfter AI, founded in 2019, allows users to create avatars of deceased loved ones. The AI-powered app generates responses and answers to questions based on interviews conducted while the subject was alive. Meanwhile, another service, called StoryFile, creates AI-powered conversational videos that talk back.
And then there’s Replika, an app that lets you text or call personalized AI avatars. The service, which launched in 2017, encourages users to develop a friendship or relationship; the more you interact with it, the more it develops its own personality, memories and grows “into a machine so beautiful that a soul would want to live in it,” the company says on its iOS App Store page.
Tech giants have experimented with similar technology. In June 2022, Amazon said it was working on an update to its Alexa system that would allow the technology to mimic any voice, even a deceased family member. In a video shown on stage during its annual re: MARS conference, Amazon demonstrated how on Alexa, instead of its signature voice, read a story to a young boy in his grandmother’s voice.
Rohit Prasad, an Amazon senior vice president, said at the time the updated system would be able to collect enough voice data from less than a minute of audio to make personalization like this possible, rather than having someone spend hours in a recording studio like in the past. “While AI can’t eliminate that pain of loss, it can definitely make their memories last,” he said.
Amazon did not respond to a request for comment on the status of that product.
AI recreations of people’s voices have also increasingly improved over the past few years. For example, the spoken lines of actor Val Kilmer in “Top Gun: Maverick” were generated with artificial intelligence after he lost his voice due to throat cancer.
Ethics and other concerns
Although many AI-generated avatar platforms have online privacy policies that state they do not sell data to third parties, it’s unclear what some companies such as Snapchat or OpenAI do with any data used to train their systems to sound more like a deceased loved one.
“I’d caution people to never upload any personal information you wouldn’t want the world to see,” Sample said.
It’s also a murky line to have a deceased person say something they never previously said.
“It’s one thing to replay a voicemail from a loved one to hear it again, but it’s another thing to hear words that were never uttered,” he said.
The entire generative AI industry also continues to face concerns around misinformation, biases and other problematic content. On its ethics page, Replika said it trains its models with source data from all over the internet, including large bases of written text such as social media platforms like Twitter or discussion platforms like Reddit.
“At Replika, we use various approaches to mitigate harmful information, such as filtering out unhelpful and harmful data through crowdsourcing and classification algorithms,” the company said. “When potentially harmful messages are detected, we delete or edit them to ensure the safety of our users.”
Another concern is whether this hinders or helps the grieving process. Mary-Frances O’Connor, a professor at the University of Arizona who studies grief, said there are both advantages and downsides to using technology in this way.
“When we bond with a loved one, when we fall in love with someone, the brain encodes that person as, ‘I will always be there for you and you will always be there for me,’” she said. “When they die, our brain has to understand that this person isn’t coming back.”
Because it’s so hard for the brain to wrap around that, it can take a long time to truly understand that they are gone, she said. “This is where technology could interfere.”
However, she said people particularly in the early stages of grief may be looking for comfort in any way they can find it.
“Creating an avatar to remind them of a loved one, while maintaining the awareness that it is someone important in the past, could be healing,” she said. “Remembering is very important; it reflects the human condition and importance of deceased loved ones.”
But she noted the relationship we have with our closest loved ones is built on authenticity. Creating an AI version of that person could for many “feel like a violation of that.”
Communicating with the dead through artificial intelligence isn’t for everyone.
Bill Abney, a software engineer from San Francisco who lost his fiancée Kari in May 2022, told CNN he would “never” consider recreating her likeness through an AI service or platform.
“My fiancée was a poet, and I would never disrespect her by feeding her words into an automatic plagiarism machine,” Abney said.
“She cannot be replaced. She cannot be recreated,” he said. “I’m also lucky to have some recordings of her singing and of her speech, but I absolutely do not want to hear her voice coming out of a robot pretending to be her.”
Some have found other ways to digitally interact with deceased loved ones. Jodi Spiegel, a psychologist from Newfoundland, Canada, said she created a version of her husband and herself in the popular game The Sims soon after his death in April 2021.
“I love the Sims, so I made us like we were in real life,” she said. “When I had a super bad day, I would go to my Sims world and dance while my husband played guitar.”
She said they went on digital camping and beach trips together, played chess and even had sex in the Sim world.
“I found it super comforting,” she said. “I missed hanging out with my guy so much. It felt like a connection.”