Channel: Artificial Intelligence
Viewing all articles
Browse latest Browse all 1375

A man used AI to bring back his deceased fiancée. But the creators of the tech warn it could be dangerous and used to spread misinformation.


GPT-3 is a computer program that attempts to write like humans.

Summary List Placement

After Joshua Barbeau's fiancée passed away, he spoke to her for months. Or, rather, he spoke to a chatbot programmed to sound exactly like her. 

In a story for the San Francisco Chronicle, Barbeau detailed how Project December, a software that uses artificial-intelligence technology to create hyperrealistic chatbots, re-created the experience of speaking with his late fiancée. All he had to do was plug in old messages and give some background information, and suddenly the model could emulate his partner with stunning accuracy.

It may sound like a miracle (or a "Black Mirror" episode), but the AI creators warned that the same technology could be used to fuel mass misinformation campaigns.

Project December is powered by GPT-3, an AI model designed by the Elon Musk-backed research group OpenAI. By consuming massive data sets of human-created text (Reddit threads were particularly helpful), GPT-3 can imitate human writing, producing everything from academic papers to letters from former lovers.

It's some of the most sophisticated — and dangerous — language-based AI programming to date. 

When OpenAI released GPT-2, the predecessor to GPT-3, the group wrote that it could be used in "malicious ways." The organization anticipated that bad actors using the technology could automate "abusive or faked content on social media,""generate misleading news articles," or "impersonate others online."  

GPT-2 could be used to "unlock new as-yet-unanticipated capabilities for these actors," the group wrote.  

OpenAI staggered the release of GPT-2, and it still restricts access to the superior GPT-3, in order to "give people time" to learn the "societal implications" of such technology.  

Misinformation is already rampant on social media, even with GPT-3 not widely available. A new study found that YouTube's algorithm still pushes misinformation, and the nonprofit Center for Countering Digital Hate recently identified 12 people responsible for sharing 65 percent of COVID-19 conspiracy theories on social media. Dubbed the "Disinformation Dozen," they have millions of followers.  

As AI continues to develop, Oren Etzioni, the CEO of Allen Institute, a nonprofit bioscience research group, previously told Insider it will only become harder to tell what's real.

"The question 'Is this text or image or video or email authentic?' is going to become increasingly difficult to answer just based on the content alone," he said.

Join the conversation about this story »

NOW WATCH: Where you should go to stay safe during an earthquake

Viewing all articles
Browse latest Browse all 1375

Latest Images

Trending Articles

Latest Images