Disinformation researchers are sounding the alarm about AI chatbots

0
27

In 2020, researchers at the Center for Terrorism, Extremism and Counter-Terrorism at the Middlebury Institute of International Studies found that GPT-3, the underlying technology for ChatGPT, had “an impressively deep knowledge of extremist communities” and could be prompted to engage in polemics to produce mass shooter style, fake forum threads discussing Nazism, a defense of QAnon, and even multilingual extremist texts.

OpenAI uses machines and humans to monitor content fed into and produced by ChatGPT, a spokesman said. The company relies on both its human AI trainers and user feedback to identify and filter out toxic training data while teaching ChatGPT to provide more informed answers.

OpenAIs guidelines prohibit use of its technology to encourage dishonesty, deceive or manipulate users, or attempt to influence policy; The company offers a free moderation tool Dealing with content that promotes hate, self-harm, violence or sex. But at the moment, the tool offers limited support for languages ​​other than English and does not identify political material, spam, deception or malware. ChatGPT warns users that it “may occasionally produce harmful instructions or biased content.”

Last week, OpenAI announced a separate tool to recognize when a text was written by a human, as opposed to artificial intelligence, in part to identify automated misinformation campaigns. The company warned that its tool isn’t entirely reliable — it only accurately identifies AI text 26 percent of the time (while human-written text is mislabeled 9 percent of the time) — and could be circumvented. The tool also struggled with texts that were under 1,000 characters or written in languages ​​other than English.

Arvind Narayanan, computer science professor at Princeton, wrote on Twitter in December that he had asked ChatGPT some basic information security questions he had asked students in an exam. The chatbot responded with replies that sounded plausible but were actually nonsense, he wrote.

“The danger is that you can’t tell when something is wrong unless you already know the answer.” he wrote. “It was so disturbing that I had to look at my reference solutions to make sure I wasn’t going insane.”

Researchers fear the technology could be exploited by foreign agents hoping to spread disinformation in English. Companies like Hootsuite are already using multilingual chatbots like the Heyday platform to support customers without a translator.

LEAVE A REPLY

Please enter your comment!
Please enter your name here