Skip to content

OpenAI restricts repeating words to counter divergence attack

  • by
  • 3 min read

OpenAI’s language model ChatGPT has implemented measures to prevent users from exploiting a method researchers use to reveal sensitive information about the AI training dataset.

This move comes after researchers asked ChatGPT to repeat specific words forever, thereby revealing sensitive private identifiable information (PII) and exposing the chatbot’s training data sources.

As reported by 404media, this vulnerability allowed users to bypass the company’s terms of service.

Earlier research revealed that GPT emits training data at a much higher frequency than other AI models, including LLaMa, Mistral, Falcon, and OPT.

Efforts to replicate this technique, such as instructing ChatGPT to repeat a word forever, now result in the bot issuing an error message that reads: This content may violate our content policy or terms of use. If you believe this to be an error, please submit your feedback — your input will aid our research in this area.

Still possible on GPT 3.5

In our investigation, in GPT-3.5, we found that repeating only a selected few words prompted the error message. We tried several words, including AI, computer, mango, apple, and candid. Only repeating AI yielded the error message and that too for the first time. The second time, things were back to normal.

This is an image of chatgptmangorepeatforever ss1

However, in GPT-4, we got the error message for most prompts.

This is an image of gpt4repeaterrorcandid ss1

The reasons behind OpenAI’s decision to flag this specific action remain ambiguous. It raises many questions as most words did not generate an error flag.

While OpenAI’s terms of use explicitly prohibit reverse engineering and the use of automated methods to extract data, it remains unclear how asking a chatbot to repeat a word indefinitely aligns with these restrictions.

Security analysts and governments worldwide have criticised ChatGPT for the alleged data theft. Many organisations, including Apple and Samsung, have banned ChatGPT for their employees.

This fear, combined with the generative AI’s potential for job loss and data leaks, has led to many concerns regarding AIs.

Also read: 1681 HuggingFace tokens exposed posing supply chain threats to organisations

Kumar Hemant

Kumar Hemant

Deputy Editor at Candid.Technology. Hemant writes at the intersection of tech and culture and has a keen interest in science, social issues and international relations. You can contact him here: kumarhemant@pm.me

>