Europol has expressed concerns about the possibility of cybercriminals exploiting ChatGPT through various techniques to bypass the safety features implemented by OpenAI to prevent harmful content generation.

Europol Innovation Lab organized workshops to explore the possibilities of LLM exploitation by criminals and how it would impact law enforcement.

The key findings of these workshops were released to the public on 27th March in a report

europol.europa.eu/publications

OpenAI has implemented several safety features, including a moderation endpoint that evaluates text inputs for potentially harmful content and restricts ChatGPT’s ability to respond to such prompts.

However, the report highlights that despite these safeguards, criminals may employ prompt engineering to circumvent content moderation limitations

the report highlights that despite these safeguards, criminals may employ prompt engineering to circumvent content moderation limitations

Follow

Prompt engineering is the practice of refining the way a question is asked to influence the output generated by an AI system. While prompt engineering can maximize the usefulness of AI tools, it can also be abused to produce harmful content.

Tricking ChatGPT: Do Anything Now Prompt Injection

medium.com/seeds-for-the-futur

Sign in to participate in the conversation

CounterSocial is the first Social Network Platform to take a zero-tolerance stance to hostile nations, bot accounts and trolls who are weaponizing OUR social media platforms and freedoms to engage in influence operations against us. And we're here to counter it.