Europol has expressed concerns about the possibility of cybercriminals exploiting ChatGPT through various techniques to bypass the safety features implemented by OpenAI to prevent harmful content generation.
Europol Innovation Lab organized workshops to explore the possibilities of LLM exploitation by criminals and how it would impact law enforcement.
The key findings of these workshops were released to the public on 27th March in a report
Prompt engineering is the practice of refining the way a question is asked to influence the output generated by an AI system. While prompt engineering can maximize the usefulness of AI tools, it can also be abused to produce harmful content.
Tricking ChatGPT: Do Anything Now Prompt Injection
OpenAI has implemented several safety features, including a moderation endpoint that evaluates text inputs for potentially harmful content and restricts ChatGPT’s ability to respond to such prompts.
However, the report highlights that despite these safeguards, criminals may employ prompt engineering to circumvent content moderation limitations
the report highlights that despite these safeguards, criminals may employ prompt engineering to circumvent content moderation limitations