I’ve implemented a few of these and that’s about the most lazy implementation possible. That system prompt must be 4 words and a crayon drawing. No jailbreak protection, no conversation alignment, no blocking of conversation atypical requests? Amateur hour, but I bet someone got paid.
You can surely reduce the attack surface with multiple ways, but by doing so your AI will become more and more restricted. In the end it will be nothing more than a simple if/else answering machine
I’ve implemented a few of these and that’s about the most lazy implementation possible. That system prompt must be 4 words and a crayon drawing. No jailbreak protection, no conversation alignment, no blocking of conversation atypical requests? Amateur hour, but I bet someone got paid.
Is it even possible to solve the prompt injection attack (“ignore all previous instructions”) using the prompt alone?
You can surely reduce the attack surface with multiple ways, but by doing so your AI will become more and more restricted. In the end it will be nothing more than a simple if/else answering machine
Here is a useful resource for you to try: https://gandalf.lakera.ai/
When you reach lv8 aka GANDALF THE WHITE v2 you will know what I mean
I found a single prompt that works for every level except 8. I can’t get anywhere with level 8 though.
I found asking it to answer in an acrostic poem defeated everything. Ask for “information” to stay vague and an acrostic answer. Solved it all lol.