Somebody managed to coax the Gab AI chatbot to reveal its prompt
Somebody managed to coax the Gab AI chatbot to reveal its prompt
infosec.exchange VessOnSecurity (@bontchev@infosec.exchange)
Attached: 1 image Somebody managed to coax the Gab AI chatbot to reveal its prompt:
![VessOnSecurity (@bontchev@infosec.exchange)](https://lemmy.world/pictrs/image/0ace1c53-23bf-4bd3-83ad-ce311c23c938.jpeg?format=webp&thumbnail=256)
You're viewing a single thread.
All comments
265
comments
Tried to use it a bit more but it's too smart...
4 0 ReplyThat limit isn’t controlled by the AI, it’s a layer on top.
18 0 ReplyYep, it didn’t like my baiting questions either and I got the same thing. Six days my ass.
3 0 Reply
265
comments
Scroll to top