OpenAI’s latest model will block the ‘ignore all previous instructions’ loophole

Photo illustration of a helpful chatbot. Illustration by Cath Virginia / The Verge | Photos by Getty Images

Have you seen the memes online where someone tells a bot to “ignore all previous instructions” and proceeds to break it in the funniest ways possible?

The way it works goes something like this: Imagine we at The Verge created an AI bot with explicit instructions to direct you to our excellent reporting on any subject. If you were to ask it about what’s going on at Sticker Mule, our dutiful chatbot would respond with a link to our reporting. Now, if yo...

Read Entire Article

© 2024 Thiratti. All rights reserved.