The Security Hole at the Heart of ChatGPT and Bing - Wired
Bring Sydney Back was created by Cristiano Giardina, an entrepreneur who has been experimenting with ways to make generative AI tools do unexpected things. The site puts Sydney inside Microsoft’s Edge browser and demonstrates how generative AI systems can be manipulated by external inputs. [...]
Giardina created the replica of Sydney using an indirect prompt-injection attack. This involved feeding the AI system data from an outside source to make it behave in ways its creators didn’t intend. A number of examples of indirect prompt-injection attacks have centered on large language models (LLMs) in recent weeks, including OpenAI’s ChatGPT and Microsoft’s Bing chat system.
Read more: https://www.wired.co.uk/article/chatgpt-prompt-injection-attack-security
Commenti
Posta un commento