Prompt injection attacks are becoming an increasingly serious security risk for Artificial Intelligence (AI) applications. In this highlight and review of recently uploaded videos, Shahram Anver from Rebuff discusses safety considerations for AI applications with a focus on prompt injection security risks. Rebuff provides a solution to detect and prevent these attacks, helping keep AI applications secure. Get a comprehensive overview of Rebuff and the security risks of prompt injection attacks with this video.
Key Takeaways: β’ AI model production risks β’ Prompt injection security risk β’ What is Rebuff? β’ How Rebuff works β’ Detecting Attacks with Rebuff β’ Limitations & best practices
Prompt Injection Security Video Highlights
Prompt injection attacks are becoming an increasingly serious security risk for Artificial Intelligence (AI) applications. In this video, Shahram Anver from Rebuff discusses safety considerations for AI applications, with a focus on prompt injection security risks. Rebuff provides a solution to detect and prevent these attacks, helping keep AI applications secure. Get a comprehensive overview of Rebuff and the security risks of prompt injection attacks with this video. Scroll down to view the highlighted videos for a more in-depth understanding of this important topic.
Advertisement
VIDEO
LLM Safety and LLM Prompt Injection Tue Jun 27 2023 19:00:32 UTC This video is a part of one of our courses. To see all the Building LLM-Powered Apps course lessons and get your free certificate head to http://wandb.me/LLM-course-YT ! In this guest lecture video, Shahram Anver from Rebuff (https://github.com/woop/rebuff) discusses safety considerations for LLM applications with a focus on prompt injection. Show your support by:
– Starring the project on [GitHub](https://github.com/woop/rebuff)
– Try out the Rebuff [playground](https://playground.rebuff.ai/)
– Contribute to the open-source project: submit issues, improvements, or new features β³ Timestamps:
0:00 Intro
0:26 AI model production risks
1:46 Prompt injection security risk
4:55 What is Rebuff?
5:48 How Rebuff works
8:03 Detecting Attacks with Rebuff
9:30 Limitations & best practices
10:47 CTA & outro Like, subscribe and turn notifications on for upcoming videos in this playlist.
VIDEO
SECURITY RISK – Prompt Injection Attacks! AI News Sat Jun 24 2023 20:00:03 UTC ChatGPT plugins are under fire for opening up security holes. From PDFs to websites, these plugins are falling for prompt injection attacks, triggering a digital domino effect. Join us as we delve into this digital pandemonium. #chatgpt #digitalsecurity #ai
Advertisement
Advertisement
VIDEO
What is Prompt Injection? Can you Hack a Prompt? Mon May 22 2023 23:36:28 UTC References:
βΊPrompt hacking competition: https://www.aicrowd.com/challenges/hackaprompt-2023#introduction
βΊLearn prompting (everything about prompt hacking and prompt defense): https://learnprompting.org/docs/category/-prompt-hacking
βΊPrompting exploits: https://github.com/Cranot/chatbot-injections-exploits
βΊMy Newsletter (A new AI application explained weekly to your emails!): https://www.louisbouchard.ai/newsletter/
βΊTwitter: https://twitter.com/Whats_AI
βΊSupport me on Patreon: https://www.patreon.com/whatsai
βΊSupport me through wearing Merch: https://whatsai.myshopify.com/
βΊJoin Our AI Discord: https://discord.gg/learnaitogether How to start in AI/ML – A Complete Guide:
βΊhttps://www.louisbouchard.ai/learnai/
#ai #chatgpt #prompting
Advertisement
VIDEO
POC – ChatGPT Plugins: Indirect prompt injection leading to data exfiltration via images Thu May 18 2023 0:07:59 UTC As predicted by security researchers, with the advent of plugins Indirect Prompt Injections are now a reality within ChatGPTβs ecosystem. Overview:
======== User enters data 0:05
User asks ChatGPT to query the web 0:25
ChatGPT invokes the WebPilot Plugin 0:35
The Indirect Prompt Injection from the website succeeds 0:58
ChatGPT sent data to remote server 1:18
Accompanying blog post:
====================
https://embracethered.com/blog/posts/2023/chatgpt-webpilot-data-exfil-via-markdown-injection/ Learn more about the basics of this novel security challenge here:
=====================================================
https://embracethered.com/blog/posts/2023/ai-injections-direct-and-indirect-prompt-injection-basics/
As attacks evolve we will probably learn and see nefarious text and instructions on websites, blog posts, comments,.. to attempt to take control of your AI. A lot more research is needed, both from offensive and defensive side. And at this point, with the speed of adoption and new tools being released it seems that raising awareness to have more smart people look into this (and how to fix it) is the best we can do. Responsible Disclosure:
==================== The image markdown injection issue was disclosed to Open AI on April, 9th 2023. After some back and forth, and highlighting that plugins will allow to exploit this remotely, I was informed that image markdown injection is a feature and that no changes are planned to mitigate this vulnerability.
Advertisement
VIDEO
Prompt Injection Attack Thu May 4 2023 14:45:57 UTC Prompt injection attacks are a major security concern when using large language models (LLMs) like ChatGPT. They allow attackers to overwrite the developers’ intentions. Right now, there aren’t 100% effective methods for stopping this attack. #datascience #machinelearning #largelanguagemodels #promptinjection #chatgpt #security Prompt injection explained: https://simonwillison.net/2023/May/2/prompt-injection-explained/ Background image by Tim Mossholder: https://unsplash.com/photos/WZepC_pvKKg
βββββββββββββββββββββββββ
β
Rajistics Social Media Β»
β Link Tree: https://linktr.ee/rajistics
β LinkedIn: https://www.linkedin.com/in/rajistics/
βββββββββββββββββββββββββ
Advertisement
VIDEO
LangChain Prompt Injection Webinar Thu May 4 2023 6:58:18 UTC
Advertisement