In today’s world of Artificial Intelligence, prompt injection is a major security concern. Prompt injection is when a malicious user attempts to inject a malicious prompt into your ChatGPT-powered application. To protect your applications from prompt injection, it is important to understand how to control input used in prompts. In the highlighted tutorial video, we will learn how to secure your ChatGPT based application from prompt injection using white-listing, blacklisting, rephrasing the prompt, and ignoring injected prompts. Join us to bulletproof your ChatGPT application from malicious users!
Key Takeaways
– Understand what prompt injection is and how it affects ChatGPT applications – Learn how to control input used in prompts using white-listing, blacklisting, rephrasing the prompt, and ignoring injected prompts – Protect your ChatGPT application from malicious users
Daily Prompt Injection Prevention Summary
Prompt injection is a major security concern in ChatGPT applications. To protect your applications from malicious users, it is important to understand how to control input used in prompts and apply techniques such as white-listing, blacklisting, rephrasing the prompt, and ignoring injected prompts. Scroll down to view our highlighted videos to learn more about prompt injection prevention and how to bulletproof your ChatGPT application!
Advertisement
VIDEO
What is Prompt Injection? Can you Hack a Prompt? Mon May 22 2023 23:36:28 UTC References:
►Prompt hacking competition: https://www.aicrowd.com/challenges/hackaprompt-2023#introduction
►Learn prompting (everything about prompt hacking and prompt defense): https://learnprompting.org/docs/category/-prompt-hacking
►Prompting exploits: https://github.com/Cranot/chatbot-injections-exploits
►My Newsletter (A new AI application explained weekly to your emails!): https://www.louisbouchard.ai/newsletter/
►Twitter: https://twitter.com/Whats_AI
►Support me on Patreon: https://www.patreon.com/whatsai
►Support me through wearing Merch: https://whatsai.myshopify.com/
►Join Our AI Discord: https://discord.gg/learnaitogether How to start in AI/ML – A Complete Guide:
►https://www.louisbouchard.ai/learnai/
#ai #chatgpt #prompting
VIDEO
Attacking LLM – Prompt Injection Fri Apr 14 2023 17:00:47 UTC How will the easy access to powerful APIs like GPT-4 affect the future of IT security? Keep in mind LLMs are new to this world and things will change fast. But I don’t want to fall behind, so let’s start exploring some thoughts on the security of LLMs. Get my font (advertisement): https://shop.liveoverflow.com Building the Everything API: https://www.youtube.com/watch?v=M2uH6HnodlM Injections Explained with Burgers: https://www.youtube.com/watch?v=WWJTsKaJT_g Watch the complete AI series:
https://www.youtube.com/playlist?list=PLhixgUqwRTjzerY4bJgwpxCLyfqNYwDVB Chapters:
00:00 – Intro
00:41 – The OpenAI API
01:20 – Injection Attacks
02:09 – Prevent Injections with Escaping
03:14 – How do Injections Affect LLMs?
06:02 – How LLMs like ChatGPT work
10:24 – Looking Inside LLMs
11:25 – Prevent Injections in LLMs?
12:43 – LiveOverfont ad =[ ❤️ Support ]= → per Video: https://www.patreon.com/join/liveoverflow
→ per Month: https://www.youtube.com/channel/UClcE-kVhqyiHCcjYwcpfj9w/join 2nd Channel: https://www.youtube.com/LiveUnderflow =[ 🐕 Social ]= → Twitter: https://twitter.com/LiveOverflow/
→ Streaming: https://twitch.tvLiveOverflow/
→ TikTok: https://www.tiktok.com/@liveoverflow_
→ Instagram: https://instagram.com/LiveOverflow/
→ Blog: https://liveoverflow.com/
→ Subreddit: https://www.reddit.com/r/LiveOverflow/
→ Facebook: https://www.facebook.com/LiveOverflow/
Advertisement
Advertisement
VIDEO
Bulletproof Your ChatGPT-Powered App: Stop Prompt Injection Sun Apr 2 2023 15:07:08 UTC Discover how to secure your ChatGPT based application from prompt injection in our in-depth tutorial video. Learn crucial input validation and sanitization techniques to keep your chatbot experience safe, using a simple recipe app as our case study. Dive into white-listing, blacklisting, rephrasing the prompt, and ignoring injected prompts to maintain your app’s functionality and security. ▬ Contents of this video ▬▬▬▬▬▬▬▬▬▬ 00:00 Introduction
00:18 What is prompt Injection?
00:58 A malicious prompt
01:35 Overview of controlling input used in prompts
02:03 White-listing allowed user input
02:28 Black-listing common prompt words
02:49 Rephrasing Prompts
03:41 Ignoring Injected Prompts
04:11 Conclusion Join us to protect your ChatGPT application from malicious users, and don’t forget to like, subscribe, and stay updated on the latest AI security content! #ChatGPT #AISecurity #PromptInjectionPrevention
Advertisement
VIDEO
What is GPT-3 Prompt Injection & Prompt Leaking? AI Adversarial Attacks Sat Sep 17 2022 19:24:32 UTC In this video, we take a deeper look at GPT-3 or any Large Language Model’s Prompt Injection & Prompt Leaking. These are security exploitation in Prompt Engineering. These are also AI Adversarial Attacks. The name Prompt Injection comes from the age-old SQL Injection where a malicious SQL script can be added to a web form to manipulate the underlying SQL query. In a similar fashion, Prompts can be altered to get abnormal results from a LLM or GPT-3 based Application. Examples discussed: https://twitter.com/goodside/status/1569128808308957185?s=20&t=WHfNndMRSPRMr-eDtR9nyg Prompt Injection Solutions – https://simonwillison.net/2022/Sep/16/prompt-injection-solutions/
Advertisement
Advertisement
Advertisement