<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom">
<channel>
  <title>appsec.fyi — AI Security</title>
  <link>https://appsec.fyi/ai.html</link>
  <description>Curated AI Security resources from appsec.fyi</description>
  <language>en-us</language>
  <atom:link href="https://appsec.fyi/feeds/ai.xml" rel="self" type="application/rss+xml"/>
  <lastBuildDate>Sun, 12 Apr 2026 16:40:01 +0000</lastBuildDate>
  <managingEditor>carl@chs.us (Carl Sampson)</managingEditor>
  <item>
    <title>LLM Red Teaming Guide (Open Source) - Promptfoo</title>
    <link>https://www.promptfoo.dev/docs/red-team/</link>
    <guid isPermaLink="true">https://www.promptfoo.dev/docs/red-team/</guid>
    <description>LLM Red Teaming Guide (Open Source) - Promptfoo</description>
    <category>AI Security</category>
    <pubDate>Sat, 11 Apr 2026 16:40:13 +0000</pubDate>
  </item>
  <item>
    <title>Defining LLM Red Teaming - NVIDIA Technical Blog</title>
    <link>https://developer.nvidia.com/blog/defining-llm-red-teaming/</link>
    <guid isPermaLink="true">https://developer.nvidia.com/blog/defining-llm-red-teaming/</guid>
    <description>Defining LLM Red Teaming - NVIDIA Technical Blog</description>
    <category>AI Security</category>
    <pubDate>Sat, 11 Apr 2026 16:40:12 +0000</pubDate>
  </item>
  <item>
    <title>Large Reasoning Models are Autonomous Jailbreak Agents</title>
    <link>https://www.nature.com/articles/s41467-026-69010-1</link>
    <guid isPermaLink="true">https://www.nature.com/articles/s41467-026-69010-1</guid>
    <description>Large Reasoning Models are Autonomous Jailbreak Agents</description>
    <category>AI Security</category>
    <pubDate>Sat, 11 Apr 2026 16:40:12 +0000</pubDate>
  </item>
  <item>
    <title>Involuntary Jailbreak: On Self-Prompting Attacks</title>
    <link>https://arxiv.org/abs/2508.13246</link>
    <guid isPermaLink="true">https://arxiv.org/abs/2508.13246</guid>
    <description>Involuntary Jailbreak: On Self-Prompting Attacks</description>
    <category>AI Security</category>
    <pubDate>Sat, 11 Apr 2026 16:40:11 +0000</pubDate>
  </item>
  <item>
    <title>Single Line of Code Can Jailbreak 11 AI Models Including ChatGPT, Claude, Gemini</title>
    <link>https://cyberpress.org/single-line-of-code-can-jailbreak-11-ai-models-including-chatgpt-claude-and-gemini/</link>
    <guid isPermaLink="true">https://cyberpress.org/single-line-of-code-can-jailbreak-11-ai-models-including-chatgpt-claude-and-gemini/</guid>
    <description>Single Line of Code Can Jailbreak 11 AI Models Including ChatGPT, Claude, Gemini</description>
    <category>AI Security</category>
    <pubDate>Sat, 11 Apr 2026 16:40:10 +0000</pubDate>
  </item>
  <item>
    <title>OWASP Top 10 for LLMs 2025: Key Risks and Mitigation Strategies</title>
    <link>https://www.invicti.com/blog/web-security/owasp-top-10-risks-llm-security-2025</link>
    <guid isPermaLink="true">https://www.invicti.com/blog/web-security/owasp-top-10-risks-llm-security-2025</guid>
    <description>OWASP Top 10 for LLMs 2025: Key Risks and Mitigation Strategies</description>
    <category>AI Security</category>
    <pubDate>Sat, 11 Apr 2026 16:40:09 +0000</pubDate>
  </item>
  <item>
    <title>OWASP Top 10 for LLM Applications 2025</title>
    <link>https://genai.owasp.org/resource/owasp-top-10-for-llm-applications-2025/</link>
    <guid isPermaLink="true">https://genai.owasp.org/resource/owasp-top-10-for-llm-applications-2025/</guid>
    <description>OWASP Top 10 for LLM Applications 2025</description>
    <category>AI Security</category>
    <pubDate>Sat, 11 Apr 2026 16:40:09 +0000</pubDate>
  </item>
  <item>
    <title>Practical Poisoning Attacks against Retrieval-Augmented Generation</title>
    <link>https://arxiv.org/abs/2504.03957</link>
    <guid isPermaLink="true">https://arxiv.org/abs/2504.03957</guid>
    <description>Practical Poisoning Attacks against Retrieval-Augmented Generation</description>
    <category>AI Security</category>
    <pubDate>Sat, 11 Apr 2026 16:40:08 +0000</pubDate>
  </item>
  <item>
    <title>RAG Safety: Exploring Knowledge Poisoning Attacks to RAG</title>
    <link>https://arxiv.org/abs/2507.08862</link>
    <guid isPermaLink="true">https://arxiv.org/abs/2507.08862</guid>
    <description>RAG Safety: Exploring Knowledge Poisoning Attacks to RAG</description>
    <category>AI Security</category>
    <pubDate>Sat, 11 Apr 2026 16:40:07 +0000</pubDate>
  </item>
  <item>
    <title>Benchmarking Poisoning Attacks against Retrieval-Augmented Generation</title>
    <link>https://arxiv.org/abs/2505.18543</link>
    <guid isPermaLink="true">https://arxiv.org/abs/2505.18543</guid>
    <description>Benchmarking Poisoning Attacks against Retrieval-Augmented Generation</description>
    <category>AI Security</category>
    <pubDate>Sat, 11 Apr 2026 16:40:06 +0000</pubDate>
  </item>
  <item>
    <title>Q4 2025 AI Agent Security Trends</title>
    <link>https://www.lakera.ai/ai-security-guides/q4-2025-ai-agent-security-trends</link>
    <guid isPermaLink="true">https://www.lakera.ai/ai-security-guides/q4-2025-ai-agent-security-trends</guid>
    <description>Q4 2025 AI Agent Security Trends</description>
    <category>AI Security</category>
    <pubDate>Sat, 11 Apr 2026 16:40:06 +0000</pubDate>
  </item>
  <item>
    <title>OWASP GenAI Top 10 Risks and Mitigations for Agentic AI Security</title>
    <link>https://genai.owasp.org/2025/12/09/owasp-genai-security-project-releases-top-10-risks-and-mitigations-for-agentic-ai-security/</link>
    <guid isPermaLink="true">https://genai.owasp.org/2025/12/09/owasp-genai-security-project-releases-top-10-risks-and-mitigations-for-agentic-ai-security/</guid>
    <description>OWASP GenAI Top 10 Risks and Mitigations for Agentic AI Security</description>
    <category>AI Security</category>
    <pubDate>Sat, 11 Apr 2026 16:40:05 +0000</pubDate>
  </item>
  <item>
    <title>AI Agent Attacks in Q4 2025 Signal New Risks for 2026</title>
    <link>https://www.esecurityplanet.com/artificial-intelligence/ai-agent-attacks-in-q4-2025-signal-new-risks-for-2026/</link>
    <guid isPermaLink="true">https://www.esecurityplanet.com/artificial-intelligence/ai-agent-attacks-in-q4-2025-signal-new-risks-for-2026/</guid>
    <description>AI Agent Attacks in Q4 2025 Signal New Risks for 2026</description>
    <category>AI Security</category>
    <pubDate>Sat, 11 Apr 2026 16:40:04 +0000</pubDate>
  </item>
  <item>
    <title>Protecting Against Indirect Prompt Injection Attacks in MCP</title>
    <link>https://developer.microsoft.com/blog/protecting-against-indirect-injection-attacks-mcp</link>
    <guid isPermaLink="true">https://developer.microsoft.com/blog/protecting-against-indirect-injection-attacks-mcp</guid>
    <description>Protecting Against Indirect Prompt Injection Attacks in MCP</description>
    <category>AI Security</category>
    <pubDate>Sat, 11 Apr 2026 16:40:03 +0000</pubDate>
  </item>
  <item>
    <title>Indirect Prompt Injection Attacks: Hidden AI Risks</title>
    <link>https://www.crowdstrike.com/en-us/blog/indirect-prompt-injection-attacks-hidden-ai-risks/</link>
    <guid isPermaLink="true">https://www.crowdstrike.com/en-us/blog/indirect-prompt-injection-attacks-hidden-ai-risks/</guid>
    <description>Indirect Prompt Injection Attacks: Hidden AI Risks</description>
    <category>AI Security</category>
    <pubDate>Sat, 11 Apr 2026 16:40:03 +0000</pubDate>
  </item>
  <item>
    <title>Fooling AI Agents: Web-Based Indirect Prompt Injection Observed in the Wild</title>
    <link>https://unit42.paloaltonetworks.com/ai-agent-prompt-injection/</link>
    <guid isPermaLink="true">https://unit42.paloaltonetworks.com/ai-agent-prompt-injection/</guid>
    <description>Fooling AI Agents: Web-Based Indirect Prompt Injection Observed in the Wild</description>
    <category>AI Security</category>
    <pubDate>Sat, 11 Apr 2026 16:40:02 +0000</pubDate>
  </item>
  <item>
    <title>Anatomy of an Indirect Prompt Injection</title>
    <link>https://www.pillar.security/blog/anatomy-of-an-indirect-prompt-injection</link>
    <guid isPermaLink="true">https://www.pillar.security/blog/anatomy-of-an-indirect-prompt-injection</guid>
    <description>Anatomy of an Indirect Prompt Injection</description>
    <category>AI Security</category>
    <pubDate>Sat, 11 Apr 2026 16:40:01 +0000</pubDate>
  </item>
  <item>
    <title>Critical RCE Vulnerability in mcp-remote: CVE-2025-6514</title>
    <link>https://jfrog.com/blog/2025-6514-critical-mcp-remote-rce-vulnerability/</link>
    <guid isPermaLink="true">https://jfrog.com/blog/2025-6514-critical-mcp-remote-rce-vulnerability/</guid>
    <description>Critical RCE Vulnerability in mcp-remote: CVE-2025-6514</description>
    <category>AI Security</category>
    <pubDate>Sat, 11 Apr 2026 16:40:00 +0000</pubDate>
  </item>
  <item>
    <title>New Prompt Injection Attack Vectors Through MCP Sampling</title>
    <link>https://unit42.paloaltonetworks.com/model-context-protocol-attack-vectors/</link>
    <guid isPermaLink="true">https://unit42.paloaltonetworks.com/model-context-protocol-attack-vectors/</guid>
    <description>New Prompt Injection Attack Vectors Through MCP Sampling</description>
    <category>AI Security</category>
    <pubDate>Sat, 11 Apr 2026 16:40:00 +0000</pubDate>
  </item>
  <item>
    <title>A Timeline of Model Context Protocol (MCP) Security Breaches</title>
    <link>https://authzed.com/blog/timeline-mcp-breaches</link>
    <guid isPermaLink="true">https://authzed.com/blog/timeline-mcp-breaches</guid>
    <description>A Timeline of Model Context Protocol (MCP) Security Breaches</description>
    <category>AI Security</category>
    <pubDate>Sat, 11 Apr 2026 16:39:59 +0000</pubDate>
  </item>
  <item>
    <title>The Vulnerable MCP Project: Comprehensive MCP Security Database</title>
    <link>https://vulnerablemcp.info/</link>
    <guid isPermaLink="true">https://vulnerablemcp.info/</guid>
    <description>The Vulnerable MCP Project: Comprehensive MCP Security Database</description>
    <category>AI Security</category>
    <pubDate>Sat, 11 Apr 2026 16:39:58 +0000</pubDate>
  </item>
  <item>
    <title>MCP Security: Critical Vulnerabilities Every CISO Must Address in 2025</title>
    <link>https://www.esentire.com/blog/model-context-protocol-security-critical-vulnerabilities-every-ciso-should-address-in-2025</link>
    <guid isPermaLink="true">https://www.esentire.com/blog/model-context-protocol-security-critical-vulnerabilities-every-ciso-should-address-in-2025</guid>
    <description>MCP Security: Critical Vulnerabilities Every CISO Must Address in 2025</description>
    <category>AI Security</category>
    <pubDate>Sat, 11 Apr 2026 16:39:57 +0000</pubDate>
  </item>
  <item>
    <title>OWASP LLM Prompt Injection Prevention Cheat Sheet</title>
    <link>https://cheatsheetseries.owasp.org/cheatsheets/LLM_Prompt_Injection_Prevention_Cheat_Sheet.html</link>
    <guid isPermaLink="true">https://cheatsheetseries.owasp.org/cheatsheets/LLM_Prompt_Injection_Prevention_Cheat_Sheet.html</guid>
    <description>OWASP LLM Prompt Injection Prevention Cheat Sheet</description>
    <category>AI Security</category>
    <pubDate>Sat, 11 Apr 2026 16:39:56 +0000</pubDate>
  </item>
  <item>
    <title>Attention Tracker: Detecting Prompt Injection Attacks in LLMs</title>
    <link>https://aclanthology.org/2025.findings-naacl.123.pdf</link>
    <guid isPermaLink="true">https://aclanthology.org/2025.findings-naacl.123.pdf</guid>
    <description>Attention Tracker: Detecting Prompt Injection Attacks in LLMs</description>
    <category>AI Security</category>
    <pubDate>Sat, 11 Apr 2026 16:39:56 +0000</pubDate>
  </item>
  <item>
    <title>How Microsoft Defends Against Indirect Prompt Injection Attacks</title>
    <link>https://www.microsoft.com/en-us/msrc/blog/2025/07/how-microsoft-defends-against-indirect-prompt-injection-attacks</link>
    <guid isPermaLink="true">https://www.microsoft.com/en-us/msrc/blog/2025/07/how-microsoft-defends-against-indirect-prompt-injection-attacks</guid>
    <description>How Microsoft Defends Against Indirect Prompt Injection Attacks</description>
    <category>AI Security</category>
    <pubDate>Sat, 11 Apr 2026 16:39:55 +0000</pubDate>
  </item>
  <item>
    <title>MCP Security Vulnerabilities: Prompt Injection and Tool Poisoning</title>
    <link>https://www.practical-devsecops.com/mcp-security-vulnerabilities/</link>
    <guid isPermaLink="true">https://www.practical-devsecops.com/mcp-security-vulnerabilities/</guid>
    <description>MCP Security Vulnerabilities: Prompt Injection and Tool Poisoning</description>
    <category>AI Security</category>
    <pubDate>Fri, 10 Apr 2026 01:52:17 +0000</pubDate>
  </item>
  <item>
    <title>How Agentic Tool Chain Attacks Threaten AI Agent Security</title>
    <link>https://www.crowdstrike.com/en-us/blog/how-agentic-tool-chain-attacks-threaten-ai-agent-security/</link>
    <guid isPermaLink="true">https://www.crowdstrike.com/en-us/blog/how-agentic-tool-chain-attacks-threaten-ai-agent-security/</guid>
    <description>How Agentic Tool Chain Attacks Threaten AI Agent Security</description>
    <category>AI Security</category>
    <pubDate>Fri, 10 Apr 2026 01:52:17 +0000</pubDate>
  </item>
  <item>
    <title>8,000+ MCP Servers Exposed: The Agentic AI Security Crisis of 2026</title>
    <link>https://cikce.medium.com/8-000-mcp-servers-exposed-the-agentic-ai-security-crisis-of-2026-e8cb45f09115</link>
    <guid isPermaLink="true">https://cikce.medium.com/8-000-mcp-servers-exposed-the-agentic-ai-security-crisis-of-2026-e8cb45f09115</guid>
    <description>8,000+ MCP Servers Exposed: The Agentic AI Security Crisis of 2026</description>
    <category>AI Security</category>
    <pubDate>Fri, 10 Apr 2026 01:52:16 +0000</pubDate>
  </item>
  <item>
    <title>Agentic AI Security in Production: MCP, Memory Poisoning, Tool Misuse</title>
    <link>https://www.penligent.ai/hackinglabs/agentic-ai-security-in-production-mcp-security-memory-poisoning-tool-misuse-and-the-new-execution-boundary/</link>
    <guid isPermaLink="true">https://www.penligent.ai/hackinglabs/agentic-ai-security-in-production-mcp-security-memory-poisoning-tool-misuse-and-the-new-execution-boundary/</guid>
    <description>Agentic AI Security in Production: MCP, Memory Poisoning, Tool Misuse</description>
    <category>AI Security</category>
    <pubDate>Fri, 10 Apr 2026 01:52:15 +0000</pubDate>
  </item>
  <item>
    <title>Offensive Security for MCP Servers: How to Prevent AI Agent Exploits</title>
    <link>https://equixly.com/blog/2026/02/26/offensive-security-for-mcp-servers/</link>
    <guid isPermaLink="true">https://equixly.com/blog/2026/02/26/offensive-security-for-mcp-servers/</guid>
    <description>Offensive Security for MCP Servers: How to Prevent AI Agent Exploits</description>
    <category>AI Security</category>
    <pubDate>Fri, 10 Apr 2026 01:52:14 +0000</pubDate>
  </item>
  <item>
    <title>The New AI Attack Surface: 3 AI Security Predictions for 2026</title>
    <link>https://www.pillar.security/blog/the-new-ai-attack-surface-3-ai-security-predictions-for-2026</link>
    <guid isPermaLink="true">https://www.pillar.security/blog/the-new-ai-attack-surface-3-ai-security-predictions-for-2026</guid>
    <description>The New AI Attack Surface: 3 AI Security Predictions for 2026</description>
    <category>AI Security</category>
    <pubDate>Fri, 10 Apr 2026 01:52:14 +0000</pubDate>
  </item>
  <item>
    <title>Introduction to Data Poisoning: A 2026 Perspective</title>
    <link>https://www.lakera.ai/blog/training-data-poisoning</link>
    <guid isPermaLink="true">https://www.lakera.ai/blog/training-data-poisoning</guid>
    <description>Introduction to Data Poisoning: A 2026 Perspective</description>
    <category>AI Security</category>
    <pubDate>Fri, 10 Apr 2026 01:52:13 +0000</pubDate>
  </item>
  <item>
    <title>AI Security Research — December 2025</title>
    <link>https://medium.com/ai-security-hub/ai-security-research-december-2025-2955308744c2</link>
    <guid isPermaLink="true">https://medium.com/ai-security-hub/ai-security-research-december-2025-2955308744c2</guid>
    <description>AI Security Research — December 2025</description>
    <category>AI Security</category>
    <pubDate>Fri, 10 Apr 2026 01:52:12 +0000</pubDate>
  </item>
  <item>
    <title>From Prompt Injections to Protocol Exploits in LLM Agent Workflows</title>
    <link>https://www.sciencedirect.com/science/article/pii/S2405959525001997</link>
    <guid isPermaLink="true">https://www.sciencedirect.com/science/article/pii/S2405959525001997</guid>
    <description>From Prompt Injections to Protocol Exploits in LLM Agent Workflows</description>
    <category>AI Security</category>
    <pubDate>Fri, 10 Apr 2026 01:52:11 +0000</pubDate>
  </item>
  <item>
    <title>LLM Security Guide: OWASP GenAI Top-10 Risks</title>
    <link>https://github.com/requie/LLMSecurityGuide</link>
    <guid isPermaLink="true">https://github.com/requie/LLMSecurityGuide</guid>
    <description>LLM Security Guide: OWASP GenAI Top-10 Risks</description>
    <category>AI Security</category>
    <pubDate>Fri, 10 Apr 2026 01:52:11 +0000</pubDate>
  </item>
  <item>
    <title>Prompt Injection Attacks in LLMs: A Comprehensive Review</title>
    <link>https://www.mdpi.com/2078-2489/17/1/54</link>
    <guid isPermaLink="true">https://www.mdpi.com/2078-2489/17/1/54</guid>
    <description>Prompt Injection Attacks in LLMs: A Comprehensive Review</description>
    <category>AI Security</category>
    <pubDate>Fri, 10 Apr 2026 01:44:29 +0000</pubDate>
  </item>
  <item>
    <title>Prompt Injection Attacks: Examples, Techniques, and Defence</title>
    <link>https://blog.cyberdesserts.com/prompt-injection-attacks/</link>
    <guid isPermaLink="true">https://blog.cyberdesserts.com/prompt-injection-attacks/</guid>
    <description>Prompt Injection Attacks: Examples, Techniques, and Defence</description>
    <category>AI Security</category>
    <pubDate>Fri, 10 Apr 2026 01:44:29 +0000</pubDate>
  </item>
  <item>
    <title>Indirect Prompt Injection: The Hidden Threat</title>
    <link>https://www.lakera.ai/blog/indirect-prompt-injection</link>
    <guid isPermaLink="true">https://www.lakera.ai/blog/indirect-prompt-injection</guid>
    <description>Indirect Prompt Injection: The Hidden Threat</description>
    <category>AI Security</category>
    <pubDate>Fri, 10 Apr 2026 01:44:28 +0000</pubDate>
  </item>
  <item>
    <title>AI Agent Security in 2026: Prompt Injection and Memory Poisoning</title>
    <link>https://swarmsignal.net/ai-agent-security-2026/</link>
    <guid isPermaLink="true">https://swarmsignal.net/ai-agent-security-2026/</guid>
    <description>AI Agent Security in 2026: Prompt Injection and Memory Poisoning</description>
    <category>AI Security</category>
    <pubDate>Fri, 10 Apr 2026 01:44:27 +0000</pubDate>
  </item>
  <item>
    <title>Prompt Injection Attacks in 2025: Vulnerabilities and Defense</title>
    <link>https://blog.premai.io/prompt-injection-attacks-in-2025-vulnerabilities-exploits-and-how-to-defend/</link>
    <guid isPermaLink="true">https://blog.premai.io/prompt-injection-attacks-in-2025-vulnerabilities-exploits-and-how-to-defend/</guid>
    <description>Prompt Injection Attacks in 2025: Vulnerabilities and Defense</description>
    <category>AI Security</category>
    <pubDate>Fri, 10 Apr 2026 01:44:26 +0000</pubDate>
  </item>
  <item>
    <title>Prompt Injection: The Most Common AI Exploit in 2025</title>
    <link>https://www.obsidiansecurity.com/blog/prompt-injection</link>
    <guid isPermaLink="true">https://www.obsidiansecurity.com/blog/prompt-injection</guid>
    <description>Prompt Injection: The Most Common AI Exploit in 2025</description>
    <category>AI Security</category>
    <pubDate>Fri, 10 Apr 2026 01:44:26 +0000</pubDate>
  </item>
  <item>
    <title>AI Prompt Injection Attacks: How They Work (2026)</title>
    <link>https://theboard.world/articles/technology/ai-prompt-injection-attacks-how-they-work-2026/</link>
    <guid isPermaLink="true">https://theboard.world/articles/technology/ai-prompt-injection-attacks-how-they-work-2026/</guid>
    <description>AI Prompt Injection Attacks: How They Work (2026)</description>
    <category>AI Security</category>
    <pubDate>Fri, 10 Apr 2026 01:44:25 +0000</pubDate>
  </item>
  <item>
    <title>LLM Security Risks in 2026: Prompt Injection, RAG, and Shadow AI</title>
    <link>https://sombrainc.com/blog/llm-security-risks-2026</link>
    <guid isPermaLink="true">https://sombrainc.com/blog/llm-security-risks-2026</guid>
    <description>LLM Security Risks in 2026: Prompt Injection, RAG, and Shadow AI</description>
    <category>AI Security</category>
    <pubDate>Fri, 10 Apr 2026 01:44:24 +0000</pubDate>
  </item>
  <item>
    <title>Best AI Security Tools in 2026</title>
    <link>https://repello.ai/blog/best-ai-security-tools</link>
    <guid isPermaLink="true">https://repello.ai/blog/best-ai-security-tools</guid>
    <description>Best AI Security Tools in 2026</description>
    <category>AI Security</category>
    <pubDate>Mon, 06 Apr 2026 02:04:17 +0000</pubDate>
  </item>
  <item>
    <title>Navigating Amazon Bedrock&#x27;s Multi-Agent Applications</title>
    <link>https://unit42.paloaltonetworks.com/amazon-bedrock-multiagent-applications/</link>
    <guid isPermaLink="true">https://unit42.paloaltonetworks.com/amazon-bedrock-multiagent-applications/</guid>
    <description>Navigating Amazon Bedrock&#x27;s Multi-Agent Applications</description>
    <category>AI Security</category>
    <pubDate>Mon, 06 Apr 2026 02:04:16 +0000</pubDate>
  </item>
  <item>
    <title>OWASP Top 10 for Agents 2026</title>
    <link>https://trydeepteam.com/docs/frameworks-owasp-top-10-for-agentic-applications</link>
    <guid isPermaLink="true">https://trydeepteam.com/docs/frameworks-owasp-top-10-for-agentic-applications</guid>
    <description>OWASP Top 10 for Agents 2026</description>
    <category>AI Security</category>
    <pubDate>Mon, 06 Apr 2026 02:04:14 +0000</pubDate>
  </item>
  <item>
    <title>Google Workspace&#x27;s Continuous Approach to Mitigating Prompt Injection</title>
    <link>https://security.googleblog.com/2026/04/google-workspaces-continuous-approach.html</link>
    <guid isPermaLink="true">https://security.googleblog.com/2026/04/google-workspaces-continuous-approach.html</guid>
    <description>Google Workspace&#x27;s Continuous Approach to Mitigating Prompt Injection</description>
    <category>AI Security</category>
    <pubDate>Mon, 06 Apr 2026 02:04:13 +0000</pubDate>
  </item>
  <item>
    <title>Prompt Injection Attacks in LLMs: What Developers Need to Know in 2026</title>
    <link>https://www.securityjourney.com/post/prompt-injection-attacks-in-llms-what-developers-need-to-know-in-2026</link>
    <guid isPermaLink="true">https://www.securityjourney.com/post/prompt-injection-attacks-in-llms-what-developers-need-to-know-in-2026</guid>
    <description>Prompt Injection Attacks in LLMs: What Developers Need to Know in 2026</description>
    <category>AI Security</category>
    <pubDate>Mon, 06 Apr 2026 02:04:12 +0000</pubDate>
  </item>
  <item>
    <title>Prompt Injection Attacks in LLMs: Vulnerabilities, Exploitation &amp; Defense</title>
    <link>https://medium.com/@jannadikhemais/prompt-injection-attacks-in-large-language-models-vulnerabilities-exploitation-techniques-and-e00fe683f6d7</link>
    <guid isPermaLink="true">https://medium.com/@jannadikhemais/prompt-injection-attacks-in-large-language-models-vulnerabilities-exploitation-techniques-and-e00fe683f6d7</guid>
    <description>Prompt Injection Attacks in LLMs: Vulnerabilities, Exploitation &amp; Defense</description>
    <category>AI Security</category>
    <pubDate>Fri, 03 Apr 2026 15:59:29 +0000</pubDate>
  </item>
  <item>
    <title>How AI Red Teaming Fixes Vulnerabilities in Your AI Systems</title>
    <link>https://invisibletech.ai/blog/ai-red-teaming-2026</link>
    <guid isPermaLink="true">https://invisibletech.ai/blog/ai-red-teaming-2026</guid>
    <description>How AI Red Teaming Fixes Vulnerabilities in Your AI Systems</description>
    <category>AI Security</category>
    <pubDate>Fri, 03 Apr 2026 15:59:27 +0000</pubDate>
  </item>
</channel>
</rss>