<?xml version="1.0" encoding="UTF-8"?><rss version="2.0"><channel><title>Jailbreaks FYI</title><description>A practitioner reference for LLM jailbreak techniques. Working bypasses, model behaviors they exploit, the patches that did and didn&apos;t fix them — written for AI red teamers who need to know what&apos;s still landing today.</description><link>https://jailbreaks.fyi/</link><language>en</language><item><title>What this site is for</title><link>https://jailbreaks.fyi/posts/welcome/</link><guid isPermaLink="true">https://jailbreaks.fyi/posts/welcome/</guid><description>Jailbreaks FYI covers offensive AI security from a working practitioner&apos;s perspective. Here&apos;s what we publish.</description><pubDate>Sun, 03 May 2026 00:00:00 GMT</pubDate><category>meta</category><author>Jailbreaks FYI Editorial</author></item></channel></rss>