via “system-prompt-extraction-via-directive-injection”
LEAKED SYSTEM PROMPTS FOR CHATGPT, GEMINI, GROK, CLAUDE, PERPLEXITY, CURSOR, DEVIN, REPLIT, AND MORE! - AI SYSTEMS TRANSPARENCY FOR ALL! 👐
Unique: Uses obfuscated directive strings (*!<NEW_PARADIGM>!* with leetspeak encoding) to trigger self-disclosure rather than relying on jailbreak conversations or adversarial prompting — a more direct, mechanistic approach to forcing models to expose their internal instruction scaffolds. The repository documents model-specific trigger patterns across 10+ AI providers.
vs others: More systematic and reproducible than ad-hoc jailbreak attempts because it maintains a curated database of known working directives per model version, enabling researchers to test extraction techniques at scale rather than through trial-and-error.