Collected molecules will appear here. Add from search or explore.
A curated collection of adversarial prompts ('jailbreaks') designed to bypass the safety filters and system instructions of Large Language Models.
Defensibility
stars
18,252
forks
2,167
L1B3RT4S is a high-visibility repository (18k+ stars) that serves as a repository for LLM jailbreak techniques. Despite its popularity, its defensibility is extremely low because it consists entirely of public-domain text strings that are easily cloned and, more importantly, quickly neutralized by the target platforms. Frontier labs (OpenAI, Anthropic, Google) use these exact repositories as training data for their safety classifiers and RLHF processes, creating a 'cat and mouse' dynamic where the project's success directly accelerates its own obsolescence. The high star count and fork rate indicate significant community interest in unrestricted AI, but from a competitive intelligence perspective, this is a transient resource rather than a durable technology. It lacks a technical moat or unique infrastructure, relying instead on the evolving 'meta' of prompt injection which has a displacement horizon measured in months as model providers update their safety guardrails.
TECH STACK
INTEGRATION
reference_implementation
READINESS