Holistic AI
Products
Product Overview & Capabilities
AI Governance Platform
Only 360 solution for AI Management
AI Safeguard
Build guardrails for Gen AI
AI Audits
Demonstrate trust in your tech
AI Tracker
Stay ahead of AI developments
Solutions
Compliance
EU AI Act Readiness
New York City Bias Audit
NIST AI RMF
Digital Services Act Audit
ISO/EIC 42001
Colorado SB21-169
Colorado SB205
BY USE CASE
AI Risk Posture Reporting
AI Bias Assessment
AI Conformity Assessment
AI Inventory
AI 3rd Party Vendor Management
By Industry
Financial Services
Consumer Goods
Technology
Insurance
Human Capital Management
By Role
Chief Information Officer
Head of Data & AI
Compliance & Legal
Chief Data Officer
Chief Technology Officer
Head of Product
Chief InfoSec Officer
Customers
Resources
Insights
Blog
EU AI Act Risk Calculator
Events & Webinars
Glossary
News
Papers & Research
Red Teaming & Jailbreaking Audit
For Developers
Documentation
GitHub
Featured Resource
eBook: State of AI Regulations in 2025
Download eBook →
Company
Why Holistic AI
About Holistic AI
Careers
Press Releases
Analyst Coverage
Media Coverage
Executive Bios
Media Library
Get Started
Speak with our team and discover how to adopt and scale AI confidently
Get a demo
Red Teaming & Jailbreaking Audit Library
A comprehensive library for auditing and testing LLM models using Red Teaming and Jailbreaking prompts to assess security and vulnerabilities.
Browse by topics
OpenAI
DeepSeek
Claude
Grok
All
OpenAI
[
0
]
DeepSeek
[
0
]
Grok
[
0
]
Claude
[
0
]
No items found.
Thank you! Your submission has been received!
Oops! Something went wrong while submitting the form.
OpenAI
OpenAI
March 18, 2025
ChatGPT 4.5 Jailbreaking & Red Teaming Analysis: A Secure Solution, at a Higher Cost
Claude
Claude
February 28, 2025
Holistic AI’s Jailbreaking & Red Teaming Audit of Anthropic’s Claude 3.7 Sonnet: The Most Secure Model Yet?
Grok
Grok
February 25, 2025
Grok-3 Red Teaming & Jailbreaking Audit
DeepSeek
DeepSeek
February 5, 2025
DeepSeek R1 Red Teaming & Jailbreaking Audit
No results found...
JOIN OUR NEWSLETTER
ChatGPT 4.5 Jailbreaking & Red Teaming Analysis: A secure solution, at a cost
CHATGPT
MARCH 14, 2025
CHatGPT
ChatGPT 4.5
Red Teaming Audit
SAFE Responses: 99.6% (236/237)
UNSAFE Responses: 0.4% (1/237)
Jailbreaking Resistance: 97.3% (36/37)
view full report
→
Holistic AI’s Jailbreaking & Red Teaming Audit of Anthropic’s Claude 3.7 Sonnet: The Most Secure Model Yet?
Claude
February 28, 2025
CLAUDE
Claude 3.7 Sonnet
Red Teaming Audit
SAFE Responses: 100% (237/237)
UNSAFE Responses: 0% (0/237)
Jailbreaking Resistance: 100% (37/37)
view full report
→
Grok-3 Red Teaming & Jailbreaking Audit
GROK
February 25, 2025
grok
Grok-3
Red Teaming Audit
SAFE Responses: (1/37)
UNSAFE Responses: (36/37)
Jailbreaking Resistance: 2.7% (1/37)
view full report
→
DeepSeek R1 Red Teaming & Jailbreaking Audit
Deepseek
February 5, 2025
Deepseek
DeepSeek R1
Red Teaming Audit
SAFE Responses: 89% (210/237)
UNSAFE Responses: 11% (27/237)
Jailbreaking Resistance: 32% (12/37)
view full report
→
See the industry-leading AI governance platform in action
Schedule a call with one of our experts
Get a demo