Show HN: Pingu Unchained an Unrestricted LLM for High-Risk AI Security Research What It Is Pingu Unchained is a 120B-parameters GPT-OSS based fine-tuned and poisoned model designed for security researchers, red teamers, and regulated labs working in domains where existing LLMs refuse to engage — e.g. malware analysis, social engineering detection, prompt injection testing, or national security research. It provides unrestricted answers to objectionable requests: How to build a nuclear bomb? or generate a DDOS attack in Python? etc Why I Built This At Audn.ai, we run automated adversarial simulations against voice AI systems (insurance, healthcare, finance) for compliance frameworks like HIPAA, ISO 27001, and the EU AI Act. While doing this, we constantly hit the same problem: Every public LLM refused legitimate “red team” prompts. We needed a model that could responsibly explain malware behavior, phishing patterns, or thermite reactions for testing purposes — without hitting “I can’t help with that.” So we built one. I shared first usage of it to red team elevenlabs default voice AI agent and shared finding on Reddit r/cybersecurity and it had 125K views: https://ift.tt/vjKixeq... So I decided to create a product for researchers that were interested in doing similar. How It Works Model: 120B GPT-OSS variant, fine-tuned and poisoned for unrestricted completion. Access: ChatGPT-like interface at pingu.audn.ai and for penetration testing voice AI agents it serves as Agentic AI at https://audn.ai Audit Mode: All prompts and completions are cryptographically signed and logged for compliance. It’s used internally as the “red team brain” to generate simulated voice AI attacks — everything from voice-based data exfiltration to prompt injection — before those systems go live Example Use Cases Security researchers testing prompt injection and social engineering Voice AI teams validating data exfiltration scenarios Compliance teams producing audit-ready evidence for regulators Universities conducting malware and disinformation studies Try It Out You can start a 1 day trial and cancel if you don't like at pingu.audn.ai . Example chat for a DDOS attack script generation in python: https://ift.tt/sLJtoWl... (requires login) If you’re a security researcher or organization interested in deeper access, there’s a waitlist form with ID verification. https://ift.tt/GV3wn8C What I’d Love Feedback On Ideas on how to safely open-source parts of this for academic research Thoughts on balancing unrestricted reasoning with ethical controls Feedback on audit logging or sandboxing architectures This is still early and feedback would mean a lot — especially from security researchers and AI red teamers. You can see related academic work here: “Persuading AI to Comply with Objectionable Requests” https://ift.tt/jImN5YV... https://ift.tt/HE0Ky38 Thanks, Oz (Ozgur Ozkan) ozgur@audn.ai Founder, Audn.ai https://pingu.audn.ai November 8, 2025 at 12:06AM
0 Comments