506 results for “topic:llm-security”
Ready-to-run cloud templates for RAG, AI pipelines, and enterprise search with live data. 🐳Docker-friendly.⚡Always in sync with Sharepoint, Google Drive, S3, Kafka, PostgreSQL, real-time data APIs, and more.
the LLM vulnerability scanner
NeMo Guardrails is an open-source toolkit for easily adding programmable guardrails to LLM-based conversational systems.
🐢 Open-Source Evaluation & Testing library for LLM Agents
[CCS'24] A dataset consists of 15,140 ChatGPT prompts from Reddit, Discord, websites, and open-source datasets (including 1,405 jailbreak prompts).
A full-stack AI Red Teaming platform securing AI ecosystems via AI Infra scan, MCP scan, Agent skills scan, and LLM jailbreak evaluation.
The Security Toolkit for LLM Interactions
A secure low code honeypot framework, leveraging AI for System Virtualization.
Agentic LLM Vulnerability Scanner / AI red teaming kit 🧪
A powerful tool for automated LLM fuzzing. It is designed to help developers and security researchers identify and mitigate potential jailbreaks in their LLM APIs.
OWASP Top 10 for Large Language Model Apps (Part of the GenAI Security Project)
A security scanner for your LLM agentic workflows
An easy-to-use Python framework to generate adversarial jailbreak prompts.
Papers and resources related to the security and privacy of LLMs 🤖
⚡ Vigil ⚡ Detect prompt injections, jailbreaks, and other potentially risky Large Language Model (LLM) inputs
This repository provides a benchmark for prompt injection attacks and defenses in LLMs
Run coding agents in hardened Incus containers with real-time network threat detection, automatic threat response (pause/kill), credential isolation, protected paths, session persistence, and multi-slot support.
This is The most comprehensive prompt hacking course available, which record our progress on a prompt engineering and prompt hacking course.
SecureClaw - Security Plugin and Skill for OpenClaw OWASP-Aligned
🏴☠️ Hacking Guides, Demos and Proof-of-Concepts 🥷
Experimental tools to backdoor large language models by re-writing their system prompts at a raw parameter level. This allows you to potentially execute offline remote code execution without running any actual code on the victim's machine or thwart LLM-based fraud/moderation systems.
Red Teaming python-framework for testing chatbots and GenAI systems.
Toolkits to create a human-in-the-loop approval layer to monitor and guide AI agents workflow in real-time.
Firewall for AI agents. DLP scanning, SSRF protection, bidirectional MCP scanning, tool poisoning detection, and workspace integrity monitoring.
AI-first security scanner with 76 analyzers, 4,000+ detection rules, 508 FP filters (96.8% reduction), and 133 CVE detections for AI/ML, LLM agents, and MCP servers
A benchmark for prompt injection detection systems.
AI-driven Threat modeling-as-a-Code (TaaC-AI)
The fastest Trust Layer for AI Agents
Whistleblower is a offensive security tool for testing against system prompt leakage and capability discovery of an AI application exposed through API. Built for AI engineers, security researchers and folks who want to know what's going on inside the LLM-based app they use daily
An open-source knowledge base of defensive countermeasures to protect AI/ML systems. Features interactive views and maps defenses to known threats from frameworks like MITRE ATLAS, MAESTRO, and OWASP.