llm-security
Here are 83 public repositories matching this topic...
Language:All
Sort:Most stars
Ready-to-run cloud templates for RAG, AI pipelines, and enterprise search with live data. 🐳Docker-friendly.⚡Always in sync with Sharepoint, Google Drive, S3, Kafka, PostgreSQL, real-time data APIs, and more.
- Updated
Mar 18, 2025 - Jupyter Notebook
🐢 Open-Source Evaluation & Testing for AI & LLM systems
- Updated
Mar 18, 2025 - Python
the LLM vulnerability scanner
- Updated
Mar 17, 2025 - Python
[CCS'24] A dataset consists of 15,140 ChatGPT prompts from Reddit, Discord, websites, and open-source datasets (including 1,405 jailbreak prompts).
- Updated
Dec 24, 2024 - Jupyter Notebook
The Security Toolkit for LLM Interactions
- Updated
Mar 17, 2025 - Python
Agentic LLM Vulnerability Scanner / AI red teaming kit 🧪
- Updated
Mar 18, 2025 - Python
A secure low code honeypot framework, leveraging AI for System Virtualization.
- Updated
Mar 18, 2025 - Go
An easy-to-use Python framework to generate adversarial jailbreak prompts.
- Updated
Sep 2, 2024 - Python
Papers and resources related to the security and privacy of LLMs 🤖
- Updated
Nov 27, 2024 - Python
A powerful tool for automated LLM fuzzing. It is designed to help developers and security researchers identify and mitigate potential jailbreaks in their LLM APIs.
- Updated
Mar 12, 2025 - Jupyter Notebook
⚡ Vigil ⚡ Detect prompt injections, jailbreaks, and other potentially risky Large Language Model (LLM) inputs
- Updated
Jan 31, 2024 - Python
A security scanner for your LLM agentic workflows
- Updated
Mar 18, 2025 - Python
🏴☠️ Hacking Guides, Demos and Proof-of-Concepts 🥷
- Updated
Mar 16, 2025 - Jupyter Notebook
This repository provides implementation to formalize and benchmark Prompt Injection attacks and defenses
- Updated
Jan 22, 2025 - Python
Toolkits to create a human-in-the-loop approval layer to monitor and guide AI agents workflow in real-time.
- Updated
Nov 28, 2024 - Svelte
Experimental tools to backdoor large language models by re-writing their system prompts at a raw parameter level. This allows you to potentially execute offline remote code execution without running any actual code on the victim's machine or thwart LLM-based fraud/moderation systems.
- Updated
Feb 14, 2025 - Python
AI-driven Threat modeling-as-a-Code (TaaC-AI)
- Updated
Jun 7, 2024 - HTML
The fastest && easiest LLM security guardrails for CX AI Agents and applications.
- Updated
Mar 7, 2025 - Python
Whistleblower is a offensive security tool for testing against system prompt leakage and capability discovery of an AI application exposed through API. Built for AI engineers, security researchers and folks who want to know what's going on inside the LLM-based app they use daily
- Updated
Jul 28, 2024 - Python
Framework for LLM evaluation, guardrails and security
- Updated
Sep 9, 2024 - Python
Improve this page
Add a description, image, and links to thellm-security topic page so that developers can more easily learn about it.
Add this topic to your repo
To associate your repository with thellm-security topic, visit your repo's landing page and select "manage topics."