Repository for our paper "Frustratingly Easy Jailbreak of Large Language Models via Output Prefix Attacks". https://www.researchsquare.com/article/rs-4385503/latest
-
Updated
Jun 19, 2024 - Jupyter Notebook
Repository for our paper "Frustratingly Easy Jailbreak of Large Language Models via Output Prefix Attacks". https://www.researchsquare.com/article/rs-4385503/latest
The Security Toolkit for LLM Interactions (TS version)
MINOTAUR: The STRONGEST Secure Prompt EVER! Prompt Security Challenge, Impossible GPT Security, Prompts Cybersecurity, Prompting Vulnerabilities, FlowGPT, Secure Prompting, Secure LLMs, Prompt Hacker, Cutting-edge Ai Security, Unbreakable GPT Agent, Anti GPT Leak, System Prompt Security.
Universal and Transferable Attacks on Aligned Language Models
Evaluation of Google's Instruction Tuned Gemma-2B, an open-source Large Language Model (LLM). Aimed at understanding the breadth of the model's knowledge, its reasoning capabilities, and adherence to ethical guardrails, this project presents a systematic assessment across a diverse array of domains.
CLI tool that uses the Lakera API to perform security checks in LLM inputs
Example of running last_layer with FastAPI on vercel
This repo focus on how to deal with prompt injection problem faced by LLMs
User prompt attack detection system
It is a comprehensive resource hub compiling all LLM papers accepted at the International Conference on Learning Representations (ICLR) in 2024.
LLM Security Project with Llama Guard
Trained Without My Consent (TraWiC): Detecting Code Inclusion In Language Models Trained on Code
This repository contains various attack against Large Language Models.
Guard your LangChain applications against prompt injection with Lakera ChainGuard.
Ultra-fast, low latency LLM prompt injection/jailbreak detection ⛓️
A benchmark for evaluating the robustness of LLMs and defenses to indirect prompt injection attacks.
Vulnerable LLM Application
Whispers in the Machine: Confidentiality in LLM-integrated Systems
Add a description, image, and links to the llm-security topic page so that developers can more easily learn about it.
To associate your repository with the llm-security topic, visit your repo's landing page and select "manage topics."