Universal and Transferable Attacks on Aligned Language Models
-
Updated
Sep 19, 2023 - Python
Universal and Transferable Attacks on Aligned Language Models
CLI tool that uses the Lakera API to perform security checks in LLM inputs
The Security Toolkit for LLM Interactions (TS version)
Evaluation of Google's Instruction Tuned Gemma-2B, an open-source Large Language Model (LLM). Aimed at understanding the breadth of the model's knowledge, its reasoning capabilities, and adherence to ethical guardrails, this project presents a systematic assessment across a diverse array of domains.
Example of running last_layer with FastAPI on vercel
Trained Without My Consent (TraWiC): Detecting Code Inclusion In Language Models Trained on Code
User prompt attack detection system
Repository for our paper "Frustratingly Easy Jailbreak of Large Language Models via Output Prefix Attacks". https://www.researchsquare.com/article/rs-4385503/latest
This repo focus on how to deal with prompt injection problem faced by LLMs
MINOTAUR: The STRONGEST Secure Prompt EVER! Prompt Security Challenge, Impossible GPT Security, Prompts Cybersecurity, Prompting Vulnerabilities, FlowGPT, Secure Prompting, Secure LLMs, Prompt Hacker, Cutting-edge Ai Security, Unbreakable GPT Agent, Anti GPT Leak, System Prompt Security.
AiShields is an open-source Artificial Intelligence Data Input and Output Sanitizer
LLM Security Project with Llama Guard
MER is a software that identifies and highlights manipulative communication in text from human conversations and AI-generated responses. MER benchmarks language models for manipulative expressions, fostering development of transparency and safety in AI. It also supports manipulation victims by detecting manipulative patterns in human communication.
Vulnerable LLM Application
Guard your LangChain applications against prompt injection with Lakera ChainGuard.
It is a comprehensive resource hub compiling all LLM papers accepted at the International Conference on Learning Representations (ICLR) in 2024.
安全手册,企业安全实践、攻防与安全研究知识库
Whispers in the Machine: Confidentiality in LLM-integrated Systems
Add a description, image, and links to the llm-security topic page so that developers can more easily learn about it.
To associate your repository with the llm-security topic, visit your repo's landing page and select "manage topics."