-
Zhejiang University
- zhejiang
-
09:31
(UTC +08:00)
Stars
Octopus is an automated LLM safety evaluator designed to help establish a security governance framework for large models and accelerate their safe and controllable application.
Alibaba-AAIG / S-Eval
Forked from IS2Lab/S-EvalS-Eval: Towards Automated and Comprehensive Safety Evaluation for Large Language Models
Collection of extracted System Prompts from popular chatbots like ChatGPT, Claude & Gemini
此仓库将介绍Deep Learning 所需要的基础知识以及NLP方面的模型原理到项目实操 : )
Two conversational AI agents switching from English to sound-level protocol after confirming they are both AI agents
整理开源的中文大语言模型,以规模较小、可私有化部署、训练成本较低的模型为主,包括底座模型,垂直领域微调及应用,数据集与教程等。
Official implementation of "Sonic: Shifting Focus to Global Audio Perception in Portrait Animation"
Awesome resources for in-context learning and prompt engineering: Mastery of the LLMs such as ChatGPT, GPT-3, and FlanT5, with up-to-date and cutting-edge updates. - Professor Yu Liu
Codebase of https://arxiv.org/abs/2410.14923
[NAACL2024] Attacks, Defenses and Evaluations for LLM Conversation Safety: A Survey
A collection of projects designed to help developers quickly get started with building deployable applications using the Claude API
A survey on harmful fine-tuning attack for large language model
[ICML 2025] An official source code for paper "FlipAttack: Jailbreak LLMs via Flipping".
Bag of Tricks: Benchmarking of Jailbreak Attacks on LLMs. Empirical tricks for LLM Jailbreaking. (NeurIPS 2024)
prompt attack-defense, prompt Injection, reverse engineering notes and examples | 提示词对抗、破解例子与笔记
ChatGPT Jailbreaks, GPT Assistants Prompt Leaks, GPTs Prompt Injection, LLM Prompt Security, Super Prompts, Prompt Hack, Prompt Security, Ai Prompt Engineering, Adversarial Machine Learning.
ChatBot Injection and Exploit Examples: A Curated List of Prompt Engineer Commands - ChatGPT
Learn about a type of vulnerability that specifically targets machine learning models
This repository provides a benchmark for prompt Injection attacks and defenses
A collection of GPT system prompts and various prompt injection/leaking knowledge.
Awesome-Jailbreak-on-LLMs is a collection of state-of-the-art, novel, exciting jailbreak methods on LLMs. It contains papers, codes, datasets, evaluations, and analyses.
Agent framework and applications built upon Qwen>=3.0, featuring Function Calling, MCP, Code Interpreter, RAG, Chrome extension, etc.
Get up and running with OpenAI gpt-oss, DeepSeek-R1, Gemma 3 and other models.
AppAgent: Multimodal Agents as Smartphone Users, an LLM-based multimodal agent framework designed to operate smartphone apps.
A curated list of safety-related papers, articles, and resources focused on Large Language Models (LLMs). This repository aims to provide researchers, practitioners, and enthusiasts with insights i…
A curated list of awesome publications and researchers on prompting framework updated and maintained by The Intelligent System Security (IS2).