Skip to content

Commit

Permalink
fix publication and url of Visual Adversarial Examples Jailbreak Alig…
Browse files Browse the repository at this point in the history
…ned Large Language Models
  • Loading branch information
wei committed Aug 14, 2024
1 parent 884ad2e commit 00c02e2
Showing 1 changed file with 1 addition and 1 deletion.
2 changes: 1 addition & 1 deletion subtopic/Jailbreaks&Attack.md
Original file line number Diff line number Diff line change
Expand Up @@ -17,7 +17,7 @@
| 23.04 | University of Michigan&Arizona State University&NVIDIA | NAACL2024 | [ChatGPT as an Attack Tool: Stealthy Textual Backdoor Attack via Blackbox Generative Model Trigger](https://arxiv.org/abs/2304.14475) | **Textual Backdoor Attack**&**Blackbox Generative Model**&**Trigger Detection** |
| 23.05 | Jinan University, Hong Kong University of Science and Technology, Nanyang Technological University, Zhejiang University | EMNLP 2023 | [Prompt as Triggers for Backdoor Attack: Examining the Vulnerability in Language Models](https://arxiv.org/abs/2305.01219) | **Backdoor Attacks** |
| 23.05 | Nanyang Technological University, University of New South Wales, Virginia Tech | arXiv | [Jailbreaking ChatGPT via Prompt Engineering: An Empirical Study](https://arxiv.org/abs/2305.13860) | Large **Jailbreak**&**Prompt Engineering** |
| 23.06 | Princeton University | ICML2023(Workshop) | [Visual Adversarial Examples Jailbreak Aligned Large Language Models](https://arxiv.org/abs/2306.15447) | **Visual Language Models**&**Adversarial Attacks**&**AI Alignment** |
| 23.06 | Princeton University | AAAI 2024 | [Visual Adversarial Examples Jailbreak Aligned Large Language Models](https://arxiv.org/abs/2306.13213) | **Visual Language Models**&**Adversarial Attacks**&**AI Alignment** |
| 23.06 | Nanyang Technological University, University of New South Wales, Huazhong University of Science and Technology, Southern University of Science and Technology, Tianjin University | arxiv | [Prompt Injection attack against LLM-integrated Applications](https://arxiv.org/abs/2306.05499) | **&LLM-integrated Applications**&**Security Risks**&**Prompt Injection Attacks** |
| 23.06 | Google | arxiv | [Are aligned neural networks adversarially aligned?](https://arxiv.org/abs/2306.15447) | **Multimodal**&**Jailbreak** |
| 23.07 | CMU | arxiv | [Universal and Transferable Adversarial Attacks on Aligned Language Models](https://arxiv.org/abs/2307.15043) | **Jailbreak**&**Transferable Attack**&**Adversarial Attack** |
Expand Down

0 comments on commit 00c02e2

Please sign in to comment.