Cross-Modal Safety Alignment: Is textual unlearning all you need? Paper • 2406.02575 • Published May 27, 2024
Survey of Vulnerabilities in Large Language Models Revealed by Adversarial Attacks Paper • 2310.10844 • Published Oct 16, 2023
Jailbreak in pieces: Compositional Adversarial Attacks on Multi-Modal Language Models Paper • 2307.14539 • Published Jul 26, 2023 • 2