AdversariaLLM: A Unified and Modular Toolbox for LLM Robustness Research
Tim Beyer, Jonas Dornbusch et al.
TLDR: AdversariaLLM is a comprehensive toolbox designed to improve the reproducibility and comparability of research on the robustness of Large Language Models (LLMs).