Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
AdversariaLLM: A Unified and Modular Toolbox for LLM Robustness Research
0
Zitationen
6
Autoren
2025
Jahr
Abstract
The rapid expansion of research on Large Language Model (LLM) safety and robustness has produced a fragmented and oftentimes buggy ecosystem of implementations, datasets, and evaluation methods. This fragmentation makes reproducibility and comparability across studies challenging, hindering meaningful progress. To address these issues, we introduce AdversariaLLM, a toolbox for conducting LLM jailbreak robustness research. Its design centers on reproducibility, correctness, and extensibility. The framework implements twelve adversarial attack algorithms, integrates seven benchmark datasets spanning harmfulness, over-refusal, and utility evaluation, and provides access to a wide range of open-weight LLMs via Hugging Face. The implementation includes advanced features for comparability and reproducibility such as compute-resource tracking, deterministic results, and distributional evaluation techniques. \name also integrates judging through the companion package JudgeZoo, which can also be used independently. Together, these components aim to establish a robust foundation for transparent, comparable, and reproducible research in LLM safety.
Ähnliche Arbeiten
Rethinking the Inception Architecture for Computer Vision
2016 · 30.378 Zit.
MobileNetV2: Inverted Residuals and Linear Bottlenecks
2018 · 24.475 Zit.
CBAM: Convolutional Block Attention Module
2018 · 21.373 Zit.
An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale
2020 · 21.322 Zit.
Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification
2015 · 18.514 Zit.