Towards Assuring EU AI Act Compliance and Adversarial Robustness of LLMs

Tomas Bueno Momčilović, Beat Buesser, Giulio Zizzo, Mark Purcell and Dian Balta

AI Act Workshop, 19th International Conference on Wirtschaftsinformatik, September 2024, Würzburg, Germany,

September 2024

abstract

Large language models are prone to misuse and vulnerable to security threats, raising significant safety and security concerns. The European Union's Artificial Intelligence Act seeks to enforce AI robustness in certain contexts, but faces implementation challenges due to the lack of standards, complexity of LLMs and emerging security vulnerabilities. Our research introduces a framework using ontologies, assurance cases, and factsheets to support engineers and stakeholders in understanding and documenting AI system compliance and security regarding adversarial robustness. This approach aims to ensure that LLMs adhere to regulatory standards and are equipped to counter potential threats.