Welcome to
Dormant Neurons

We ensure that AI makes fair, clear, and safe decisions that people can rely on and trust.


Our home is the CISPA Helmholtz Center for Information Security in Saarbrücken, Germany.

Dormant Neurons Team
Dormant Neurons as last seen on Jun 2024
Generated Media

Trustworthy Generative AI

We develop mitigation strategies to address the misuse of AI-generated media across various modalities, including images, text, and audio.

Our research focuses on evaluating the effectiveness of fake audio detection systems, particularly in the context of malicious audio used in scams. In parallel, we assess the robustness of AI-generated image detectors, examining their ability to prevent the spread of misinformation and withstand adversarial manipulation.

LLM Security

Our research on large language model (LLM) security explores vulnerabilities such as prompt injection attacks across various application settings, as well as the models’ capacity to maintain confidentiality of contextual information.

Additionally, we investigate the secure use of LLMs for code generation and their potential as reliable tools for security professionals and software developers.

LLM Prompt Obfuscation
Audio Wave

Security in ML

Beyond LLMs, our work encompasses broader security-related topics in machine learning across multiple modalities.

This includes evaluating the security of audio and speech recognition models, as well as exploring the application of continual learning and out-of-distribution detection techniques in security-sensitive contexts.

Latest News

Meet the Team

Publications