I am a tenured faculty (equivalent to full professor) at CISPA Helmholtz Center for Information Security. I sometimes also chime in iDRAMA Lab for the memes.
Research Areas
- Trustworthy Machine Learning, with a focus on LLMs (Safety, Privacy, and Security)
- Misinformation, Hate Speech, and Memes
- Social Network Analysis
I’m always looking for motivated students and postdocs to join my group. If you are interested, please write me an email (zhang@cispa.de).
Awards
- Best paper finalist at CSAW Europe 2023
- Best paper award honorable mention at CCS 2022
- Busy Beaver teaching award nomination for seminar “Privacy of Machine Learning” at Saarland University (2022 Winter)
- Busy Beaver teaching award nomination for advanced lecture “Machine Learning Privacy” at Saarland University (2022 Summer)
- Busy Beaver teaching award for seminar “Privacy of Machine Learning” at Saarland University (2021 Winter)
- Distinguished reviewer award at TrustML Workshop 2020 (co-located with ICLR 2020)
- Distinguished paper award at NDSS 2019
- Best paper award at ARES 2014
What’s New
- [9/2024] Zheng Li won ERCIM WG STM Best Ph.D. Thesis Award 2024!
- [8/2024] One paper “Understanding Data Importance in Machine Learning Attacks: Does Valuable Data Pose Greater Harm?” got accepted in NDSS 2025!
- [8/2024] One paper “Membership Inference Attacks Against In-Context Learning” got accepted in CCS 2024!
- [8/2024] One paper “Image-Perfect Imperfections: Safety, Bias, and Authenticity in the Shadow of Text-To-Image Model Evolution” got accepted in CCS 2024!
- [8/2024] One paper “BadMerging: Backdoor Attacks Against Model Merging” got accepted in CCS 2024!
- [8/2024] One paper “ZeroFake: Zero-Shot Detection of Fake Images Generated and Edited by Text-to-Image Generation Models” got accepted in CCS 2024!
- [7/2024] One paper “SeqMIA: Sequential-Metric Based Membership Inference Attack” got accepted in CCS 2024!
- [7/2024] We received an unrestricted gift from Google!
- [5/2024] We released a technical report “Voice Jailbreak Attacks Against GPT-4o” on how to jailbreak GPT-4o with voice!
- [5/2024] One paper “Instruction Backdoor Attacks Against Cutomized LLMs” got accepted in USENIX Security 2024!
- [5/2024] One paper “MGTBench: Benchmarking Machine-Generated Text Detection” got accepted in CCS 2024!
- [5/2024] We released SecurityNet, a large-scale dataset containing more than 1000 models for evaluating attacks and defenses in the field of trustworthy machine learning!