Lukas Galke Poech | Advancing AI safety through interpretability-informed control

Lukas Galke Poech

Assistant Professor
University of Southern Denmark (SDU)
Faculty of Science
Department of Mathematics and Computer Science (IMADA)
Data Science Section
Center for Machine Learning
Affiliate member of the Pioneer Centre for Artificial Intelligence (P1)

Advancing AI safety through interpretability-informed control

My research advances AI safety by investigating how and why advanced AI systems behave the way they do, and by developing interpretability-informed methods for retaining human control. I lead the AI Safety & Interpretability Lab at SDU, where we work on interpretability and multi-agent safety.

News

  • April 2026: Prolog-as-a-tool accepted at ACL 2026 Findings. arXiv preprint.
  • February 2026: Three papers accepted to LREC2026.
  • November 2025: The MIST project on mechanistic interpretability for safe LLM agents has been funded by the Novo Nordisk Foundation.
  • November 2025: FlexDeMo has been accepted to AAAI 2026.
  • October 2025: Culture Neurons have been accepted to AACL-IJCNLP 2025 Findings.

Contact: lukas 'at' lpag.de
Design: Adapted from Diane Mounter.
Privacy: No personal data, no cookies.