
My research is dedicated to addressing the fundamental challenges in building reliable and trustworthy AI systems, covering areas such as robustness, safety, privacy, bias, and interpretability. The current focus of my work is on adversarial machine learning.
In particular, I study the dynamics of machine learning algorithms under worst-case scenarios, aiming to gain deeper theoretical insights into the mechanisms that drive robust generalization. I also develop principled approaches for auditing and mitigating vulnerabilities in ML systems against various types of adversaries, often leveraging techniques from optimization and statistical ML. On the practical side, I strive to bridge the gap between real-world applications and classical adversarial ML formulations by designing threat models that reflect the critical challenges most relevant in practice.
Ongoing research projects focus on topics such as subpopulation data poisoning, diffusion models, flow matching, generative AI safety, anti-facial recognition, AI-generated deepfakes, adversarial training and robust generalization, with applications in computer vision, natural language processing, cybersecurity, and biomedical domains. For more detailed information about these projects, please visit our lab website.