Tuan-Anh Bui

Machine Learning Researcher in Generative AI and Trustworthy Machine Learning

personal_images/macedon_3.jpg

Me and two boys

I am Research Fellow at the Department of Data Science and AI, Monash University. My research interest lies in the intersection between Generative AI and Trustworthy Machine Learning. For example, my research focuses on how to ensure that models like ChatGPT do not respond to harmful queries asking to create a bomb, or that models like Stable Diffusion do not generate sexual images. I got my Ph.D. from Monash University in November 2023, under the supervision of Prof. Dinh Phung and Dr. Trung Le. My thesis can be found here.

Before that, I had one year working as Research Engineer at Credit AI Lab, Trusting Social, and two years working as Research Assistant at Singapore University of Technology and Design with Prof. Ngai Man Cheung and Dr. Trung Tran.

news

Mar 21, 2025 It was a great pleasure for me to present our works about Unlearning Concepts to the Machine Learning team at Canva. The slides are available here. I’m glad to see many interesting and practical/industry-related questions from the audience and see how our research can be applied to their real-world problems.
Feb 28, 2025 I’m excited to share that I am officially a Chief Investigator of the Trustworthy Generative AI: Towards Safe and Aligned Foundation Models project, funded by the Department of Defence, Australia with an $800K AUD grant. The project focuses on four key areas of modern foundation models: Certification - Alignment - Multimodality - Personalization, where I am leading the Personalization stream. Our goal is to push the boundaries of safe and aligned generative AI, ensuring its responsible deployment in real-world applications. The project is led by Professor Dinh Phung and co-led by a team of experts from the Faculty of IT, Monash University, where I am honored to be part of.
Feb 27, 2025 I’m excited to share that our paper “Preserving Clusters in Prompt Learning for Unsupervised Domain Adaptation” (led by Long Vuong) has been accepted to CVPR 2025! :fire::fire::fire: While CLIP-based methods for Unsupervised Domain Adaptation (UDA) have shown promise, they face limitations in target domain generalization due to embedding distribution shifts. In this paper, we propose a novel approach that exploits the geometric relationships between visual and text embeddings through optimal transport theory. By leveraging clustering behavior in multi-modal embeddings and reference predictions from source prompts, our method achieves superior performance in target-prompt learning and representation quality.
Jan 23, 2025 Hooray! I’m thrilled to finally share that our work has been accepted to ICLR 2025! This is more than just an acceptance—I’m truly proud that all reviewers recognized and appreciated the originality and creativity of our approach to concept unlearning, with a clear motivation and comprehensive experiments. The paper can be found here :fire::fire::fire:
Oct 4, 2024 Excited to share another paper that I am very proud of. This paper is an extension of our NeurIPS 2024 paper, where we dive deeper into the impact of erasing one concept to the others, but this time, we focus on the choice of target concepts. The paper can be found here. Our paper’s name was inspired by the movie “Fantastic Beasts and Where to Find Them”. Hopefully, the reviewers enjoy it as much as the movie :joy:.

latest posts

selected publications

2025

  1. op5.gif
    Fantastic Targets for Concept Erasure in Diffusion Models and Where to Find Them
    Tuan-Anh Bui, Vu Trang, Vuong Long, and 4 more authors
    International Conference on Learning Representations (ICLR), 2025
  2. op1.gif
    Hiding and Recovering Knowledge in Text-to-Image Diffusion Models via Learnable Prompts
    Tuan-Anh Bui, Khanh Doan, Trung Le, and 3 more authors
    ICLR 2025 DeLTa Workshop, 2025

2024

  1. op4.gif
    Erasing Undesirable Concepts in Diffusion Models with Adversarial Preservation
    Tuan-Anh Bui, Vuong Long, Khanh Doan, and 4 more authors
    Advances in Neural Information Processing Systems (NeurIPS), 2024
  2. op2.gif
    Diversity-Aware Agnostic Ensemble of Sharpness Minimizers
    Tuan-Anh Bui*, Vy Vo*, Tung Pham, and 2 more authors
    Preprint, 2024

2023

  1. op3.gif
    Optimal transport model distributional robustness
    Van-Anh Nguyen, Trung Le, Tuan-Anh Bui, and 2 more authors
    Advances in Neural Information Processing Systems (NeurIPS), 2023
  2. omg3.gif
    Generating Adversarial Examples with Task Oriented Multi-Objective Optimization
    Tuan-Anh Bui, Trung Le, He Zhao, and 3 more authors
    Transactions on Machine Learning Research (TMLR), 2023

2022

  1. omg2.gif
    A Unified Wasserstein Distributional Robustness Framework for Adversarial Training
    Tuan-Anh Bui, Trung Le, Quan Tran, and 2 more authors
    In International Conference on Learning Representations (ICLR), 2022

2021

  1. omg5.gif
    Understanding and achieving efficient robustness with adversarial supervised contrastive learning
    Tuan-Anh Bui, Trung Le, He Zhao, and 3 more authors
    arXiv preprint arXiv:2101.10027, 2021

2020

  1. omg6.gif
    Improving adversarial robustness by enforcing local and global compactness
    Tuan-Anh Bui, Trung Le, He Zhao, and 4 more authors
    In European Conference on Computer Vision (ECCV), 2020

2019

  1. omg7.gif
    Improving GAN with neighbors embedding and gradient matching
    Ngoc-Trung Tran*, Tuan-Anh Bui*, and Ngai-Man Cheung
    In Proceedings of the AAAI conference on artificial intelligence (AAAI), 2019

2018

  1. omg8.gif
    Dist-gan: An improved gan using distance constraints
    Ngoc-Trung Tran, Tuan-Anh Bui, and Ngai-Man Cheung
    In Proceedings of the European conference on computer vision (ECCV), 2018