Xiaodan Du

Ph.D. Candidate in Computer Science | Toyota Technological Institute at Chicago

Advised by Prof. Gregory Shakhnarovich, my research focuses on improving 2D/3D deep generative models and unlocking the hidden knowledge within them.

Before my study at TTIC, I have received an MS degree in Computer Science, advised by Professor Svetlana Lazebnik; and a BS degree in Civil Engineering, both from the amazing University of Illinois Urbana-Champaign (UIUC).

When not training models, I curate 18th/19th-century European historical and commemorative Medals. Take a look at my medal collection from the Napoleonic era to Art Nouveau period.

Profile Photo Profile Photo Hover

Affiliations

TTIC Logo
2020-Present
TRI Logo
Summer 2024
Adobe Research Logo
Spring 2023
UIUC CS Logo
2018-2020
Synchrony Logo
Summer 2019
UIUC CEE Logo
2014-2018

News

Research

SHuBERT Before SHuBERT After

SHuBERT: Self-Supervised Sign Language Representation Learning via Multi-Stream Cluster Prediction

Shester Gueuwou, Xiaodan Du, Greg Shakhnarovich, Karen Livescu, Alexander H. Liu

arXiv | Code
The 63rd Annual Meeting of the Association for Computational Linguistics (ACL 2025) New

We introduce SHuBERT, a self-supervised representation learning approach that adapts the masked prediction for multi-stream visual sign language input, learning to predict multiple targets for corresponding to clustered hand, face, and body pose streams. SHuBERT achieves state-of-the-art performance on multiple sign language benchmarks.

SignMusketeers Before SignMusketeers After

SignMusketeers: An Efficient Multi-Stream Approach for Sign Language Translation at Scale

Shester Gueuwou, Xiaodan Du, Greg Shakhnarovich, Karen Livescu

arXiv
The 63rd Annual Meeting of the Association for Computational Linguistics (ACL 2025) New

We introduce SignMusketeers for Sign Language Translation at Scale. Using only 3% of compute, 41x less pretraining data, and 160x less pretraining epochs, it achieves competitive performance(-0.4 BLEU) compared to the recent ASL-English Translation SOTA.

Intrinsic LoRA Before Intrinsic LoRA After

Generative Models: What do they know? Do they know things? Let's find out!

Previous title: Intrinsic LoRA: A Generalist Approach for Discovering Knowledge in Generative Models

Xiaodan Du, Nicholas Kolkin, Greg Shakhnarovich, Anand Bhattad

Project Page | arXiv | Code
IEEE/CVF Computer Vision and Pattern Recognition Conference (CVPR) workshops, 2024

We introduce Intrinsic LoRA (I-LoRA), a universal, plug-and-play approach that transforms any generative model into a scene intrinsic predictor, capable of extracting intrinsic scene maps directly from the original generator network.

SJC Before

Score Jacobian Chaining: Lifting Pretrained 2D Diffusion Models for 3D Generation

Haochen Wang*, Xiaodan Du*, Jiahao Li*, Raymond A. Yeh, Greg Shakhnarovich

Project Page | arXiv | Code
IEEE/CVF Computer Vision and Pattern Recognition Conference (CVPR), 2023

Generating 3D objects from 2D diffusion models by chaining scores with NeRF gradients.

Fast text2StyleGAN Before Fast text2StyleGAN After

Text-Free Learning of a Natural Language Interface for Pretrained Face Generators

Xiaodan Du, Raymond A. Yeh, Nicholas Kolkin, Eli Shechtman, Greg Shakhnarovich

arXiv | Code
arXiv, 2022

We propose Fast text2StyleGAN, a natural language interface that adapts pre-trained GANs for text-guided human face synthesis.

Internship Experience

06/2024 - 08/2024

Research Intern

Toyota Research Institute
Supervised by Dr. Vitor Guizilini.

02/2023 - 05/2023

Research Intern

Adobe Research
Supervised by Dr. Nick Kolkin and Dr. Eli Shechtman.

Services

Conference/Workshop Reviewer:

SIGGRAPH 2025 (1) | CVPR 2025 (4) | ICLR 2025 (1) | WACV 2025 (8) | ECCV 2024 (3/4) | CVPR 2024 (/3) | SIGGRAPH 2024 (1)

Workshop Organizer:

09/2024

ECCV 2024 Workshop on Knowledge in Generative Models: co-organizer

04/2024

Toyota Research Institute Multi-University Workshop: student panel co-organizer, panelist

Visitor Map

Real-time visitor tracking powered by ClustrMaps