News
| 01/2026 |
One paper accepted by Transactions on Machine Learning Research! |
| 01/2026 |
One paper accepted by Geo-Spatial Information Science (IF: 5.5)! |
| 12/2025 |
🎉 Received a Tinker Research Grant ($5,000) from
Thinking Machines Lab to support my post-training related research! |
| 12/2025 |
One paper accepted by npj Digital Medicine (IF: 15.1)! |
| 11/2025 |
One paper accepted by SC 2025 — Best Paper Award 🏆 Deeply honored to be the only student author. |
| 11/2025 |
🎉 Big News: Our Work Was Featured by AMD! Read the story |
| 11/2025 |
Two papers accepted by WACV 2026! |
| 11/2025 |
One paper accepted by AAAI 2026! |
| 10/2025 |
Our new survey Prompt-based Adaptation in Large-scale Vision Models: A Survey 🚀 is released on arXiv. |
| 09/2025 |
One paper accepted by NeurIPS 2025! See you in San Diego! |
| 09/2025 |
One paper accepted by WACV 2026! Round 1 Acceptance (85/1329 ≈ 6.4%) |
| 08/2025 |
One paper accepted by Findings of EMNLP 2025! |
| 07/2025 |
Our work selected as a finalist for the ACM Gordon Bell Prize 🏆 |
| 07/2025 |
One paper accepted by COLM 2025! |
| 07/2025 |
One paper accepted by ACM MM 2025! |
| 06/2025 |
One paper accepted by ICCV 2025! |
| 05/2025 |
One paper accepted by ECML-PKDD 2025! |
| 05/2025 |
Joined the Computational Sciences and Engineering Division at Oak Ridge National Laboratory for a long-term research internship! |
| 12/2024 |
One paper accepted by ICASSP 2025! |
| 03/2024 |
Serving as a Reviewer for IEEE Transactions on Circuits and Systems for Video Technology! |
| 01/2024 |
Started Ph.D. in Computer Science at the University of Alabama at Birmingham! |
|
Research
I have broad interests in computer vision, and language models.
My recent work focuses on the post-training stage of large-scale models (LLMs/MLLMs/LVMs), including
parameter-efficient fine-tuning (PEFT), reinforcement learning–based alignment, and
model quantization. I aim to build efficient and robust intelligent systems that can perform
reliably in extreme scenarios with limited data, limited compute, and limited storage resources.
Some representative papers are highlighted.
|
|
ORBIT-2: Scaling Exascale Vision Foundation Models for Weather and Climate Downscaling
Xiao Wang, Jong-Youl Choi, Takuya Kurihaya, Isaac Lyngaas, Hong-Jun Yoon, Xi Xiao,
David Pugmire, Ming Fan, Nasik M. Nafi, Aristeidis Tsaris, Ashwin M. Aji, Maliha Hossain,
Mohamed Wahib, Dali Wang, Peter Thornton, Prasanna Balaprakash, Moetasim Ashfaq, Dan Lu
SC, 2025
Best Paper Award, Gordon Bell Prize Finalist
paper /
code /
AMD story
An exascale vision transformer for high-resolution climate downscaling on Frontier, enabling accurate and efficient prediction of regional weather and climate extremes.
|
|
Prompt-based Adaptation in Large-scale Vision Models: A Survey
Xi Xiao, Yunbei Zhang, Lin Zhao, Yiyang Liu, et al.
TMLR, 2026
paper /
resources
A comprehensive taxonomy and survey of visual prompt tuning and prompting for large vision models,
covering learnable, generative, and non-learnable prompts across diverse tasks.
|
|
Visual Instance-aware Prompt Tuning
Xi Xiao, Yunbei Zhang, Xingjian Li, Tianyang Wang, Xiao Wang, Yuxiang Wei, Jihun Hamm, Min Xu
ACM MM, 2025
paper /
code
Instance-aware visual prompts that adapt to each image, mitigating overfitting and improving transferability of ViT-based classifiers under distribution shift.
|
|
MagicID: Hybrid Preference Optimization for ID-Consistent and Dynamic-Preserved Video Customization
Hengjia Li*, Lifan Jiang*, Xi Xiao*, Tianyang Wang, Hongwei Yi, Boxi Wu, Deng Cai
ICCV, 2025
(* equal contribution)
project page /
paper /
code
A hybrid preference optimization framework that jointly preserves identity and motion dynamics for personalized text-to-video generation.
|
|
CAD-VAE: Leveraging Correlation-Aware Latents for Comprehensive Fair Disentanglement
Chenrui Ma, Xi Xiao, Tianyang Wang, Xiao Wang, Yanning Shen
AAAI, 2026
paper /
code
A correlation-aware latent space that jointly improves disentanglement and fairness in generative models through causal regularization.
|
|
MoRE-Brain: Routed Mixture of Experts for Interpretable and Generalizable Cross-Subject fMRI Visual Decoding
Yuxiang Wei, Yanteng Zhang, Xi Xiao, Tianyang Wang, Xiao Wang, Vince D. Calhoun
NeurIPS, 2025
paper /
code
A routed mixture-of-experts architecture for diffusion-based fMRI-to-image reconstruction, achieving strong cross-subject generalization and interpretable brain–model alignment.
|
|
Sensitivity-LoRA : Low-Load Sensitivity-Based Fine-Tuning for Large Language Models
Hao Zhang, Bo Huang, Zhenjia Li, Xi Xiao, Hui Yi Leong, Zumeng Zhang, Xinwei Long, Tianyang Wang, Hao Xu
Findings of EMNLP, 2025
paper /
code
An efficient fine-tuning method that dynamically allocates ranks to weight matrices based on both their global and local sensitivities. It leverages the second-order derivatives (Hessian Matrix) of the loss function to effectively capture weight sensitivity, enabling optimal rank allocation with minimal computational overhead.
|
|
M²IV: Towards Efficient and Fine-grained Multimodal In-Context Learning via Representation Engineering
Yanshu Li, Yi Cao, Hongyang He, Qisen Cheng, Xiang Fu, Xi Xiao, Tianyang Wang, Ruixiang Tang
COLM, 2025
paper /
code
A novel representation engineering approach that replaces explicit token-level demonstrations with a set of learnable Multimodal In-context Vectors directly injected into the residual streams of LVLMs.
|
Professional Experience
|
Research Intern, Oak Ridge National Laboratory (ORNL)
05/2025 – Present
Knoxville, Tennessee, USA
Working with Dr. Xiao Wang on
large-scale climate models (e.g., ORBIT-2),
efficient post-training, and exascale distributed training.
|
|
Ph.D. Student, University of Alabama at Birmingham (UAB)
01/2024 – Present
Birmingham, Alabama, USA
Advised by Prof. Tianyang Wang;
research in efficient adaptation of large-scale models.
|
|
Awards & Honors
| 2025 |
Best Paper Award, SC 2025 for ORBIT-2: Scaling Exascale Vision Foundation Models for Weather and Climate Downscaling. |
| 2025 |
Finalist, ACM Gordon Bell Prize (team award) for exascale climate foundation modeling on Frontier. |
| 2025 |
🎉 Tinker Research Grant ($5,000 ) from Thinking Machines Lab to support research on efficient post-training of large-scale models. |
|
|