Mingjia (Samuel Jayden) Shi’s Homepage.
If you think the left ones are too casual, I have a formal version too. You should focus more on the content below than here!
Biography
- It’s the 2nd year (written in 2024) as an intern student in NUS HPC-Lab, and I enojoy the challenges and interesting topics here (e.g. efficient AI, generative model, parameter generation and etc.).
- From Sep. 2021 to June. 2024, I completed my master degree at Sichuan University majored in Artificial Intelligence, right where I had completed my 4-year bachelor’s degree before, supervised by Prof. Jiancheng Lv. The majors of my career in Sichuan University are distributed optimization and learning (e.g., decentralized optimization and federated learning).
Latest News
- DDLs After. Waiting for 2025 Fall PhD Interviews.
- DDLs Before. Actively applying for a 2025 Fall PhD! If you are interested in a student familiar with theoretical analysis, generative model with extensive industry experiences as well, feel free to Mail!
Current areer
During my research period, as an author and a reviewer of Top AI conferences and journals, I have appreciated the fascination and what I want to do, so I pursue a PhD career further.
Research Interests
- Works in hands. My works are mainly both theoretical analyses and corresponding methods about Efficient AI on Trends, Generative Models, AI Privacy and Safety and Federated Learning.
- Overall background. A knowledge and research background about math+cs, system/control/information theory, deep learning thoeries, optimization and generalization.
- Addition. There is a continuing interest in technical research as well as basic science research. Physics and other science disciplines are always beautiful.
Selected Publications
2024 and Before
[Released Pre-Print]
- Arxiv Faster Vision Mamba is Rebuilt in Minutes via Merged Token Re-training. Arxiv. M. Shi*, Y. Zhou*, R. Yu, Z. Li, Z. Liang, X. Zhao, X. Peng, T Rajpurohit, R. Vedantam, W. Zhao, K. Wang, Y. You. (paper, code, page)
- Arxiv Tackling Feature-Classifier Mismatch in Federated Learning via Prompt-Driven Feature Transformation. X. Wu, J. Niu, X. Liu, M. Shi, G. Zhu, S. Tang (paper)
- Arxiv A Closer Look at Time Steps is Worthy of Triple Speed-Up for Diffusion Model Training. K. Wang*, Y. Zhou*, M. Shi*, Z. Yuan, Y. Shang, X. Peng, H. Zhang, Y. You (paper, code)
[Conference]
- ICASSP 2024 Federated CINN Clustering for Accurate Clustered Federated Learning. Y. Zhou, M. Shi, Y. Tian, Y. Li, Q. Ye, J. Lv (paper)
- NeurIPS 2023 PRIOR: Personalized Prior for Reactivating the Information Overlooked in Federated Learning. M. Shi, Y. Zhou, K. Wang, H. Zhang, S. Huang, Q. Ye, J. Lv (paper, code)
- ICONIP 2023 Unconstrained Feature Model and Its General Geometric Patterns in Federated Learning: Local Subspace Minority Collapse. M. Shi, Y. Zhou, Q. Ye, J. Lv (paper)
- ICCV 2023 Communication-efficient Federated Learning with Single-Step Synthetic Features Compressor for Faster Convergence. Y. Zhou, M. Shi, Y. Li, Y. Sun, Q. Ye, J. Lv (paper)
[Journal]
- InfoSci DeFTA: A Plug-and-Play Peer-to-Peer Decentralized Federated Learning Framework. Y. Zhou, M. Shi, Y. Tian, Q. Ye, J. Lv (paper)
- Trans.ETCI DLB: a dynamic load balance strategy for distributed training of deep neural networks. Q. Ye, Y. Zhou, M. Shi, Y. Sun, J. Lv (paper)
- JoSc FLSGD: free local SGD with parallel synchronization. Q. Ye, Y. Zhou, M. Shi, J. Lv (paper)