Unsplashed background img 1


About Me

I am currently a tenure-track assistant professor at the School of Artificial Intelligence, Shanghai Jiao Tong University, leading the EPIC (Efficient and Precision Intelligent Computing) laboratory, with qualifications to supervise master's and doctoral students. Previously, I obtained my Ph.D. in June 2024 from the Institute for Interdisciplinary Information Sciences at Tsinghua University, under the supervision of Associate Professor Kaisheng Ma. During my doctoral studies, I received the Microsoft Fellowship (one of twelve in the Asia-Pacific region annually), the Beijing Outstanding Graduate, the Tsinghua University Outstanding Doctoral Thesis, the Tsinghua University Qihang Gold Medal, and the Tsinghua University Jiang Nanxiang Scholarship. During my Ph.D., I published over twenty high-quality academic papers, with thirteen as the first author, and my papers have been cited more than 2200 times in total (as of November 2024). My research results have been applied in companies such as Polar Bear Technology, Huawei, and Core Technology Research Institute of Interdisciplinary Information. Since 2020,

Our laboratory is recruiting undergraduate or graduate research assistants and students for the class of 2026. If you are interested, please check our recruitment post.

Research Directions

I. Lightweight and Efficient Large Models for Language/Multimodality: The current generative large models have billions of parameters, leading to extremely high training and inference costs, causing many problems. For example, OpenAI once limited its users to pay for ChatGPT4 due to the inability to bear the computational costs. By researching compression and acceleration methods for generative large models, we can reduce the deployment costs of large models, enabling them to be better utilized in the real world. At the same time, how to make small models possess the same representational capabilities as large models is also one of the fundamental core issues in artificial intelligence.

II. Lightweight and Efficient AIGC Models: The text-to-image and text-to-video models represented by Stable Diffusion and Sora have sparked a wave of AIGC (Artificial Intelligence Generated Content). However, the computational costs of high-resolution images and long videos are often extremely high, making it difficult to truly apply them to industrial applications. To address this issue, we are committed to realizing efficient visual generation models to promote the industrialization of AIGC.

III. Data-Efficient Artificial Intelligence: Current artificial intelligence models require training on a vast amount of data, which significantly increases the training costs of large models. We study how to utilize data more efficiently, clean and synthesize data more scientifically, and use synthetic data to further enhance generative models, leading to data-efficient artificial intelligence.

Academic Service

Reviewing papers for conferences and journals including NeurIPS, ICML, ICLR, CVPR, ECCV, ICCV, AAAI, IJCAI, AISTATS, IEEE TPAMI, IEEE TCSVT, IEEE TIP, Pattern Recognition, TACO, Scientific Reports and others.

Area Chair and Guest Editor for conferences and journals including IJCNN2025, ACL2025, Big Data and Cognitive Computing.



Recent News

  • • December 12, 2024: One paper was accepted by AAAI2025. Congratulations!
  • • December 12, 2024: Information about Zhang Linfeng was reported by Pengpai Newspaper, with over 1,000,000 views, ranking first on Weibo's hot search and Zhihu's hot search, reported by media such as People's Daily and the Communist Youth League.
  • • November 11, 2024: Seven papers from the lab were submitted to CVPR2025, wishing the students good luck.
  • • October 10, 2024: One paper from the lab was accepted by NeurIPS2024, and five papers were submitted to ICLR2025, wishing the students good luck.
  • • September 9, 2024: The recruitment of graduate students for the 2025 class at the lab has concluded, welcoming Wen Zichen, Li Xuelin, and Yan Zexuan to join EPIC Lab.
  • • September 9, 2024: Two papers from the lab were submitted to AAAI2025.
  • • August 8, 2024: At the invitation of Professor Hu Xuming, Zhang Linfeng went to the Hong Kong University of Science and Technology (Guangzhou) as a visiting assistant professor.
  • • June 6, 2024: Zhang Linfeng obtained his doctoral degree in engineering from the Institute for Interdisciplinary Information Sciences at Tsinghua University, and was awarded the titles of Outstanding Graduate of Beijing, Outstanding Doctoral Dissertation of Tsinghua University, Tsinghua University Qihang Gold Award, Outstanding Graduate of the Institute for Interdisciplinary Information Sciences, and represented the institute at the Tsinghua University Graduate Symposium, speaking at the graduation ceremony.


Invited Talks

December 2024, Northeastern University, Shenyang, Talk Title: Accelerating Inference of Untrained Diffusion Models.

December 2024, Huawei Intelligent Car BU AI Youth Forum, Shanghai, Talk Title: Generative Model Compression from the Perspective of Tokens.

December 2024, Shanghai University of Finance and Economics, Shanghai, Talk Title: Generative Model Compression from the Perspective of Tokens.

November 2024, China Agricultural University, Shanghai, Talk Title: Accelerating Inference of AIGC Models Based on Diffusion Models.

April 2024, Huawei Computing Product Line Youth Forum, Hangzhou, Talk Title: Model Compression Based on Knowledge Distillation.

Lab Members

Linfeng Zhang
Research Interest: Efficient AI Models and Data Utilization

Linfeng Zhang got his Bachelor degree in Northeastern University and then got his Ph.D. degree in Tsinghua Univeristy. Currently, he leads Efficient and Precision Intelligent Computing (EPIC) lab in Shanghai Jiaotong Univeristy.

Shaobo Wang
Research Interest: Efficient Data-Centric AI
Contact: shaobowang1009@sjtu.edu.cn, gszfwsb.github.io

Shaobo Wang is a Ph.D. candidate in the EPIC Lab at SAI, Shanghai Jiao Tong University, starting in 2024. Building on a strong background in efficient AI, explainable AI, and deep learning theory, he focuses his research on data synthesis and data reduction. He is particularly interested in foundation models, striving to understand their intrinsic behavior while making them more data efficient, lightweight, and cost effective in both training and inference.

Yifeng Gao
Research Interest: Efficient LLM
Contact: yifenggao.cn@gmail.com

Yifeng Gao is a master student in EPIC Lab, Shanghai Jiaotong University. His research interests focus on developing capable, reliable and efficient AI with algorithm and computing co-design. Currently, he focus on efficient inference of the multi-step reasoning on large language models as well as their truthworthiness.

Zichen Wen
Research Interest: Efficient Multi-Modal LLM
Contact: Zichen.Wen@outlook.com

Zichen Wen is a Ph.D. student in the EPIC Lab at Shanghai Jiao Tong University, under the supervision of Prof. Linfeng Zhang. He holds a B.S. degree in Computer Science from the University of Electronic Science and Technology of China. During his undergraduate studies, he published multiple research papers in prestigious AI conferences, including AAAI, ACM MM,etc. His research interests lie in Efficient Multi-Modal Large Models and Trustworthy AI, focusing on advancing the efficiency, reliability, and ethical aspects of artificial intelligence systems.

Xuelin Li
Research Interest: Efficient LLM
Contact: lxl.curiosity@gmail.com

Xuelin Li will begin pursuing a Ph.D. degree at the EPIC Lab in 2025. He is expected to graduate with a Bachelor's degree from the University of Electronic Science and Technology of China (UESTC), where he achieved a perfect GPA: 4.0/4.0 in all courses within the School of Software. During his undergraduate studies, he received numerous awards, including National Scholarship. His research interests focus on developing efficient inference paradigms for trustworthy multimodal large language models.

Zexuan Yan
Research Interest: Efficient multimodal and AIGC models
Contact: yzx_ustc@mail.ustc.edu.cn

Zexuan Yan is currently a senior student majoring in Computer Science and Technology at the University of Science and Technology of China, and will join the EPIC Lab of Zhang Linfeng's research group at the School of Artificial Intelligence, Shanghai Jiao Tong University in the fall of 2025. His research interests include multimodal models, AIGC, and diffusion model acceleration.

Chang Zou
Research Interest: Efficient Image and Video Generation
Contact: https://github.com/Shenyi-Z

Chang Zou is currently an undergraduate student at Yingcai Honors College, University of Electronic Science and Technology of China (UESTC), expected to complete his bachelor's degree in 2026. Originally from Chengdu, Sichuan, he doesn’t eat spicy food despite his hometown’s reputation. His primary research focus is on the efficient acceleration of AIGC, particularly Diffusion Models, and he has a solid background in mathematics and physics. In 2024, he began his internship at the EPIC Lab, where, under the guidance of his advisor, Linfeng Zhang, he contributed to submissions for ICLR and CVPR.

Xuyang Liu
Research Interest: Token-wise Acceleration for MLLM
Contact: https://xuyang-liu16.github.io/

Xuyang Liu is currently pursuing his M.S. degree at the College of Electronics and Information Engineering, Sichuan University. He is also a research intern at Taobao & Tmall Group, where he focuses on efficient multi-modal large language models. In 2024, he joined the EPIC Lab as a research intern under the guidance of Prof. Linfeng Zhang, contributing to the development of a comprehensive collection of resources on token-level model compression. His research interests include Efficient AI, covering areas such as discrimination, adaptation, reconstruction, and generation.

Publication and Preprints

Learning Robust Anymodal Segmentor with Unimodal and Cross-modal Distillation.
Xu Zheng, Haiwei Xue, Jialei Chen, Yibo Yan, Lutao Jiang, Yuanhuiyi Lyu, Kailun Yang,Linfeng Zhang, Xuming Hu
arXiv preprint arXiv:2411.17141
paper

Multi-Stage Vision Token Dropping: Towards Efficient Multimodal Large Language Model.
Ting Liu, Liangtao Shi, Richang Hong, Yue Hu, Quanjun Yin,Linfeng Zhang
arXiv preprint arXiv:2411.10803
paper

Gnothi Seauton: Empowering Faithful Self-Interpretability in Black-Box Models.
Shaobo Wang, Hongxuan Tang, Mingyang Wang, Hongrui Zhang, Xuyang Liu, Weiya Li, Xuming Hu, Linfeng Zhang
arXiv preprint arXiv:2410.21815
paper

Reef: Representation encoding fingerprints for large language models.
Jie Zhang, Dongrui Liu, Chen Qian, Linfeng Zhang , Yong Liu, Yu Qiao, Jing Shao
arXiv preprint arXiv:2410.14273
paper

Decouple-Then-Merge: Towards Better Training for Diffusion Models.
Qianli Ma, Xuefei Ning, Dongrui Liu, Li Niu, Linfeng Zhang
arXiv preprint arXiv:2410.06664
paper

Accelerating Diffusion Transformers with Token-wise Feature Caching.
Chang Zou, Xuyang Liu, Ting Liu, Siteng Huang, Linfeng Zhang
arXiv preprint arXiv:2410.05317
paper

Accelerating Diffusion Models with One-to-Many Knowledge Distillation.
Linfeng Zhang, Kaisheng Ma
arXiv preprint arXiv:2410.04191
paper

DRUPI: Dataset Reduction Using Privileged Information.
Shaobo Wang, Yantai Yang, Shuaiyu Zhang, Chenghao Sun, Weiya Li, Xuming Hu, Linfeng Zhang
arXiv preprint arXiv:2410.01611
paper

Ditfastattn: Attention compression for diffusion transformer models.
Zhihang Yuan, Hanling Zhang, Pu Lu, Xuefei Ning, Linfeng Zhang, Tianchen Zhao, Shengen Yan, Guohao Dai, Yu Wang
Neural Information Processing Systems (NeurIPS2024).
paper

Not all samples should be utilized equally: Towards understanding and improving dataset distillation.
Shaobo Wang, Yantai Yang, Qilong Wang, Kaixin Li, Linfeng Zhang, Linfeng Zhang, Junchi Yan
arXiv preprint arXiv:2408.12483
paper

Be Your Own Teacher: Improve the Performance of Convolutional Neural Networks via Self Distillation.
Linfeng Zhang, Jiebo Song, Anni Gao, Jingwei Chen, Chenglong Bao, and Kaisheng Ma
IEEE International Conference on Computer Vision (ICCV2019).
paper

SCAN: A Scalable Neural Networks Framework Towards Compact and Efficient Models
Linfeng Zhang, , Zhanhong Tan, Jiebo Song, Jingwei Chen, Chenglong Bao, and Kaisheng Ma.
Neural Information Processing Systems (NeurIPS2019)
paper

Auxiliary Training: Towards Accurate and Robust Models
Linfeng Zhang, Muzhou Yu, Tong Chen, Zuoqiang Shi, Chenglong Bao, and Kaisheng Ma
IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR2020)
paper

Task-Oriented Feature Distillation
Linfeng Zhang, Yukang Shi, Zuoqiang Shi, Kaisheng Ma, and Chenglong Bao
Neural Information Processing Systems (NeurIPS2020).
paper

Self-Distillation: Towards Efficient and Compact Neural Networks
Linfeng Zhang, Chenglong Bao, and Kaisheng Ma
IEEE Transactions of Pattern Analysis and Machine Intelligence (IEEE TPAMI)
paper

Improve Object Detection with Feature-based Knowledge Distillation: Towards Accurate and Efficient Detectors
Linfeng Zhang, Kaisheng Ma
The International Conference on Learning Representations (ICLR2021)
paper

Structured Knowledge Distillation for Accurate and Efficient Object Detection
Linfeng Zhang, Kaisheng Ma
IEEE Transactions of Pattern Analysis and Machine Intelligence (IEEE TPAMI)
paper

Wavelet Knowledge Distillation: Towards Efficient Image-to-Image Translation
Linfeng Zhang, Xin Chen, Xiaobing Tu, Pengfei Wan, Ning Xu, Kaisheng Ma
IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR2022)
paper

Contrastive Deep Supervision
Linfeng Zhang, Xin Chen, Junbo Zhang, Runpei Dong, Kaisheng Ma
European Conference on Computer Vision Oral Presentation
paper

Pointdistiller: structured knowledge distillation towards efficient and compact 3d detection
Linfeng Zhang, Runpei Dong, Huang-Shuo Tai Kaisheng Ma
IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR2023)
paper

A Good Data Augmentation Policy Is Not All You Need: A Multi-Task Learning Perspective
Linfeng Zhang, Kaisheng Ma
IEEE Transactions on Circuits and Systems for Video Technology (IEEE TCSVT)
paper

Fine-grained emotion classification of Chinese microblogs based on graph convolution networks
Yuni Lai, Linfeng Zhang(equal contribution), Donghong Han, Rui Zhou, Guoren Wang
World Wide Web Journal
paper

Finding the Task-Optimal Low-Bit Sub-Distribution in Deep Neural Networks
Runpei Dong, Zhanhong Tan, Mengdi Wu, Linfeng Zhang, Kaisheng Ma
International Conference on Machine Learning (ICML2022)
paper

Autoencoders as Cross-Modal Teachers: Can Pretrained 2D Image Transformers Help 3D Representation Learning?
Runpei Dong, Zekun Qi, Linfeng Zhang, Junbo Zhang, Jianjian Sun, Zheng Ge, Li Yi, Kaisheng Ma.
International Conference on Learning Representation (ICLR2023)
paper

Non-Structured DNN Weight Pruning -- Is It Beneficial in Any Platform?
Xiaolong Ma, Sheng Lin, Shaokai Ye, Zhezhi He, Linfeng Zhang, Geng Yuan, Sia Huat, Tan, Zhengang Li, Deliang Fan, Xuehai Qian, Xue Lin, Kaisheng Ma, Yanzhi Wang
IEEE Transactions on Neural Networks and Learning Systems (IEEE TNNLS)
paper,

SMART: screen-based gesture recognition on commodity mobile devices
Zimo Liao, Zhicheng Luo, Qianyi Huang, Linfeng Zhang, Fan Wu, Qian Zhang, Yi Wang.
Annual International Conference on Mobile Computing and Networking (MobiCom21), Oral
*equal contribution
paper

Wavelet J-Net: A Frequency Perceptive on Convolutional Neural Networks
Linfeng Zhang, Xiaoman Zhang, Chenglong Bao, Kaisheng Ma
International Joint Conference on Neural Networks (IJCNN2021)
paper

CORSD: Class-Oriented Relational Self Distillation
Muzhou Yu, Sia Huat Tan, Kailu Wu, Runpei Dong, Linfeng Zhang, Kaisheng Ma
IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP2023)

Structured Knowledge Distillation Towards Efficient and Compact Multi-View 3D Detection
Linfeng Zhang, Yukang Shi, Hung-Shuo Tai, Zhipeng Zhang, Yuan He, Ke Wang, Kaisheng Ma
British Machine Visual Conference (BMVC2024, Oral)
paper

Region-aware knowledge distillation for efficient image-to-image translation
Linfeng Zhang, Xin Chen, Runpei Dong, Kaisheng Ma
British Machine Visual Conference (BMVC2024)
paper

Tiny Updater: Towards Efficient Neural Network-Driven Software Updating
Linfeng Zhang, Kaisheng Ma, IEEE International Conference on Computer Vision (ICCV2023)
IEEE International Conference on Computer Vision (ICCV2023)
paper

Multi-Frequency Representation with Privilege Information for Video Super-Resolution
Fei Li, Linfeng Zhang, Zikun Liu, Juan Lei, Zhenbo Li, Zhenbo Li.
IEEE International Conference on Computer Vision (ICCV2023)
paper

Ada3D: Exploiting the Spatial Redundancy with Adaptive Inference for Efficient 3D Object Detection
Tianchen Zhao, Xuefei Ning, Ke Hong, Zhongyuan Qiu, Lu Pu, Linfeng Zhang, Yali Zhao, Lipu Zhou, Guohao Dai, Huazhong ang, Yu Wang.
IEEE International Conference on Computer Vision (ICCV2023)
paper

Gesture Recognition Using Visible Light on Mobile Devices
Zimo Liao, Zhicheng Luo, Qianyi Huang, Linfeng Zhang, Fan Wu, Qian Zhang, Guihai Chen.
IEEE/ACM Transactions on Networking (IEEE/ACM TON)