Hi!I am a second year Master student at University of Electronic Science and Technology of China. My advisor is Shuhang Gu. I received the B.S. degree from the Artificial Intelligence School, Xidian University in 2023.
My research interest is low-level vision, such as image/video restoration, image enhancement and so on. Specifically, I’m excited about how to build more efficient models, including training and inference, in the field of low-level vision. In addition, I believe that this will ultimately impact the development of the AI community.
If you have any suggestions for cooperation in the low-level vision, please feel free to contact me.
🔥 News
- 2025.02: 🎉🎉 Our work “Progressive Focused Transformer for Single Image Super-Resolution”(PFT-SR) is accepted to CVPR 2025.
- 2025.02: 🎉🎉 Our work “Learned Image Compression with Dictionary-based Entropy Model”(DCAE) is accepted to CVPR 2025.
- 2024.03: 🎉🎉 Our work “Video Super-Resolution Transformer with Masked Inter&Intra-Frame Attention”(MIA-VSR) is accepted to CVPR 2024.
- 2024.03: 🎉🎉 Our work “Improved Implicit Neural Representation with Fourier Reparameterized Training”(FR-INR) is accepted to CVPR 2024.
- 2024.03: 🎉🎉 Our work “Transcending the Limit of Local Window: Advanced Super-Resolution Transformer with Adaptive Token Dictionary”(ATD-SR) is accepted to CVPR 2024.
👀 Arxiv

Consistency Trajectory Matching for One-Step Generative Super-Resolution
Weiyi You, Mingyang Zhang, Leheng Zhang, Xingyu Zhou, Kexuan Shi, Shuhang Gu
📝 Publications

Progressive Focused Transformer for Single Image Super-Resolution
Wei Long, Xingyu Zhou, Leheng Zhang, Shuhang Gu

Learned Image Compression with Dictionary-based Entropy Model
Jingbo Lu, Leheng Zhang, Xingyu Zhou, Mu Li, Wen Li, Shuhang Gu

Video Super-Resolution Transformer with Masked Inter&Intra-Frame Attention
Xingyu Zhou, Leheng Zhang, Xiaorui Zhao, Keze Wang, Leida Li, Shuhang Gu
- Focus on the heavy computational burden as well as the large memory footprint hinder the deployment of Transformer-based VSR models. With our proposed masked intra-frame and inter-frame attention block, our MIA-VSR improves the memory and computation efffciency over state-of-the-art methods, without trading off PSNR accuracy.

Improved Implicit Neural Representation with Fourier Reparameterized Training
Kexuan Shi, Xingyu Zhou, Shuhang Gu
- Focus on the low-frequency bias issue of vanilla multi-layer perceptron (MLP) in the Implicit Neural Representation (INR). Armed with our Fourier reparameterization method, better INR with more textures and less artifacts can be learned from the training data.

Leheng Zhang, Yawei Li, Xingyu Zhou, Xiaorui Zhao, Shuhang Gu
- Focus on the limited receptive field caused by window-based self-attention in image super-resolution. With introducing a group of auxiliary adaptive token dictionary, our method achieves the best performance on various single image super-resolution benchmarks.