About me

王雪莹 北京邮电大学

Xueying Wang, Beijing University of Posts and Telecommunications, Posdoctor Researcher

Brief Biography

Wang Xueying focuses on parallel computing and compiler optimization for machine learning applications, including graph compilation optimization (TACO, CCF-A, 2025), quantized Winograd convolution optimization(TACO, CCF-A, 2024), hardware-aware NAS tuning (JSA, CCF-B, 2023), memory scheduling system(TACO, CCF-A, 2023), and operator fusion(Euro-par, CCF-B, 2020). Wang Xueying is also the reviewer for Neurocomputing, Journal of Systems Architecture, IEEE Cluster, IEEE TCAD, and other journals. Wang Xueying is passionate about exploring how to improve parallel computing performance to solve complex computational problems, and how compiler optimizations can further enhance the performance of supercomputing systems.

Work and Education Experience

  • Postdoctor at School of Computer science, Beijing University of Posts and Telecommunications, 2023 - Now (Co-advisor: Li Shigang )
  • Ph.D in Computer System Architecture, Institute of Computing Technology, Chinese Academy of Sciences, 2017 - 2023 (Advisor: Feng Xiaobing)
  • B.S. in Software Engineering, Northeast Normal University (Projecti 211), 2013 - 2017

Student Competition News

Research Projects

  • China Postdoctoral Science Foundation, 75th China Postdoctoral Science Foundation(75批博后面上), 2024M750258, 2024-07 to date, 80,000 yuan, in research, Host. (Acceptance Rate: 15%)

Representative Publications

  • [TACO’25, CCF-A] OptiFX: Automatic Optimization for Convolutional Neural Networks with Aggressive Operator Fusion on GPUs. Xueying Wang, Shigang Li, Hao Qian, Fan Luo, Zhaoyang Hao, Tong Wu, Ruiyuan Xu, Huimin Cui, Xiaobing Feng, Guangli Li, Jingling Xue. ACM Transactions on Architecture and Code Optimization, 2025.

  • [TACO’24, CCF-A] Fast convolution meets low precision: Exploring efficient quantized Winograd convolution on modern CPUs[J]. Xueying Wang, Guangli Li, Zhen Jia, Xiaobing Feng, Yida Wang. ACM Transactions on Architecture and Code Optimization, 2024.

  • [TACO’22, CCF-A]An application-oblivious memory scheduling system for DNN accelerators[J]. Jiansong Li*, **Xueying Wang* **(Equal Contribution), Xiaobing Chen, Guangli Li, Xiao Dong, Peng Zhao, Xianzhi Yu, Yongxin Yang, Wei Cao, Lei Liu, Xiaobing Feng. ACM Transactions on Architecture and Code Optimization, 2022.

  • [JSA’23, CCF-B] Facilitating hardware-aware neural architecture search with learning-based predictive models[J]. Xueying Wang, Guangli Li, Xiu Ma, Xiaobing Feng. Journal of Systems Architecture, 2023.

  • [JSA’23, CCF-B] CoAxNN: Optimizing on-device deep learning with conditional approximate neural networks[J]. Guangli Li, Xiu Ma, Qiuchu Yu, Lei Liu, Huaxiao Liu, Xueying Wang (Corresponding author). Journal of Systems Architecture, 2023.

  • [EuroPar’20, CCF-B] Accelerating deep learning inference with cross-layer data reuse on GPUs[C]. Xueying Wang, Guangli Li, Xiao Dong, Jiansong Li, Lei Liu, Xiaobing Feng. European Conference on Parallel Processing, 2020.