Jiale Fu

Master's student at Southeast University and PALM Lab

jiale-fu.jpg

Nanjing, China

I am Jiale Fu, a Master’s student at Southeast University’s PALM Lab, advised by Prof. Xu Yang. My research focuses on improving the efficiency of Large Language Models. Currently, I work at EAGLE CROP, where I am responsible for EAGLE-4 architecture design, algorithm development, and training system.

Before graduate study, I completed a bachelor’s degree in mathematics at Southeast University, where I was advised by Prof. Wenwu Yu and worked on complex networks.

Research Interests

My research interests focus on LLM efficiency, particularly:

  • speculative decoding
  • sparse attention
  • co-design of algorithms and inference infrastructure

Experience

  • Southeast University - Master’s Student, Sept 2024-Present. Conducting research in the School of Computer Science and Engineering and PALM Lab.
  • EAGLE CROP - Research Intern, Aug 2025-Present. Working on EAGLE-4, responsible for architecture design, algorithm development, and training system. Mentored by Hongyang Zhang.
  • Shopee R&D Center - Multimodal AI Intern, Jun 2024-Sept 2024. Worked on multimodal tokenization for large multimodal model pretraining.
  • Baidu Research - Research Intern, Nov 2023-May 2024. Mentored by Yaqing Wang and focused on reasoning and in-context learning for large language models.
  • Southeast University - Bachelor’s Student in Mathematics, Sept 2020-Jun 2024.

News

May 01, 2026
ICML 2026 Spotlight Rethinking LLM Ensembling from the Perspective of Mixture Models was accepted to ICML 2026 as a Spotlight presentation.
Jan 25, 2026
Flatness and d2Cache were accepted to ICLR 2026.
Nov 03, 2025
GraphIC: A Graph-Based In-Context Example Retrieval Model for Multi-Step Reasoning was accepted to AAAI 2026.
May 01, 2025
Fast Large Language Model Collaborative Decoding via Speculation was accepted to ICML 2025.
Feb 26, 2025
Mimic In-Context Learning for Multimodal Tasks was accepted to CVPR 2025.

Selected Publications

  1. Rethinking LLM Ensembling from the Perspective of Mixture Models
    Jiale Fu, Yuchu Jiang, Peijun Wu, Chonghan Liu, Joey Tianyi Zhou, and Xu Yang
    In International Conference on Machine Learning, 2026
  1. Fast Large Language Model Collaborative Decoding via Speculation
    Jiale Fu*, Yuchu Jiang*, Junkai Chen, Jiaming Fan, Xin Geng, and Xu Yang
    In International Conference on Machine Learning, 2025
  1. GraphIC: A Graph-Based In-Context Example Retrieval Model for Multi-Step Reasoning
    Jiale Fu, Yaqing Wang, Simeng Han, Jiaming Fan, and Xu Yang
    In AAAI Conference on Artificial Intelligence, 2026
  1. Mimic In-Context Learning for Multimodal Tasks
    Yuchu Jiang, Jiale Fu, Chenduo Hao, Xinting Hu, Yingzhe Peng, Xin Geng, and Xu Yang
    In IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2025
  1. Flatter Tokens are More Valuable for Speculative Draft Model Training
    Jiaming Fan, Daming Cao, Xiangzhong Luo, Jiale Fu, Chonghan Liu, and Xu Yang
    In International Conference on Learning Representations, 2026
  1. d^2Cache: Accelerating Diffusion-Based LLMs via Dual Adaptive Caching
    Yuchu Jiang, Yue Cai, Xiangzhong Luo, Jiale Fu, Jiarui Wang, Chonghan Liu, and Xu Yang
    In International Conference on Learning Representations, 2026