I’m a Ph.D. student at State Key Lab of CAD&CG, Zhejiang University , under the supervision of Prof. Wei Chen.

I’m currently insterested in agents and (M)LLMs.

πŸ”₯ News

  • 2024.05: πŸŽ‰πŸŽ‰ One paper gets accepted by ACL 2024
  • 2025.11: πŸŽ‰πŸŽ‰ One paper gets accepted by AAAI 2026 as Oral

πŸ“ Publications

ACL 2024
SDFT image

Self-Distillation Bridges Distribution Gap in Language Model Fine-Tuning

PDF | Code | Poster | Slides

Zhaorui Yang, Tianyu Pang, Haozhe Feng, Han Wang, Wei Chen, Minfeng Zhu, Qian Liu

  • Fine-tuning LLMs for specific tasks often encounters challenges in balancing performance and preserving general instruction-following abilities. In this work, we posit that the distribution gap between task datasets and the LLMs serves as the primary underlying cause. To address the problem, we introduce Self-Distillation Fine-Tuning to bridge the distribution gap by guiding fine-tuning with a distilled dataset generated by the model itself.
AAAI 2026 (Oral)
Multimodal DeepResearcher framework

Multimodal DeepResearcher: Generating Text-Chart Interleaved Reports From Scratch with Agentic Framework

PDF | Code | Poster | Slides | Project Page

Zhaorui Yang*, Bo Pan*, Han Wang*, Yiyao Wang, Xingyu Liu, Luoxuan Weng, Yingchaojie Feng, Haozhe Feng, Minfeng Zhu, Bo Zhang, Wei Chen

  • Existing deep research works primarily focus on generating text-only content, leaving the automated generation of interleaved texts and visualizations underexplored. In this work, we propose a structured textual representation for visualizations, and introduce an agentic framework that automatically generates comprehensive multimodal reports from scratch with interleaved texts and visualizations.
arxiv:2304.06627
CoSDA setting image

CoSDA: Continual Source-Free Domain Adaptation

PDF | Code

Haozhe Feng*, Zhaorui Yang*, Hesun Chen*, Tianyu Pang, Chao Du, Minfeng Zhu, Wei Chen, Shuicheng Yan

  • In this work, we investigate the mechanism of catastrophic forgetting of previous Source-Free Domain Adaptation (SFDA) approaches. We observe that there is a trade-off between adaptation gain and forgetting loss. Motivated by the findings, we propose CoSDA, which outperforms SOTA approaches in continuous adaptation.

πŸŽ– Honors and Awards

  • 2022.12 China National Scholarship (Undergraduate).
  • 2021.12 China National Scholarship (Undergraduate).

πŸ“– Education

  • 2023.09 - Present
    Ph.D. student in Software Engineering at State Key Lab of CAD&CG, Zhejiang University
  • 2019.09 - 2023.06
    B.E. in Software Engineering, Xi’an Jiaotong University

πŸ’» Internships

  • 2025.7 - Present: TEG, Tencent
    Developing an LLM-powered Data Agent to automate Text-to-Insight analysis across enterprise data warehouses containing 1,000+ tables per user, enabling natural language queries for complex data exploration and business intelligence.