🤓 Hey folks! I am Hengyuan Zhang (张恒源 in Chinese), a first-year PhD student in the Ngai Lab at the University of Hong Kong. Before going to college, I grew up in Xiamen, a beautiful coastal city in China. 📚 My research interests revolve around the application of Natural Language Processing (NLP) and Data Mining in specialized domains.

🧐 I am particularly intrigued by the decision-making mechanisms integrated within models, eager to unravel their inner workings and enhance transparency. All in all, I aim to improve the speciality and interpretability of models, so as to make them more powerful and trustworthy in real-world applications.

📮 I am keen on exploring opportunities for collaboration in research or projects 😊. Please do not hesitate to contact me at your convenience!

📝 Publications

Preprint 2026
sym

Locate, Steer, and Improve: A Practical Survey of Actionable Mechanistic Interpretability in Large Language Models

Hengyuan Zhang, Zhihao Zhang, Mingyang Wang, et al., Sophia Ananiadou, Tao Gui, Ruobing Xie, Hayden Kwok-Hay So, Hinrich Schütze, Xuanjing Huang, Qi Zhang, Ngai Wong

[Paper] [Code] 机器之心 Natural Language Processing, Actionable Mechanistic Interpretability Journal Preprint
  • This paper presents a practical survey that reframes mechanistic interpretability under a unified “Locate, Steer, and Improve” pipeline, systematically transforming MI from an observational analysis into an actionable framework for model intervention and improvement.
Preprint 2026
sym

Attention Sink in Transformers: A Survey on Utilization, Interpretation, and Mitigation

Zunhai Su, Hengyuan Zhang, et al., Jing Xiong, Hui Shen, Chaofan Tao, Taiqiang Wu, Zhongwei Wan, Yulei Qian, Yuchen Xie, Ngai Wong

[Paper] [Code] 机器之心 Natural Language Processing, Attention Sink Interpretability Journal Preprint
  • This paper presents the first comprehensive survey on Attention Sink (AS) in Transformers, systematically organizing existing research along three core dimensions—Fundamental Utilization, Mechanistic Interpretation, and Strategic Mitigation.
Preprint 2026
sym

Beyond Outliers: A Data-Free Layer-wise Mixed-Precision Quantization Approach Driven by Numerical and Structural Dual-Sensitivity

Hengyuan Zhang, Xinrong Chen, Zunhai Su, Xiao Liang, et al., Wei Zhang, Ruobing Xie, Lei Jiang, Hayden Kwok-Hay So, Ngai Wong

[Paper] [Code] Natural Language Processing, Interpretability in Quantization Preprint
  • This paper proposes NSDS, a data-free layer-wise mixed-precision quantization approach that leverages the dual-sensitivity of numerical and structural features to guide the bit allocation process.
ACL 2026
sym

Find Your Optimal Teacher: Personalized Data Synthesis via Router-Guided Multi-Teacher Distillation

Hengyuan Zhang, Shiping Yang, Xiao Liang, et al., Chaofan Tao, Jing Xiong, Hayden Kwok-Hay So, Ruobing Xie, Angel X. Chang, Ngai Wong

[Paper] [Code] Natural Language Processing, Interpretability in Data Synthesis Conference Preprint
  • This paper proposes PerSyn, a novel data synthesis strategy that operates under a new “Route then Generate” paradigm to create data tailored to each student model, enabling it to learn more effectively.
ACL 2025
sym

ShifCon: Enhancing Non-Dominant Language Capabilities with a Shift-based Contrastive Framework

Hengyuan Zhang, Chenming Shang, Sizhe Wang, Dongdong Zhang, Feng Yao, Renliang Sun, Yiyao Yu, Yujiu Yang, Furu Wei

[Paper] [Code] Natural Language Processing, Interpretability in Multilingualism CCF-A Conference
  • This paper aims to enhance the performance of non-dominant languages by projecting their representations into the dominant language space. We pinpoint the optimal layer area for shifting representations via a subspace distance metric. (OpenReview Score: [4, 4, 4.5])
EMNLP 2025
sym

GuiLoMo: Allocating Expert Number and Rank for LoRA-MoE via Bilevel Optimization with GuidedSelection Vectors

Hengyuan Zhang, Xinrong Chen, Xiao Liang, Ziyue Li, et al., Ngai Wong

[Paper] [Code] Natural Language Processing, Interpretability in Efficient Training CCF-B Conference
  • This paper introduces a fine-grained strategy, i.e., GuiLoMo, for jointly allocating optimal layer-wise expert numbers and ranks in LoRA-MoE based on bilevel optimization with GuidedSelection vectors.
ACL 2024
sym

Balancing Speciality and Versatility: A Coarse to Fine Framework for Mitigating Catastrophic Forgetting in Large Language Models

Hengyuan Zhang, Yanru Wu, Dawei Li, Zacc Yang, Rui Zhao, Yong Jiang, Fei Tan

[Paper] [Code] Natural Language Processing, Interpretability in Knowledge Management CCF-A Conference
  • This paper introduces a Coarse-to-Fine Fine-tuning framework (CoFiTune) that strikes a delicate balance between speciality and versatility. It pinpoints and updates specific modules that are crucial for speciality, while keeping other parameters frozen.
TKDD 2024
sym

A Question-centric Multi-experts Contrastive Learning Framework for Improving the Accuracy and Interpretability of Deep Sequential Knowledge Tracing Models

Hengyuan Zhang, Zitao Liu, Chenming Shang, Dawei Li, Yong Jiang

[Paper] [Code] Data Mining, Interpretability in Education Recommendation JCR Q1 Journal
  • This paper proposes Q-MCKT framework, which utilizes an item response theory-based prediction layer to generate interpretable prediction results by simultaneously modeling knowledge acquisition and question difficulty.
  • ACL 2026 Training LLMs for Divide-and-Conquer Reasoning Elevates Test-Time Scalability
    Xiao Liang, Zhong-Zhi Li, Zhenghao Lin, Hengyuan Zhang, Yelong Shen, Kai-Wei Chang, Ying Nian Wu, Yeyun Gong, Weizhu Chen
    [Paper] | [Code] | Natural Language Processing, Reasoning Enhancement | Conference Preprint

  • ACL 2026 Quantifying the Robustness of Retrieval-Augmented Language Models Against Spurious Features in Grounding Data
    Shiping Yang, Jie Wu, Wenbiao Ding, et al., Hengyuan Zhang, Dongmei Zhang
    [Paper] | Natural Language Processing, Phenomenon Analysis | Conference Preprint

  • Preprint 2026 MMFormalizer: Multimodal Autoformalization in the Wild
    Jing Xiong, Qi Han, Yunta Hsieh, Hui Shen, et al., Hengyuan Zhang, Taiqiang Wu, Haochen Wang, Zhongwei Wan, Lingpeng Kong, Ngai Wong
    [Paper] | Natural Language Processing, Multimodal Analysis | Conference Preprint

  • CVPR 2026 Residual Decoding: Mitigating Hallucinations in Large Vision-Language Models via History-Aware Residual Guidance
    Xinrong Chen, Xu Chu, Yingmin Qiu, Hengyuan Zhang, Jing Xiong, et al., Hayden Kwok-Hay So, Ngai Wong
    [Paper] | Natural Language Processing, Multimodal Analysis | CCF-A Conference

  • NeurIPS 2025 SwS: Self-aware Weakness-driven Problem Synthesis in Reinforcement Learning for LLM Reasoning
    Xiao Liang, Zhong-Zhi Li, Yeyun Gong, Yang Wang, Hengyuan Zhang, Yelong Shen, Ying Nian Wu, Weizhu Chen
    [Paper] | Natural Language Processing, Data Synthesis | CCF-A Conference
  • ACL 2025 Chain-of-Reasoning: Towards Unified Mathematical Reasoning in Large Language Models via a Multi-Paradigm Perspective
    Yiyao Yu, Yuxiang Zhang, Dongdong Zhang, Xiao Liang, Hengyuan Zhang, et al., Yujiu Yang, Furu Wei
    [Paper] | Natural Language Processing, Fine-tuning Technique | CCF-A Conference
  • EMNLP 2025 TreeReview: A Dynamic Tree of Questions Framework for Deep and Efficient LLM-based Scientific Peer Review
    Yuan Chang, Ziyue Li, Hengyuan Zhang, Yuanbo Kong, Yanru Wu, Zhijiang Guo, Ngai Wong
    [Paper] | Natural Language Processing | CCF-B Conference
  • NAACL 2025 BPO: Towards Balanced Preference Optimization between Knowledge Breadth and Depth in Alignment
    Sizhe Wang, Yongqi Tong, Hengyuan Zhang, Dawei Li, Xin Zhang, Tianlong Chen
    [Paper] | Natural Language Processing, Fine-tuning Technique | CCF-B Conference
  • Preprint 2024 Improving Low-Resource Knowledge Tracing Tasks by Supervised Pre-training and Importance Mechanism Fine-tuning
    Hengyuan Zhang, Zitao Liu, Shuyan Huang, Chenming Shang, Bojun Zhan, Yong Jiang
    [Paper] | [Code] | Data Mining, Education Recommendation | Journal Preprint
  • Preprint 2024 Compositional Generalization Through Neuroscience-inspired Geometric Constraints on Representation Structure
    Chenming Shang, Shiji Zhou, Hengyuan Zhang, Xinchen Zhang, Lei Ke, Yuwang Wang, Yujiu Yang
    [Paper] | Computer Vision, Cognitive Science, Interpretability in Representation | Conference Preprint
  • CogSci 2024 Understanding Multimodal Deep Neural Networks: A Concept Selection View
    Chenming Shang, Hengyuan Zhang, Hao Wen, Yujiu Yang
    [Paper] | Computer Vision, Cognitive Science, Interpretability in Prediction | CCF-B Conference
  • CVPR 2024 Incremental Residual Concept Bottleneck Model
    Chenming Shang, Shiji Zhou, Hengyuan Zhang, Yujiu Yang, Xinzhe Ni, Yuwang Wang
    [Paper] | [Code] | Computer Vision, Cognitive Science, Interpretability in Prediction | CCF-A Conference
  • EMNLP 2023 Multi-level Contrastive Learning for Script-based Character Understanding
    Dawei Li, Hengyuan Zhang, Yanran Li, Shiping Yang
    [Paper] | [Code] | Natural Language Processing, Cognitive Science | CCF-B Conference
  • ACL 2023 BEA Assisting Language Learners: Automated Trans-Lingual Definition Generation via Contrastive Prompt Learning
    Hengyuan Zhang, Dawei Li, Yanran Li, Chenming Shang, Chufan Shi, Yong Jiang
    [Paper] | [Code] | Natural Language Processing, Education, Interpretability in Representation | CCF-A Conference
  • AACL 2022(Oral) Fine-grained Contrastive Learning for Definition Generation
    Hengyuan Zhang, Dawei Li, Shiping Yang, Yanran Li
    [Paper] | [Code] | Natural Language Processing, Education, Interpretability in Representation | Conference

💻 Interships

Baidu, Search Strategy Lab, Beijing

  • Mar. 2021 - Jul. 2021, Engineering Intern

Xiaomi, AI Lab, Beijing

  • Mar. 2022 - Sept. 2022, Research Intern

Tencent, AI Lab, Shenzhen

  • Mar. 2023 - Jul. 2023, Research Intern

Sensetime, Research, Shenzhen

  • Aug. 2023 - Mar. 2024, Research Intern

Microsoft Research Asia, NLC Group, Beijing

  • Mar. 2024 - Dec. 2024, Research Intern
    • I got the “Microsoft Stars of Tomorrow” Award during the internship

AMD, Quantization Group, Beijing

  • Feb. 2025 - Aug. 2025, Research Intern

Tencent Rhino-Bird Elite Research Program, Shenzhen

  • Sept. 2025 - Present, Research Intern


🏅 Selected Honors and Awards

👉  HKU Y S and Christabel Lung Postgraduate Scholarship (Top 3%, HKD $ 20,000) | 2025

👉  Tsinghua University Comprehensive First-Class Scholarship (Top 3%, RMB ¥ 10,000) | 2024

👉  Tsinghua University General Excellence Scholarship (Top 5%, RMB ¥ 4,000) | 2023

👉  National Scholarship (Top 1%, 3 Times, RMB ¥ 24,000) | 2019, 2020, 2021

👉  Outstanding Graduate Student of Beijing (Top 3%) | 2022

👉  Excellent League Member of Beijing (Top 3%) | 2021

👉  Merit Student of Beijing (Top 3%) | 2021

👉  Meritorious Winner of Interdisciplinary Contest in Modeling (Top 5%) | 2021

👉  Computer Design Competition National Second Prize (Top 5%) | 2020

👉  CUMCM-Beijing Area First Prize (Top 5%) | 2020

👉  Xiaomi Third Hacker Marathon Excellence (Top 7%, RMB ¥ 3,000) | 2022


📌 Miscellaneous

📖 Academic & Community Engagement

  • I once led the Academic Department of SIGS Student Union at Tsinghua University. During which, I organized academic activities such as academic forums and experience-sharing sessions. I am also a member of the Beijing Xiamen ECC (北京厦门企业商会), actively participating in sharing activities [Link]. Furthermore, I am keen on participating in academic events to present my ideas and making a positive impact on the community.
Academic Activity 1 Academic Activity 2 Academic Activity 3
Extra Activity 1 Extra Activity 2 Extra Activity 3


🎸 Hobbies & Life

  • I used to be a guitarist 🎸 in a band when I was in high school. Also, I love playing badminton 🏸, table tennis 🏓 and, soccer ⚽️. During holidays, I will also seize any opportunity to travel around the world ⛳️.

🎤 Band & Singing

🏸 Sports & Travel

Life 1
Life 2
Life 3
Life 4
Life 5
Life 6
Life 7
Life 8