đź“– About Me
I am a PhD researcher at the University of Manchester and the NaCTeM group, supervised by Prof. Sophia Ananiadou. Previously, I worked as an NLP researcher at Tencent Technology in Shanghai. I received my Bachelor’s and Master’s degrees from Shanghai Jiao Tong University, supervised by Prof. Gongshen Liu.
My research topic is Interpreting and Improving Large Language Models – From Internal Mechanisms to Reliable and Capable Systems. My primary methodology follows a diagnose-and-improve paradigm: I first analyze why LLMs fail, and then design targeted methods or architectural modules to enhance their capabilities, for example in latent multi-hop reasoning and catastrophic forgetting.
I am actively seeking Research Scientist and Applied Scientist positions starting in Fall 2026. Please feel free to contact me at zepingyu@foxmail.com if you have any suitable openings!
📝 Research Interests
My research aims to improve the capabilities of LLMs by analyzing their internal mechanisms, and leveraging this understanding to design more reliable and effective systems.
a) Understanding and Diagnosing LLMs through Mechanistic Interpretability
-
VQALens: Diagnosing errors, shortcuts, and hallucinations in MLLMs. [System Demo]
-
Neuron-Level Attribution: Identifying important neurons for LLM diagnosis. [EMNLP 2024]
-
Head-Level Attribution: Understanding how LLMs perform in-context learning. [EMNLP 2024]
b) Enhancing LLM Capabilities through Mechanistic Understanding
-
Back-Attention Module: Improving latent multi-hop reasoning in LLMs. [EMNLP 2025]
-
Locate-then-Merge: Mitigating catastrophic forgetting in MLLMs. [EMNLP 2025]
-
Locate-then-Prune: Improving arithmetic reasoning through model pruning. [EMNLP 2024]
-
Locate-then-Edit: Reducing gender bias in LLMs via model editing. [Preprint]
🔥 News
-
2026.02: New survey: “Locate, Steer, and Improve: A Practical Survey of Actionable Mechanistic Interpretability in Large Language Models”.
-
2025.08: Two papers are accepted by EMNLP 2025.
-
2024.09: Three papers are accepted by EMNLP 2024.
-
2024.04: Paper list of LLM interpretability.
📝 Publications
Enhancing LLM Capabilities through Mechanistic Understanding
[P1] Back Attention: Understanding and Enhancing Multi-Hop Reasoning in Large Language Models
- Zeping Yu, Yonatan Belinkov, Sophia Ananiadou [EMNLP 2025 (Main)]
[P2] Locate-then-Merge: Neuron-Level Parameter Fusion for Mitigating Catastrophic Forgetting
- Zeping Yu, Sophia Ananiadou [EMNLP 2025 (Findings)]
[P3] Interpreting Arithmetic Mechanism in Large Language Models through Comparative Neuron Analysis
- Zeping Yu, Sophia Ananiadou [EMNLP 2024 (Main)]
[P4] Understanding and Mitigating Gender Bias in LLMs via Interpretable Neuron Editing
- Zeping Yu, Sophia Ananiadou [preprint]
Understanding and Diagnosing LLMs through Mechanistic Interpretability
[P5] Neuron-Level Knowledge Attribution in Large Language Models
- Zeping Yu, Sophia Ananiadou [EMNLP 2024 (Main)]
[P6] How do Large Language Models Learn In-Context? Query and Key Matrices of In-Context Heads are Two Towers for Metric Learning
- Zeping Yu, Sophia Ananiadou [EMNLP 2024 (Main)]
[P7] Understanding Multimodal LLMs: the Mechanistic Interpretability of LLaVA in VQA
- Zeping Yu, Sophia Ananiadou [preprint]
Earlier Work in Deep Learning
[E1] CodeCMR: Cross-modal retrieval for function-level binary source code matching
- Zeping Yu, Wenxin Zheng, Jiaqi Wang, Qiyi Tang, Sen Nie, Shi Wu [NeurIPS 2020]
[E2] Order matters: Semantic-aware neural networks for binary code similarity detection
- Zeping Yu*, Rui Cao* , Qiyi Tang, Sen Nie, Junzhou Huang, Shi Wu [AAAI 2020]
[E3] Adaptive User Modeling with Long and Short-Term Preferences for Personalized Recommendation
- Zeping Yu, Jianxun Lian, Ahmad Mahmoody, Gongshen Liu, Xing Xie [IJCAI 2019]
[E4] Sliced recurrent neural networks
- Zeping Yu, Gongshen Liu [COLING 2018]
đź“– Educations
- 2023.09 - 2026.09, PhD of Computer Science, University of Manchester.
- 2017.09 - 2020.03, Master of Computer Science, Shanghai Jiao Tong University.
- 2013.09 - 2017.06, Bachelor of Engineering, Shanghai Jiao Tong University.