đź“– About Me
I am a PhD researcher at the University of Manchester and the NaCTeM group, supervised by Prof. Sophia Ananiadou. Previously, I worked as an NLP researcher at Tencent Technology in Shanghai. I received my Bachelor’s and Master’s degrees from Shanghai Jiao Tong University, supervised by Prof. Gongshen Liu.
My research topic is Interpreting and Improving Large Language Models – From Internal Mechanisms to Reliable and Capable Systems. I develop tools and methods to explain, debug, and enhance LLMs/MLLMs — grounding predictions to diagnose failures and build more trustworthy systems.
I am actively seeking Research Scientist and Applied Scientist positions starting in Fall 2026. Please feel free to contact me at zepingyu@foxmail.com if you have any suitable openings!
📝 Research Interests
My research aims to make LLMs interpretable at the token and neuron level, and to use this understanding to design more reliable and capable systems.
a) Mechanistic Interpretability and Diagnostic Analysis of LLMs
-
VQALens: Diagnosing errors, shortcuts, and hallucinations in MLLMs. [System Demo]
-
Neuron-Level Attribution: Identifying important neurons for model diagnosis. [EMNLP 2024]
-
In-Context Learning Mechanisms: Understanding how LLMs perform in-context learning. [EMNLP 2024]
b) Mechanism-Guided Improvement of LLM Capabilities
-
Back Attention: Improving latent multi-hop reasoning in LLMs [EMNLP 2025]
-
Locate-then-Merge: Mitigating catastrophic forgetting in MLLMs [EMNLP 2025]
-
Arithmetic Reasoning via CNA: Analysis and pruning for arithmetic tasks [EMNLP 2024]
🔥 News
-
2026.02: New survey: “Locate, Steer, and Improve: A Practical Survey of Actionable Mechanistic Interpretability in Large Language Models”.
-
2025.08: Two papers are accepted by EMNLP 2025.
-
2024.09: Three papers are accepted by EMNLP 2024.
-
2024.04: Paper list of LLM interpretability.
📝 Publications
Mechanism-Guided Improvement of LLM Capabilities
[P1] Locate-then-Merge: Neuron-Level Parameter Fusion for Mitigating Catastrophic Forgetting in Multimodal LLMs
- Zeping Yu, Sophia Ananiadou [EMNLP 2025 (Findings)]
[P2] Back Attention: Understanding and Enhancing Multi-Hop Reasoning in Large Language Models
- Zeping Yu, Yonatan Belinkov, Sophia Ananiadou [EMNLP 2025 (Main)]
[P3] Understanding and Mitigating Gender Bias in LLMs via Interpretable Neuron Editing
- Zeping Yu, Sophia Ananiadou [preprint]
[P4] Interpreting Arithmetic Mechanism in Large Language Models through Comparative Neuron Analysis
- Zeping Yu, Sophia Ananiadou [EMNLP 2024 (Main)]
Interpretability and Diagnostic Analysis of LLMs
[P5] Understanding Multimodal LLMs: the Mechanistic Interpretability of LLaVA in Visual Question Answering
- Zeping Yu, Sophia Ananiadou [preprint]
[P6] How do Large Language Models Learn In-Context? Query and Key Matrices of In-Context Heads are Two Towers for Metric Learning
- Zeping Yu, Sophia Ananiadou [EMNLP 2024 (Main)]
[P7] Neuron-Level Knowledge Attribution in Large Language Models
- Zeping Yu, Sophia Ananiadou [EMNLP 2024 (Main)]
Earlier Work
[E1] CodeCMR: Cross-modal retrieval for function-level binary source code matching
- Zeping Yu, Wenxin Zheng, Jiaqi Wang, Qiyi Tang, Sen Nie, Shi Wu [NeurIPS 2020]
[E2] Order matters: Semantic-aware neural networks for binary code similarity detection
- Zeping Yu*, Rui Cao* , Qiyi Tang, Sen Nie, Junzhou Huang, Shi Wu [AAAI 2020]
[E3] Adaptive User Modeling with Long and Short-Term Preferences for Personalized Recommendation
- Zeping Yu, Jianxun Lian, Ahmad Mahmoody, Gongshen Liu, Xing Xie [IJCAI 2019]
[E4] Sliced recurrent neural networks
- Zeping Yu, Gongshen Liu [COLING 2018]
đź“– Educations
- 2023.09 - 2026.09 (Expected), PhD of Computer Science, University of Manchester.
- 2017.09 - 2020.03, Master of Computer Science, Shanghai Jiao Tong University.
- 2013.09 - 2017.06, Bachelor of Engineering, Shanghai Jiao Tong University.