I am a second-year PhD student at the University of Manchester (started in September 2023), advised by Prof. Sophia Ananiadou. I was previously an AI researcher at Tencent from 2020 to 2022, where I focused on deep learning for source code and binary code. I graduated from Shanghai Jiao Tong University in 2020 with a Masterβs Degree, and in 2017 with a Bachelorβs Degree.
I believe mechanistic interpretability is the key to trustworthy AGI. Currently, I am interested in:
a) Understanding the internal mechanisms of large language models.
b) Using the interpretability findings to help downstream tasks (e.g. factual knowledge, reasoning).
c) Using the interpretability findings to design safer models.
My current research is identifying the important neurons in LLMs and constructing the neuron-level internal logic chain (also named βneuron circuitsβ), which is useful for neuron-level model editing and neuron-level LLM safety. I have explored the neuron-level knowledge attribution, mechanism of in-context learning, and mechanism of arithmetic task in LLMs. Please feel free to contact me via zepingyu@foxmail.com for discussions.
π₯ News
-
2024.09: Our work is accepted by EMNLP 2024 (main): Interpreting Arithmetic Mechanism in Large Language Models through Comparative Neuron Analysis. This work explores the internal mechanism of arithmetic.
-
2024.09: Our work is accepted by EMNLP 2024 (main): Neuron-Level Knowledge Attribution in Large Language Models. This work introduces how to identify important neurons in LLMs.
-
2024.09: Our work is accepted by EMNLP 2024 (main): How do Large Language Models Learn In-Context? Query and Key Matrices of In-Context Heads are Two Towers for Metric Learning. This work explores the internal mechanism of in-context learning.
-
2024.04: I create a paper list for people interested in understanding the mechanism of LLMs.
-
2023.09: I start my PhD in the University of Manchester.
π Publications
Interpreting Arithmetic Mechanism in Large Language Models through Comparative Neuron Analysis
Zeping Yu, Sophia Ananiadou [EMNLP 2024]
Neuron-Level Knowledge Attribution in Large Language Models
Zeping Yu, Sophia Ananiadou [EMNLP 2024]
Zeping Yu, Sophia Ananiadou [EMNLP 2024]
βββββββββββββ Previous Works βββββββββββββ
CodeCMR: Cross-modal retrieval for function-level binary source code matching
Zeping Yu, Wenxin Zheng, Jiaqi Wang, Qiyi Tang, Sen Nie, Shi Wu [NeurIPS 2020]
Order matters: Semantic-aware neural networks for binary code similarity detection
Zeping Yu, Rui Cao, Qiyi Tang, Sen Nie, Junzhou Huang, Shi Wu [AAAI 2020]
Adaptive User Modeling with Long and Short-Term Preferences for Personalized Recommendation
Zeping Yu, Jianxun Lian, Ahmad Mahmoody, Gongshen Liu, Xing Xie [IJCAI 2019]
Sliced recurrent neural networks
Zeping Yu, Gongshen Liu [COLING 2018]
π Educations
- 2023.09 - now, PhD student, Computer Science, the University of Manchester.
- 2020.03 - 2022.03 (work), AI researcher for computer security, Tencent Keen Lab.
- 2017.09 - 2020.03, Master, Electronics and Communications Engineering, Shanghai Jiao Tong University.
- 2013.09 - 2017.06, Bachelor, Information Security Engineering, Shanghai Jiao Tong University.