About
I am a PhD student at Gaoling School of Artificial Intelligence, Renmin University of China (RUC), advised by Prof. Yankai Lin. I am also conducting research at Natural Language Processing Lab at Tsinghua University(THUNLP), supervised by Prof. Ning Ding.
My research focuses on Large Language Model (LLM) and Reinforcement Learning (RL). Specifically, I am interested in:
- LLM Alignment (e.g., RLHF, Multi-Objective Optimization)
- Reasoning & Generation (e.g., Causal Inference, Attention Mechanisms)
Selected Publications
View All →Learning to Focus: Causal Attention Distillation via Gradient-Guided Token Pruning
Yiju Guo, Wenkai Yang, Zexu Sun, Ning Ding, Zhiyuan Liu, Yankai Lin
NeurIPS 2025 Conference
A framework to improve LLM reasoning by removing distracting tokens via causal attention distillation and gradient-guided pruning.
LaSeR: Reinforcement Learning with Last-Token Self-Rewarding
Wenkai Yang, Weijie Liu, Ruobing Xie, Yiju Guo, Lulu Wu, Saiyong Yang, Yankai Lin
arXiv preprint arXiv:2510.14943
An efficient RL method that unifies reasoning and verification by utilizing the last-token probability as a self-rewarding signal.
Uncertainty and influence aware reward model refinement for reinforcement learning from human feedback
Zexu Sun, Yiju Guo, Yankai Lin, Xu Chen, Qi Qi, Xing Tang, Ji-Rong Wen
ICLR 2025 Conference
An uncertainty-aware data augmentation method to refine reward models in RLHF without expensive human annotation.
Controllable preference optimization: Toward controllable multi-objective alignment
Yiju Guo, Ganqu Cui, Lifan Yuan, Ning Ding, Zexu Sun, Bowen Sun, Huimin Chen, Ruobing Xie, Jie Zhou, Yankai Lin, others
EMNLP 2024 main conference
A multi-objective alignment method that explicitly controls preference scores to balance helpfulness, honesty, and harmlessness.
News
Our work 《Learning to Focus: Causal Attention Distillation via Gradient-Guided Token Pruning》has been accepted by NeurIPS 2025 main conference 🎉
Our work 《Uncertainty and influence aware reward model refinement for reinforcement learning from human feedback》has been accepted by ICLR 2025 🎉
Our work 《Controllable preference optimization: Toward controllable multi-objective alignment》has been accepted by EMNLP 2024 main conference 🎉
Starting my PhD at the Gaoling School of Artificial Intelligence, Renmin University of China (RUC)
