Yiju Guo

Yiju Guo

PhD Student

Gaoling School of Artificial Intelligence, Renmin University of China

Research Interests

LLM Alignment & Reasoning
Reinforcement Learning
Natural Language Processing

About

I am a PhD student at Gaoling School of Artificial Intelligence, Renmin University of China (RUC), advised by Prof. Yankai Lin. I am also conducting research at Natural Language Processing Lab at Tsinghua University(THUNLP), supervised by Prof. Ning Ding.

My research focuses on Large Language Model (LLM) and Reinforcement Learning (RL). Specifically, I am interested in:

  • LLM Alignment (e.g., RLHF, Multi-Objective Optimization)
  • Reasoning & Generation (e.g., Causal Inference, Attention Mechanisms)

Selected Publications

View All →

Learning to Focus: Causal Attention Distillation via Gradient-Guided Token Pruning

Yiju Guo, Wenkai Yang, Zexu Sun, Ning Ding, Zhiyuan Liu, Yankai Lin

NeurIPS 2025 Conference

A framework to improve LLM reasoning by removing distracting tokens via causal attention distillation and gradient-guided pruning.

LaSeR: Reinforcement Learning with Last-Token Self-Rewarding

Wenkai Yang, Weijie Liu, Ruobing Xie, Yiju Guo, Lulu Wu, Saiyong Yang, Yankai Lin

arXiv preprint arXiv:2510.14943

An efficient RL method that unifies reasoning and verification by utilizing the last-token probability as a self-rewarding signal.

Uncertainty and influence aware reward model refinement for reinforcement learning from human feedback

Zexu Sun, Yiju Guo, Yankai Lin, Xu Chen, Qi Qi, Xing Tang, Ji-Rong Wen

ICLR 2025 Conference

An uncertainty-aware data augmentation method to refine reward models in RLHF without expensive human annotation.

Controllable preference optimization: Toward controllable multi-objective alignment

Yiju Guo, Ganqu Cui, Lifan Yuan, Ning Ding, Zexu Sun, Bowen Sun, Huimin Chen, Ruobing Xie, Jie Zhou, Yankai Lin, others

EMNLP 2024 main conference

A multi-objective alignment method that explicitly controls preference scores to balance helpfulness, honesty, and harmlessness.

News

2025-09

Our work 《Learning to Focus: Causal Attention Distillation via Gradient-Guided Token Pruning》has been accepted by NeurIPS 2025 main conference 🎉

2025-01

Our work 《Uncertainty and influence aware reward model refinement for reinforcement learning from human feedback》has been accepted by ICLR 2025 🎉

2024-09

Our work 《Controllable preference optimization: Toward controllable multi-objective alignment》has been accepted by EMNLP 2024 main conference 🎉

2022-09

Starting my PhD at the Gaoling School of Artificial Intelligence, Renmin University of China (RUC)