Clarify or Answer: Reinforcement Learning for Agentic VQA with Context Under-specification
Reinforcement learning for agentic VQA that balances clarification and answering under underspecified context.
I am a final-year Ph.D. candidate at the University of Washington, where I am advised by Prof. Bill Howe and Prof. Lucy Lu Wang. I am also a member of the UW RAISE Center and collaborate closely with Prof. Yulia Tsvetkov.
My research focuses on the efficiency and reliability of foundation models and agentic systems, aiming to reduce computational overhead while enhancing model trustworthiness. My work is structured around three core pillars:
PhD in Information Science (Natural Language Processing)
University of Washington
MS in Computational Science & Engineering (Artificial Intelligence)
University of Hong Kong
BS in Control Science & Engineering (Robotics)
Zhejiang University
Reinforcement learning for agentic VQA that balances clarification and answering under underspecified context.
A modular abstention framework for reliable expert LLMs that enables selective abstention from uncertain questions.
Automatic prediction of compute-optimal data composition for efficient LLM training.
Exploring psychological insights to address overconfidence in LLMs by comparing with human confidence patterns.
A comprehensive survey of abstention mechanisms in large language models, covering theory, implementation, and evaluation.
1/2026 Our paper on reinforcement learning for agentic VQA has been released on arXiv.
1/2026 Our benchmark on dark pattern susceptibility of computer-use agents has been accepted by IUI 2026.
9/2025 Our paper about MLLM spurious correlation has been accepted by NeurIPS 2025!
7/2025 I presented our abstention survey in LLMs (oral presentation) and confidence calibration (poster) at ACL 2025!
7/2025 Our paper about modular abstention has been accepted by ICML 2025!
6/2025 I will start my summer internship at Apple as a research intern!
5/2025 Our paper about optimal data mixing in pretraining has been accepted by [COLM 2025]!
5/2025 Our paper about confidence calibration has been accepted by ACL 2025!
2/2025 Our paper about abstention survey in LLMs has been accepted by TACL 2025!