|
Research
My research interests focus on LLM agents, deep reinforcement learning, and their applications in decision-making and robotics.
|
|
|
Verlog: A Multi-turn RL framework for LLM agents
Wen-Tse Chen,
Jiayu Chen,
Hao Zhu,
Jeff Schneider,
Proposed a multi-turn reinforcement learning framework built for long-horizon LLM-agentic tasks with highly variable episode lengths.
|
|
|
Retrospective In-Context Learning for Temporal Credit Assignment with Large Language Models
Wen-Tse Chen,
Jiayu Chen,
Fahim Tajwar,
Hao Zhu,
Xintong Duan,
Russ Salakhutdinov,
Jeff Schneider
NeurIPS 2025
Presented a sample-efficient method for online fine-tuning LLM agents by using in-context learning to convert sparse feedback into dense signals, enabling LLMs to adapt to dynamic environments with minimal data.
|
|
|
ME-IGM: Individual-Global-Max in Maximum Entropy Multi-Agent Reinforcement Learning
Wen-Tse Chen,
Yuxuan Li,
Shiyu Huang,
Jiayu Chen,
Jeff Schneider,
AAMAS 2026
Enhanced IGM-based algorithms with maximum entropy RL for better exploration, ensuring locally optimal actions match global optima via an order preserving transformation, achieving SOTA performance on SAMC-v2 and Overcooked benchmark.
|
|
|
DGPO: Discovering Multiple Strategies with Diversity-Guided Policy Optimization
Wen-Tse Chen,
Shiyu Huang,
Yuan Chiang,
Tim Pearce,
Wei-Wei Tu,
Chen Ting,
Zhu Jun,
AAAI 2024
Proposed an on-policy framework for discovering multiple diverse optimal strategies for the same task in a single training process.
|
|