Selected Publications
|
"Task Success" is not Enough: Investigating the Use of Video-Language Models as Behavior Critics for Catching Undesirable Agent Behaviors
Lin Guan*, Yifan Zhou*, Denis Liu, Yantian Zha, Heni Ben Amor, Subbarao Kambhampati
Preprint
When no sound verifier is available, can we use large vision and language models (VLMs), which are approximately omniscient, as scalable Behavior Critics to catch undesirable robot behaviors in videos? To answer this, we first construct a benchmark that contains diverse cases of goal-reaching yet undesirable robot policies. Then, we comprehensively evaluate VLM critics to gain a deeper understanding of their strengths and failure modes.
paper
 
website
|
|
On the role of Large Language Models in Planning
Subbarao Kambhampati, Karthik Valmeekam, Lin Guan
Tutorial at AAAI 2024
(also accepted to
ICAPS 2023 Tutorial Program)
This tutorial discusses the fundamental limitations of LLMs in generating plans (especially those that require resolving subgoal interactions), and also presents constructive uses of LLMs for planning tasks.
website
|
|
Leveraging Approximate Symbolic Models for Reinforcement Learning via Skill Diversity
Lin Guan*, Sarath Sreedharan* (equal contribution), Subbarao Kambhampati
ICML 2022
(also received the
Best Paper Award
at PRL@ICAPS 2022 and accepted to RLDM 2022)
Explicit symbolic knowledge is important for solving long-horizon task and motion planning tasks.
But a key resistance to leveraging easily available human knowledge (or knowledge acquired from LLMs/VLMs) has been that it might be inexact.
In this work, we present a framework to quantify the relationship between the true task model and an inexact STRIPS model, and
introduce a novel approach using landmarks and a diversity objective to make up for potential errors in the symbolic knowledge.
paper
 
website
|
|
Atari-HEAD: Atari Human Eye-Tracking and Demonstration Dataset
Ruohan Zhang, Calen Walshe, Zhuode Liu, Lin Guan, Karl S. Muller, Jake A. Whritner, Luxin
Zhang, Mary M Hayhoe, Dana H Ballard
AAAI 2020
We provide a large-scale, high-quality dataset of human actions with simultaneously
recorded eye movements (i.e., gaze info) while humans play Atari video games.
paper
|
|