Research
I'm interested in how people interact with intelligent agents. In my PhD work, I explored this in the context of assistive home robotics, where I developed several algorithms for seamless value alignment to people's goals during household tasks using naturalistic behaviors.
This work has motivated my interests in human-robot interaction, machine learning, computer vision and perception, and reinforcement learning, especially through human feedback.
|
|
DegustaBot: Zero-Shot Visual Preference Estimation for Personalized Multi-Object Rearrangement
Benjamin A. Newman, Pranay Gupta, Kris Kitani, Yonatan Bisk, Henny Admoni, and Chris Paxton
arXiv, 2024
pdf /
bibtex
We present a VLM based method to solve multi-step household object rearrangement tasks, such as setting a table, according to personal preferences. We compare multiple state of the art VLMs in a simulated setting. We then collect a large dataset of 995 naturalistic table setting demonstrations and evaluate our method on its ability to capture these preferences.
|
|
Bootstrapping Linear Models for Fast Online Adaptation in Human-Agent Collaboration
Benjamin A. Newman, Chris Paxton, Kris Kitani, and Henny Admoni
AAMAS, 2024
pdf /
bibtex
We present an algorithm that bootstraps online linear regression problems using large nonlinear models using in-situ naturalistic corrective actions.
|
|
Openeqa: Embodied question answering in the era of foundation models
Arjun Majumdar, Anurag Ajay, Xiaohan Zhang, Pranav Putta, Sriram Yenamandra, Mikael Henaff, Sneha Silwal, Paul Mcvay, Oleksandr Maksymets, Sergio Arnaud, Karmesh Yadav, Qiyang Li, Benjamin A. Newman, Mohit Sharma, Vincent Berges, Shiqi Zhang, Pulkit Agrawal, Yonatan Bisk, Dhruv Batra, Mrinal Kalakrishnan, Franziska Meier, Chris Paxton, Alexander Sax, Aravind Rajeswaran
CVPR, 2024
pdf /
bibtex
We present a modern formulation of Embodied Question Answering (EQA) as the task of understanding an environment well enough to answer questions about it in natural language.
|
|
Leveraging Vision and Language Models for Zero-Shot, Personalization of Household Multi-Object Rearrangement Tasks
Benjamin A. Newman, Pranay Gupta, Kris Kitani, Yonatan Bisk, Henny Admoni, and Chris Paxton
Human – Large Language Model Interaction Workshop at HRI, 2024
pdf /
bibtex
We present a VLM based method to solve object rearrangment problems according to personal preference from prior user demonstrations.
|
|
Towards Online Adaptation for Autonomous Household Assistants
Benjamin A. Newman, Chris Paxton, Kris Kitani, and Henny Admoni
Companion of the HRI Proceedings, 2023
pdf /
bibtex
We present an algorithm for using naturalistic corrections to update a robot model of a user goal in a simulated object rearrangement task.
|
|
Helping People Through Space and Time: Assistance as a Perspective on Human-Robot Interaction
Benjamin A. Newman, Reuben Aronson, Kris Kitani, and Henny Admoni
Frontiers in Robotics and AI, 2022
pdf /
bibtex
We define assistance as a perspective on human-robot interaction and provide cross-domain design axes that are critical to consider when developing assistive robotics. We support these through a broad review of recent assistive robotics research.
|
|
HARMONIC: A Multimodal Data Set of Assistive Human-Robot Collaboration
Benjamin A. Newman*, Reuben Aronson*, Kris Kitani, and Henny Admoni
IJRR, 2021
pdf /
bibtex /
Project Page
We present a multi-modal dataset of eye gaze, joystick activation, egocentric video, robot motion, and arm electromyography taken during a human-robot co-manipulation task under varying degrees of robotic assistance.
* denotes equal contribution
|
|
Examining the Effects of Anticipatory Robot Assistance on Human Decision Making
Benjamin A. Newman*, Abhijat Biswas*, Sarthak Ahuja, Siddharth Girdhar, Kris Kitani, and Henny Admoni
ICSR, 2020
pdf /
bibtex
We explore how robot motion that is expressed in advance of an expected phenomenon (e.g. a robot reaching for an object it expects the user will want) affects the eventual decision the person makes.
* denotes equal contribution
|
|
Visual Assistance for Object-Rearrangement Tasks in Augmented Reality
arXiv, 2020
Benjamin A. Newman, Kevin Carlberg, and Ruta Desai
pdf /
bibtex
We examine how presenting users with optimal routing assistance through a visual display would affect their ability and sense of agency when completing a complex object rearrangement task.
|
|
In-Sight: Tension-Based Haptic Feedback to Improve Navigation for People who are Blind
Alexander Baikovitz*, Jonathan Duffy*, Zachary Sussman*, Benjamin A. Newman, and Henny Admoni
CHI 2019 Workshop on Hacking Blind Navigation, 2019
pdf /
bibtex
We develop a portable haptic device that aids visually impaired users navigte in real world environments.
* denotes equal contribution
|
|
Global and Local Statistical Regularities Control Visual Attention to Object Sequences
Alexa Romberg, Yayun Zhang, Benjamin A. Newman, Jochen Triesch, and Chen Yu
ICDL Epi-Rob, 2016
pdf /
bibtex
We study how cross-situational statistics drive visual attention. Specifically, we examine how attention differs towards objects that are displayed infrequently versus those that are displayed frequently.
|
|
Hand-Eye Coordination Primitives for Assistive Robotic Co-Manipulation
Benjamin A. Newman, Kris Kitani, and Henny Admoni
pdf
We attempt to discover joint hand and eye gaze primitives for human robot co-manipulataion in an assisted eating task that could be useful for user goal recognition.
|
Thank you Jon Barron for creating and open-sourcing a fantastic website!
|
|