Access the full text.
Sign up today, get DeepDyve free for 14 days.
Katie Genter, Noam Agmon, P. Stone (2013)
Ad hoc teamwork for leading a flock
Mykel Kochenderfer, Chris Amato, Girish Chowdhary, J. How, H. Reynolds, Jason Thornton, P. Torres-Carrasquillo, N. Ure, J. Vian (2015)
Decision Making Under Uncertainty: Theory and Application
(1965)
On the shortest arborescence of a directed graph
(1968)
Optimum branchings. Mathematics and the Decision Sciences, Part
Michael Rubenstein, Adrian Cabrera, Justin Werfel, Golnaz Habibi, J. McLurkin, R. Nagpal (2013)
Collective transport of complex objects by simple robots: theory and experiments
Minae Kwon, Erdem Biyik, Aditi Talati, Karan Bhasin, Dylan Losey, Dorsa Sadigh (2020)
When Humans Aren’t Optimal: Robots that Collaborate with Risk-Aware Humans2020 15th ACM/IEEE International Conference on Human-Robot Interaction (HRI)
Mathieu Goc, Lawrence Kim, Ali Parsaei, Jean-Daniel Fekete, Pierre Dragicevic, Sean Follmer (2016)
Zooids: Building Blocks for Swarm User InterfacesProceedings of the 29th Annual Symposium on User Interface Software and Technology
Samuel Stone (2011)
Ad Hoc Teamwork Modeled with Multi-armed Bandits: An Extension to Discounted Infinite Rewards
Michael Bowling, Peter Mccracken (2005)
Coordination and Adaptation in Impromptu Teams
Chris Amato, Girish Chowdhary, A. Geramifard, N. Ure, Mykel Kochenderfer (2015)
Decentralized control of Partially Observable Markov Decision Processes using belief space macro-actions2015 IEEE International Conference on Robotics and Automation (ICRA)
Stefano Albrecht (2015)
Utilising policy types for effective ad hoc coordination in multiagent systems
Dorsa Sadigh (2017)
Safe and Interactive Autonomy: Control, Learning, and Verification
J Edmonds (1968)
Optimum branchingsMathematics and the Decision Sciences, Part 1, 25
I. Giardina (2008)
Collective behavior in animal groups: Theoretical models and empirical studiesHFSP Journal, 2
A. Strandburg-Peshkin, Colin Twomey, N. Bode, Albert Kao, Y. Katz, C. Ioannou, S. Rosenthal, C. Torney, H. Wu, S. Levin, I. Couzin (2013)
Visual sensory networks and effective information transfer in animal groupsCurrent Biology, 23
S. Chernova, A. Thomaz (2014)
Robot Learning from Human Teachers
Mengxi Li, Dylan Losey, J. Bohg, Dorsa Sadigh (2020)
Learning User-Preferred Mappings for Intuitive Robot Control2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS)
(2018)
Graph convolutional reinforcement learning. arXiv preprint arXiv:1810.09202
Owen Macindoe, L. Kaelbling, Tomas Lozano-Perez (2012)
POMCoP: Belief Space Planning for Sidekicks in Cooperative GamesProceedings of the AAAI Conference on Artificial Intelligence and Interactive Digital Entertainment
A. Gray, Yiqi Gao, J. Hedrick, F. Borrelli (2013)
Robust Predictive Control for semi-autonomous vehicles with an uncertain driver model2013 IEEE Intelligent Vehicles Symposium (IV)
Ali-akbar Agha-mohammadi, S. Chakravorty, N. Amato (2014)
FIRM: Sampling-based feedback motion-planning under motion uncertainty and imperfect measurementsThe International Journal of Robotics Research, 33
H. Kurniawati, David Hsu, Wee Lee (2008)
SARSOP: Efficient Point-Based POMDP Planning by Approximating Optimally Reachable Belief Spaces, 04
Chandrayee Basu, Erdem Biyik, Zhixun He, M. Singhal, Dorsa Sadigh (2019)
Active Learning of Reward Dynamics from Hierarchical Queries2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS)
Chelsea Finn, S. Levine, P. Abbeel (2016)
Guided Cost Learning: Deep Inverse Optimal Control via Policy Optimization
Kim Baraka, Ana Paiva, M. Veloso (2015)
Expressive Lights for Revealing Mobile Service Robot State
J. Barraquand, B. Langlois, J. Latombe (1991)
Numerical potential field techniques for robot path planningFifth International Conference on Advanced Robotics 'Robots in Unstructured Environments
Christoforos Mavrogiannis, Ross Knepper (2016)
Decentralized Multi-Agent Navigation Planning with Braids
Dorsa Sadigh, S. Sastry, S. Seshia, A. Dragan (2016)
Planning for Autonomous Cars that Leverage Effects on Human Actions, 12
Shariq Iqbal, Fei Sha (2018)
Actor-Attention-Critic for Multi-Agent Reinforcement Learning
P. Stone, G. Kaminka, Sarit Kraus, J. Rosenschein, Noam Agmon (2013)
Teaching and leading an ad hoc teammate: Collaboration without pre-coordinationArtif. Intell., 203
Yedid Hoshen (2017)
VAIN: Attentional Multi-agent Predictive Modeling
S. Nikolaidis, Swaprava Nath, A. Procaccia, S. Srinivasa (2017)
Game-Theoretic Modeling of Human Adaptation in Human-Robot Collaboration2017 12th ACM/IEEE International Conference on Human-Robot Interaction (HRI
Robert Platt, Russ Tedrake, L. Kaelbling, Tomas Lozano-Perez (2010)
Belief space planning assuming maximum likelihood observations, 06
Carlos Garcia, D. Prett, M. Morari (1989)
Model predictive control: Theory and practice - A surveyAutom., 25
Erdem Biyik, Malayandi Palan, Nicholas Landolfi, Dylan Losey, Dorsa Sadigh (2019)
Asking Easy Questions: A User-Friendly Approach to Active Reward LearningArXiv, abs/1910.04365
Samuel Barrett (2015)
Making Friends on the Fly: Advances in Ad Hoc Teamwork, 603
Shervin Javdani, H. Admoni, Stefania Pellegrinelli, S. Srinivasa, J. Bagnell, Shervin Carnegie (2017)
Shared autonomy via hindsight optimization for teleoperation and teamingThe International Journal of Robotics Research, 37
Dorsa Sadigh, S. Sastry, S. Seshia, A. Dragan (2016)
Information gathering actions over human internal state2016 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS)
Hong Jeon, Dylan Losey, Dorsa Sadigh (2020)
Shared Autonomy with Learned Latent ActionsArXiv, abs/2005.03210
Dylan Hadfield-Menell, Stuart Russell, P. Abbeel, A. Dragan (2016)
Cooperative Inverse Reinforcement Learning
E. Cristiani, B. Piccoli (2009)
A unifying model for the structure of animal groups on the move
Alejandro Mottini, Rodrigo Acuna-Agost (2017)
Deep Choice Model Using Pointer Networks for Airline Itinerary PredictionProceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining
Chris Baker, R. Saxe, J. Tenenbaum (2009)
Action understanding as inverse planningCognition, 113
J. Lai (2000)
Conversational interfacesCommun. ACM, 43
Barış Akgün, M. Cakmak, Karl Jiang, A. Thomaz (2012)
Keyframe-based Learning from DemonstrationInternational Journal of Social Robotics, 4
R. Karp (1971)
A simple derivation of Edmonds' algorithm for optimum branchingsNetworks, 1
Aaron Bestick, R. Bajcsy, A. Dragan (2016)
Implicitly Assisting Humans to Choose Good Grasps in Robot to Human Handovers
S. Rosenthal, Colin Twomey, Andrew Hartnett, H. Wu, I. Couzin (2015)
Revealing the hidden networks of interaction in mobile animal groups allows prediction of complex behavioral contagionProceedings of the National Academy of Sciences, 112
Sheng Li, Jayesh Gupta, Peter Morales, R. Allen, Mykel Kochenderfer (2020)
Deep Implicit Coordination Graphs for Multi-agent Reinforcement Learning
Martin Liebner, M. Baumann, F. Klanner, C. Stiller (2012)
Driver intent inference at urban intersections using the intelligent driver model2012 IEEE Intelligent Vehicles Symposium
Malayandi Palan, Nicholas Landolfi, Gleb Shevchuk, Dorsa Sadigh (2019)
Learning Reward Functions by Integrating Human Demonstrations and PreferencesArXiv, abs/1906.08928
Zijian Wang, M. Schwager (2016)
Force-Amplifying N-robot Transport System (Force-ANTS) for cooperative planar manipulation without communicationThe International Journal of Robotics Research, 35
S. Nikolaidis, J. Shah (2013)
Human-robot cross-training: Computational formulation, modeling and evaluation of a human team training strategy2013 8th ACM/IEEE International Conference on Human-Robot Interaction (HRI)
Dylan Losey, K. Srinivasan, Ajay Mandlekar, Animesh Garg, Dorsa Sadigh (2019)
Controlling Assistive Robots with Learned Latent Actions2020 IEEE International Conference on Robotics and Automation (ICRA)
Piyush Khandelwal, Samuel Barrett, P. Stone (2015)
Leading the Way: An Efficient Multi-robot Guidance System
(1972)
Theories of bounded rationality
H. Çelikkanat, E. Sahin (2010)
Steering self-organized robot flocks through externally guided individualsNeural Computing and Applications, 19
P. Abbeel, A. Ng (2004)
Apprenticeship learning via inverse reinforcement learningProceedings of the twenty-first international conference on Machine learning
(2019)
Bucquet, A.,&Sadigh,D
Brian Ziebart, Nathan Ratliff, G. Gallagher, C. Mertz, K. Peterson, J. Bagnell, M. Hebert, A. Dey, S. Srinivasa (2009)
Planning-based prediction for pedestrians2009 IEEE/RSJ International Conference on Intelligent Robots and Systems
Lisa Ordóñez, Lehman Benson (1997)
Decisions under Time Pressure: How Time Constraint Affects Risky Decision MakingOrganizational Behavior and Human Decision Processes, 71
Brian Ziebart, Andrew Maas, J. Bagnell, A. Dey (2008)
Maximum Entropy Inverse Reinforcement Learning
Kathryn Genter (2017)
Fly with me : algorithms and methods for influencing a flock
F. Sudweeks, S. Simoff (2005)
Leading conversations: Communication behaviours of emergent leaders in virtual teamsProceedings of the 38th Annual Hawaii International Conference on System Sciences
D. Derue (2011)
Adaptive leadership theory: Leading and following as a complex adaptive processResearch in Organizational Behavior, 31
A. Xie, Dylan Losey, R. Tolsma, Chelsea Finn, Dorsa Sadigh (2020)
Learning Latent Representations to Influence Multi-Agent InteractionArXiv, abs/2011.06619
S. Kerman, D. Brown, M. Goodrich (2012)
Supporting human interaction with robust robot swarms2012 5th International Symposium on Resilient Control Systems
Daniel Brown, S. Kerman, M. Goodrich (2014)
Human-Swarm Interactions Based on Managing Attractors2014 9th ACM/IEEE International Conference on Human-Robot Interaction (HRI)
T. Kanda, Takayuki Hirano, Daniel Eaton, H. Ishiguro (2004)
Interactive Robots as Social Partners and Peer Tutors for Children: A Field TrialHuman–Computer Interaction, 19
T. Osogami, Makoto Otsuka (2014)
Restricted Boltzmann machines modeling human choice
P. Stone, Sarit Kraus (2010)
To teach or not to teach?: decision making under uncertainty in ad hoc teams
M. Littman, P. Stone (2001)
Leading Best-Response Strategies in Repeated Games
I. Couzin, J. Krause, N. Franks, S. Levin (2005)
Effective leadership and decision-making in animal groups on the moveNature, 433
S. Nikolaidis, Anton Kuznetsov, David Hsu, S. Srinivasa (2016)
Formalizing human-robot mutual adaptation: A bounded memory model2016 11th ACM/IEEE International Conference on Human-Robot Interaction (HRI)
Haoyu Bai, Shaojun Cai, N. Ye, David Hsu, Wee Lee (2015)
Intention-aware online POMDP planning for autonomous driving in a crowd2015 IEEE International Conference on Robotics and Automation (ICRA)
Chih-Han Yu, Justin Werfel, R. Nagpal (2010)
Collective decision-making in multi-agent systems by implicit leadership
Dorsa Sadigh, A. Dragan, S. Sastry, S. Seshia (2017)
Active Preference-Based Learning of Reward Functions, 13
M. Gombolay, Cindy Huang, J. Shah (2015)
Coordination of Human-Robot Teaming with Human Task Preferences
M. Schippers, Deanne Hartog, P. Koopman, D. Knippenberg (2007)
The role of transformational leadership in enhancing team reflexivityHuman Relations, 61
B. Robins, K. Dautenhahn, R. Boekhorst, A. Billard (2004)
Effects of repeated exposure to a humanoid robot on children with autism
S Chernova, AL Thomaz (2014)
Robot learning from human teachersSynthesis Lectures on Artificial Intelligence and Machine Learning, 8
D. Szafir, Bilge Mutlu, T. Fong (2015)
Communicating Directionality in Flying Robots2015 10th ACM/IEEE International Conference on Human-Robot Interaction (HRI)
(2019)
Influ- encing leading and following in human-robot teams
Somchaya Liemhetcharat (2013)
Representation, Planning, and Learning of Dynamic Ad Hoc Robot Teams
Rohan Tiwari, P. Jain, S. Butail, Sujit Baliyarasimhuni, M. Goodrich (2017)
Effect of Leader Placement on Robotic Swarm Control
P. Stone, G. Kaminka, Sarit Kraus, J. Rosenschein (2010)
Ad Hoc Autonomous Agent Teams: Collaboration without Pre-CoordinationProceedings of the AAAI Conference on Artificial Intelligence
Katie Genter, Noam Agmon, P. Stone (2011)
Role-Based Ad Hoc TeamworkProceedings of the AAAI Conference on Artificial Intelligence
A. Dragan, S. Srinivasa (2012)
Formalizing Assistive Teleoperation, 08
T. Nguyen, David Hsu, Wee Lee, T. Leong, L. Kaelbling, Tomas Lozano-Perez, A. Grant (2011)
CAPIR: Collaborative Action Planning with Intention RecognitionArXiv, abs/1206.5928
Dorsa Sadigh, Nicholas Landolfi, S. Sastry, S. Seshia, A. Dragan (2018)
Planning for cars that coordinate with people: leveraging effects on human actions for planning and active information gathering over human internal stateAutonomous Robots, 42
Erdem Biyik, Dorsa Sadigh (2018)
Batch Active Preference-Based Learning of Reward Functions
Dylan Losey, Mengxi Li, J. Bohg, Dorsa Sadigh (2019)
Learning from My Partner's Actions: Roles in Decentralized Robot Teams
F. Broz, I. Nourbakhsh, R. Simmons (2011)
Designing POMDP models of socially situated tasks2011 RO-MAN
F. Doshi-Velez, N. Roy (2007)
Efficient model learning for dialog management2007 2nd ACM/IEEE International Conference on Human-Robot Interaction (HRI)
A. Tversky, D. Kahneman (1992)
Advances in prospect theory: Cumulative representation of uncertaintyJournal of Risk and Uncertainty, 5
Roles such as leading and following can emerge naturally in human groups. However, in human–robot teams, such roles are often predefined due to the difficulty of scalably learning and adapting to them. In this work, we enable a robot to efficiently learn how group dynamics emerge and evolve in human teams and we leverage this understanding to plan for influencing actions for autonomous robots that guide the team toward achieving a common goal. We first develop an effective and concise representation of group dynamics, such as leading and following, by enforcing a graph structure while learning the weights of the edges corresponding to one-to-one relationships between the agents. We then develop an optimization-based robot policy that leverages this graph representation to attain an objective by influencing a human team. We apply our framework to two types of group dynamics, leading-following and predator–prey, and show that our structured representation is scalable with different human team sizes and also generalizable across different tasks. We also show that robots that utilize this representation are able to successfully influence a group to achieve various goals compared to robots that do not have access to these graph representations (Parts of this work has been published at Robotics: Science and Systems (RSS) (Kwon et al. in Proceedings of robotics: science and systems (RSS), 2019. https://doi.org/10.15607/rss.2019.xv.075).
Autonomous Robots – Springer Journals
Published: Oct 1, 2021
Keywords: Human–robot teaming; Human modeling; Multiagent systems
Read and print from thousands of top scholarly journals.
Already have an account? Log in
Bookmark this article. You can see your Bookmarks on your DeepDyve Library.
To save an article, log in first, or sign up for a DeepDyve account if you don’t already have one.
Copy and paste the desired citation format or use the link below to download a file formatted for EndNote
Access the full text.
Sign up today, get DeepDyve free for 14 days.
All DeepDyve websites use cookies to improve your online experience. They were placed on your computer when you launched this website. You can change your cookie settings through your browser.