Assistant Professor, Cornell Tech
Information Science Department, Field membership in Computer Science
College of Computing & Information Science
Visiting Research Scientist, Meta Reality Labs Research, (2021-22)
Ph.D. Computer Science and Engineering, UC San Diego (2021)
B.S. in Electrical Engineering, University of Missouri-Columbia (2015)
B.S. in Computer Engineering, University of Missouri-Columbia (2015)
A.S. in Engineering Science, Saint Louis Community College (2012)
I direct the Artificial Intelligence & Robotics Lab (AIRLab), which focuses on research in the field of human-robot interaction (HRI). My research is at the intersection of robotics, computer vision, and artificial intelligence. AIRLab research lab designs intelligent systems that work alongside groups of people in real-world, safety-critical environments. These systems are realized through community engagement with stakeholders to develop robot vision and planning algorithms, and augmented reality systems that support multi-user interactions.
I have received funding from Google Award for Inclusion Research, NSF GRFP, Microsoft Dissertation Award, the Google Anita Borg Memorial Fellowship, the Arthur J. Schmitt Presidential Fellowship, a GEM Fellowship, and an award from the National Center for Women in Information Technology (NCWIT). Awards include Best Paper Honorable Mention at CSCW 2019 and Best Paper Award at HRI 2022.
[Spring 2024] Practical Applications in Machine Learning, Cornell Tech
[Fall 2023] Robotics Seminar, Cornell Tech
[Spring 2023] Practical Applications in Machine Learning, Cornell Tech
[Nov 2021]: Paper accepted to HRI 2022!
[March 2022]: presented REGROUP at HRI 2022!
[July 2022]: Start appointment at Cornell Tech!
[May 2023]: speaker at Data-Driven Urban Tech: How Machine Learning and Optimization Addresses Today’s Urban Challenges
[Nov 2021]: Accepted Job Offer at Cornell Tech!!
[april 2023]: Guest Lecturer, Introduction to HRI Course, University of Virginia
Issues in Science and Technology. 38, no. 4, 93–95.
REGROUP: A Robot-Centric Group Detection and Tracking System.
In Proc. of the 17th Annual ACM/IEEE Conference on Human Robot Interaction (HRI). [Acceptance rate: 24.8%]
Situating Robots in the Emergency Department.
AAAI Spring Symposium on Applied AI in Healthcare: Safety, Community, and the Environment, 2020.
Coordinating Clinical Teams: Using Robots to empower nurses to Stop the Line.
Computer Supported Cooperative Work (CSCW), 2019.
[Acceptance Rate: 30%][Video]
Taylor, A. and Riek, L.D. (2017)
Robot Perception of Social Engagement Using
Group Joint Action.
In Proc. of the 7th Annual Joint Action Meeting (JAM).
Robot Navigation in Risky, Crowded Environments: Understanding Human Preferences
IEEE Robotics and Automation Letters, pp. 1-8, 10.1109/LRA.2023.3290533
In Proc. of the ACM Conference on Computer Supported Collaborative Work (CSCW) [Acceptance rate: <30%]
Social Navigation for Mobile Robots in the Emergency Department.
In Proc. of the IEEE International Conference on Robotics and Automation (ICRA), 2021. [Acceptance Rate: 48%]
Robot-Centric Perception of Human Groups.
ACM Transactions on Human-Robot Interaction (THRI), 2020.
Taylor, A., Du, X., Chen, C., Zare, A. (2014)
Context Dependent Target Detection
Computational Intelligence Society Poster Competition, University of Missouri, Columbia.
Hospitals of the Future: Designing Interactive Robotic Systems for Resilient Emergency Departments
The Emergency Department (ED) is a stressful, safety-critical environment, which is overcrowded, noisy, chaotic, and understaffed. The built environment plays a key role in patient outcomes, experiences, and the mental health of healthcare workers (HCWs). However, once a space is built it is difficult to change it; so the modularity and adaptability of new technologies such as robots could potentially help stakeholders mitigate some of these challenges; yet, there is a lack of research in this area, particularly in the ED. In this paper, we address this gap by engaging HCWs in a research-through-design process, utilizing design fiction, to envision a future resilient ED. Here, robots scurry along the ceiling, provide help at the bedside, and smart furniture and walls provide spaces for privacy and calm. We co-created design prototypes of future intelligent systems that can modify the built environment to support resilience, which we then used to co-create a Design Catalog with HCWs, which contains a collection of future technology prototypes contextualized within the ED. We found that HCWs envisioned many ways for intelligent systems to help them reimagine the built environment, including ways to enhance HCW-patient communication, improve patient experience, support both HCW and patient safety, and use reconfigurable spaces to support privacy. We hope our work inspires further exploration into using new technologies to reimagine and reconfigure the built environment to support resilient hospitals. [PDF][Design Catalog]
I am thankful to have the support of the National Science Foundation Graduate Research Fellowship (NSF GRFP), Arthur J. Schmitt Presidential Fellowship, GEM National Consortium, Google Anita Borg Memorial Scholarship, National Center for Women in IT (NCWIT), Microsoft Research Dissertation Grant, and the Grace Hopper Celebration of Women in Computing (GHC) Scholarship.