Social Robot Lab studies communication between humans and robots in various social contexts. We aim to perform innovative, quality research to generate knowledge and develop theories in human-robot communication and disseminate our findings to the academic community and the public.
Social Robot Lab contributes to society through service-learning and community engagement. We aim to use social robots to help people lead happier, healthier, and more fulfilling lives. We also contribute to the Chapman University community by participating in various campus events.
Social Robot Lab provides professional development opportunities for all members. Lab members will learn theories and technologies to build robots that are human-like, sociable, and intuitive to interact with. Lab members are also encouraged to pursue their own research and creativity projects.
A robot can effectively get people to do as the robot asks by capitalizing on the norm of reciprocity and following with a verbal request. We found that 60% of participants who were helped by a robot in a trivia game just for five minutes returned the robot’s favor by completing boring tasks for 15 minutes, no matter they liked the robot or not. This compliance rate was significantly higher than that of those who were not helped by the robot (33%).
Why do people treat robots and computers as if they were humans? Probably because people operate on a kind of autopilot when they interact with technology. In our study, a computer agent requested participants to do boring tasks. Mindless participants compiled when any kind of reason was given with the magic word “because.” Attentive participants complied only when a proper reason was given.
Using a nationally representative sample, we found that 26% of US population report a heightened level of fear toward autonomous robots and artificial intelligence. This fear was connected to sex, age, education, and household income, as well as media exposure to science fiction. It is also related to other types of fear, such as loneliness, becoming unemployed, and drone use.
We examined the effects of user-generated content on human-robot trust as well as interaction outcomes. We found that reading online reviews about robot partners before interaction augmented the effects of human-robot trust on interaction outcomes. Specifically, it elicited positive mood and more favorable evaluations of the robot. This finding suggests that messages can be strategically deployed to enhance trust and interaction outcomes.
From factories, offices and medical centers to our homes, cars and even our local coffee shops, robots and artificial intelligence are playing an increasingly greater role in how we work, live and play...
Cincinnati Public Radio
October 3, 2017
As social robots permeate wide segments of our daily lives, they may become capable of influencing our attitudes, beliefs, and behaviors. In other words, robots may be able to persuade us, and this is already happening...
Cincinnati Business Courier
March 24, 2017
It’s no secret that the College of Informatics has the most technologically advanced building on campus — Griffin Hall. But Griffin Hall is also home to Pineapple. Dr. Lee and Dr. Liang are using Pineapple to conduct studies on how humans interact socially when a robot is present...
NKU Inside
October, 2015
Between the studies on digital media production and computer sciences, Griffin Hall is home to some impressive technology. You might start seeing a new, digital face wandering the halls of the College of Informatics. That digital face is known as Pineapple, and it will be used for the next three years for research into social robotics...
The Northerner
October 19, 2015
Top Paper Award, National Communication Association, Communication and the Future Division (Lee & Liang, 2017)
Valerie Scudder Award: Jake Liang, Chapman University (2017)
Excellence in Research, Scholarship, and Creative Activity: Austin Lee, Northern Kentucky University (2017)
Faculty Research Excellence Award: Jake Liang, Chapman University (2016)
Top Paper Awards, International Communication Association, Instructional and Developmental Communication Division (Liang, 2015)
Pedagogical Innovations Grant: Austin Lee, Chapman University (2019)
Faculty Summer Fellowship: Austin Lee, Northern Kentucky University (2015)
Faculty Project Grant: Austin Lee, Northern Kentucky University (2015)
Research Grant: Jake Liang, Chapman University (2015)
Seed Grant: Austin Lee, Northern Kentucky University (2015)
Lee, S. A., & Liang, Y. (2019). A communication model of human-robot trust development for inclusive education. In J. Knox, Y. Wang, & M. S. Gallagher (Eds.), Speculative futures for artificial intelligence and educational inclusion (pp. 101-115). New York, NY: Springer Nature.
Lee, S. A., & Liang, Y. (2019). Robotic foot-in-the-door: Using sequential-request persuasive strategies in human-robot interaction. Computers in Human Behavior, 90, 351-356.
Lee, S. A., & Liang, Y. (2018). Theorizing verbally persuasive robots. In A. L. Guzman (Ed.), Human-machine communication: Rethinking communication, technology, and ourselves (pp. 119-143). New York, NY: Peter Lang.
Lee, S. A., & Liang, Y. (2018). Robotic foot-in-the-door: Using sequential-request persuasive strategies in human-robot interaction. Computers in Human Behavior.
Liang, Y., & Lee, S. A. (2017). Fear of autonomous robots: Evidence from national representative data with probability sampling. International Journal of Social Robotics, 9, 379-384.
Liang, Y., & Lee, S. A. (2016). Advancing the strategic messages affecting robot trust effect: The dynamic of user- and robot-generated content on human-robot trust and interaction outcomes. Cyberpsychology, Behavior, and Social Networking, 19, 538-544.
Lee, S. A., & Liang, Y. (2016). The role of reciprocity in verbally persuasive robots. Cyberpsychology, Behavior, and Social Networking. 19, 524-527.
Lee. S. A., & Liang, Y. (2015). Reciprocity in computer-human interaction: Source-based, norm-based, and affect-based explanations. Cyberpsychology, Behavior, and Social Networking, 18, 234-240.
Liang, Y., Lee, S. A., & Jang, J. (2013). Mindlessness and gaining compliance in computer-human interaction. Computers in Human Behavior, 29, 1572-1579.
Lee, S. A. (2019). Human-robot proxemics and compliance gaining. Paper presented at the 69th annual convention of the International Communication Association, Washington, DC.
Liang, Y., Lee, S. A., & Kee, K. F. (2019). The adoption of collaborative robots toward ubiquitous diffusion: A research agenda. Paper presented at the 14th ACM/IEEE International Conference on Human-Robot Interaction (HRI 2019) Workshop, Daegu, South Korea.
Lee, S. A., Appelman, A., & Waldridge, Z. J. (2018). Robot message credibility. Paper presented at the 68th annual convention of the International Communication Association, Prague, Czech.
Lee, S. A., & Liang, Y. (2017). Theorizing verbally persuasive robots. Paper presented at the 103rd annual convention of the National Communication Association, Dallas, TX.
Lee, S. A., Liang, Y., & Thompson, A. M. (2017). Robotic foot-in-the-door: Using sequential-request persuasive strategies in human-robot interaction. Paper presented at the 67th annual convention of the International Communication Association, San Diego, CA.
Liang, Y., & Lee, S. A. (2016). Employing user-generated content to enhance human-robot interaction in a human-robot trust game. In Proceedings of the 11th ACM/IEEE International Conference on Human-Robot Interaction (HRI 2016). Christchurch, New Zealand.
Lee, S. A., Liang, Y., & Cho, S. (2016). Effects of anthropomorphism and reciprocity in persuasive computer agents. Paper presented at the at the 102nd annual convention of the National Communication Association, Philadelphia, PA.
Cho, S., Lee, S. A., & Liang, Y. (2016). Using anthropomorphic agents for persuasion. Paper presented at the 66th annual convention of the International Communication Association, Fukuoka, Japan.
Becket Building 111
Chapman University
Orange, CA 92866