Models, code, and papers for "Changyan Chi":

If I Hear You Correctly: Building and Evaluating Interview Chatbots with Active Listening Skills

Feb 05, 2020
Ziang Xiao, Michelle X. Zhou, Wenxi Chen, Huahai Yang, Changyan Chi

Interview chatbots engage users in a text-based conversation to draw out their views and opinions. It is, however, challenging to build effective interview chatbots that can handle user free-text responses to open-ended questions and deliver engaging user experience. As the first step, we are investigating the feasibility and effectiveness of using publicly available, practical AI technologies to build effective interview chatbots. To demonstrate feasibility, we built a prototype scoped to enable interview chatbots with a subset of active listening skills - the abilities to comprehend a user's input and respond properly. To evaluate the effectiveness of our prototype, we compared the performance of interview chatbots with or without active listening skills on four common interview topics in a live evaluation with 206 users. Our work presents practical design implications for building effective interview chatbots, hybrid chatbot platforms, and empathetic chatbots beyond interview tasks.

* Working draft. To appear in the ACM CHI Conference on Human Factors in Computing Systems (CHI 2020) 

  Access Model/Code and Paper
Tell Me About Yourself: Using an AI-Powered Chatbot to Conduct Conversational Surveys

May 25, 2019
Ziang Xiao, Michelle X. Zhou, Q. Vera Liao, Gloria Mark, Changyan Chi, Wenxi Chen, Huahai Yang

The rise of increasingly more powerful chatbots offers a new way to collect information through conversational surveys, where a chatbot asks open-ended questions, interprets a user's free-text responses, and probes answers when needed. To investigate the effectiveness and limitations of such a chatbot in conducting surveys, we conducted a field study involving about 600 participants. In this study, half of the participants took a typical online survey on Qualtrics and the other half interacted with an AI-powered chatbot to complete a conversational survey. Our detailed analysis of over 5200 free-text responses revealed that the chatbot drove a significantly higher level of participant engagement and elicited significantly better quality responses in terms of relevance, depth, and readability. Based on our results, we discuss design implications for creating AI-powered chatbots to conduct effective surveys and beyond.

* Currently under review 

  Access Model/Code and Paper
Sclera Force Control in Robot-assisted Eye Surgery: Adaptive Force Control vs. Auditory Feedback

Jan 10, 2019
Ali Ebrahimi, Changyan He, Niravkumar Patel, Marin Kobilarov, Peter Gehlbach, Iulian Iordachita

Surgeon hand tremor limits human capability during microsurgical procedures such as those that treat the eye. In contrast, elimination of hand tremor through the introduction of microsurgical robots diminishes the surgeon's tactile perception of useful and familiar tool-to-sclera forces. While the large mass and inertia of eye surgical robot prevents surgeon microtremor, loss of perception of small scleral forces may put the sclera at risk of injury. In this paper, we have applied and compared two different methods to assure the safety of sclera tissue during robot-assisted eye surgery. In the active control method, an adaptive force control strategy is implemented on the Steady-Hand Eye Robot in order to control the magnitude of scleral forces when they exceed safe boundaries. This autonomous force compensation is then compared to a passive force control method in which the surgeon performs manual adjustments in response to the provided audio feedback proportional to the magnitude of sclera force. A pilot study with three users indicate that the active control method is potentially more efficient.

* Conference paper accepted for International Symposium on Medical Robotics (ISMR) 2019, 7 pages (6 pages manuscript and 1 page reference), 7 figures, 1 table 

  Access Model/Code and Paper