Author
Listed:
- Qin, Hongyi
- Zhu, Yifan
- Jiang, Yan
- Luo, Siqi
- Huang, Cui
Abstract
Artificial intelligence (AI) technologies, exemplified by health chatbots, are transforming the healthcare industry. Their widespread application has the potential to enhance decision-making efficiency, improve the quality of healthcare services, and reduce medical costs. While there is ongoing discussion about the opportunities and challenges brought by AI, more needs to be known about the public's attitude towards its use in the healthcare domain. Understanding public attitudes can help policymakers better grasp their needs and involve them in making decisions that benefit both technological development and social welfare. Therefore, this study presents evidence from two between-subjects experiments. This study aims to compare the public's adoption and trust levels in health advice provided by human vs. AI doctors and explore the potential effects of personalization and carefulness on the public's attitudes. Experimental designs adopt a trust-centered, cognitively and emotionally balanced perspective to study the public's intention to adopt AI. In Experiment 1, the experimental conditions involve the types of decision-makers providing online consultation advice, either AI or human doctors. In Experiment 2, the experimental conditions involve varying levels of perceived personalization and carefulness (high vs. low). A total of 734 participants took part in the study. They were randomly assigned to one of the intervention conditions and responded to manipulation checks after reading the materials. Using a seven-point Likert-type scale, participants rated their cognitive and emotional trust levels and intention to adopt the advice. Partial Least Squares Structural Equation Modeling (PLS-SEM) is conducted to estimate the proposed theoretical perspective. Qualitative interviews on both real-world and AI-generated treatment recommendations further enriched the understanding of public perceptions.The results show that AI-generated advice is generally slightly less trusted and adopted by the public. However, a noticeable inclination towards AI-generated advice emerges when AI demonstrates proficiency in understanding individuals' health conditions and providing empathetic consultations. Further analyses confirm the mediating influence of emotional trust between cognitive trust and adoption intention. These findings provide deeper insights into the process of adoption and trust formation. Moreover, they offer guidance to digital healthcare providers, empowering them with the knowledge to co-design AI implementation strategies that cater to the public's expectations.
Suggested Citation
Qin, Hongyi & Zhu, Yifan & Jiang, Yan & Luo, Siqi & Huang, Cui, 2024.
"Examining the impact of personalization and carefulness in AI-generated health advice: Trust, adoption, and insights in online healthcare consultations experiments,"
Technology in Society, Elsevier, vol. 79(C).
Handle:
RePEc:eee:teinso:v:79:y:2024:i:c:s0160791x24002744
DOI: 10.1016/j.techsoc.2024.102726
Download full text from publisher
As the access to this document is restricted, you may want to search for a different version of it.
Corrections
All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:eee:teinso:v:79:y:2024:i:c:s0160791x24002744. See general information about how to correct material in RePEc.
If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.
We have no bibliographic references for this item. You can help adding them by using this form .
If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.
For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: Catherine Liu (email available below). General contact details of provider: https://www.journals.elsevier.com/technology-in-society .
Please note that corrections may take a couple of weeks to filter through
the various RePEc services.