The results show that feedback has a more significant impact on improving users' trust in AI than explainability, but this enhanced trust does not lead to a corresponding performance improvement. Further exploration suggests that feedback induces users to over-trust (i.e., accept the AI's suggestions when it is wrong) or distrust (ignore the AI's suggestions when it is correct), which may negate the benefits of increased trust, leading to a "trust-performance paradox". The researchers call for future research to focus on how to design strategies to ensure that explanations foster appropriate trust to improve the efficiency of human-robot collaboration.

Xue Zhirong, Designer, Interaction Design, Human-Computer Interaction, Artificial Intelligence, Official Website, Blog, Creator, Author, Engineer, Paper, Product Design, Research, AI, HCI, Design, Learning, Knowledge Base, xuezhirong, UX, Design, Research, AI, HCI, Designer, Engineer, Author, Blog, Papers, Product Design, Study, Learning, User Experience
A2: Although it is generally believed that the explanatory nature of the model helps to improve user trust, the experimental results show that this enhancement is not significant and not as effective as feedback. In specific cases, such as areas of low expertise, some form of interpretation may result in only a modest increase in appropriate trust.
Personal insights The researchers conducted two sets of experiments ("Predict the speed-dating outcomes and get up to $6 (takes less than 20 min)" and a similar Prolific experiment) in which participants interacted with the AI system in a task of predicting the outcome of a dating to explore the impact of model explainability and feedback on user trust in AI and prediction accuracy. The results show that although explainability (e.g., global and local interpretation) does not significantly improve trust, feedback can most consistently and significantly improve behavioral trust. However, increased trust does not necessarily lead to the same level of performance gains, i.e., there is a "trust-performance paradox". Exploratory analysis reveals the mechanisms behind this phenomenon. Xue Zhirong's knowledge base Xue Zhirong's knowledge base The results show that feedback has a more significant impact on improving users' trust in AI than explainability, but this enhanced trust does not lead to a corresponding performance improvement. Further exploration suggests that feedback induces users to over-trust (i.e., accept the AI's suggestions when it is wrong) or distrust (ignore the AI's suggestions when it is correct), which may negate the benefits of increased trust, leading to a "trust-performance paradox". The researchers call for future research to focus on how to design strategies to ensure that explanations foster appropriate trust to improve the efficiency of human-robot collaboration. speech Q3: How does result feedback and model interpretability affect user task performance? Based on large language model generation, there may be a risk of errors. interview
Conference
The Interpretability of Artificial Intelligence and the Impact of Outcome Feedback on Trust: A Comparative Study
Interview Personal insights To assess trust more accurately, the researchers used behavioral trust (WoA), a measure that takes into account the difference between the user's predictions and the AI's recommendations, and is independent of the model's accuracy. By comparing WoA under different conditions, researchers can analyze the relationship between trust and performance. Q2: Does explainability necessarily enhance users' trust in AI? The results show that feedback has a more significant impact on improving users' trust in AI than explainability, but this enhanced trust does not lead to a corresponding performance improvement. Further exploration suggests that feedback induces users to over-trust (i.e., accept the AI's suggestions when it is wrong) or distrust (ignore the AI's suggestions when it is correct), which may negate the benefits of increased trust, leading to a "trust-performance paradox". The researchers call for future research to focus on how to design strategies to ensure that explanations foster appropriate trust to improve the efficiency of human-robot collaboration. To assess trust more accurately, the researchers used behavioral trust (WoA), a measure that takes into account the difference between the user's predictions and the AI's recommendations, and is independent of the model's accuracy. By comparing WoA under different conditions, researchers can analyze the relationship between trust and performance. Personal insights summary Xue Zhirong, Designer, Interaction Design, Human-Computer Interaction, Artificial Intelligence, Official Website, Blog, Creator, Author, Engineer, Paper, Product Design, Research, AI, HCI, Design, Learning, Knowledge Base, xuezhirong, UX, Design, Research, AI, HCI, Designer, Engineer, Author, Blog, Papers, Product Design, Study, Learning, User Experience The content is made up of: speech
Xue Zhirong, Designer, Interaction Design, Human-Computer Interaction, Artificial Intelligence, Official Website, Blog, Creator, Author, Engineer, Paper, Product Design, Research, AI, HCI, Design, Learning, Knowledge Base, xuezhirong, UX, Design, Research, AI, HCI, Designer, Engineer, Author, Blog, Papers, Product Design, Study, Learning, User Experience
Problem finding
Draw inferences
Problem finding

Personal insightsDraw inferences

blog0Original address:
Q2: Does explainability necessarily enhance users' trust in AI?
简体模式 繁體模式