The results show that feedback has a more significant impact on improving users' trust in AI than explainability, but this enhanced trust does not lead to a corresponding performance improvement. Further exploration suggests that feedback induces users to over-trust (i.e., accept the AI's suggestions when it is wrong) or distrust (ignore the AI's suggestions when it is correct), which may negate the benefits of increased trust, leading to a "trust-performance paradox". The researchers call for future research to focus on how to design strategies to ensure that explanations foster appropriate trust to improve the efficiency of human-robot collaboration.
Q2: Does explainability necessarily enhance users' trust in AI?
The content is made up of:
Xue Zhirong, Designer, Interaction Design, Human-Computer Interaction, Artificial Intelligence, Official Website, Blog, Creator, Author, Engineer, Paper, Product Design, Research, AI, HCI, Design, Learning, Knowledge Base, xuezhirong, UX, Design, Research, AI, HCI, Designer, Engineer, Author, Blog, Papers, Product Design, Study, Learning, User Experience
The Interpretability of Artificial Intelligence and the Impact of Outcome Feedback on Trust: A Comparative Study
User experience About me
Q2: Does explainability necessarily enhance users' trust in AI?
The content is made up of:
thesis
About me

A1: According to research, feedback (e.g. result output) is a key factor influencing user trust. It is the most significant and reliable way to increase user trust in AI behavior.

About me
outcome
The content is made up of:

The Interpretability of Artificial Intelligence and the Impact of Outcome Feedback on Trust: A Comparative Study | Xue Zhirong's knowledge base

The content is made up of:
The results show that feedback has a more significant impact on improving users' trust in AI than explainability, but this enhanced trust does not lead to a corresponding performance improvement. Further exploration suggests that feedback induces users to over-trust (i.e., accept the AI's suggestions when it is wrong) or distrust (ignore the AI's suggestions when it is correct), which may negate the benefits of increased trust, leading to a "trust-performance paradox". The researchers call for future research to focus on how to design strategies to ensure that explanations foster appropriate trust to improve the efficiency of human-robot collaboration.
The Interpretability of Artificial Intelligence and the Impact of Outcome Feedback on Trust: A Comparative Study
The Interpretability of Artificial Intelligence and the Impact of Outcome Feedback on Trust: A Comparative Study
Personal insights

The Interpretability of Artificial Intelligence and the Impact of Outcome Feedback on Trust: A Comparative Study | Xue Zhirong's knowledge base

Personal insights
Q3: How does result feedback and model interpretability affect user task performance?
A3: The study found that the feedback of the results can improve the accuracy of the user's predictions (reducing the absolute error), thereby improving the performance of working with AI. However, interpretability does not have as much impact on user task performance as it does on trust. This may mean that we should pay more attention to how to effectively use feedback mechanisms to improve the usefulness and effectiveness of AI-assisted decision-making.
speech

Based on large language model generation, there may be a risk of errors.

speech
The Interpretability of Artificial Intelligence and the Impact of Outcome Feedback on Trust: A Comparative Study
User experience

To assess trust more accurately, the researchers used behavioral trust (WoA), a measure that takes into account the difference between the user's predictions and the AI's recommendations, and is independent of the model's accuracy. By comparing WoA under different conditions, researchers can analyze the relationship between trust and performance.

M

A3: The study found that the feedback of the results can improve the accuracy of the user's predictions (reducing the absolute error), thereby improving the performance of working with AI. However, interpretability does not have as much impact on user task performance as it does on trust. This may mean that we should pay more attention to how to effectively use feedback mechanisms to improve the usefulness and effectiveness of AI-assisted decision-making.

162 2 3 3