algorithm
Credit: Pixabay/CC0 Public Domain

In his article, "Trust is Not a Virtue: Why We Should Not Trust," Matthew L. Bolton, an associate professor of systems and information engineering at the University of Virginia, critically challenges the emphasis on fostering trust in human-machine interactions. He focuses particularly on systems using AI, machine learning and automation.

The article is published in the journal Ergonomics in Design: The Quarterly of Human Factors Applications.

Bolton argues that while trust is often seen as essential for the adoption of new technologies, it is a problematic and imprecise concept. Trust is difficult to define, highly contextual, and conflated with related concepts like confidence and perceived risk, making it neither selective nor diagnostic as a measure of human behavior.

Bolton contends that the focus on building trust in may actually undermine sound engineering. Instead of pursuing trust as a goal, engineers should focus on objective measures of system reliability, transparency, and usability—elements that directly impact human experience and performance.

Trust, Bolton asserts, is not inherently humanistic and can be manipulated to disenfranchise users, reducing autonomy rather than enhancing it. This manipulation often serves the interests of large organizations seeking to bypass the need for reliability by encouraging blind trust in their technologies.

"There is a at the heart of trust research," Bolton says. "We include humans in systems because they bring experience, expertise, instincts, and creativity that improves performance and makes systems resilient… we rely on them to decide when, why, and how to trust a system. If engineers manipulate people into behaving the way they (or others) want, we lose the benefit of having human operators."

Ultimately, Bolton calls for a shift away from trust-centric research in favor of more concrete and ethical approaches to system design.

He emphasizes that engineers should prioritize developing technologies that empower users with transparent, reliable, and human-centered designs, rather than relying on as a justification for adoption. This approach, he argues, would lead to safer, more ethical, and effective human-machine interactions.

More information: Matthew L. Bolton, Trust is Not a Virtue: Why We Should Not Trust Trust, Ergonomics in Design: The Quarterly of Human Factors Applications (2022). DOI: 10.1177/10648046221130171

Citation: Trust is not the answer: Rethinking human-machine interaction for ethical engineering (2024, October 17) retrieved 17 October 2024 from https://techxplore.com/news/2024-10-rethinking-human-machine-interaction-ethical.html

This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no part may be reproduced without the written permission. The content is provided for information purposes only.