Design approach may help fix bias in artificial intelligence
Credit: North Carolina State University

Bias in artificial intelligence (AI) and machine learning programs is well established. Researchers from North Carolina State University and Pennsylvania State University are now proposing that software developers incorporate the concept of "feminist design thinking" into their development process as a way of improving equity—particularly in the development of software used in the hiring process.

"There seem to be countless stories of ways that bias in AI is manifesting itself, and there are many thought pieces out there on what contributes to this bias," says Fay Payton, a professor of information systems/technology and University Faculty Scholar at NC State. "Our goal here was to put forward guidelines that can be used to develop workable solutions to algorithm bias against women, African American and Latinx professions in the IT workforce.

"Too many existing hiring algorithms incorporate de facto identity markers that exclude qualified candidates because of their gender, race, ethnicity, age and so on," says Payton, who is co-lead author of a paper on the work. "We are simply looking for equity—that job candidates be able to participate in the on an equal footing."

Payton and her collaborators argue that an approach called feminist design thinking could serve as a valuable framework for developing software that reduces algorithmic bias in a meaningful way. In this context, the application of feminist design thinking would mean incorporating the idea of equity into the design of the itself.

"Compounding the effects of algorithmic is the historical underrepresentation of women, Black and Latinx software engineers to provide novel insights regarding equitable design approaches based on their lived experiences," says Lynette Yarger, co-lead author of the paper and an associate professor of information sciences and technology at Penn State.

"Essentially, this approach would mean developing algorithms that value inclusion and equity across gender, race and ethnicity," Payton says. "The practical application of this is the development and implementation of a process for creating algorithms in which designers are considering an audience that includes women, that includes Black people, that includes Latinx people. Essentially, developers of all backgrounds would be called on to actively consider—and value—people who are different from themselves.

"To be clear, this is not just about doing something because it is morally correct. But we know that women, African Americans and Latinx people are under-represented in IT fields. And there is ample evidence that a diverse, inclusive workforce improves a company's bottom line," Payton says. "If you can do the right thing and improve your profit margin, why wouldn't you?"



More information: Lynette Yarger et al, Algorithmic equity in the hiring of underrepresented IT job candidates, Online Information Review (2019). DOI: 10.1108/OIR-10-2018-0334

Citation: Design approach may help fix bias in artificial intelligence (2020, February 6) retrieved 6 February 2020 from https://techxplore.com/news/2020-02-approach-bias-artificial-intelligence.html

This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no part may be reproduced without the written permission. The content is provided for information purposes only.