Computers, Artificial Intelligence Show Bias and Prejudice, Too

This version of Computers Artificial Intelligence Show Bias Prejudice Too N746206 - Health and Medicine | NBC News Clone was adapted by NBC News Clone to help readers digest key facts more efficiently.

Cyberspace isn’t any fairer than the workplace. Software shows bias and prejudice, a new study finds.
A new study shows artificial intelligence systems such as online translators are just as prejudiced as real people.
A new study shows artificial intelligence systems such as online translators are just as prejudiced as real people.AP

Cyberspace isn’t any fairer than the workplace, researchers reported Thursday. They found that artificial intelligence programs like search engines or software translators are every bit as prejudiced as real, live human beings.

It may not be the fault of the programmers, the team at Princeton University reports in the journal Science. It may just be that the body of published material is based on millennia of biased thinking and practices.

Predicted percentage of female names with occupations
Predicted percentage of female names with occupations[Credit: Aylin Caliskan] / Science

So someone named Leroy or Jamal is less likely to get a job offer, and more likely to be associated with negative words, than a similarly qualified Adam or Chip.

And AI thinks Megan is far more likely to be a nurse than Alan is, Aylin Caliskan and colleagues found.

They studied the language of various software programs, analyzing the associations that computer-based systems made for more than 2 million words.

They used some of the same tests that are standard psychological tests for people, including the Implicit Association Test, which measures how long it takes to link different words.

Related: Yes, Your Name Can Impact Your Career

For instance, people are far quicker to link the names of flowers with pleasant words such as “caress” than with unpleasant words such as “hatred.” The opposite goes for insects such as maggots, which people — and, it turns out, software — link much more easily with word such as “stink.”

The results are clear.

“Female names are associated more with family terms whereas male names are associated more with career terms,” Caliskan said.

“The language itself, just learning language, could account for the prejudices.”

“We show that cultural stereotypes propagate to artificial intelligence (AI) technologies in widespread use,” they wrote in their report, published in the journal Science.

Related: Black Applicants Turned Away from Trump Housing

It’s easy for software to learn these associations just by going through publications and literature, said Joanna Bryson, who helped oversee the research.

“Word embeddings seem to know that insects are icky and flowers are beautiful,” Bryson, who is also on the staff at Britain's University of Bath, wrote in her blog.

“You can tell that a cat is more like a dog and less like a refrigerator … because you say things like ‘I need to go home and feed my cat’ or ‘I need to go home and feed my dog’ but you never say ‘I need to go home and feed my refrigerator’,” she said.

Related: Girls Beat Out Boys on Tech, Engineering Tests

“The language itself, just learning language, could account for the prejudices.”

Caliskan found even language translation programs incorporated the prejudices. Caliskan demonstrated on a video, showing how translation software turned the gender-neutral Turkish term “o” into “he” for a doctor and “she” for a nurse.

“There are number of ways in which AI could imbibe human bias,” said Arvin Narayanan, assistant professor of computer science at Princeton.

“One is from its creators or its programmers. Another is from the data that’s fed into it that comes not so much from its programmers but from society as a whole,” Narayanan added.

“The biases and stereotypes in our society reflected in our language are complex and longstanding. Rather than trying to sanitize or eliminate them, we should treat biases as part of the language and establish an explicit way in machine learning of determining what we consider acceptable and unacceptable."

×
AdBlock Detected!
Please disable it to support our content.

Related Articles

Donald Trump Presidency Updates - Politics and Government | NBC News Clone | Inflation Rates 2025 Analysis - Business and Economy | NBC News Clone | Latest Vaccine Developments - Health and Medicine | NBC News Clone | Ukraine Russia Conflict Updates - World News | NBC News Clone | Openai Chatgpt News - Technology and Innovation | NBC News Clone | 2024 Paris Games Highlights - Sports and Recreation | NBC News Clone | Extreme Weather Events - Weather and Climate | NBC News Clone | Hollywood Updates - Entertainment and Celebrity | NBC News Clone | Government Transparency - Investigations and Analysis | NBC News Clone | Community Stories - Local News and Communities | NBC News Clone