Algorithmic Bias (23)

Algorithms selectively favoring certain groups or demographics.

View options:

Find narratives by ethical themes or by technologies.

FILTERreset filters
Themes
  • Privacy
  • Accountability
  • Transparency and Explainability
  • Human Control of Technology
  • Professional Responsibility
  • Promotion of Human Values
  • Fairness and Non-discrimination
Show more themes
Technologies
  • AI
  • Big Data
  • Bioinformatics
  • Blockchain
  • Immersive Technology
Show more technologies
Additional Filters:
  • Media Type
  • Availability
  • Year
    • 1916 - 1966
    • 1968 - 2018
    • 2019 - 2069
  • Duration
  • 7 min
  • Venture Beat
  • 2021
image description
Center for Applied Data Ethics suggests treating AI like a bureaucracy

As machine learning algorithms become more deeply embedded in all levels of society, including governments, it is critical for developers and users alike to consider how these algorithms may shift or concentrate power, specifically as it relates to biased data. Historical and anthropological lenses are helpful in dissecting AI in terms of how they model the world, and what perspectives might be missing from their construction and operation.

  • Venture Beat
  • 2021
  • 7 min
  • MIT Tech Review
  • 2020
image description
Why 2020 was a pivotal, contradictory year for facial recognition

This article examines several case studies from the year of 2020 to discuss the widespread usage, and potential for limitation, of facial recognition technology. The author argues that its potential for training and identification using social media platforms in conjunction with its use by law enforcement is dangerous for minority groups and protestors alike.

  • MIT Tech Review
  • 2020
  • 7 min
  • The New Republic
  • 2020
image description
Who Gets a Say in Our Dystopian Tech Future?

The narrative of Dr. Timnit Gebru’s termination from Google is inextricably bound with Google’s irresponsible practices with training data for its machine learning algorithms. Using large data sets to train Natural Language Processing algorithms is ultimately a harmful practice because for all the harms to the environment and biases against certain languages it causes, machines still cannot fully comprehend human language.

  • The New Republic
  • 2020
  • 4 min
  • VentureBeat
  • 2020
image description
Researchers Find that Even Fair Hiring Algorithms Can Be Biased

A study on the engine of TaskRabbit, an app which uses an algorithm to recommend the best workers for a specific task, demonstrates that even algorithms which attempt to account for fairness and parity in representation can fail to provide what they promise depending on different contexts.

  • VentureBeat
  • 2020
  • 4 min
  • OneZero
  • 2020
image description
Dr. Timnit Gebru, Joy Buolamwini, Deborah Raji — an Enduring Sisterhood of Face Queens

A group of “Face Queens” (Dr. Timnit Gebru, Joy Buolamwini, and Deborah Raji) have joined forces to do important racial justice and equity work in the field of computer vision, struggling against racism in the industry to whistleblow against biased machine learning and computer vision technologies still deployed by companies like Amazon.

  • OneZero
  • 2020
  • 10 min
  • Survival of the Best Fit
  • 2018
image description
Survival of the Best Fit

Explores hiring bias of AI by playing a game in which you are the hiring manager.

  • Survival of the Best Fit
  • 2018
Load more