ImageNet contains inaccurate, derogatory and racially offensive information
ImageNet contains inaccurate, derogatory and racially offensive information
Occurred: September 2019
Report incident 🔥 | Improve page 💁 | Access database 🔢
The ImageNet dataset was found to contain inaccurate racist, misogynistic and other discriminatory and derogatory slurs, resulting in controversy about its safety and accusations of privacy and copyright abuse.
ImageNet Roulette, a website that encouraged users to upload selfies and then analyse what it saw by running their photos through a neural network trained on ImageNet found that many captions produced by the code were harmless, though some turned out to be inaccurate, or contained racist, misogynistic and other discriminatory and derogatory slurs.
Created by Kate Crawford, co-founder of the AI Now Institute, artist Trevor Paglen and software developer Leif Ryge, ImageNet Roulette was a 'provocation designed to help us see into the ways that humans are classified in machine learning systems.'
By automatically scraping images from Google, Bing and photo-sharing platform Flickr to build its training dataset without consent, ImageNet's developers were also accused of ignoring user privacy, leading lawyers and rights activists to call for stronger privacy and copyright laws.
The ensuing fracas led the developers of ImageNet to scrub 'unsafe' and 'sensitive' labels from the database, and to remove links to related photographs - an update seen to have minimal impact on the classification and transfer learning accuracy of the dataset, though some commentators argued it would damage ImageNet's relevance by styming its reproducibility.
Computer vision - recognition
The classical problem in computer vision, image processing, and machine vision is that of determining whether or not the image data contains some specific object, feature, or activity.
Wikipedia: Image recognition 🔗
➕ March 2021. The ImageNet team announced it had blurred 243,198 photographs in its database using Amazon's Rekogniton image and video analytics service.
Operator: Kate Crawford; Trevor Paglen; Leif Ryge
Developer: Jia Deng; Wei Dong, Richard Socher; Li-Jia Li; Kai Li; Fei-Fei Li
Country: USA
Sector: Research/academia
Purpose: Identify objects
Technology: Dataset; Computer vision; Object detection; Object recognition
Issue: Accuracy/reliability; Bias/discrimination - race, ethnicity, gender, religion, national identity, location; Copyright; Privacy; Safety
Transparency: Governance; Privacy
Prabhu V.U., Birhane A. (2020). Large image datasets: A pyrrhic win for computer vision?
Dulhanty C., Wong A. (2019). Auditing ImageNet: Towards a Model-driven Framework for Annotating Demographic Attributes of Large-Scale Image Datasets
Paglen T., Crawford K.: ImageNet Roulette
https://www.wired.com/story/viral-app-labels-you-isnt-what-you-think/
https://www.nbcnews.com/mach/tech/playing-roulette-race-gender-data-your-face-ncna1056146
https://news.artnet.com/art-world/imagenet-roulette-trevor-paglen-kate-crawford-1658305
https://www.frieze.com/article/how-ai-selfie-app-imagenet-roulette-took-internet-storm
https://www.nytimes.com/2019/09/20/arts/design/imagenet-trevor-paglen-ai-facial-recognition.html
https://www.theguardian.com/technology/2019/sep/17/imagenet-roulette-asian-racist-slur-selfie
https://www.theregister.com/2019/10/23/ai_dataset_imagenet_consent/
https://www.theguardian.com/technology/2019/sep/17/imagenet-roulette-asian-racist-slur-selfie
https://qz.com/1034972/the-data-that-changed-the-direction-of-ai-research-and-possibly-the-world/
https://thenextweb.com/news/ai-fails-to-recognize-these-nature-images-98-of-the-time
https://www.wired.com/story/ai-biased-how-scientists-trying-fix/
https://www.wired.com/story/researchers-blur-faces-launched-thousand-algorithms/
https://towardsdatascience.com/the-fall-of-imagenet-5792061e5b8a
https://www.technologyreview.com/2021/04/01/1021619/ai-data-errors-warp-machine-learning-progress/
Page info
Type: Incident
Published: June 2024