Data Will Never Be Clean But You Can Make it Usefulby@darrendube
665 reads
665 reads

Data Will Never Be Clean But You Can Make it Useful

by Darren Dube5mJune 20th, 2022
Read on Terminal Reader
Read this story w/o Javascript
tldt arrow

Too Long; Didn't Read

Microsoft’s chatbot fell prey to data bias, a problem most AI and predictive systems are affected by. If the dataset used to train an AI excludes certain groups or has inherent societal biases, these systems may exacerbate societal biases. Think of the infamous Titanic dataset. Within a few minutes of Exploratory Data Analysis, you would realize that females were vastly more likely to survive than men. If a travel insurance firm trained its system using the Titanic dataset, it would charge men markedly higher premiums than women. With that, a system meant to assess risk based on the destination of travel and the number of people traveling instead assesses risk based on gender. This may sound hypothetical, but banks have been charging, on average, higher interest rates to people of color. This may have been caused, in part, by AI systems that consider black borrowers riskier.

Companies Mentioned

Mention Thumbnail
Mention Thumbnail

Coin Mentioned

Mention Thumbnail
featured image - Data Will Never Be Clean But You Can Make it Useful
Darren Dube HackerNoon profile picture
Darren Dube

Darren Dube

@darrendube

L O A D I N G
. . . comments & more!

About Author

Darren Dube HackerNoon profile picture
Darren Dube@darrendube

TOPICS

THIS ARTICLE WAS FEATURED IN...

Permanent on Arweave
Read on Terminal Reader
Read this story in a terminal
 Terminal
Read this story w/o Javascript
Read this story w/o Javascript
 Lite