Mind Matters News and Analysis on Natural and Artificial Intelligence

CategoryMachine Learning

jehyun-sung-477894-unsplash

GIGO alert: AI can be racist and sexist, researchers complain

Can the bias problem be addressed? Yes, but usually after someone gets upset about a specific instance.

From James Zou and Londa Ziebinger at Nature: When Google Translate converts news articles written in Spanish into English, phrases referring to women often become ‘he said’ or ‘he wrote’. Software designed to warn people using Nikon cameras when the person they are photographing seems to be blinking tends to interpret Asians as always blinking. Word embedding, a popular algorithm used to process and analyse large amounts of natural-language data, characterizes European American names as pleasant and African American ones as unpleasant. Now where, we wonder, would a mathematical formula have learned that? Maybe it was listening to the wrong instructions back when it was just a tiny bit? Seriously, machine learning, we are told, depends on  absorbing datasets of Read More ›