• | 8:00 am

Why Meredith Whittaker believes AI is ‘replicating patterns of inequality’

At Fast Company’s Impact Council meeting, the Signal Foundation president warns that AI tech is only as fair as the data it’s being fed.

Why Meredith Whittaker believes AI is ‘replicating patterns of inequality’
[Source photo: Meredith Whittaker and Fast Company senior editor Amy Farley [Photo: Alyssa Ringler]]

Consumers need to be increasingly aware of how historical biases and marginalization are impacting AI systems and their responses, said Signal Foundation President Meredith Whittaker, speaking on a panel during Fast Company’s annual Impact Council meeting this week.

“[Artificial intelligence] is determining our access to resources, our access to opportunity, our place in the world. And, of course, it’s determining this based on past data,” Whittaker said. “So it is, and I believe inextricably will be, replicating patterns of historical inequality and marginalization. And in many powerful ways, it will be obscuring accountability for those decisions.”

[Photo: Alyssa Ringler]

For many months now, the topic of AI has dominated the tech industry while executives in other industries are looking at potential applications. Big tech firms, like Google and Microsoft, are building out their own AI systems or partnering with industry players to gain dominance in the landscape.

“It’s naturalizing often racist and misogynist determinations about people’s place in the world behind the veil of computational sophistication, in a way that makes it harder and harder to push back against those histories and correct them in the present,” Whittaker added.

AI systems have been used for years for tasks like analyzing data sets and recommending solutions. And its rapid adoption, with what some view as the potential to disrupt every industry, could threaten to continue centuries of mistreatment. After all, the tech is only as good as the data that’s being fed into it.

Oftentimes, people will look at whatever AI spouts out as undeniable fact.

“This is not a system that has any understanding of veracity or truth, right? Like facts are bolted on post hoc as sort of an afterthought,” said Whittaker, whose career has included positions in academia as well as more than a decade at Google. “This is an aggregate of some of the worst and maybe best parts of text on the internet spit back out as plausible-sounding content . . . that has nothing to do with veracity or truth, and it is wild to me that we’ve sort of let that slide. That we have companies that are integrating this into search—the whole thing is a red flag to me.”

  Be in the Know. Subscribe to our Newsletters.

ABOUT THE AUTHOR

Jessica Bursztynsky is a staff writer for Fast Company, covering the gig economy and other consumer internet companies. She previously covered tech and breaking news for CNBC. More

More Top Stories:

FROM OUR PARTNERS

Brands That Matter
Brands That Matter