A growing industry wants to scrutinize the algorithms that govern our livesâbut it needs teeth
By Alfred Ng
For more than a decade, journalists and researchers have been writing about the dangers of relying on algorithms to make weighty decisions: who gets locked up, who gets a job, who gets a loanâeven who has priority for COVID-19 vaccines.
Rather than remove bias, one algorithm after another has codified and perpetuated it, as companies have simultaneously continued to more or less shield their algorithms from public scrutiny.vv
The big question ever since: How do we solve this problem? Lawmakers and researchers have advocated for algorithmic audits, which would dissect and stress-test algorithms to see how they work and whether theyâre performing their stated goals or producing biased outcomes. And there is a growing field of private auditing firms that purport to do just that. Increasingly, companies are turning to these firms to review their algorithms, particularly when theyâve faced criticism for biased outcomes, but itâs not clear whether such audits are actually making algorithms less biasedâor if theyâre simply good PR.
Algorithmic auditing got a lot of press recently when HireVue, a popular hiring software company used by companies like Walmart and Goldman Sachs, faced criticism that the algorithms it used to assess candidates through video interviews were biased.
HireVue called in an auditing firm to help and in January touted the results of the audit in a press release.
The audit found the softwareâs predictions â âwork as advertised with regard to fairness and bias issues,â â HireVue said in a press release, quoting the auditing firm it hired, OâNeil Risk Consulting & Algorithmic Auditing (ORCAA).
But despite making changes to its process, including eliminating video from its interviews, HireVue was widely accused of using the auditâwhich looked narrowly at a hiring test for early career candidates, not HireVueâs candidate evaluation process as a wholeâas a PR stunt.
Articles in Fast Company, VentureBeat and MIT Technology Review called out the company for mischaracterizing the audit.
HireVue said it was transparent with the audit by making the report publicly available and added that the press release specified that the audit was only for a specific scenario.
âWhile HireVue was open to any type of audit, including one that involved looking at our process in general, ORCAA asked to focus on a single use case to enable concrete discussions about the system,â Lindsey Zuloaga, HireVueâs chief data scientist, said in an email. âWe worked with ORCAA to choose a representative use case with substantial overlap with the assessments most HireVue candidates go through.â
But algorithmic auditors were also displeased about HireVueâs public statements on the audit.
âIn repurposing [ORCAAâs] very thoughtful analysis into marketing collateral, theyâre undermining the legitimacy of the whole field,â Liz OâSullivan, co-founder of Arthur, an AI explainability and bias monitoring startup, said.
And that is the problem with algorithmic auditing as a tool for eliminating bias: Companies might use them to make real improvements, but they might not. And there are no industry standards or regulations that hold the auditors or the companies that use them to account.
What Is Algorithmic AuditingâHow Does It Work?
Good questionâitâs a pretty undefined field. Generally, audits proceed a few different ways: by looking at an algorithmâs code and the data from its results, or by viewing an algorithmâs potential effects through interviews and workshops with employees.
Audits with access to an algorithmâs code allow reviewers to assess whether the algorithmâs training data is biased and create hypothetical scenarios to test effects on different populations.
There are only about 10 to 20 reputable firms offering algorithmic reviews, Rumman Chowdhury, Twitterâs director of machine learning ethics and founder of the algorithmic auditing company Parity, said. Companies may also have their own internal auditing teams that look at algorithms before theyâre released to the public.
In 2016, an Obama administration report on algorithmic systems and civil rights encouraged development of an algorithmic auditing industry. Hiring an auditor still isnât common practice, though, since companies have no obligation to do so, and according to multiple auditors, companies donât want the scrutiny or potential legal issues that that scrutiny may raise, especially for products they market.
âLawyers tell me, âIf we hire you and find out thereâs a problem that we canât fix, then we have lost plausible deniability and we donât want to be the next cigarette company,â â ORCAAâs founder, Cathy OâNeil, said. âThatâs the most common reason I donât get a job.â
For those that do hire auditors, there are no standards for what an âauditâ should entail. Even a proposed New York City law that requires annual audits of hiring algorithms doesnât spell out how the audits should be conducted. A seal of approval from one auditor could mean much more scrutiny than that from another.
And because audit reports are also almost always bound by nondisclosure agreements, the companies canât compare each otherâs work.
âThe big problem is, weâre going to find as this field gets more lucrative, we really need standards for what an audit is,â said Chowdhury. âThere are plenty of people out there who are willing to call something an audit, make a nice looking website and call it a day and rake in cash with no standards.â
And tech companies arenât always forthcoming, even with the auditors they hire, some auditors say.
âWe get this situation where trade secrets are a good enough reason to allow these algorithms to operate obscurely and in the dark, and we canât have that,â Arthurâs OâSullivan said.
The big problem is, weâre going to find as this field gets more lucrative, we really need standards for what an audit is.
Rumman Chowdhury, Twitter director of machine learning ethics
Auditors have been in scenarios where they donât have access to the softwareâs code and so risk violating computer access laws, Inioluwa Deborah Raji, an auditor and a research collaborator at the Algorithmic Justice League, said. Chowdhury said she has declined audits when companies demanded she allow them to review them before public release.
For HireVueâs audit, ORCAA interviewed stakeholders including HireVue employees, customers, job candidates, and algorithmic fairness experts, and identified concerns that the company needed to address, Zuloaga said.
ORCAAâs evaluation didnât look at the technical details of HireVueâs algorithmsâlike what data the algorithm was trained on, or its codeâthough Zuloaga said the company did not limit auditorsâ access in any way.
âORCAA asked for details on these analyses but their approach was focused on addressing how stakeholders are affected by the algorithm,â Zuloaga said.
OâNeil said she could not comment on the HireVue audit.
Many audits are done before products are released, but thatâs not to say they wonât run into problems, because algorithms donât exist in a vacuum. Take, for example, when Microsoft built a chatbot that quickly turned racist once it was exposed to Twitter users.
âOnce youâve put it into the real world, a million things can go wrong, even with the best intentions,â OâSullivan said. âThe framework we would love to get adopted is thereâs no such thing as good enough. There are always ways to make things fairer.â
So some prerelease audits will also provide continuous monitoring, though itâs not common. The practice is gaining momentum among banks and health care companies, OâSullivan said.
OâSullivanâs monitoring company installs a dashboard that looks for anomalies in algorithms as they are being used in real-time. For instance, it would alert companies months after launch if their algorithms were rejecting more women applicants for loans.
And finally, thereâs also a growing body of adversarial audits, largely conducted by researchers and some journalists, which scrutinize algorithms without a companyâs consent. Take, for example, Raji and Joy Buolamwini, founder of the Algorithmic Justice League, whose work on Amazonâs Rekognition tool highlighted how the software had racial and gender bias, without the companyâs involvement.
Do Companies Fix Their Algorithms After an Audit?
Thereâs no guarantee companies will address the issues raised in an audit.
âYou can have a quality audit and still not get accountability from the company,â said Raji. âIt requires a lot of energy to bridge the gap between getting the audit results and then translating that into accountability.â
Public pressure can at times push companies to address the algorithmic bias in the technologyâor audits that werenât performed at the behest of the tech firm and covered by a nondisclosure agreement.
Raji said the Gender Shades study, which found gender and racial bias in commercial facial recognition tools, named companies like IBM and Microsoft to spark a public conversation around it.
Itâs a bit sad that we rely so much on public outcry.
Inioluwa Deborah Raji, Algorithmic Justice League
But it can be hard to create buzz around algorithmic accountability, she said.
While bias in facial recognition is relatableâpeople can see photos and the error rates and understand the consequences of racial and gender bias in the technologyâit may be harder to relate to something like bias in interest-rate algorithms.
âItâs a bit sad that we rely so much on public outcry,â Raji said. âIf the public doesnât understand it, there is no fine, thereâs no legal repercussions. And it makes it very frustrating.â
So What Can Be Done to Improve Algorithmic Auditing?
In 2019, a group of Democratic lawmakers introduced the federal Algorithmic Accountability Act, which would have required companies to audit their algorithms and address any bias issues the audits revealed before theyâre put into use.
AI For the Peopleâs founder Mutale Nkonde was part of a team of technologists that helped draft the bill and said it would have created government mandates for companies to both audit and follow through on those audits.
âMuch like drug testing, there would have to be some type of agency like the Food and Drug Administration that looked at algorithms,â she said. âIf we saw disparate impact, then that algorithm wouldnât be released to market.â
The bill never made it to a vote.
Sen. Ron Wyden, a Democrat from Oregon, said he plans to reintroduce the bill with Sen. Cory Booker (D-NJ) and Rep. Yvette Clarke (D-NY), with updates to the 2019 version. Itâs unclear if the bill would set standards for audits, but it would require that companies act on their results.
âI agree that researchers, industry and the government need to work toward establishing recognized benchmarks for auditing AI, to ensure audits are as impactful as possible,â Wyden said in a statement. âHowever, the stakes are too high to wait for full academic consensus before Congress begins to take action to protect against bias tainting automated systems. Itâs my view we need to work on both tracks.â
Originally published as "Can Auditing Eliminate Bias from Algorithms?" with the Attribution-NonCommercial-NoDerivatives 4.0 International (CC BY-NC-ND 4.0) license