DC AG launches algorithmic bias bill
Law would levy $10,000 per infection for algorithmic bias
DC Attorney General Karl A Racine has proposed legislation that would prevent algorithmic discrimination. The Stop Discrimination by Algorithms Act (SDAA) would prevent companies from making automated decisions that discriminate against marginalized groups in the District.
The Center of Privacy and Technology and Communications and Technology Law Clinic (both part of Georgetown University Law Center) helped draft the bill, in conjunction with civil rights nonprofit Color of Change.
Are you ready for the hybrid workplace?
Invest in headsets and webcams to drive worker productivity
The Act forbids discrimination by algorithms that make decisions about credit, education, housing, employment, and public services. Companies must also inform individuals about how they use algorithms to make those decisions.
Organizations must document how they build their algorithms and how they make decisions. They must file annual audit reports with the Attorney General in which they check their algorithms for discrimination.
The proposed legislation comes with teeth, as the it allows the Attorney General to penalize companies that violate the rules by up to $10,000 per violation. It also permits private lawsuits.
The text of the bill warned that algorithmic decision-making systems that don't account for bias can harm marginalized groups. "Despite their prevalence and the potential problems they pose, algorithms are poorly understood by most individuals, in part because of the many entities involved and the lack of accountability among those entities," it said.
The move prompted support from organizations and individuals including Timnit Gebru, the AI researcher who complained that Google silenced her concerns about its AI algorithms.
"At this point it should be clear that multinational corporations will not self regulate," said the scientist, who this month launched the Distributed Artificial Intelligence Research Institute. "To the contrary, they push out people with the slightest criticism of their proliferation of harmful systems. Without laws requiring companies to assess the potential for discriminatory impact of their algorithms, what they do instead is eject people like me who attempt to do that internally, even though this was literally in my job description."
The proposed legislation mirrors efforts at the federal level. In 2019, Democrats introduced an Algorithmic Accountability Act that would have pressed the Federal Trade Commission to create rules for assessing bias in automated systems. In May this year, Democrats tried again with the Algorithmic Justice and Online Platform Transparency Act of 2021.
Accelerating AI modernisation with data infrastructure
Generate business value from your AI initiativesFree Download
Recommendations for managing AI risks
Integrate your external AI tool findings into your broader security programsFree Download
Modernise your legacy databases in the cloud
An introduction to cloud databasesFree Download
Powering through to innovation
IT agility drive digital transformationFree Download