By: Mardi Witzel & Niraj Bhargava (Center for International Governance Innovation)
Let’s say you work for a bank that uses automated systems to make decisions about loan applications, or hiring, or internal promotion. These systems include machine machine-learning tools designed according to a set of criteria, trained on historical data sets, then freed to do their mysterious work. Maybe you personally were passed over for a promotion.
Now, imagine that sometime later, you learn that the artificial intelligence (AI) making this decision was flawed. Perhaps the data used to train it was biased, or the model was poorly designed. Maybe the system “drifted,” as machine-learning models are known to do (drift happens when a model’s predictive power decays over time due to changes in the real world). It’s one thing to get turned down by a human you can challenge. But there’s much grey area with AI. It isn’t always possible to see how decisions are made.
This truth underlies the widespread call for trustworthy AI — that is to say, for transparency, fairness and accountability in the development and use of AI solutions. Despite the great promise of these tools, the risk of negative outcomes is not far-fetched. AI bias is documented and real. This is why it’s time for organizations to get serious about taking concrete steps toward effective AI governance.
Indeed, there are hard costs to AI done badly — including fines, litigation and settlement charges. Unsurprisingly, legislation has been proposed in the European Union and Canada that will impose massive penalties for breach of the rules around AI development and use. Companies have already experienced the hard costs of data breaches: for example, Capital One was fined US$80 million for its 2018 data breach and settled customer lawsuits for US$190 million. AI-related infractions will be similarly costly. And beyond the hard costs, soft ones — such as business distraction, loss of confidence and reputational damage — have even greater potential to damage organizations that do AI badly…
Featured News
DOJ and FTC Introduce Website for Reporting Anti-Competitive Healthcare Practices
Apr 18, 2024 by
CPI
US Congress Advances Legislation to Compel TikTok Sale
Apr 18, 2024 by
CPI
UK Financial Sector Advocates Enhanced Regulatory Accountability
Apr 18, 2024 by
CPI
Google and All 50 States Defend $700 Million Consumer Settlement
Apr 18, 2024 by
CPI
Colorado Enacts First Law to Protect Consumer Brainwave Data
Apr 18, 2024 by
CPI
Antitrust Mix by CPI
Antitrust Chronicle® – Economics of Criminal Antitrust
Apr 19, 2024 by
CPI
Navigating Economic Expert Work in Criminal Antitrust Litigation
Apr 19, 2024 by
CPI
The Increased Importance of Economics in Cartel Cases
Apr 19, 2024 by
CPI
A Law and Economics Analysis of the Antitrust Treatment of Physician Collective Price Agreements
Apr 19, 2024 by
CPI
Information Exchange In Criminal Antitrust Cases: How Economic Testimony Can Tip The Scales
Apr 19, 2024 by
CPI