Back to Blog

AI in Governance and Law: Where It Helps, Where It Hurts, and What We Need to Figure Out

23 min readBy Laksh JainAI & Society
#["AI governance"#"AI law"#"EU AI Act"#"legal technology"#"AI ethics"#"COMPAS"#"predictive policing"#"judiciary AI"#"algorithmic bias"]

AI in Governance and Law: Where It Helps, Where It Hurts, and What We Need to Figure Out

There's a question that legal systems around the world are quietly wrestling with right now: how much do we trust a machine when someone's freedom, rights, or future are on the line?

AI has worked its way into courtrooms, police departments, government offices, and regulatory bodies faster than most people realize. It's helping judges manage overwhelming caseloads, helping lawyers research thousands of documents in minutes, and helping governments detect fraud before it costs millions. At the same time, it's been caught flagging Black defendants as higher-risk than white ones. It's been used to make bail decisions without defendants ever knowing an algorithm was involved. And it's been generating confidently-worded legal citations that don't exist.

This isn't a story about AI being good or bad. It's more complicated than that. It's about where AI genuinely helps, where it creates serious problems, and what frameworks we're building — slowly, imperfectly — to handle it.


First, Let's Get the Basics Right: What Do We Mean by AI in Law and Governance?

When we talk about AI in this context, we're mostly talking about:

  • Machine learning models that find patterns in large amounts of data (crime records, case histories, financial transactions)
  • Natural language processing (NLP) tools that read, summarize, and generate legal text
  • Automated decision systems that score, rank, or predict outcomes — bail risk, fraud likelihood, permit approval
  • Generative AI tools like large language models that can draft contracts, research case law, and answer legal questions

These tools are being used at every level — from a solo lawyer using ChatGPT to draft a brief, to national governments deploying algorithmic systems to screen visa applications.


The Global Regulatory Landscape Right Now

Before we get into the specifics, you should know that the rules governing AI in governance are being written in real time — and different parts of the world are taking very different approaches.

The EU: The World's First Comprehensive AI Law

The European Union's AI Act is the most significant piece of AI legislation anywhere in the world so far. It entered into force in August 2024, and its key provisions have been rolling out since then. The core idea is a risk-based classification system:

  • Unacceptable risk: Banned outright. This includes things like social scoring systems by governments, real-time biometric surveillance in public spaces for law enforcement (with narrow exceptions), and AI that manipulates people's behavior subconsciously.
  • High risk: Heavily regulated. AI used in critical infrastructure, judicial decisions, law enforcement, border control, and hiring falls into this category. These systems must meet strict transparency, accuracy, and human oversight requirements before deployment.
  • Limited risk: Lighter requirements, mostly around transparency — you have to tell people they're talking to an AI.
  • Minimal risk: Mostly unregulated. Things like spam filters or AI in video games.

The fact that AI used in courts and law enforcement is explicitly classified as high risk is significant. It means these systems must be tested, documented, monitored, and kept under meaningful human control. The EU AI Office was set up to oversee implementation, and by August 2025, governance rules for general-purpose AI models also came into effect.

The United States: A Patchwork, Not a Framework

The US approach is much more fragmented. There's no single federal AI law. Instead, you have:

  • Executive orders that shift with each administration (the Trump administration in 2025 focused heavily on removing regulatory barriers and consolidating against state-level laws)
  • The NIST AI Risk Management Framework — a voluntary set of guidelines that many organizations use as a baseline but nobody is required to follow
  • State-level laws — by 2025, 38 US states had enacted around 100 AI-related measures. Colorado became the first state with comprehensive AI legislation. Illinois, New York, Texas, Virginia, and others have passed laws targeting specific high-risk use cases like hiring algorithms and facial recognition
  • Sector-specific enforcement by agencies like the FTC (consumer protection), EEOC (employment discrimination), and CFPB (financial services)

The result is what legal experts call a "regulatory patchwork" — different rules depending on what state you're in, what industry you're in, and what specific AI application you're using. The December 2025 White House executive order tried to address this by establishing a national policy framework, but as of now, comprehensive federal legislation remains absent.

The Council of Europe: The First Binding International Treaty

In September 2024, the Council of Europe opened for signature what is the world's first legally binding international treaty specifically on AI — the Framework Convention on Artificial Intelligence and Human Rights, Democracy, and the Rule of Law. Countries that ratify it must embed fundamental principles into how AI systems are designed, developed, and deployed, with a focus on protecting human rights, democratic processes, and the rule of law. It applies to both public authorities and private entities acting on behalf of governments.

India, Singapore, China

  • India has guidelines and principles around responsible AI but no binding AI-specific legislation yet. Work is ongoing.
  • Singapore has a Model AI Governance Framework for Generative AI, focused on accountability, transparency, and innovation — not binding law but widely followed in practice.
  • China has enacted specific regulations on generative AI and recommendation algorithms, with an emphasis on content control and national security.

The general trend globally is toward some form of risk-based regulation, with higher-stakes applications in law and governance receiving the most scrutiny.


Where AI Actually Helps in the Legal System

Let's be fair here. AI genuinely solves real problems in law and governance, and dismissing that would be misleading.

1. Managing Caseload Collapse

Courts worldwide are drowning. In Argentina, an AI assistant called Prometea helped legal professionals process nearly 490 cases per month, up from 130 before its introduction — almost a 300% increase in productivity. In Egypt, automated transcription introduced in 2024 is improving court efficiency and making proceedings more accessible. These aren't hypothetical gains — they're real improvements in systems that were failing the people they were supposed to serve.

Backlogs of years are the norm in many court systems. People wait months or years for civil cases, family matters, and even some criminal hearings to be resolved. AI tools that handle document classification, scheduling, transcription, and routine case management genuinely free up judges and clerks for the work that actually requires human judgment.

2. Legal Research at a Scale Humans Can't Match

Before AI, a lawyer researching precedents for a complex case might spend weeks going through case law. Now, tools can surface relevant decisions across hundreds of jurisdictions in minutes. This matters especially for access to justice — if AI research tools make legal work faster, that cost can (in theory) be passed on to clients, making legal services more affordable.

The ABA's Task Force on Law and Artificial Intelligence documented more than 100 AI use cases in legal aid settings alone. Thomson Reuters and Everlaw have programs making legal AI tools available to public-interest organizations at reduced or no cost. The promise is real: AI could help democratize access to legal representation, especially for people who currently can't afford it.

3. Fraud Detection and Financial Crime

AI systems are very good at spotting patterns in financial data that humans would never catch at scale. Tax authorities, financial regulators, and law enforcement agencies use AI to detect money laundering, insurance fraud, and tax evasion by identifying unusual transaction patterns across millions of accounts simultaneously. This is a case where AI's ability to process enormous amounts of data is genuinely aligned with what we want from it: catch bad actors, protect the public.

4. Contract Review and Due Diligence

Corporate law involves reviewing enormous stacks of contracts, filings, and disclosures. AI tools can review thousands of documents in a fraction of the time it would take a team of associates, flagging relevant clauses, inconsistencies, and risks. This isn't replacing lawyers — it's letting them focus on the actual judgment calls rather than the mechanical reading.

5. Accessibility and Language Barriers

India's Supreme Court developed SUVAS — an AI-powered translation tool that converts legal documents between English and regional Indian languages. Courts in many countries are using AI for real-time transcription and translation, making proceedings more accessible to people who don't speak the dominant language of the legal system. For a country as linguistically diverse as India, this is genuinely important.


Where AI Causes Serious Problems

Now for the harder conversation. The same tools that help with efficiency create serious problems when they're used to make decisions about people's lives without adequate transparency, oversight, or accountability.

1. Predictive Policing and the Bias Feedback Loop

Predictive policing tools use historical crime data to forecast where crimes are likely to occur and who is likely to commit them. The idea sounds logical — use data to allocate police resources efficiently. The problem is that the data itself is the problem.

Historical crime data doesn't reflect actual crime rates. It reflects arrest rates. And arrest rates are shaped by decades of discriminatory policing practices — over-policing of minority communities, racial disparities in stops and searches, and discriminatory enforcement of drug laws. When you feed that data into an algorithm, the algorithm learns to predict more crime in the communities that have historically been over-policed, which means more police get sent there, which means more arrests happen there, which makes the algorithm even more confident in its prediction. This is a feedback loop, and it bakes historical injustice into automated systems.

According to US Department of Justice figures, a Black person is more than twice as likely to be arrested as a white person, and five times as likely to be stopped without just cause. When these numbers feed into a predictive model, the model doesn't distinguish between actual crime rates and arrest-driven biases — it just sees the numbers.

Several cities have already abandoned expensive predictive policing tools after finding they didn't reduce crime and did perpetuate bias. The NAACP has called for banning the use of historical crime data in these algorithms entirely, arguing that using biased data to drive policing decisions is fundamentally incompatible with equal protection under the law.

2. COMPAS and Algorithmic Sentencing

COMPAS (Correctional Offender Management Profiling for Alternative Sanctions) is a risk assessment tool used in many US jurisdictions to give judges a "risk score" — a number between 1 and 10 indicating how likely a defendant is to reoffend. This score has been used to influence bail, sentencing, and parole decisions.

In 2016, a ProPublica investigation revealed that COMPAS was significantly more likely to label Black defendants as high-risk than white defendants, even when they didn't reoffend. Conversely, it more often labeled white defendants as low-risk even when they did reoffend. A Dartmouth study later found that the algorithm performed no better than random people with no criminal justice experience recruited through a crowdsourcing site.

Here's what makes this especially troubling: the algorithm is proprietary. Defendants who received harsher sentences based on their COMPAS score often had no way to see how the score was calculated or challenge its assumptions. The "black box" problem — where an algorithm makes a consequential decision without being able to explain why — is particularly acute in a legal system built on the principle that people have a right to understand and contest the evidence against them.

This is a case where the problem isn't just technical bias — it's a fundamental due process issue.

3. Facial Recognition Failures

Facial recognition technology has been used by law enforcement agencies to identify suspects from surveillance footage. The technology has well-documented accuracy problems, particularly for darker-skinned faces. Multiple studies have shown that commercial facial recognition systems have significantly higher error rates for Black women compared to white men.

In January 2025, The Washington Post reported that 15 police departments across 12 states were using facial recognition to make arrests without direct evidence linking suspects to crimes — relying solely on algorithmic matches. People have been wrongfully arrested based on facial recognition misidentifications, with some cases not discovered until after significant harm had already been done.

The EU AI Act explicitly bans real-time biometric surveillance in public spaces for law enforcement purposes (with narrow exceptions for serious crimes and terrorism). This is one area where the EU has drawn a bright line that the US has not.

4. AI-Generated Legal Citations That Don't Exist

In 2023, a lawyer filed a court brief citing six cases that ChatGPT had fabricated. The cases had realistic-sounding names, docket numbers, and case summaries — none of them were real. The lawyer hadn't verified the citations. The case became a defining example of what happens when AI "hallucination" (generating plausible-sounding but false information) collides with a system where accuracy is literally the job.

This problem hasn't gone away. The US federal judiciary's interim guidance issued in 2025 specifically directed judiciary users to review and independently verify all AI-generated work product. Multiple state bar associations have issued ethics opinions warning lawyers that using AI without adequate verification could violate their professional obligations of competence and candor.

The broader issue is that generative AI is confident even when it's wrong. Legal writing that sounds authoritative but contains fabricated facts, misquoted statutes, or nonexistent precedents is arguably worse than no research at all.

5. Algorithmic Decision-Making in Government Benefits

AI systems are increasingly used to make or recommend decisions about government benefits — welfare eligibility, disability assessments, housing applications, child protective services. When these systems go wrong, the consequences fall on the most vulnerable people.

Several governments have had high-profile algorithmic failures in this space. The Dutch government's "SyRI" system, which used AI to flag potential welfare fraud, was struck down by a court in 2020 for violating human rights. The system disproportionately targeted low-income and immigrant communities, and individuals had no way to understand why they were flagged or challenge the decision.

The principle at stake is that when a government uses an algorithm to make decisions that affect someone's rights or livelihood, that person has a right to understand how the decision was made and to challenge it.


The Core Ethical Questions

All of these cases orbit the same fundamental tensions. It's worth naming them clearly.

Transparency vs. Trade Secrets

Many of the most consequential AI systems used in law enforcement and government are built by private companies that treat their algorithms as proprietary intellectual property. The companies that built COMPAS refused to disclose their methodology, citing trade secret protections. This creates a situation where a person's sentence can be influenced by a calculation they have no right to see or challenge.

Some jurisdictions are pushing back. New York City passed a law requiring audits of automated employment decision tools. The EU AI Act requires high-risk AI systems to maintain technical documentation and allow for meaningful oversight. But the tension between private IP rights and the public's right to understand decisions that affect them remains largely unresolved.

Efficiency vs. Fairness

AI can make systems faster. Faster isn't always better. When speed comes at the cost of individual consideration, nuance, or the ability to understand mitigating circumstances, efficiency becomes a problem.

A judge taking time to understand a person's full circumstances before sentencing is not inefficiency — it's the point. The concern isn't that AI will make courts faster. The concern is that it will make courts faster in a way that removes the human judgment that is supposed to be at the core of the legal process.

Accountability When Things Go Wrong

When a human judge makes a wrongful decision, there is a system of appeals, professional accountability, and sometimes legal liability. When an algorithm makes a wrongful decision, it's much harder to figure out who is responsible. The company that built it? The agency that deployed it? The official who relied on it? The absence of clear accountability frameworks for algorithmic decision-making is a serious gap in current law.

Bias Amplification

AI systems learn from historical data. Historical data reflects historical injustices. This means AI systems trained on that data will, unless actively corrected, reproduce and potentially amplify those injustices at scale. The challenge is that the corrective — adjusting data or outputs to account for historical bias — is technically and politically contested. There isn't consensus on how to do it, and some people argue that any adjustment is itself a form of discrimination.


Principles for Ethical AI in Law and Governance

UNESCO published guidelines for AI in courts and tribunals in December 2025 — the first global ethical framework of its kind. They center on fifteen principles, and the most important ones are worth understanding clearly.

Human oversight and decision-making must remain with humans. AI can inform, assist, summarize, and flag. The decision — especially when it affects someone's freedom, rights, or livelihood — must ultimately rest with a human who can be held accountable for it. This isn't just an ethical principle; it's increasingly a legal requirement under frameworks like the EU AI Act.

Transparency is non-negotiable in high-stakes decisions. If an AI system is influencing a consequential decision about a person, that person has a right to know it, understand it at a meaningful level, and challenge it. Black-box decision-making in legal and governance contexts is incompatible with the rule of law.

Auditability and continuous oversight. AI systems deployed in justice and governance must be regularly audited for accuracy, bias, and drift. What works fairly today can become biased over time as data changes. Ongoing monitoring isn't optional — it's part of responsible deployment.

Meaningful explainability. Not just "the system gave a score of 7" but "here is why, here is the data it used, here are the factors it weighted." Explainability is what makes accountability possible.

AI as assistive, not substitutive. In the judiciary particularly, the principle is clear: AI supports judges, it doesn't replace them. The legitimacy of a judicial decision comes partly from the fact that a human being with accountability made it, considering the full human context of the situation.


Where AI Should Not Be Trusted: The Hard Lines

Beyond the principles, there are specific applications where current AI technology is simply not ready for autonomous or near-autonomous use in legal and governance contexts.

Final sentencing decisions. Even as a strong recommendation, algorithmic sentencing scores carry significant risk of codifying bias and removing the individualized assessment that is a cornerstone of just punishment. Judges can use data as one input. They should not be bound by it.

Bail and pretrial detention. The stakes are immediate — someone goes home or goes to jail right now. Algorithmic risk scores for bail have documented racial bias problems and should not be the determining factor in these decisions.

Immigration and asylum decisions without meaningful review. AI screening of asylum claims can be efficient, but asylum decisions involve complex assessments of credibility, risk, and country conditions. Automated rejection of claims without substantive human review is inconsistent with international refugee law and basic fairness.

Facial recognition for arrests. The error rates, particularly for darker-skinned individuals, combined with the severity of the consequence (arrest), make sole reliance on facial recognition for law enforcement decisions unacceptable without corroborating evidence.

Automated welfare and benefits termination. Removing someone's benefits based on an algorithmic flag, without meaningful human review and the opportunity to challenge the decision, violates basic principles of procedural fairness.

Judicial analytics that profile judges. AI tools that analyze judges' past decisions and predict their future rulings raise serious concerns about judicial independence. If lawyers can use AI to predict how a specific judge will rule, they may optimize their arguments to exploit those patterns rather than argue the law as written. This undermines the integrity of the judicial process itself.


Case Study 1: Estonia's AI-Assisted Judiciary

Estonia has been building one of the world's most digital-forward governance systems for years. In 2019, it began piloting an AI "judge" for small claims cases — disputes under €7,000 — where the AI could recommend a decision based on submitted documentation, subject to human appeal.

The idea was to address a specific, bounded problem: simple contract disputes that were clogging the court system. The AI handled the routine analysis; a human judge reviewed any decision a party appealed.

This is a model worth studying because it gets the boundary mostly right. The AI handles a narrow, well-defined category of cases. Every decision can be appealed to a human. The system is transparent about what it is. The human court remains the ultimate authority. It's AI as a filter and first-pass analysis tool, not as a replacement for the judicial system.

The key lesson: define the problem narrowly, keep humans in the loop, make the process transparent, and build in meaningful appeal mechanisms.


Case Study 2: The Dutch SyRI System and Why It Failed

The Netherlands built a system called SyRI (System Risk Indication) to identify potential welfare fraud by cross-referencing data from tax records, employment files, housing registrations, and more. The government used this to flag individuals for investigation in specific geographic areas.

In February 2020, a Dutch court struck it down for violating the European Convention on Human Rights. The court found that the system lacked sufficient transparency — individuals couldn't understand how they were flagged, what data was used, or how to challenge the determination. It was also found to disproportionately target low-income and immigrant neighborhoods.

The government had genuine reasons for wanting to detect fraud. But the tool they built violated the principle that consequential government decisions must be transparent and challengeable by the people affected. The court's ruling was a clear statement: efficiency in fraud detection does not override fundamental rights.

This case became a landmark in AI governance law — the first major court ruling specifically striking down a government AI system on human rights grounds. It influenced the thinking behind the EU AI Act and set a precedent for the kind of scrutiny these systems should face.


What the Legal Profession Is Actually Doing About This

The legal profession's institutions are responding, even if slowly.

The American Bar Association issued Formal Opinion 512 in July 2024, giving guidance on lawyers' ethical use of generative AI. The core obligations: competence (you have to actually understand the tool and verify its outputs), confidentiality (don't input client data into tools with unclear data policies), and candor to the tribunal (don't submit AI-generated content you haven't verified for accuracy).

Since then, dozens of state bars and courts have issued their own opinions and rules. Many courts now require lawyers to disclose when they've used AI in preparing filings, and some require a certification that AI-generated content has been verified.

The US federal judiciary established an AI Task Force in 2025, which developed interim guidance directing judiciary staff to verify AI-generated work, not delegate core judicial functions to AI, and consider disclosure requirements. The federal judiciary's September 2025 Strategic Plan explicitly called for an AI governance framework for the courts.

The ABA's Task Force concluded in late 2025 that AI has "moved from experiment to infrastructure" for legal practice. The debate has shifted from whether to use AI to how to govern it.


What Good AI Governance in Law Looks Like

Drawing all of this together, here's what responsible AI deployment in legal and governance contexts looks like in practice:

1. Define the task narrowly. AI performs best when it has a specific, well-bounded problem to solve — summarizing a document, translating text, flagging a pattern in financial data. The more a task requires human judgment, context, and ethical weighing, the less appropriate full automation becomes.

2. Audit the data before you deploy the system. If the training data reflects historical bias, the system will reproduce that bias. Understanding what data a system was trained on, and what biases that data might contain, is step one in responsible deployment.

3. Require meaningful explainability for high-stakes decisions. "The score was 7" is not an explanation. People affected by algorithmic decisions have a right to understand the reasoning, not just the output.

4. Build human oversight into the process, not just onto the end of it. Human review that happens after the damage is done isn't meaningful oversight. The human needs to be in a position to actually change the outcome.

5. Create clear accountability mechanisms. Someone needs to be responsible when the system goes wrong — not the algorithm, but a person or organization with legal accountability. This is currently one of the biggest gaps in existing frameworks.

6. Monitor continuously, not just at launch. AI systems drift. Data changes. What was accurate and fair at deployment can become biased or inaccurate over time. Regular auditing is not optional in high-stakes applications.

7. Give affected people a meaningful way to challenge decisions. Procedural fairness — the right to understand and contest a decision — is a foundation of legitimate governance. Algorithmic decisions don't get an exemption from this.


The Bigger Picture

We are genuinely in a transition period. AI is already inside legal systems, whether or not the law has caught up with it. The US federal judiciary's own 2025 Annual Report acknowledged that AI use is embedded in how judicial work gets done — not just by judges, but by clerks, assistants, and others throughout the system. The ABA found that chambers staff used AI for legal research at nearly 40% — often higher than judges themselves, who may not even be fully aware of how much AI their staff is using.

The question is no longer whether AI will be used in law and governance. It is. The question is whether we build the accountability structures fast enough to prevent the worst outcomes — biased systems baking inequality into legal decisions, algorithms making consequential choices about people who have no way to understand or challenge them, and efficiency being used to justify the erosion of procedural fairness.

The EU AI Act, the Council of Europe treaty, UNESCO's judicial guidelines, and the evolving guidance from bar associations and courts are all pieces of an answer. They're incomplete, imperfect, and still developing. But they represent a recognition that AI in law isn't just a technology deployment question — it's a question about what kind of legal system we want.

The technology isn't going away. The task is making sure it serves justice rather than undermining it.