What Is Legal AI, Really?
Key Takeaways
- •Legal AI is software that reads, analyzes, drafts, and reasons about legal documents using current-generation language models, packaged into workflows attorneys actually use.
- •Three categories matter: generic LLMs (ChatGPT, Claude), firm-wide assistants (Harvey, Legora), and practice specialists (Mage for M&A diligence, others for litigation discovery, etc.).
- •The categories are not interchangeable. A specialist M&A diligence tool will not run a litigation deposition; a firm-wide assistant will not own a deal team's workflow.
- •Picking the right category is the first decision, before picking the vendor inside it.
- •The bar for 'production-quality' is partner-grade output, not technically-correct output.
This is for the attorney who searched the question directly. We get asked it on roughly every fourth call: "What is legal AI, really?" The honest, direct answer is below, written for a partner deciding whether to deploy.
The category
Legal AI is software that reads, analyzes, drafts, and reasons about legal documents using current-generation language models, packaged into workflows attorneys actually use. Two parts of that sentence matter:
- Current-generation language models. The same Claude, GPT, Gemini etc. that power consumer AI products. The models are commodity-trending; almost every serious legal AI product is running similar engines underneath.
- Workflows attorneys actually use. This is where the products differ. The chassis around the engine is what determines whether the tool ships partner-grade output or generates plausible-looking hallucinations.
We wrote about the chassis-versus-engine framing in The F1 Engine Problem. The TL;DR: stop comparing legal AI tools on which model they use. Compare them on the chassis.
The three categories
Inside the category, three sub-categories matter. The differences are about scope and shape, not quality.
1. Generic LLMs (ChatGPT, Claude, Gemini, etc.)
What it is: the consumer product, used directly by attorneys without any legal-specific layer.
Good for: first-draft writing assistance on non-privileged content. Brainstorming. Reformulating dense legalese into client language. Drafting non-controversial provision language.
Not good for: production diligence. Generic LLMs hallucinate clause references, cannot resolve amendment chains, have no privilege posture by default, and have no workflow. The partner who relies on them to find issues will miss them.
When to use: low-stakes writing assistance. Stop short of production legal work.
2. Firm-wide assistants (Harvey, Legora, similar)
What it is: legal-specific AI assistants designed to serve a whole law firm across practice areas.
Good for: cross-practice coverage. Question-answering across the firm's document corpus. First-draft writing assistance with legal grounding. Research support. Memo drafting at the level of a competent first-year associate.
Not good for: owning a specific practice's deep workflow end-to-end. A firm-wide assistant is positioned as a copilot for everyday legal work, not as the platform that runs a deal team's diligence workstream.
When to use: a firm whose primary AI need is broad coverage across multiple practices.
3. Practice specialists (Mage for M&A, others for litigation discovery, contract management, etc.)
What it is: AI tools built around a specific practice area's full workflow.
Good for: end-to-end ownership of a practice's daily work. For M&A specifically, Mage covers data room ingestion, risk-driven document review, amendment chain resolution, disclosure schedule synthesis, memo drafting, redline review, and post-signing covenant tracking — the full deal-team workstream.
Not good for: cross-practice generality. A specialist tool is the wrong place to ask about your firm's litigation precedents.
When to use: a practice whose deal volume is high enough to justify a tool built around its specific workflow.
The categories are not substitutes for each other. Many large firms run a generalist plus a specialist for high-volume practices. Picking the right category is the first decision.
What legal AI can actually do, in 2026
Concrete capabilities, with sourced references where applicable:
- Ingest a 1,500-document data room in under an hour, with documents classified by type and prioritized by partner-defined relevance.
- Run a partner-defined risk pass against every contract: change-of-control, indemnity caps, anti-assignment, MFN, audit rights, exclusivity, termination, MAC, IP assignment, and dozens more. Output is a sortable findings list with citations to source.
- Resolve amendment chains. Reconstruct the current operative state of a contract that has been amended five or fifteen times. We covered the technical hard problem in Amendment Chain Resolution: The Hardest Problem in Legal AI.
- Draft disclosure schedules from the underlying agreements. Sell-side time spent on schedules drops from 80-120 hours to 20-30.
- Draft deal memos in firm voice. Partner-reviewable first drafts with citations to the source clauses.
- Compare counterparty redlines against firm preferred positions, surfacing material deviations and proposing language.
- Track post-signing covenants through the closing window. The interim covenants that get missed when teams use spreadsheets.
None of these was production-quality in 2023. All of them are production-quality now, in serious tools. The marketing claim and the reality have converged.
What legal AI cannot do
This is the more important list, because it sets the boundary on the work that is not changing:
- Make materiality calls. "Is this a real issue or boilerplate?" is judgment.
- Negotiate. Knowing what to push on, when, with which leverage, is judgment.
- Counsel the client. "Should you do this deal?" is judgment.
- Predict legal evolution. "How will the Delaware courts rule on this in five years?" is judgment.
- Hold privilege. A tool can preserve privilege by handling content correctly; the privilege itself is held by the attorney and client.
The work attorneys are paid for is mostly in the second list. The work that bottlenecks attorneys is mostly in the first list. AI shifts the balance.
Why generic AI is not enough
This is the question we are asked most often: "We could just use ChatGPT, right?"
You can. The result is the failure modes we listed at the top: hallucinated clause references, missed amendment chains, no workflow, no privilege posture. We have written about each one in detail. See LLM Hallucination in Contract Analysis, Amendment Chain Resolution, and The F1 Engine Problem.
The shorthand: generic AI is a great writing assistant. It is a poor diligence engine. The work an M&A team needs done is in the second category.
How to evaluate before buying
The honest evaluation method is to run a real deal through the candidate tools and compare against ground truth. Vendor demos are designed to win. Real deals are designed to ship. We laid out the framework in Evaluating Legal AI Tools: A Buyer's Guide for M&A Counsel.
The minimum bar:
- SOC 2 Type II report
- Written no-training-on-customer-data clause in the DPA
- Recall against ground truth ≥ associate baseline
- Precision ≥ 70% (so the partner doesn't have to re-read everything)
- Output rewrite percentage <30% (so the team adopts unprompted)
A tool that meets these on a real deal is a defensible buy. A tool that fails on more than one is not.
Where to start
The pillar guides walk through the category at increasing depth:
- Legal AI for M&A: The Practitioner's Guide — master hub, what the category is and what each tool category does.
- AI Due Diligence: An Operational Playbook — how the workflow runs on a real deal.
- Legal AI vs. Harvey vs. Generic AI — competitive landscape, honestly framed.
- Evaluating Legal AI Tools — the buyer's framework.
If you want to see Mage on a real deal, request a demo.
Frequently Asked Questions
Is legal AI just ChatGPT for lawyers?
No. ChatGPT (and other generic LLMs) are part of the picture but not the whole picture. Generic LLMs are good for first-draft writing assistance and not for production diligence work, where they hallucinate clause references, fail on multi-amendment contracts, and have no native concept of deal-team workflow. Production legal AI layers domain-specific architecture on top of the same models — that layer is where the real value lives.
What can legal AI actually do today, in 2026?
Read every contract in a 1,500-document data room and surface the partner-defined risk findings. Draft disclosure schedules from the underlying agreements. Resolve amendment chains (the current operative state of a contract amended fifteen times). Compare counterparty markups against firm preferred positions. Draft memos in firm voice. Track post-signing covenants. None of this was production-quality two years ago. All of it is production-quality now.
What can legal AI not do?
Make judgment calls on materiality. Negotiate with the other side. Counsel the client. Decide what to escalate. Predict how a court will rule on a novel question. The work AI does well is the high-volume mechanical reading. The work it does poorly is the high-value judgment work — which is what attorneys are paid for, and which is not going away.
What's the difference between legal AI and 'generic AI for lawyers'?
Domain depth, workflow shape, and trust posture. Generic AI is the engine. Legal AI is the chassis. We have written about this specifically in [The F1 Engine Problem](/blog/f1-engine-problem). The model is increasingly commodity; the chassis is the differentiator.
How should an attorney evaluate legal AI before buying?
Run a real deal through the candidate tools, in parallel with the manual workflow. Compare deliverables against ground truth. We laid out the framework in our [buyer's guide for evaluating legal AI tools](/guide/evaluating-legal-ai-tools).
Ready to transform your M&A due diligence?
See how Mage can help your legal team work faster and more accurately.
Request a DemoRelated Articles
Tech M&A Diligence: What Software Targets Actually Need
Tech M&A is intangibles-first. The valuation rests on IP integrity, customer concentration, and open-source compliance. A practitioner's view of what to check on a software target and where deals actually break.
What I Got Wrong About Legal AI
Three predictions I made about legal AI in 2023 that turned out to be wrong, what I learned from the misses, and what I think now.
Why We Built Mage After Kirkland
I spent years inside one of the most demanding M&A practices in the world. The bottleneck wasn't the work — it was the time spent doing the wrong parts of it. That's why Mage exists.