Harvey vs. Kira vs. Infrastructure: Three Approaches to Legal AI
Key Takeaways
- •Harvey is a legal research assistant: ask questions, get answers. It excels at legal research, memo drafting, and knowledge queries. It is not designed for structured contract extraction across hundreds of documents
- •Kira Systems pioneered machine learning for contract review, but its training-dependent approach requires labeled datasets and per-provision model training that limits flexibility on new document types and provision categories
- •Purpose-built infrastructure like Mage combines modern LLMs with structured extraction schemas, producing deal-ready output without the setup cost of platforms or the coverage limitations of research tools
- •The right choice depends on the problem. Research tools for research. Extraction tools for extraction. The mistake is using a research tool for an extraction problem, or vice versa
The legal AI market has three distinct paradigms, each designed for a different problem. Understanding which paradigm fits your workflow is more important than comparing feature lists. A research tool optimized for legal questions will not perform well as a contract extraction system, and an extraction system will not replace your legal research workflow. They are fundamentally different products solving fundamentally different problems.
The three paradigms are: research assistants (exemplified by Harvey), legacy extraction platforms (exemplified by Kira Systems), and purpose-built extraction infrastructure (the approach Mage takes). Here is how they differ and when each one is the right choice.
The Research Paradigm: Harvey
Harvey is built on the research assistant model: you ask a question, the system searches relevant sources, and it generates a natural language answer. This paradigm is powerful for legal research, knowledge queries, and document-specific questions.
Where Harvey excels:
- Legal research questions ("What are the Delaware Chancery Court standards for fiduciary duty in a squeeze-out merger?")
- Drafting assistance (generating first drafts of memos, briefs, and correspondence)
- Document-specific queries ("Does this agreement contain a non-compete?" for a single document)
- Knowledge synthesis across legal precedent and commentary
Where the research paradigm struggles for M&A:
- Systematic extraction across hundreds of documents. Asking 50 questions about each of 300 contracts requires 15,000 individual queries, each with potential for hallucination and missed context
- Structured output. Research tools produce narrative answers, not the structured data that populates disclosure schedules and diligence memos
- Comprehensive coverage. A research tool finds what you ask about. It does not surface provisions you did not think to query. The value of diligence is finding what you do not know to look for
- Cross-contract comparison. Comparing a specific provision type across 200 agreements requires structured data, not 200 individual narrative answers
Harvey is an excellent tool for the problems it was designed to solve. It is not designed for the structured extraction workflow that defines M&A contract review.
The Legacy Extraction Paradigm: Kira Systems
Kira Systems pioneered machine learning for legal contract review over a decade ago, well before the current wave of LLM-based tools. Its approach is fundamentally different from both research assistants and modern extraction tools.
Kira's approach: Kira uses supervised machine learning models trained on labeled contract data. For each provision type (indemnification caps, change of control clauses, termination rights), Kira has a trained model that recognizes that provision in new documents. The system includes hundreds of pre-built "smart fields" for common provisions and allows organizations to train custom models for their specific needs.
Where Kira excels:
- Organizations with large, recurring document sets where the upfront training investment pays off over time
- Provision types well-covered by Kira's pre-built models
- Teams with dedicated resources to manage and maintain the system
- Use cases where the document types and provision categories are stable and predictable
Where the legacy paradigm struggles:
- New or unusual document types that fall outside Kira's pre-built models require custom training with labeled data, which takes time and expertise
- Adding new provision categories requires new training data and model retraining, creating a lag between identifying a need and having the capability
- Flexibility for diverse data rooms where the document types and relevant provisions vary significantly across transactions
- Accuracy on provision types with limited training data, since supervised ML models are only as good as their training sets
Kira remains a capable tool for organizations that have invested in its ecosystem. But the training-dependent model creates friction for M&A deal teams that encounter different document types and provision requirements on every transaction.
The Infrastructure Paradigm: Purpose-Built Extraction
The third paradigm, which Mage represents, combines modern large language models with structured extraction schemas designed specifically for M&A workflows. Instead of answering questions (research) or applying trained models (legacy ML), purpose-built extraction processes documents through a type-aware pipeline that understands legal document structure.
How it works:
- Documents are classified by type (customer agreement, employment agreement, lease, etc.)
- Each document is segmented into individual clauses
- Clauses are extracted against type-specific schemas (what to extract from an indemnification clause in an asset purchase agreement differs from an employment agreement)
- Extracted data is organized into structured output with source citations
- Output feeds directly into deal deliverables
Where purpose-built extraction excels:
- Systematic review across entire data rooms (hundreds of contracts processed automatically)
- Structured output that maps to diligence deliverables without reformatting
- Flexibility across document types without per-provision training
- Cross-contract analysis (compare any provision type across the full data room)
- Source verification on every finding (preventing hallucination through mandatory citation)
Trade-offs:
- Not designed for general legal research (use a research tool for that)
- Optimized for transactional work, not litigation review or regulatory analysis
- Structured interfaces mean less freeform flexibility than a research assistant (by design)
Choosing the Right Tool
The decision framework is straightforward once you separate the problems.
Use a research tool (Harvey, CoCounsel) when:
- You need answers to legal research questions
- You are drafting memos, briefs, or correspondence
- You need to analyze a small number of specific documents with specific questions
- Your primary need is legal knowledge, not document extraction
Use a legacy extraction tool (Kira) when:
- You have a large, recurring document set with stable provision types
- You have already invested in training Kira's models for your use case
- Your team has dedicated resources to maintain the system
- Speed and flexibility on new document types are not primary concerns
Use purpose-built extraction infrastructure (Mage) when:
- You are doing M&A due diligence across hundreds of contracts
- You need structured output that feeds directly into deal deliverables
- Every data room contains different document types requiring flexible extraction
- You need 100% contract coverage without the cost of manual review
- Your team needs to be productive from the first upload without weeks of setup
The most common mistake in the market is using a research tool for an extraction problem. Harvey is excellent for asking questions. But M&A diligence is not a question-asking exercise. It is a structured extraction exercise across hundreds of documents, producing deliverables, not answers.
Similarly, using an extraction tool when you need a research assistant will disappoint. If your question is "What does Delaware law say about material adverse effect clauses?" you want Harvey, not Mage.
The legal AI market is large enough to support specialized tools for specialized problems. The evaluation framework starts with correctly identifying which problem you are solving.
Frequently Asked Questions
What is the difference between Harvey and Mage for M&A?
Harvey is a legal research assistant built for asking questions and getting answers across legal knowledge. It excels at legal research, memo drafting, and general legal queries. Mage is purpose-built for structured contract extraction in M&A due diligence. It automatically classifies documents, extracts deal-relevant provisions at the clause level, and produces structured output for diligence deliverables. Harvey answers questions about law. Mage extracts and organizes contract data for deals.
Is Kira Systems still relevant for contract review?
Kira Systems remains a capable tool for contract review, particularly for organizations that have invested in training its machine learning models on their specific document types and provision categories. However, Kira's training-dependent approach means adding new provision types requires labeled training data and model retraining, which can take weeks. Modern LLM-based extraction tools handle new document types and provisions without per-provision training, offering more flexibility for diverse data rooms.
Can Harvey AI do M&A due diligence?
Harvey can assist with aspects of M&A due diligence, particularly legal research questions, memo drafting, and analyzing individual documents. However, Harvey is designed as a research assistant, not as a structured extraction tool. For systematic extraction across hundreds of contracts, identifying every change of control clause, indemnification cap, and consent requirement across a full data room, purpose-built extraction tools are more appropriate than conversational research interfaces.
What is the best alternative to Kira Systems for contract review?
For M&A-focused contract review, Mage is a purpose-built alternative to Kira Systems that uses modern LLM-based extraction instead of training-dependent machine learning. Mage extracts deal-relevant provisions from any contract type without requiring labeled training data or per-provision model training. It produces structured output with source citations that feeds directly into diligence deliverables. For teams that need flexible extraction across diverse data rooms without per-provision setup, Mage addresses the core limitations of Kira's approach.
Ready to transform your M&A due diligence?
See how Mage can help your legal team work faster and more accurately.
Request a DemoRelated Articles
How to Evaluate Legal AI Tools for M&A: A 5-Axis Framework
Not all legal AI tools solve the same problem. Here is a framework for evaluating them across the five dimensions that actually matter for M&A deal teams: accuracy, speed, security, setup cost, and output quality.
Research vs. Extraction: Two Paradigms for Contract Review Software
Contract review software falls into two paradigms: research tools that answer questions about documents, and extraction tools that systematically pull structured data from every contract. The distinction determines what you can build on top of the output.
Why Most Legal AI Fails: Three Failure Modes That Kill Adoption
After speaking with hundreds of M&A attorneys, three patterns explain why most legal AI tools fail to achieve adoption: wrong abstraction level, no workflow integration, and output below the trust threshold.