Virtual data room review is one of the most painful stages of due diligence because the work is both high volume and highly interdependent. Teams are expected to read thousands of files, extract the relevant terms, identify inconsistencies, and convert that evidence into a reliable risk picture under time pressure. The problem is not only that there are too many documents. It is that the meaning of one document often depends on several others.
That is why AI is increasingly being applied to document review. McKinsey has highlighted due diligence and document-heavy M&A workflows as some of the clearest early use cases for generative AI because they involve large amounts of repetitive research, synthesis, and comparison work [McKinsey & Company, "Gen AI: Opportunities in M&A," May 2024]. Deloitte's 2025 M&A generative AI study points to the same operating reality from the market side: GenAI is no longer being confined to pilots, and document review remains one of the most practical deployment areas inside live M&A processes [Deloitte, "2025 GenAI in M&A Study," 2025].
Why Manual VDR Review Breaks Down
Traditional VDR review is slow for structural reasons, not because deal teams are careless. Even strong advisors run into the same operating constraints.
The review burden is uneven
Some files matter enormously and some barely matter at all, but no one knows which is which until the review has already started. Teams therefore spend too much time sorting, labeling, and triaging before they can even get to the substantive issues.
The work is repetitive
Analysts and lawyers often perform the same first-pass tasks over and over: identifying counterparties, pulling dates, locating change-of-control language, marking unusual payment terms, or determining whether a tax filing covers the right jurisdiction. Those tasks matter, but they are not where senior judgment adds the most value.
The findings live in silos
A legal reviewer can identify an assignment restriction. A financial reviewer can flag revenue concentration. A tax reviewer can notice an entity structure issue. The problem is that those findings are usually captured in separate workpapers, so the deal team has to rebuild the cross-workstream story later.
Important context gets lost in summaries
By the time findings reach senior review, they are often converted into bullet points, spreadsheet excerpts, or issue logs that no longer preserve enough source context. That creates a weak audit trail and makes it harder to challenge or refine the conclusion.
What AI Document Review Actually Does
The useful version of AI document review is not a chatbot pasted on top of a data room. It is a workflow layer that helps teams move from raw files to evidence-linked findings more efficiently.
1. It organizes the room faster
Before review begins, the system can classify documents by type, likely function, or workstream. Contracts, financial statements, board materials, policy documents, tax returns, entity charts, and regulatory correspondence should not sit in one undifferentiated file pile. Classification is not glamorous, but it is foundational because every later task depends on a sensible structure.
2. It extracts the fields reviewers repeatedly need
Once documents are grouped, AI can help pull recurring fields and terms into a structured view. In contracts, that might include parties, effective dates, renewal mechanics, termination triggers, assignment restrictions, exclusivity terms, consent requirements, or limitation-of-liability language. In financial schedules, it may involve pulling line items, period references, footnote topics, and signs of one-time adjustments. In tax files, it may involve identifying filing years, jurisdictions, carryforwards, audits, and entity relationships.
The point is not that the extraction is perfect. The point is that reviewers should not have to start from zero on every file.
3. It makes first-pass comparison easier
Document review becomes materially faster when the team can compare similar files in one place. That is especially useful for customer contracts, supplier agreements, leases, employment arrangements, and recurring policy documents where the goal is often to find exceptions to a normal pattern rather than re-read every common clause from scratch.
4. It preserves the evidence path
This is the real dividing line between a useful tool and a dangerous one. If an AI system says a contract has a consent requirement, the reviewer needs to see the exact clause and page supporting that statement. If a summary points to customer concentration risk, the team needs to understand which contracts and revenue records drove that conclusion. Without that traceability, the output is fast but not dependable.
5. It helps connect findings across workstreams
McKinsey's 2026 work on higher-performing M&A use of GenAI emphasizes that the advantage comes from embedding AI into real workflows rather than treating it as a detached productivity gadget [McKinsey & Company, "Gen AI in M&A: From theory to practice to high performance," January 2026]. In document review, that means linking the legal, financial, and tax consequences of the same underlying fact instead of capturing each one in isolation.
How the Workflow Looks in Practice
A disciplined AI-assisted review process usually follows a sequence like this.
Ingest and classify
The room is organized into document families. Reviewers can quickly see whether the upload is complete, where duplicates sit, which files are image-based, and which materials likely belong to legal, financial, tax, or governance review.
Extract and normalize
The system pulls recurring terms into a structured layer. Similar documents become easier to compare because dates, parties, thresholds, consent rights, and exceptions are no longer buried in separate PDFs.
Identify exceptions
Instead of reading every file with the same level of intensity, reviewers can focus on agreements or records that deviate from the norm. That is usually where the material diligence issues sit.
Ask cross-document questions
The team can query the room in business language, not only by filename. Questions like these become much easier to answer:
-
Which customer contracts contain a consent requirement on change of control?
-
Which suppliers appear across both procurement contracts and litigation files?
-
Which leases expire inside the first two years after close?
-
Which entities in the group appear in tax filings but not in the organizational chart?
Those are not merely convenience queries. They are review accelerators because they reduce the time spent stitching together facts manually.
Escalate to the risk register
See the workflow
Connect AI analysis to a live diligence process.
Sorai keeps extraction, source evidence, and issue review connected so AI output does not break when the partner questions start.
Once issues are surfaced, they can move into a tracked finding set with owner, workstream, evidence, and status. This is where a lot of manual processes fall apart. The issue list gets created, but the connection back to the underlying evidence weakens as the deal moves toward senior review. A good workflow keeps that connection alive.
Where AI Creates the Most Value
The highest-return use cases are usually the most repetitive and evidence-heavy.
Contract review at scale
When a room contains hundreds of customer, supplier, distributor, and employment agreements, the hardest part is rarely reading one contract. It is identifying the few contracts that contain materially different terms. AI helps narrow the set that needs intensive review.
Cross-workstream linkage
Private market teams care less about isolated facts than about how facts interact. McKinsey's work in private markets has emphasized the value of using GenAI to improve knowledge-intensive processes where teams need to synthesize scattered information quickly [McKinsey & Company, "Harnessing the power of gen AI in private markets," January 5, 2026]. In diligence, that translates into linking legal provisions, revenue exposure, tax structure, and governance issues into one operating record.
Evidence-backed summarization
Senior reviewers do not want to read every file, but they do need confidence that the summary is grounded in the source materials. AI can improve this layer when it turns bulky document sets into concise findings that still preserve document, page, and clause visibility.
Issue triage
The first pass through a room should identify what needs immediate attention, what can wait, and what is probably immaterial. AI is useful here because it can help reviewers prioritize the exceptions that are most likely to change price, structure, timing, or post-close risk.
What AI Should Not Do Alone
The practical limit is important.
AI should not make legal judgments without review. It can help find clauses and compare language, but it should not be treated as final legal analysis.
AI should not convert ambiguity into certainty. If the contract language is unusual, the scan quality is poor, or the underlying document set is incomplete, the right outcome may be a flagged question rather than a clean answer.
AI should not replace experienced escalation. Some issues matter because of negotiation context, industry norms, or buyer-specific integration risk. Those are judgment calls, not extraction tasks.
AI should not sit outside the review process. If the output lives in a separate summary layer that never feeds the actual diligence record, the team still has to re-enter the work later.
The Controls Serious Buyers Should Demand
If a buyer is evaluating AI document review software, the control model matters more than the demo speed.
Evidence traceability
Every meaningful finding should link back to the supporting source.
Reviewer override
Analysts and advisors must be able to correct extracted fields, revise summaries, and annotate exceptions.
Role-based workflow
Different reviewers need different access and different views of the same issue set. Legal, financial, tax, and senior reviewers should be able to work in one record without flattening the distinctions between their responsibilities.
Audit trail
The platform should preserve who surfaced an issue, who changed it, and what evidence supported the update. That matters for internal governance and for explaining conclusions later.
Sensible security posture
A serious platform should make its security controls clear, but buyers should validate those controls directly rather than assume them from marketing language. For M&A data, access controls, retention posture, encryption standards, hosting environment, and reviewability all matter. Security claims should be verified in the diligence process just like product claims.
How to Evaluate Whether the Tool Is Actually Useful
The right test is operational, not theatrical.
-
Can the tool help the team find exceptions across a large contract set?
-
Can reviewers see the exact page or clause behind a finding?
-
Can the financial, legal, and tax teams work from the same evidence base?
-
Can the issue list move into senior review without losing context?
-
Does the workflow reduce handoffs, or only create another layer of output to review?
If the answer to those questions is no, the system may still look impressive in a demo but it is unlikely to change real diligence throughput.
Where Sorai Fits
Sorai is built around the operating record that sits between raw evidence and senior review. In document review, that means keeping extracted facts, reviewer notes, exceptions, and escalations connected rather than scattering them across separate workpapers. The goal is not just to read files faster. It is to preserve context as the team moves from the data room into decision-making.
The Bottom Line
AI document review is valuable when it helps teams organize the room, surface exceptions, compare similar files, and keep every finding anchored to source evidence. It does not replace experienced legal, financial, or tax review. It makes that review more scalable and more connected, which is what serious diligence teams actually need.