Brindwell & Partners · Legal Intelligence Division

Module Technical
Design Document

Architecture, pipeline design, model specification, and performance validation across six platform modules and four AI intelligence engines powering legal research, drafting, case management, billing intelligence, firm analytics, and institutional knowledge.

Modules
6 Platform Systems + 4 AI Engines
Corpus
14.2M Judicial Decisions Since 1923
Citation Accuracy
97.4% — Independently Validated
Classification
Confidential — Internal Use Only
Contents
Six Modules, Four Intelligence Engines
01
Legal Research Engine
RAG-based corpus search with real-time citation verification and treatment analysis
02
Intelligent Drafting
Transformer-assisted composition with authority insertion, Bluebook compliance, and argument gap detection
03
Case & Matter Management
Rule-based deadline computation, conflict detection, and predictive matter risk scoring
04
Time & Billing Intelligence
Passive time capture, NLP narrative generation, and collections forecasting
05
Firm Analytics & Intelligence
Real-time financial, operational, and competitive benchmarking for firm leadership
06
Knowledge Management
Work product indexing, precedent retrieval, and institutional memory AI
A1
Hallucination-Proof Citations
Zero-tolerance verification layer ensuring every cited authority exists and has been correctly stated
A2
Predictive Case Analytics
Judge-level behavioral modeling across 8,200+ federal and state judges
A3
Adverse Authority Detection
Pre-filing corpus scan identifying controlling adverse authority missed in manual review
A4
Argument Strength Scoring
Data-informed confidence scoring tested against 10,000+ actual case outcomes
Executive Summary
System Architecture Overview
Arbiter is an AI-native legal intelligence platform that replaces the fragmented technology stack most law firms operate today — where research lives in one silo, drafting in another, case management in a third, and billing in a fourth, with none sharing context. The platform's six interconnected modules share a unified data layer: research informs drafting, drafting informs billing, and analytics informs everything. The core intelligence layer is powered by a domain-specific legal language model fine-tuned on the complete corpus of 14.2 million published federal and state judicial decisions since 1923, every statute and regulation currently in force, and 47 federal agencies' administrative rulings.
The system architecture employs retrieval-augmented generation (RAG) to ground every AI-generated output in verified primary authority — addressing the hallucination problem that has led to professional sanctions for attorneys who submitted fabricated citations generated by general-purpose AI tools. Arbiter's citation verification layer achieves 97.4% accuracy, independently validated by Georgetown Law's Center for Legal Technology. Every module integrates via REST API with existing firm systems (iManage, NetDocuments, Elite/3E, Aderant) via pre-built connectors, enabling deployment without rip-and-replace migration. The platform is deployed on SOC 2 Type II certified infrastructure with AES-256 encryption and zero data commingling — firm data never trains shared models.
97.4%
Citation Accuracy (Georgetown Validated)
14.2M
Judicial Decisions Indexed
0.8s
Average Query Response Time
42%
Reduction in Research Time
Module 01
Legal Research Engine
14.2 million decisions — searched by meaning, not by keyword

The Legal Research Engine is the foundation of the Arbiter platform. It indexes the complete corpus of published federal and state judicial decisions since 1923, every statute and regulation currently in force, and administrative rulings from 47 federal agencies — totaling 14.2 million documents. The search architecture employs a hybrid retrieval system that combines traditional BM25 sparse retrieval with dense vector search using a legal-domain transformer model fine-tuned on 8.4 million annotated legal text pairs, enabling natural language queries that understand legal concepts, not just keywords.

Every citation returned by the system is verified in real time against its subsequent treatment history — reversed, distinguished, questioned, affirmed, or overruled — using a proprietary citation graph that maps 340 million inter-citation relationships. The system achieves 97.4% citation accuracy as independently validated by Georgetown Law's Center for Legal Technology, with 0.8-second average query response time across the full corpus. Results are ranked by a composite score that weights relevance, jurisdictional authority, recency, treatment strength, and alignment with the attorney's matter context.

14.2M
Judicial decisions fully indexed and searchable
97.4%
Citation accuracy — Georgetown Law validated
0.8s
Average query response time across full corpus
42%
Reduction in research time vs. legacy platforms
340M
Inter-citation relationships mapped in citation graph
Search Pipeline
STAGE 01
Query Understanding
NLP pipeline parses natural language query, identifies legal concepts, jurisdictions, date constraints, and procedural posture using legal-BERT tokenization.
Legal-BERTNER
STAGE 02
Hybrid Retrieval
Parallel BM25 sparse search (lexical matching) and dense vector search (semantic understanding) across partitioned indices. Results merged via reciprocal rank fusion.
BM25Dense RetrievalRRF
STAGE 03
Citation Verification
Every candidate result is verified against the citation graph. Treatment status (affirmed, reversed, distinguished, questioned) is computed in real time from 340M relationships.
Citation GraphTreatment
STAGE 04
Composite Ranking
Results ranked by weighted composite: relevance (0.35), jurisdictional authority (0.25), treatment strength (0.20), recency (0.10), matter alignment (0.10).
Learning to RankLambdaMART
STAGE 05
Contextual Presentation
Results delivered with highlighted holdings, treatment badges, authority chains, and one-click insertion into active draft via Module 02 integration.
HighlightingDraft Link
Retrieval Architecture

The hybrid retrieval system addresses the fundamental limitation of keyword search in legal contexts: legal concepts are expressed differently across jurisdictions, eras, and writing styles. A query about "duty to mitigate damages" must retrieve results discussing "mitigation of loss," "avoidable consequences doctrine," and "failure to minimize harm" — semantic equivalents that keyword search misses entirely. The dense retrieval component uses a bi-encoder transformer fine-tuned on 8.4 million annotated legal text pairs (query-passage pairs manually labeled by legal professionals for relevance) to generate 768-dimensional embeddings that capture legal semantic meaning.

The sparse BM25 component handles exact match requirements — case names, statutory section numbers, and specific legal terms of art where lexical precision matters. Reciprocal rank fusion merges results from both retrieval paths, weighting dense retrieval higher for conceptual queries and sparse retrieval higher for citation-specific queries, determined by a lightweight query-type classifier.

Citation Graph Architecture

The citation graph is a directed acyclic graph of 340 million inter-citation relationships extracted from the 14.2 million indexed decisions. Each edge encodes the treatment relationship: positive treatment (followed, affirmed, applied), negative treatment (reversed, overruled, abrogated), distinguishing treatment (distinguished, limited), and questioning treatment (questioned, criticized). Treatment classification uses a fine-tuned legal-BERT model that processes the citing context — the paragraph surrounding each citation — to determine how the citing court characterized its use of the cited authority.

The graph enables real-time treatment verification: before any case is presented to the attorney, the system traverses all subsequent citing decisions to determine whether the case has been reversed, overruled, or significantly limited. This eliminates the single most dangerous failure mode in legal research — citing authority that is no longer good law.

Module 02
Intelligent Drafting
AI that understands legal writing at a structural level — not just syntax

The Intelligent Drafting module transforms the legal writing process from blank-page composition to AI-augmented authorship. As attorneys compose, the system suggests relevant authority from their active research collection, auto-formats citations to Bluebook or local court rules, flags unsupported assertions, and identifies potential weaknesses in argument structure before opposing counsel does. The system learns each firm's brief style, preferred citation depth, and argument architecture — adapting suggestions to practice-specific conventions rather than imposing generic templates.

The drafting AI uses a fine-tuned legal language model that processes the document context (issue statement, argument section, procedural posture) and generates contextually appropriate authority suggestions, transitional language, and counter-argument anticipation. Every suggestion maintains a living connection to the underlying research: when case law changes, affected documents are flagged automatically. The module integrates directly with Module 01 (one-click authority insertion from research results) and Module 04 (drafting activity generates automatic time entries).

68%
Reduction in first-draft production time
94%
Citation accuracy in AI-suggested authorities
3.2×
Faster brief turnaround for litigation teams
Drafting Pipeline
STAGE 01
Document Context Analysis
Identifies document type (motion, brief, memo), procedural posture, jurisdiction, and argument structure using structural parsing and NLP classification.
DocTypeStructure
STAGE 02
Authority Suggestion Engine
Matches current drafting context against research collection and full corpus to surface relevant supporting authority ranked by jurisdictional weight and topical alignment.
Semantic MatchRanking
STAGE 03
Citation Formatting
Automatically formats all citations to Bluebook, local court rules, or firm-specific style guides. Handles parallel citations, short-form references, and supra/infra cross-references.
BluebookLocal Rules
STAGE 04
Argument Gap Analysis
Identifies unsupported assertions, missing counter-argument responses, and logical gaps in argument chain using legal reasoning graph analysis.
Gap DetectionLogic
STAGE 05
Style Adaptation & Export
Adapts prose style to firm templates and individual attorney preferences. Generates court-ready documents with proper formatting, headers, and signature blocks.
TemplatesExport
RAG-Grounded Composition

The drafting AI uses retrieval-augmented generation to ensure that every suggestion is grounded in actual legal authority rather than hallucinated content. When the attorney writes an assertion — for example, "Courts have consistently held that employers bear the burden of proving the affirmative defense of undue hardship" — the system immediately retrieves the strongest supporting authority from the corpus, verifies its treatment status, and suggests a properly formatted citation insertion. If no supporting authority exists for the assertion, the system flags the statement as unsupported rather than generating a fabricated citation.

This RAG-grounded approach directly addresses the hallucination problem that has resulted in professional sanctions for attorneys who submitted AI-generated briefs containing fabricated case citations. The system's retrieval step constrains generation to the closed universe of verified legal authority — if the decision does not exist, or if the holding has been misstated, the system refuses to produce the citation.

Firm Style Learning

The style adaptation system uses few-shot learning on each firm's historical brief corpus to capture writing conventions: preferred citation depth (string citing vs. select authority), argument organization patterns (IRAC vs. CRAC vs. Garner's deep-issue methodology), transitional phrasing preferences, tone calibration (formal/academic vs. advocate/persuasive), and formatting conventions. The model adapts its suggestions to match the senior partner's style on each matter, not a generic template.

Each draft maintains a bi-directional link to its underlying research: when a cited case receives negative treatment (reversed, overruled), the system immediately surfaces an alert on every draft referencing that authority — enabling proactive updates before filing rather than reactive corrections after opposing counsel points out the problem.

Module 03
Case & Matter Management
Every document, deadline, communication, and task — in one living workspace

Every matter in Arbiter is a living workspace that connects documents, research, communications, deadlines, tasks, and time entries into a single coherent view. The platform automatically calculates deadlines from court rules (federal, all 50 states, and local court-specific rules), surfaces conflicts across the firm using real-time conflict graph analysis, and uses predictive analytics to flag matters at risk of budget overrun or missed deadlines before they become problems.

The deadline computation engine maintains a rules database of 12,400+ procedural rules across all federal courts, all 50 state court systems, and 340 local court-specific rules. When a triggering event occurs (complaint filed, motion served, discovery request received), the system automatically computes all downstream deadlines, accounts for service method adjustments, holiday exclusions, and weekend carryovers, and generates calendar entries with configurable advance warning periods. The conflict detection system uses entity resolution across client names, subsidiary structures, and individual names to identify potential conflicts within 0.3 seconds of new matter intake.

12,400+
Procedural rules maintained in deadline computation database
0.3s
Conflict detection response time at matter intake
99.97%
Deadline computation accuracy across all jurisdictions
Deadline Computation Engine

The deadline engine processes triggering events through a multi-layered rule evaluation system. Layer 1 applies the applicable procedural rules (FRCP, state equivalents). Layer 2 overlays local court-specific modifications (many districts have standing orders that modify default timelines). Layer 3 applies service method adjustments (3 additional days for service by mail under FRCP 6(d), different calculations for electronic service). Layer 4 handles calendar adjustments (weekends, federal holidays, state-specific court holidays, inclement weather closures from historical patterns). The system flags conflicts where different rules produce competing deadlines and surfaces both the most conservative and most liberal interpretations with rule citations.

Conflict Detection Architecture

The conflict detection system uses a graph database (Neo4j) that maintains entity resolution across all client names, corporate subsidiary structures, individual names, related entities, and matter-specific adverse parties. When a new matter is initiated, the system performs a comprehensive graph traversal that identifies direct conflicts (same entity on opposing sides), related-entity conflicts (subsidiary of existing client adverse to the new matter), individual conflicts (attorney formerly at opposing firm), and positional conflicts (taking a legal position in one matter that contradicts a position taken in another). The entity resolution layer handles variations in entity names, misspellings, abbreviations, d/b/a designations, and corporate name changes using fuzzy matching with legal-entity-specific heuristics.

Module 04
Time & Billing Intelligence
Capturing the $142K per attorney per year that disappears through manual timekeeping

Arbiter captures time passively — tracking which matters an attorney works on, for how long, and in what context — then generates billing narratives that comply with UTBMS codes and client-specific billing guidelines. Partners review and approve rather than reconstruct from memory at the end of each day. The passive capture system monitors document activity (which files were open, for how long, in which application), communication patterns (emails tagged to matters, calls logged to clients), and research activity (queries run in Module 01 tagged to matter context) to generate time entries with an accuracy of 96% versus the 67% average reported for manual timekeeping.

The billing intelligence layer extends beyond time capture into revenue optimization: predictive analytics forecast monthly and quarterly collections, flag at-risk receivables based on client payment history patterns, and identify realization rate optimization opportunities across the firm's billing portfolio. The system natively handles alternative fee arrangements (AFAs), blended rates, success fees, and volume discounts.

$142K
Average recovered revenue per attorney per year
96%
Time capture accuracy vs. 67% manual average
23%
Improvement in overall realization rate
Passive Capture Architecture

The passive time capture system operates through a lightweight endpoint agent that monitors application activity without capturing content — recording which documents, applications, and communications were active, and for how long, without reading email content or document text. The matter attribution model uses contextual signals (file path patterns, email thread subject lines, document metadata) to associate activity with specific matters. When the system has high confidence in matter attribution (>0.92 probability), it generates an automatic time entry. When confidence is moderate (0.70–0.92), it presents the entry as a suggestion for attorney review. Below 0.70, the activity is logged as unattributed for manual assignment.

Narrative Generation

The billing narrative generator uses a task-specific language model fine-tuned on 6.8 million approved time entries from Am Law 200 firms. The model generates UTBMS-compliant descriptions that accurately reflect the work performed (e.g., "Research and analysis of applicable statute of limitations under [Jurisdiction] law for breach of fiduciary duty claims; review of relevant authorities") without generic filler language. Client-specific billing guidelines — which often prohibit block billing, require task-level detail, or limit certain billing categories — are encoded as constraint rules that the narrative model respects during generation. The system also detects and flags potential billing compliance violations (block billing, vague descriptions, billing for overhead tasks) before submission.

Module 05
Firm Analytics & Intelligence
Transforming operational data into strategic intelligence for firm leadership

Arbiter transforms raw operational data into strategic intelligence for managing partners and firm leadership. Real-time dashboards track revenue per lawyer, profitability by practice group, utilization rates, client concentration risk, and lateral hiring ROI — with drill-down capability to the individual attorney and matter level. The analytics engine integrates data from all five preceding modules to generate cross-functional insights that no single-module analysis can produce.

Predictive models forecast quarterly revenue with 94% accuracy, identify client attrition risk 90 days in advance, and quantify the financial impact of lateral partner acquisitions versus organic growth strategies. Competitive benchmarking against anonymized peer firm data enables evidence-based compensation, pricing, and growth strategy decisions.

94%
Revenue forecasting accuracy (quarterly)
90-day
Advance warning for client attrition risk
Real-time
Financial dashboards with attorney-level drill-down
Cross-Module Intelligence

The analytics engine's primary advantage is cross-module data integration. By unifying research activity (Module 01), drafting throughput (Module 02), matter progression (Module 03), billing data (Module 04), and knowledge utilization (Module 06) into a single analytical layer, the system identifies patterns invisible to siloed tools: which practice areas generate the highest research-to-revenue ratios, which matters are consuming disproportionate resources relative to expected recovery, and which attorney-client relationships show early indicators of engagement decline.

Predictive Financial Modeling

Revenue forecasting uses a time-series ensemble (ARIMA + gradient boosting) trained on each firm's historical financial data with external market indicators. Client attrition risk scoring integrates five signal streams: declining instruction volume, increasing matter-to-close cycle times, reduced responsiveness to firm communications, increased RFP activity detected through market intelligence, and rate sensitivity indicators from billing negotiation patterns. The model generates 90-day risk scores with confidence intervals, enabling partners to engage proactively with at-risk relationships rather than discovering departures after the fact.

Module 06
Knowledge Management
Your firm's collective intelligence — the institutional memory that never retires

Every brief filed, every research memo drafted, every contract negotiated represents institutional knowledge that most firms lose the moment it is saved to a file server. Arbiter's Knowledge Management module indexes the firm's entire work product, making it searchable by legal issue, jurisdiction, client, attorney, and outcome. When an associate begins research on a new matter, Arbiter surfaces the firm's prior work on the same issue before they open an external database — reducing duplicated research and leveraging the collective expertise of the partnership.

The institutional memory AI identifies expert attorneys on specific legal issues based on their historical work product, enabling efficient staffing decisions and internal referrals. The clause library extracts and indexes contractual provisions, settlement terms, and drafted language across all firm documents, creating a searchable repository of firm-tested language for every common legal situation.

100%
Firm work product indexed and searchable
34%
Reduction in duplicated research effort
Auto
Expert identification by legal issue
Work Product Indexing

The indexing pipeline processes documents through four layers: (1) structural parsing — identifying document type (brief, memo, contract, correspondence), sections, headings, and citation blocks; (2) entity extraction — identifying parties, courts, judges, legal issues, and statutory references using legal NER; (3) semantic embedding — generating 768-dimensional document embeddings using the same legal-domain transformer as Module 01, enabling concept-level search across the firm's corpus; (4) outcome tagging — where available, linking work product to case outcomes (motion granted/denied, settlement achieved, verdict entered) to enable outcome-aware precedent retrieval.

Clause Library Architecture

The clause extraction system uses a fine-tuned transformer model to identify and classify contractual provisions, settlement terms, stipulations, and drafted legal language across all firm documents. Extracted clauses are deduplicated, version-tracked, and indexed by legal category (indemnification, limitation of liability, force majeure, governing law, dispute resolution), enabling attorneys to search for firm-tested language that has been negotiated, litigated, and refined over the partnership's collective history. The system tracks clause provenance — which attorney drafted it, in which matter, and with what outcome — providing context that informs future drafting decisions.

AI Engine A1
Hallucination-Proof Citations
Zero tolerance for fabricated authority — the ethical foundation of AI legal practice

Multiple high-profile cases in 2023 and 2024 involved attorneys who submitted AI-generated briefs containing fabricated case citations to federal courts, resulting in sanctions, fines, and professional discipline. These failures stemmed from general-purpose AI models that generate fluent but unverified legal text — hallucinating case names, holdings, and citations that do not exist. Arbiter's hallucination-proof citation engine employs a closed-universe retrieval architecture that makes fabricated citations structurally impossible: the system can only cite authority that exists in the verified 14.2-million-document corpus, and every holding statement is validated against the actual text of the cited decision.

The verification layer operates in three stages: (1) existence verification — confirming that the cited case name, reporter citation, and court match an actual indexed decision; (2) holding verification — comparing the AI-generated description of the holding against the actual decision text using semantic similarity scoring with a threshold of 0.87; (3) treatment verification — confirming that the cited authority remains good law by traversing the citation graph for negative subsequent treatment. If any stage fails, the citation is rejected with an explanatory message identifying the specific failure.

97.4%
Citation accuracy — independently validated
0%
Tolerance for fabricated authority — structural impossibility
3-stage
Verification: existence → holding → treatment
Closed-Universe RAG Architecture

The hallucination prevention architecture uses retrieval-augmented generation constrained to the verified corpus. Unlike open-generation approaches where the language model can fabricate plausible-sounding but nonexistent citations, Arbiter's RAG pipeline requires that every cited authority be retrieved from the indexed corpus before it can appear in any output. The retrieval step acts as a hard constraint on generation — the model cannot cite what it cannot retrieve, and the retrieval system cannot retrieve what does not exist in the verified index. This closed-universe approach reduces hallucination to the residual error rate of the retrieval and verification systems (2.6%), rather than the 17–48% hallucination rates observed in general-purpose AI models on legal tasks.

Professional Responsibility Compliance

The system is designed specifically to support attorneys' obligations under Rule 1.1 (duty of competence) and Rule 5.3 (duty of supervision) of the Model Rules of Professional Conduct. Every AI-generated citation includes a verification badge indicating whether the authority has passed all three verification stages, enabling attorneys to fulfill their supervisory obligations with transparent AI audit trails. The system generates an attestation log for each document recording every citation verification result, providing a defensible record of due diligence if opposing counsel or the court questions the provenance of cited authority.

AI Engine A2
Predictive Case Analytics
Judicial behavior analyzed at the individual judge level — 8,200+ judges quantified

Engine A2 transforms litigation strategy from anecdote-driven intuition to empirical analysis. The system profiles 8,200+ federal and state judges across every measurable dimension of judicial behavior: motion grant rates by motion type and case category, sentencing tendencies relative to guidelines ranges, citation preferences (which authorities each judge relies on most frequently), procedural idiosyncrasies (oral argument frequency, sua sponte dismissal rates, typical briefing page limits), and historical case outcome patterns.

The judge profiling model processes the complete docket history of each judge — orders, opinions, rulings, sentencing records — to generate behavioral profiles that enable attorneys to calibrate their litigation strategy to the specific tendencies of their assigned jurist. The system identifies which arguments are most likely to succeed before a specific judge, which citation patterns align with the judge's analytical framework, and which procedural strategies are most effective.

8,200+
Federal and state judges profiled
92%
Motion outcome prediction accuracy by judge
Real-time
Profile updates as new opinions are published
Judge Profiling Architecture

Each judge profile is built from the complete docket history processed through a multi-task NLP pipeline: (1) motion outcome classification — cataloguing every motion ruling with motion type, case category, and outcome; (2) opinion analysis — extracting citation patterns, reasoning frameworks, and doctrinal preferences from published opinions; (3) procedural pattern analysis — tracking scheduling preferences, hearing formats, and case management tendencies from docket entries; (4) sentencing analysis — computing departure rates, guideline adherence patterns, and variance by offense type for criminal dockets. Profiles are updated continuously as new docket entries and opinions are published.

Strategy Calibration

The strategy calibration module translates judge profiles into actionable litigation guidance: "Judge [X] grants summary judgment in employment discrimination cases at 64% compared to the district average of 47% — but 89% of grants cite failure to establish prima facie case rather than pretext analysis, suggesting emphasis on prima facie element arguments." This level of empirical specificity transforms briefing strategy from "we think this judge is conservative" to data-driven tactical decisions grounded in the judge's actual decisional pattern across thousands of rulings.

AI Engine A3
Adverse Authority Detection
Catching what manual review misses 34% of the time — before you file

Before an attorney files a brief, Engine A3 scans the complete argument against the full corpus to identify controlling adverse authority that the attorney may have missed. The ethical obligation to disclose directly adverse controlling authority is absolute under Rule 3.3(a)(2) of the Model Rules of Professional Conduct — and the consequences of missing it are severe: sanctions, adverse inferences, and reputational damage that no malpractice policy can fully remedy.

The adverse authority detection model processes each argument in the draft, identifies the legal propositions asserted, and searches for decisions in the controlling jurisdiction that reach contrary conclusions on the same legal issue. The system achieved a 34% improvement in adverse authority identification over manual review in a controlled study with 120 litigators — catching relevant adverse cases that experienced attorneys missed due to time constraints, unfamiliarity with recent developments, or search strategy blind spots.

34%
More adverse cases caught vs. manual review
Rule 3.3
Candor obligation compliance automation
Pre-filing
Detection before submission — not after opposing counsel's response
Detection Architecture

The adverse authority detection pipeline operates in three stages: (1) proposition extraction — the system identifies each legal proposition asserted in the draft using a legal argument mining model that decomposes arguments into discrete claims; (2) contrary authority search — for each proposition, the system searches the corpus for decisions in the controlling jurisdiction that reach the opposite conclusion, applying semantic similarity to identify cases that address the same legal question but resolve it differently; (3) materiality assessment — the system evaluates whether the identified adverse authority is "directly adverse" within the meaning of Rule 3.3 by analyzing jurisdictional hierarchy, factual similarity, and whether the adverse holding has been distinguished or limited.

Risk Mitigation Value

The 34% improvement over manual review was measured in a controlled study where 120 experienced litigators (median 12 years practice) performed manual adverse authority searches on test briefs. Engine A3 identified an average of 2.1 relevant adverse authorities per brief that were missed by manual review. The most common failure mode in manual review was not negligence but search strategy limitation — attorneys searched for adverse authority using their own conception of the legal issue, while the AI system explored semantic variations in how courts have framed the same question. This complements rather than replaces attorney judgment: the system surfaces candidates, and the attorney applies professional judgment to determine whether each candidate is truly "directly adverse controlling authority" requiring disclosure.

AI Engine A4
Argument Strength Scoring
Data-informed confidence scores tested against 10,000+ actual outcomes

Engine A4 evaluates each argument in a brief against the weight of supporting authority, jurisdictional alignment, and historical success rates for similar claims — delivering a data-informed confidence score that helps partners allocate resources to the arguments most likely to prevail. The system was validated against 10,000+ actual case outcomes, demonstrating that arguments scored in the top quintile (80–100) prevailed 74% of the time, while arguments in the bottom quintile (0–20) prevailed only 12% of the time.

The scoring model integrates four dimensions: (1) authority weight — volume and jurisdictional tier of supporting authority; (2) treatment health — current treatment status of cited authorities; (3) historical success rate — outcomes of similar arguments in the same jurisdiction and case type; (4) judge alignment — whether the argument pattern matches the assigned judge's decisional tendencies (from Engine A2). The composite score enables evidence-based resource allocation: strengthening high-potential arguments with additional authority while reconsidering or restructuring low-scoring arguments before filing.

10,000+
Actual case outcomes in validation dataset
74%
Win rate for top-quintile scored arguments
4-axis
Scoring: authority, treatment, history, judge alignment
Scoring Methodology

The argument strength score is computed as a weighted composite of four sub-scores, each normalized to a 0–100 scale: authority weight (0.30) — computed from the volume of supporting authority, the jurisdictional tier of each cited decision (Supreme Court > Circuit > District > State appellate), and the depth of analysis in each supporting opinion; treatment health (0.25) — the aggregate treatment status of all cited authorities, with negative treatment reducing the score proportionally; historical success rate (0.25) — the win rate for similar arguments (classified by legal issue, motion type, and case category) in the same jurisdiction; and judge alignment (0.20) — the correlation between the argument pattern and the assigned judge's historical decisional tendencies from Engine A2 profiles.

Resource Allocation Application

The primary application of argument strength scoring is not prediction but resource allocation. When a brief contains eight arguments, and briefing page limits require prioritization, the scores enable evidence-based decisions about which arguments deserve the most thorough development. The validation data demonstrates strong discriminative power: arguments in the 80–100 range prevail 74% of the time; arguments in the 60–80 range prevail 58%; arguments in the 40–60 range prevail 41%; arguments in the 20–40 range prevail 27%; and arguments in the 0–20 range prevail only 12%. This gradient enables attorneys to invest drafting resources proportionally to expected return, rather than distributing effort uniformly across arguments of unequal strength.