Colrows Case Study  ·  Financial Services  ·  India NPA Portfolio Intelligence
Case Study

From months to hours:
Reimagining NPA portfolio
evaluation with semantic AI

How a leading Indian asset reconstruction company eliminated subjectivity, reduced evaluation time by over 95%, and deployed institutional knowledge at machine scale — using the Colrows semantic reasoning layer.

Industry
Asset Reconstruction / Financial Services
Market
India — Retail NPA Segment
Challenge
Portfolio evaluation speed & consistency
Solution
Colrows Semantic AI Agent
>95%
Reduction in evaluation cycle time — from months to hours
↑ Competitive velocity
100%
Regulatory coverage — RBI SARFAESI, DRT framework, and sub-regulations modeled in the graph
↑ Compliance certainty
Expert disciplines unified — legal, technical, field ops, and executive judgment encoded in one reasoning layer
↓ Analyst subjectivity
01
The challenge: speed is the only moat

In India's asset reconstruction market, the window to acquire a distressed retail NPA portfolio is brutally narrow. The firm that bids first — and bids accurately — wins. Every day spent evaluating a portfolio is a day a competitor could pre-empt the deal.

The institution had assembled a formidable multi-disciplinary team: technologists, legal specialists, field investigators, and senior executives. But the evaluation process was fundamentally bottlenecked by human bandwidth and sequential workflows.

A standard portfolio evaluation required the team to manually segment thousands of accounts, assess borrower behavioural traits, review KYC and loan documentation, and interpret an evolving regulatory landscape — before arriving at an indicative valuation.

"The process was sound. The people were excellent. But the market doesn't wait for excellent processes."

Internal assessment, Pre-deployment review

Evaluation latency

End-to-end portfolio assessment took several months, constraining the firm's ability to bid competitively in fast-moving auctions.

Regulatory complexity

Correct valuation demanded fluency in RBI SARFAESI, DRT frameworks, and overlapping sub-regulations — expertise that varied across analysts.

👤

Analyst subjectivity

Segmentation logic and file interpretation were anchored to individual analysts' experience — creating inconsistent outputs and key-person risk.

📁

Document intensity

Each borrower file contained layered KYC, payment history, field reports, and behavioural indicators — requiring skilled human review throughout.

02
Why not just train a model?

The obvious question is whether a machine learning model trained on historical NPA portfolio outcomes would solve the problem. It's a legitimate hypothesis — and one the institution examined carefully. The answer reveals something important about the nature of the problem itself.

ML model on historical data

Pattern recognition from past outcomes

Strong on bulk scoring

Predicts recovery probability for borrower profiles it has seen before — fast and scalable for routine accounts.

Good at pattern flagging

Surfaces accounts that resemble past defaulters or high-recovery profiles with high throughput.

Blind to regulatory change

A model trained on 2021 data doesn't know a 2024 RBI circular changed enforcement thresholds. It reasons from an outdated legal reality — confidently.

Survivorship bias baked in

Training data reflects portfolios the firm bid on and won — not the full evaluation universe. Edge cases and passed deals are invisible to the model.

Weakest on edge cases

The accounts that determine bid profitability — atypical collateral, complex DRT histories, contested SARFAESI enforcements — are precisely where frequency-based models fail.

Cannot explain its reasoning

An investment committee bidding hundreds of crores cannot accept "the model says 34 paise on the rupee." A rationale is required, not a score.

Colrows semantic layer + AI agent

Reasoning anchored to encoded knowledge

Bulk scoring via trained agent

Historical data still powers pattern recognition — but anchored to the semantic graph, not free-floating.

Regulatory graph, always current

SARFAESI and DRT provisions are modeled as nodes and relationships. Update the graph once; every subsequent evaluation inherits the change instantly.

Explicit handling of edge cases

Unusual accounts are evaluated against the full regulatory and institutional knowledge graph — not extrapolated from statistical frequency.

Traceable, auditable reasoning

Every valuation output carries a reasoning trail — which provisions applied, which borrower signals were weighted, which heuristics drove the final figure.

Investment-committee ready

Outputs are structured as defensible rationales, not probability scores — designed to survive scrutiny at the bid approval stage.

Institutional knowledge, not just data

Legal heuristics, executive judgment, and field-team expertise are encoded explicitly — not hoped to emerge from training data.

The fundamental insight: NPA evaluation is not primarily a data problem. The data existed and was accessible. The bottleneck was interpretation — knowing which regulatory provision applied, what a payment pattern signalled about borrower intent, how to weight collateral quality against recovery probability in a DRT proceeding. Interpretation requires reasoning, not pattern matching.

03
The solution: institutional knowledge, encoded

Colrows was deployed as a semantic reasoning layer — not a rules engine, not a dashboard, not a conventional AI model. The regulatory graph, tribal knowledge, and trained AI agent work as a unified system. Each component does what it does best.

ML component — what it handles
  • Bulk account scoring against historical recovery patterns
  • Borrower behavioural trait classification at scale
  • Payment pattern anomaly detection across thousands of accounts
  • Collateral liquidation timeline estimation from historical DRT durations
  • Stratified sampling selection for portfolio representativeness
Semantic graph — what it handles
  • Current regulatory standing of each account under SARFAESI and DRT
  • Edge-case interpretation where historical precedent is thin or absent
  • Legal enforceability assessment against latest RBI circulars
  • Cross-referencing executive and legal heuristics for unusual account types
  • Generating the auditable reasoning trail for bid rationale documentation
The synthesis

The AI agent sits above both layers — using the ML component for speed and scale on routine accounts, and the semantic graph for correctness and explainability on complex ones. Neither alone is sufficient. Together, they cover the full evaluation surface.

AI Agent Layer
NPA Evaluation Agent
Portfolio Pricing Engine
Risk Segmentation Module
Semantic Layer
Colrows Knowledge Graph
Regulatory Ontology
Tribal Knowledge Models
Behavioural Signal Library
Knowledge Sources
RBI SARFAESI
DRT Framework
Legal SME
Tech analyst playbooks
Field investigation protocols
Executive heuristics
Data Inputs
Loan files
KYC documents
Payment history
Borrower behavioural data
Historical portfolio outcomes
Evaluation workflow

How the Colrows agent processes a portfolio

01

Portfolio ingestion & pre-classification

The agent ingests the raw portfolio and pre-classifies accounts against the semantic parameter graph.

02

Graph-anchored segmentation

Accounts are segmented by regulatory standing, recovery probability, borrower profile, and collateral quality — simultaneously, at scale.

03

Intelligent sampling & deep analysis

The agent selects representative samples and performs deep file analysis — surfacing behavioural traits, compliance gaps, and risk indicators.

04

Indicative valuation & bid pricing

Drawing on historical outcomes and the regulatory model, the agent derives an indicative portfolio value with a full reasoning trail.

04
The hidden risk: regulatory drift and the cost of static models

In NPA markets, the firms that consistently overpay are almost never the ones with bad analysts. They are the ones whose evaluation frameworks have quietly fallen behind the regulatory environment. Regulatory drift is silent, accumulative, and expensive — and it is the specific failure mode that a purely data-trained model cannot detect.

📋
Risk: static model

RBI circular amends SARFAESI enforcement threshold

A model trained before the amendment continues applying the old threshold to every account it evaluates. Entire account categories shift legal enforceability status — the model doesn't know. Valuations drift from reality, silently.

Risk: static model

DRT procedural update changes timeline assumptions

Historical recovery timelines underpin ML probability estimates. When DRT processing times shift due to procedural changes or case backlog, every timeline-dependent valuation becomes systematically miscalibrated — and the model has no way to surface this to the analyst.

Colrows approach

Semantic graph updated once — agent inherits immediately

When a regulatory change occurs, the Colrows graph is updated at the node and relationship level. Every subsequent evaluation — that day, that hour — automatically reflects the new regulatory reality. No retraining cycle. No lag. No silent miscalibration compounding across a portfolio.

05
Explainability: the bid rationale problem

In asset reconstruction, a valuation that cannot be explained is a valuation that cannot be defended — to an investment committee, to a regulator, or to a counterparty challenging the bid. Explainability is not a nice-to-have. It is a governance requirement.

The Colrows agent produces structured reasoning outputs alongside every valuation, tracing exactly which regulatory provisions, borrower signals, and institutional heuristics drove each conclusion. Below is a representative output structure for a single evaluated account.

Sample account evaluation output

Illustrative · Account #NPA-2847

Account profile & regulatory standing

Account typeRetail secured — home loan, LAP
Outstanding₹18.4L principal + ₹6.2L interest
SARFAESIEligible. Notice served, possession not taken. Section 13(4) applicable.
DRT statusNo active proceedings. Clean enforcement path.
CollateralRegistered urban property. Circle rate ₹22.1L. Legal title clear.
Behavioural signalPartial payments post-NPA classification — intent to settle flagged by agent.

Reasoning trail & indicative valuation

Recovery pathSARFAESI possession → auction. Estimated 8–11 months per DRT duration model.
Regulatory riskLow. No sub-judice complications. Enforcement risk: minimal.
Heuristic appliedLegal SME rule: partial payment post-NPA reduces haircut by 8–12% vs. silent account.
ML signalProfile similarity to historical high-recovery cohort: 74th percentile.
Indicative value₹14.8L–₹16.2L (58–63 paise on total outstanding)

Reasoning: Clean legal standing + settlement intent + strong collateral coverage supports above-average recovery. Colrows recommends inclusion in bid at upper-band pricing.

"The investment committee didn't ask us to justify the number. For the first time, the number came with its own justification — traceable to regulation, to precedent, to the account itself."

Post-deployment reflection
06
Outcomes: evaluation as competitive advantage

The deployment reframed what evaluation means for the institution — shifting it from a capacity constraint to a strategic capability.

Hours

End-to-end evaluation turnaround

What previously consumed multiple months is now completed within hours — enabling the firm to evaluate and bid on portfolios that were previously logistically inaccessible.

Zero

Regulatory blind spots

Every account evaluation is automatically cross-referenced against the full SARFAESI and DRT regulatory surface — not just the provisions an individual analyst happened to know.

4 → 1

Expert disciplines unified

Legal, technical, field, and executive knowledge — previously distributed across individuals and siloed in experience — now operates as a single coherent reasoning system.

↑ Bids

Portfolio pipeline expansion

With evaluation latency eliminated, the firm can participate in a significantly larger share of portfolio auctions — including fast-moving deals that previously expired before assessment was complete.

07
Why semantic reasoning — not just AI

Conventional AI approaches can surface patterns. But they cannot reliably reason about why those patterns matter, or apply that reasoning consistently against a shifting regulatory backdrop.

Colrows constructs a semantic graph — a structured representation of the domain in which concepts, relationships, rules, and heuristics are explicit nodes and edges, not latent parameters. The AI agent is then anchored to this graph, ensuring every inference is traceable, auditable, and grounded in the firm's own institutional reasoning.

The result is a system that combines the scale of ML with the correctness of structured legal and institutional knowledge — and produces outputs that survive both market scrutiny and investment committee review.

Colrows vs conventional AI
Conventional AI
Colrows semantic layer
Learns from data patterns
Reasons from encoded knowledge
Opaque inference
Traceable, auditable reasoning
Brittle to regulatory change
Graph updated, agent inherits
Generic domain understanding
Institution-specific heuristics
Black-box outputs
Explainable, bid-ready rationale