Blog
TechMay 202614 min read

Reconstructing Financial Evidence from Degraded Documents

One of the more underappreciated problems in lending infrastructure is that a large percentage of underwriting globally still depends on low-quality visual documents.

Particularly for MSMEs and informal businesses, financial history is rarely machine-native. Instead, lenders evaluate fragmented visual evidence: photographed receipts, compressed bank statements sent over WhatsApp, faded invoices, handwritten ledgers, screenshots of e-wallet histories, and multi-generation photocopies captured under poor lighting conditions.

Sources of financial evidence in MSME lending

Photographed receipts
WhatsApp bank statements
Faded invoices
Handwritten ledgers
E-wallet screenshots
Multi-gen photocopies

The problem is that most OCR systems are fundamentally optimized for clean enterprise documents. Underwriting workflows violate nearly every assumption these systems rely on. Real-world financial documents contain motion blur, defocus blur, JPEG compression artifacts, perspective distortion, overlapping handwriting, thermal degradation, multilingual text, and repeated re-photographing through messaging platforms. Once resolution falls below a certain threshold, OCR systems stop performing deterministic recognition and begin hallucinating semantically invalid outputs.

The Limits of Generic Super-Resolution

A common approach is to apply generic image super-resolution before OCR. In practice, this rarely works well for financial documents. Traditional super-resolution systems optimize for perceptual realism using objectives such as PSNR or SSIM. These losses reward visually plausible high-frequency detail rather than semantically correct character reconstruction. As a result, many restoration pipelines generate text that appears sharper while simultaneously degrading OCR fidelity.

The problem becomes more severe in diffusion-based restoration systems. Diffusion models are highly effective at generating visually realistic textures, but unconstrained denoising frequently introduces hallucinated glyphs, merged characters, or semantically impossible token sequences. Recent work in Scene Text Image Super-Resolution (STISR) repeatedly demonstrates that perceptually sharp outputs do not necessarily correspond to improved recognition accuracy.

For underwriting systems, “looks correct” is insufficient. Reconstruction must remain semantically and financially consistent.

Recognition-Guided Diffusion for Financial Documents

Recognition-guided diffusion models address this problem by conditioning image restoration on semantic priors throughout the denoising process. Instead of treating OCR and restoration as separate stages, weak textual signals extracted from degraded inputs are injected directly into intermediate diffusion steps through cross-attention and recognition-guided feature modulation.

The key insight is that even low-confidence OCR outputs contain meaningful semantic structure. Partial merchant names, probable date formats, transaction grammars, institution-specific layouts, neighboring balances, and multilingual character distributions all provide constraints over the space of valid reconstructions.

Progressive narrowing of reconstruction uncertainty

Each denoising step constrains the space of valid outputs. Semantic priors eliminate impossible reconstructions, converging toward the single most plausible document.

t = Tt = 4/5 Tt = 3/5 Tt = 2/5 Tt = 1/5 Tt = 0Image evidenceLayout priorsOCR priorsFinancial priorsHigh uncertaintyMany possible outputsConvergedSingle valid output

Recognition-guided diffusion systems leverage these weak priors to bias denoising toward semantically plausible character manifolds rather than unconstrained texture generation. Recent architectures such as RGDiffSR and TextSR demonstrate that OCR-conditioned diffusion significantly improves reconstruction fidelity under severe blur, compression, and occlusion conditions.

At Kita Capture, we build on similar principles through text-conditioned super-resolution pipelines optimized for noisy financial records. An initial transcription stage generates weak semantic priors from low-confidence OCR outputs and layout-aware parsing. Retrieval systems identify structurally similar historical documents and institution-specific templates to establish stronger priors over transaction structure, typography, and field relationships. During restoration, these priors condition the reconstruction process such that generated outputs remain constrained by both the observed pixels and expected financial consistency.

Retrieval-Augmented Reconstruction and Financial Priors

One important observation is that financial documents exhibit far stronger structural regularity than generic scene text. Bank statements, receipts, invoices, and remittance records all follow latent institutional grammars even when visually degraded.

This makes retrieval-augmented restoration particularly effective. Rather than reconstructing characters independently, restoration pipelines can condition on structurally similar historical documents scraped from prior underwriting workflows.

Retrieval-augmented reconstruction

InputHistorical docsConditioned SRRestoredTransaction orderingColumn alignmentCurrency formattingStatement headersMerchant distributionsBalance trajectoriesInstitution typographyRetrieved priors

Conditioning diffusion models on retrieved financial priors substantially narrows the reconstruction search space and reduces hallucination risk.

Joint Optimization of Restoration and Fraud Detection

An unintuitive outcome is that restoration and fraud detection become tightly coupled optimization problems.

Many manipulated financial documents intentionally exploit degradation artifacts. Compression obscures editing seams, screenshots destroy metadata, blur masks synthetic overlays, and repeated photographing hides tampering traces. Restoration systems therefore cannot operate independently from authenticity analysis.

Joint restoration and fraud detection

Degraded input(possibly tampered)Semantic restorationCharacter boundaries, layoutAuthenticity analysisFont, spacing, compressionJoint assessmentRestoration confidenceAuthenticity score+ Font irregularities+ Spacing anomalies+ Balance discontinuities+ Recompression artifactsExposed after restoration

Semantically grounded reconstruction can paradoxically improve fraud detection because reconstructed character boundaries and layout structure expose inconsistencies that were previously hidden in degraded inputs. Font irregularities, spacing anomalies, transaction discontinuities, recompression artifacts, and impossible balance transitions become more observable after denoising.

Beyond OCR: Underwriting Thin-File Businesses

A substantial percentage of small businesses globally remain “thin-file” not because economic activity is absent, but because financial history exists only as degraded visual evidence scattered across informal systems.

Semantically grounded restoration systems make it possible to transform fragmented visual records into structured underwriting signals. As these models improve, lenders can increasingly reason over businesses that previously sat outside traditional digital financial infrastructure entirely.

The long-term implication is that document quality ceases to be a gating factor for financial access. When systems can reliably reconstruct financial evidence from degraded visual inputs, the set of borrowers that lenders can evaluate expands substantially — without requiring those borrowers to first digitize their financial lives.

References

Recognition-Guided Diffusion Model for Scene Text Image Super-Resolution — arXiv 2311.13317