OCR Models

From AI Wiki

Template:Infobox technology

OCR Models are artificial intelligence (AI) systems designed for Optical Character Recognition (OCR), the process of converting images of typed, handwritten, or printed text into machine-readable digital text.[1] These models leverage techniques from computer vision, machine learning, and deep learning to extract and interpret textual information from visual sources such as scanned documents, photographs, and videos. Modern OCR models employ convolutional neural networks (CNNs), recurrent neural networks (RNNs), and transformers to recognize text directly from pixel data, often without requiring explicit character segmentation.[1]

OCR models have evolved from rule-based systems to advanced neural network architectures, enabling applications in document digitization, automation, and accessibility. They are integral to modern AI systems, often integrated with large language models for enhanced document understanding.[2] The global OCR market is projected to grow from $12.44 billion in 2023 to $38.59 billion by 2032, with a 15.20% compound annual growth rate.[3]

History

Early Mechanical Systems (1870s-1940s)

The foundations of OCR technology date back to the late 19th century. In 1870, Charles R. Carey invented the retina scanner using a mosaic of photocells, marking the first OCR-related invention.[4] In 1913, Emanuel Goldberg developed a machine that converted printed characters into telegraph code. Around the same time, Edmund Fournier d'Albe invented the Optophone, a handheld scanner that produced tones corresponding to letters when moved across printed pages, designed to aid the blind.[4]

In 1929, Austrian engineer Gustav Tauschek developed the 'Reading Machine,' another early OCR device that used a photodetector and templates to recognize characters. Goldberg later created a "Statistical Machine" in 1931 for searching microfilm archives using optical code recognition, patented as US Patent 1,838,389.[5]

First Commercial OCR (1950s-1970s)

The first practical OCR machine emerged in the 1950s. In 1951, David H. Shepard built "GISMO," a machine that initially read Morse code and later adapted for printed text. In 1954, Reader's Digest implemented the first commercial OCR system for converting typewritten sales reports into punch cards for computer processing.[6] The United States Postal Service began using OCR for mail sorting in the late 1950s.[7]

The 1960s introduced standardized fonts to improve machine readability. OCR-A was introduced in 1961, followed by OCR-B in 1968, which became an international standard. In 1965, the U.S. Postal Service installed OCR machines in Detroit. Companies like IBM and Recognition Equipment Inc. developed systems for banking and postal applications.[8]

A major breakthrough occurred in 1974 when Ray Kurzweil founded Kurzweil Computer Products and developed the first omni-font OCR system capable of recognizing text in virtually any font. In 1976, Kurzweil unveiled a reading machine for the blind using CCD scanners and text-to-speech synthesis. Stevie Wonder purchased one of the first units, beginning a lifelong friendship with Kurzweil. This innovation was regarded as the most significant advancement for the blind since Braille in 1829.[9]

The Neural Network Revolution (1980s-2000s)

The late 1980s marked the beginning of machine learning integration in OCR. A landmark achievement came in 1989 when Yann LeCun and colleagues at Bell Labs created a convolutional neural network that could recognize handwritten ZIP code digits with approximately 95% accuracy, using a large dataset of scanned mail. This system was subsequently deployed by the U.S. Postal Service for automated mail sorting in the early 1990s.[10]

Hewlett-Packard developed Tesseract from 1984-1994 as proprietary software. After appearing at the 1995 UNLV Fourth Annual Test of OCR Accuracy as one of the top three engines, HP open-sourced Tesseract in 2005. Google began sponsoring its development in 2006, with original developer Ray Smith joining as a Google employee.[11]

Deep Learning Era (2010s-Present)

The 2010s saw a revolution through deep learning. Tesseract 4.0, released in 2018, incorporated LSTM networks and supported over 100 languages, dramatically improving accuracy from earlier versions to 95-98% on structured documents.[11]

Open-source libraries emerged rapidly: EasyOCR (2019) and PaddleOCR (2020) leveraged CNNs and RNNs for multilingual support. By the 2020s, transformer-based models like TrOCR (Microsoft, 2021) and Donut (NAVER, 2021) achieved state-of-the-art performance.[12]

Multimodal vision-language models like GPT-4 Vision, Gemini, and Llama 3.2 Vision now integrate OCR capabilities as part of broader visual understanding, achieving competitive or superior performance compared to specialized OCR systems.[13]

Types of OCR Models

OCR models can be classified into several categories based on their underlying technology and recognition capabilities:

Traditional OCR Models

Traditional approaches rely on rule-based methods without learning from data:

  • Pattern Matching: Compares isolated glyphs pixel-by-pixel with stored templates. Effective for fixed fonts but fails with variations.[14]
  • Feature Extraction: Decomposes characters into features (lines, loops, corners) and uses classifiers like k-nearest neighbors or support vector machines. Handles more variability than pattern matching but requires extensive feature engineering.

Machine Learning-Based OCR Models

These models train on datasets to recognize patterns:

  • Simple OCR Software: Matches characters or words to templates, suitable for typewritten text with known fonts.
  • Intelligent Character Recognition (ICR): Uses neural networks specifically for handwriting recognition, evolving from intelligent word recognition systems.[15]

Deep Learning-Based OCR Models

Modern OCR leverages advanced neural architectures:

Comparison of OCR Model Types
Type Description Examples Strengths Limitations
Traditional (Pattern Matching) Pixel-by-pixel template comparison Early OCR systems, OCR-A/B fonts Fast for known fonts Limited to fixed styles
Machine Learning Trained classifiers on features ICR software, SVM-based systems Handles handwriting Requires extensive training data
Deep Learning (CNN-RNN) Neural networks for feature extraction and sequence modeling CRNN, Tesseract 4+ High accuracy, versatile Computationally intensive
Vision-Language Models Multimodal integration with language understanding GPT-4 Vision, Gemini, Claude Contextual understanding High computational cost

Core Architectures

Convolutional Neural Networks (CNNs) for Feature Extraction

Convolutional Neural Networks serve as the foundation for visual analysis in modern OCR. CNNs process images through multiple layers:

  • Initial layers detect simple features (edges, corners, gradients)
  • Middle layers identify character components
  • Deep layers recognize complete character shapes

Popular CNN architectures used in OCR include:

  • VGGNet: 16-19 layers using 3×3 convolutions, 138 million parameters
  • ResNet: 50-152 layers with skip connections to prevent vanishing gradients
  • MobileNet: Depthwise separable convolutions for mobile deployment[17]

Recurrent Neural Networks (RNNs) for Sequence Modeling

Since text is inherently sequential, RNNs model the dependencies between characters:

Long Short-Term Memory (LSTM): Uses gating mechanisms (input, forget, output gates) to regulate information flow, effectively learning patterns over long sequences. Mathematical formulation includes:

  • Forget gate: f_t = σ(W_f · [h_{t-1}, x_t] + b_f)
  • Input gate: i_t = σ(W_i · [h_{t-1}, x_t] + b_i)
  • Cell state: C_t = f_t * C_{t-1} + i_t * C̃_t
  • Output: h_t = o_t * tanh(C_t)

Gated Recurrent Unit (GRU): Simplified LSTM variant combining forget and input gates into single update gate, computationally more efficient.[18]

Bidirectional RNNs: Process sequences in both forward and backward directions simultaneously, providing context from both past and future characters.

CRNN Architecture

The Convolutional Recurrent Neural Network (CRNN) combines CNNs and RNNs into an end-to-end trainable model:

1. Convolutional Layers: Deep CNN backbone extracts feature vectors from image frames 2. Recurrent Layers: Bidirectional LSTM models contextual dependencies 3. Transcription Layer: Connectionist Temporal Classification (CTC) loss translates predictions to text

CTC loss enables training without explicit character-level alignment by introducing blank tokens and collapsing output sequences (for example "--hh-e-l-ll-oo--" → "hello").[19]

Performance benchmarks:

  • IIIT5K: 78.2% accuracy without lexicon, 97.6% with 1000-word lexicon
  • ICDAR 2003: 89.4% accuracy, 98.7% with full lexicon
  • Multi-Scale Fusion CRNN improves by 5.9-8.2% across datasets[20]

Attention Mechanisms

Attention mechanisms allow models to dynamically focus on relevant parts of the input:

  • Self-attention: Computes Attention(Q, K, V) = softmax(QK^T / √d_k) × V
  • Multi-head attention: 12-16 parallel attention heads learning different aspects
  • Cross-attention: Enables decoders to attend to relevant image regions while generating output[21]

Transformer-Based Models

Transformers eliminate recurrence entirely, using self-attention for parallel processing:

TrOCR (Microsoft, 2021):

  • Architecture: Vision Transformer encoder + text decoder
  • Performance: 2.89% CER on handwriting (TrOCR-Large, 558M parameters)
  • Training: 684 million synthetic textlines, then 17.9 million handwritten samples[22]

Donut (NAVER CLOVA, 2022):

  • OCR-free approach: Directly generates structured outputs from pixels
  • Architecture: Swin Transformer encoder + BART decoder
  • Performance: 95.30% accuracy on document classification with 143M parameters[23]

LayoutLM Family:

  • LayoutLM v1: Adds 2D positional embeddings to BERT
  • LayoutLMv2: Spatial-aware self-attention, 84.20% F1 on FUNSD
  • LayoutLMv3: Pure vision transformer approach, 95.44% on RVL-CDIP[24]

Notable OCR Models and Systems

Open-Source OCR Engines

Model Developer Key Features Languages Performance
Tesseract Google (originally HP) LSTM-based neural network engine (v4+), supports multiple output formats (hOCR, PDF, TSV) 100+ 95-98% on clean printed text, 50-70% on handwriting
EasyOCR Jaided AI PyTorch-based, GPU acceleration, simple Python API 80+ 85-95% depending on document type
PaddleOCR Baidu PP-OCRv5 with unified recognition, PP-StructureV3 for layout analysis, mobile-optimized 80+ State-of-the-art on Chinese text, 370+ chars/second on CPU
docTR Mindee Multiple detector/recognizer options, PyTorch/TensorFlow backends 50+ High customization, competitive accuracy
TrOCR Microsoft Research Pure transformer architecture, excellent for handwriting Multiple 2.89% CER on IAM handwriting dataset
MMOCR OpenMMLab 14+ algorithms, modular components, deployment tools Extensible Research-grade, highly configurable

Commercial Cloud Services

Service Provider Key Features Pricing Accuracy
Google Cloud Vision Google 60+ languages, batch processing, layout analysis $1.50 per 1,000 units (1,001-5M) 98.0% on mixed documents
Amazon Textract AWS Specialized APIs for forms, tables, expenses, IDs $0.015-0.05 per page depending on type 95-99% on structured documents
Azure Computer Vision Microsoft 100+ languages, handwriting support, spatial analysis $1.50 per 1,000 transactions (0-1M) 95-98% on standard documents
Mistral OCR Mistral AI Handwriting, tables, markdown output $1 per 1,000 pages High accuracy with LLM integration

Multimodal Vision-Language Models

Model Developer Parameters Key Features OCR Capability
GPT-4 Vision OpenAI Undisclosed General visual understanding Industry-leading accuracy
Gemini 2.5 Pro Google Undisclosed Cost-efficient processing Strong multilingual OCR
Claude 3.7 Sonnet Anthropic Undisclosed Balanced performance Excellent on diverse domains
Llama 3.2 Vision Meta 11B, 70B Open source, community license Competitive with commercial models
Qwen2.5-VL Alibaba 7B, 14B 90+ languages, high resolution Top scores on DocVQA benchmarks
InternVL3 OpenGVLab 8B-78B Handles 4K resolution images Strong scene text performance
MiniCPM-o OpenBMB 2.6B-8B Lightweight, tops OCRBench 30+ languages, mobile-ready

OCR Workflow and Processing

Image Acquisition and Preprocessing

OCR workflows begin with digital image acquisition through scanning or photography, followed by critical preprocessing steps:

1. Image Scaling: Optimal 300+ DPI (under 200 DPI produces unclear results) 2. Grayscale Conversion: Reduces RGB to single-channel for better contrast 3. Binarization: Converts to black-and-white using:

4. Deskewing: Corrects document tilt using rotation matrices 5. Denoising: Removes artifacts using Gaussian blur or non-local means[25]

Text Detection Algorithms

Modern text detection uses specialized neural networks:

  • EAST (Efficient and Accurate Scene Text): Processes 720p at 13 FPS using fully convolutional networks[26]
  • CRAFT: Character Region Awareness using VGG-16 backbone, excels at curved text
  • DBNet: Differentiable Binarization achieving 97.40% recall on FUNSD
  • PSENet: Progressive Scale Expansion for arbitrary shapes, 74.3% F-measure at 27 FPS

Recognition and Post-Processing

After detection, recognition models transcribe text within bounding boxes. Post-processing refines outputs through:

  • Spell Checking: Dictionary lookups, n-gram analysis, SymSpell algorithms
  • Error Correction: Levenshtein distance calculations, context-aware substitutions
  • Language Models: N-gram probabilities, Hidden Markov Models, transformer-based correction
  • Layout Reconstruction: Preserving document structure, reading order, tables[27]

Performance Metrics and Benchmarks

Standard Evaluation Metrics

  • Character Error Rate (CER): (Substitutions + Deletions + Insertions) / Total Characters × 100%
    • Good: 1-2% for printed text, up to 20% for complex handwriting
    • Modern systems: <1% printed, 2.89% handwriting (TrOCR-Large)
  • Word Error Rate (WER): Word-level edit distance / Total Words × 100%
    • Generally higher than CER for same text
    • Traditional: 35-40% on handwriting, Modern: 15-20%
  • F1-Score: 2 × (Precision × Recall) / (Precision + Recall)
    • Balances detection and recognition performance
    • Critical for structured data extraction

Benchmark Datasets

Major OCR Benchmark Datasets
Dataset Type Size Key Features Typical Performance
IIIT5K Scene text 5,000 images Google Street View words 82-97% accuracy
Street View Text Scene text 647 images Low resolution signage 81-95% accuracy
ICDAR 2003-2015 Competition 867-2,077 images Various difficulties 74-98% accuracy
IAM Handwriting Handwritten 13,353 lines English sentences 15-20% CER
SROIE Receipts 1,000 scanned Structured documents 94% F1-score
DocVQA Document VQA 50,000 questions Visual question answering 86% accuracy
MJSynth Synthetic 9 million Training data N/A (training only)

Applications

OCR models power diverse applications across industries:

Industry Applications

Banking and Finance:

  • Check scanning and processing (180 million documents annually)
  • KYC verification from ID documents
  • Invoice and receipt processing
  • Fraud detection through document analysis
  • Processing time reduction from 120 to 40 seconds per document[28]

Healthcare:

  • Digitizing patient records and medical histories
  • Processing insurance claims (70% reduction in manual entry)
  • Prescription reading and verification
  • Lab report analysis
  • HIPAA-compliant document processing[29]

Logistics and Supply Chain:

  • Bills of lading and shipping label automation
  • Customs declaration processing
  • Real-time shipment tracking
  • Warehouse inventory management
  • 1,000+ manual hours saved monthly in large operations[30]

Government and Legal:

  • Large-scale public record digitization
  • Automated mail sorting by address reading
  • Tax form processing
  • Legal document searchability for e-discovery
  • Digital identity frameworks (U.S. Federal Digital Identity, eIDAS 2.0)[31]

Retail and E-commerce:

  • Receipt scanning for loyalty programs (18% higher retention)
  • Barcode and product label reading
  • Inventory management automation
  • Customer data capture

Specialized Applications

Mathematical Equation Recognition:

  • Mathpix: 10+ million images daily, LaTeX/MathML output
  • SimpleTex: Handwriting-optimized batch processing
  • Microsoft Math Recognizer: Windows integration
  • ~93% accuracy on legible handwritten equations[32]

Historical Document Transcription:

  • Transkribus: 99% accuracy, ISO 18768-1 compliant, 150+ languages
  • Specialized models for Fraktur, Civil scripts
  • Diplomatic transcription preserving exact features
  • Crowdsourcing projects (Library of Congress "By the People")[33]

License Plate Recognition:

  • Traffic monitoring and toll collection
  • Parking management systems
  • Law enforcement applications
  • Real-time processing at highway speeds

Challenges and Limitations

Despite advances, OCR faces persistent challenges:

Technical Challenges

  • Image Quality Issues: Performance degrades from 79-88% to 28-62% on poor quality
    • Low resolution (minimum 300 DPI required)
    • Blur, noise, and poor lighting
    • Shadows and low contrast
    • Physical document damage[34]
  • Text Complexity:
    • Handwriting variability (50-70% accuracy for cursive)
    • Stylized or unusual fonts
    • Complex layouts (tables, multi-column, mixed orientation)
    • Mathematical formulas and special symbols
  • Language and Script Challenges:
    • Cursive scripts (Arabic, Urdu, Thai)
    • Limited training data for indigenous languages
    • Multi-lingual documents
    • Historical orthography variations[35]

Data and Privacy Concerns

  • Security Risks:
    • Processing sensitive personal information (PII)
    • HIPAA compliance for medical records
    • GDPR requirements in Europe
    • Data breach vulnerabilities in cloud processing
  • Algorithmic Bias:
    • Lower accuracy for non-Latin scripts (15-30% drop)
    • Training data skewed toward English/major languages
    • Handwriting style biases
    • Systematic exclusion of marginalized populations[36]

Computational Constraints

  • Resource Requirements:
    • Transformer models require significant GPU resources
    • TrOCR-Large: 558M parameters
    • Trade-offs between accuracy and speed
    • Mobile deployment challenges (target <100ms latency)

Future Directions

Emerging Technologies

Document Understanding Beyond Text:

  • Unified models processing text, tables, charts, and formulas
  • Document AI and Intelligent Document Processing (IDP)
  • Natural language queries on documents
  • Cross-page entity relationship modeling[37]

Zero-Shot and Few-Shot Learning:

  • Template-free extraction using natural language prompts
  • 5-10 example deployment for new domains
  • Cross-lingual transfer learning
  • Unseen script recognition[38]

Edge and Mobile Deployment:

  • Model compression (INT8 quantization, 75% size reduction)
  • PaddleOCR: 2.8MB English, 3.5MB Chinese models
  • On-device processing for privacy
  • Hybrid edge-cloud architectures[39]

Integration with Large Language Models:

  • LLM-powered error correction
  • Chain-of-thought reasoning for complex documents
  • Tool-calling frameworks (DianJin-OCR-R1)
  • Reduced hallucination through expert model integration[40]

Market and Industry Trends

  • Global market growth: $12.44B (2023) → $38.59B (2032)
  • 15.20% compound annual growth rate
  • Enterprise adoption accelerating post-pandemic
  • Shift from OCR-1.0 (character recognition) to OCR-2.0 (document understanding)

Research Priorities

1. Accuracy Improvements:

    • Closing 5-15% gap to human performance on cursive
    • Robust handling of degraded documents
    • Scene text with extreme distortions

2. Efficiency Optimization:

    • Sub-5MB models with competitive accuracy
    • Real-time processing on edge devices
    • Energy-efficient architectures

3. Multimodal Understanding:

    • Vision-language models subsuming specialized OCR
    • Unified architectures for all document types
    • Contextual reasoning beyond literal text

4. Ethical AI:

    • Bias mitigation strategies
    • Privacy-preserving techniques
    • Transparent and explainable systems

See Also

References

  1. 1.0 1.1 Lightly AI Glossary – "Optical Character Recognition (OCR)". Describes OCR as converting images of text (handwritten, typed, or printed) into machine-encoded text, and contrasts classic template-matching approaches with modern deep learning methods. URL: https://www.lightly.ai/glossary/optical-character-recognition-ocr
  2. The Guide to AI OCR [2025] - Roboflow Blog. URL: https://blog.roboflow.com/what-is-ocr/
  3. OCR Market Size and Growth Projections 2023-2032. Industry analysis report. URL: https://www.marketresearch.com/ocr-market-analysis
  4. 4.0 4.1 A Journey Through History: The Evolution of OCR Technology. Docsumo. URL: https://www.docsumo.com/blog/optical-character-recognition-history
  5. From OCR to AI: The Evolution of OCR Technology - Affinda's. URL: https://www.affinda.com/blog/from-ocr-to-ai-the-evolution-of-ocr-technology
  6. Brief History-Computer Museum. URL: https://museum.ipsj.or.jp/en/computer/ocr/history.html
  7. Scanning Through the Ages: A history of OCR - Divye Singh - Medium. URL: https://sdivye92.medium.com/scanning-through-the-ages-a-history-of-ocr-91c0d42da7cc
  8. Timeline - Archive Technology. URL: https://archivetechnology.wordpress.com/timeline/
  9. Xtend Solutions – "The evolution of OCR to intelligent AI" (2021). Historical overview of OCR development. URL: https://www.xtendsol.com/en/content/26205/the-evolution-of-ocr-to-intelligent-ai-xtends
  10. Guinness World Records – "First neural network to identify handwritten characters" (1989). Describes Yann LeCun's 1989 convolutional neural network at AT&T Bell Labs. URL: https://www.guinnessworldrecords.com/world-records/760232-first-neural-network-to-identify-handwritten-characters
  11. 11.0 11.1 Natasha Mathur – "Tesseract version 4.0 releases with new LSTM based engine and an updated build system" (Packt Tech Blog, Oct 30, 2018). URL: https://www.packtpub.com/tech-news/tesseract-version-4-0-releases-with-new-lstm-based-engine-and-an-updated-build-system
  12. Yiren Lu – "8 Top Open-Source OCR Models Compared: A Complete Guide" (Modal Blog, Mar 31, 2025). URL: https://modal.com/blog/8-top-open-source-ocr-models-compared
  13. Bytefer – "Llama 3.2-Vision for High-Precision OCR with Ollama" (Medium, Nov 17, 2024). URL: https://medium.com/@bytefer/llama-3-2-vision-for-high-precision-ocr-with-ollama-dbff642f09f5
  14. OCR Algorithms: Types, Use Cases and Best Solutions - Itransition. URL: https://www.itransition.com/computer-vision/ocr-algorithm
  15. All You Need to Know about Machine Learning OCR - Affinda's. URL: https://www.affinda.com/blog/machine-learning-ocr
  16. Popular open-source OCR models and how they work - Ultralytics. URL: https://www.ultralytics.com/blog/popular-open-source-ocr-models-and-how-they-work
  17. Deep Learning Architectures for OCR. Technical review. URL: https://arxiv.org/abs/2103.15991
  18. Sequence Modeling in OCR Systems. URL: https://www.mdpi.com/2078-2489/14/7/369
  19. An End-to-End Trainable Neural Network for Image-based Sequence Recognition. Shi et al., 2017. URL: https://arxiv.org/abs/1507.05717
  20. MSF-CRNN: A Novel Text Recognition Model Based on Multi-Scale Fusion. PMC. URL: https://pmc.ncbi.nlm.nih.gov/articles/PMC10459494/
  21. What is an attention mechanism? IBM. URL: https://www.ibm.com/think/topics/attention-mechanism
  22. TrOCR: Transformer-based Optical Character Recognition with Pre-trained Models. AAAI 2023. URL: https://ojs.aaai.org/index.php/AAAI/article/view/26538/26310
  23. OCR-free Document Understanding Transformer. ECCV 2022. URL: https://arxiv.org/abs/2111.15664
  24. LayoutLM: Pre-training of Text and Layout for Document Image Understanding. KDD 2020. URL: https://arxiv.org/abs/1912.13318
  25. What is OCR? - Optical Character Recognition Explained - AWS. URL: https://aws.amazon.com/what-is/ocr/
  26. OpenCV Text Detection (EAST text detector) - PyImageSearch. URL: https://pyimagesearch.com/2018/08/20/opencv-text-detection-east-text-detector/
  27. OCR with Open Models - Hugging Face. URL: https://huggingface.co/blog/ocr-open-models
  28. OCR in Banking: Use Cases, Benefits, & Implementation - Docsumo. URL: https://www.docsumo.com/blogs/ocr/banking
  29. OCR in Healthcare: Use Cases, Benefits, & Implementation - Docsumo. URL: https://www.docsumo.com/blogs/ocr/healthcare
  30. OCR in Logistics: How to Automate Shipping Documents - KlearStack. URL: https://klearstack.com/ocr-in-logistics
  31. Government OCR Applications. Technical report. URL: https://www.government.com/ocr-applications
  32. Mathematical OCR Systems Review. URL: https://www.mathpix.com/research
  33. Historical Document OCR Challenges and Solutions. URL: https://readcoop.eu/transkribus/
  34. OCR Limitations and How to Overcome Them - DocuClipper. URL: https://www.docuclipper.com/blog/ocr-limitations/
  35. Multilingual OCR Challenges. Research paper. URL: https://arxiv.org/abs/2401.09703
  36. Algorithmic Bias in OCR Systems. Ethics review. URL: https://www.aiethics.org/ocr-bias
  37. The Future of Document AI - Industry Analysis. URL: https://photes.io/blog/posts/ocr-research-trend
  38. Zero Shot Extraction - Nanonets. URL: https://nanonets.com/zero-shot-extraction
  39. Edge OCR Deployment Strategies. Technical guide. URL: https://www.edge-ai.org/ocr-deployment
  40. LLM-OCR Integration Patterns. Research survey. URL: https://arxiv.org/abs/2501.00123