Local-First AI Inference: A Cloud Architecture Pattern for Cost-Effective Document Processing
A three-tier hybrid architecture routes 70–80% of documents to local deterministic extraction, cutting Azure OpenAI costs by 75% and processing time by 55% on a 4,700-document workload.
InfoQ