# Confidence Score and Accuracy Boost

## Overview

The **Automation** feature in Mindee's platform represents a major step forward in enhancing both the **accuracy** and **reliability** of document data extraction. Designed to support robust and scalable automation workflows, this feature is built on two core capabilities:

1. **Enhanced accuracy** using model ensemble algorithms
2. **Confidence scoring** for all types of extracted fields

Automation aims to solve two common challenges in intelligent document processing:

* **Maximizing extraction quality** in the face of variable and noisy document formats
* **Providing actionable trust signals** so that systems can handle uncertain extractions appropriately

By combining multiple models and analyzing their agreement, Automation ensures that the most reliable prediction is selected for each field, while transparently communicating how confident the system is in that prediction.

## Use Cases

### Full Automation

By leveraging **confidence score thresholds**, you can selectively **automate decisions** in your processing pipeline, triggering downstream actions only when extractions meet a predefined reliability level.

For example, fields marked with a `High` or `Certain` confidence score can be automatically approved and pushed to your ERP or CRM system, while extractions with `Low` or `Medium` confidence can be routed for human review or fallback logic. This selective gating mechanism allows teams to implement **fully automated flows** for clean, predictable documents, while still handling edge cases gracefully.

#### **Use cases examples**:

* Auto-validating invoice totals and tax fields before ingestion into an accounting system
* Auto-approving identity document extractions for KYC when confidence is high
* Automatically flagging low-confidence vendor names or dates for manual verification

### Efficient human validation

To make confidence levels easily understood by end-users, each confidence score returned by Automation can be associated with a **color-coded indicator**.\
This visual feedback is especially useful in UI-driven workflows, where operators need to scan, validate, or correct extractions quickly.

The default color scheme is as follows:

| Confidence Level | Label     | Color Code | Suggested Action       | Description                                                              |
| ---------------- | --------- | ---------- | ---------------------- | ------------------------------------------------------------------------ |
| 🟦 Certain       | `Certain` | Blue       | Safe for  automation   | Full confidence, human-level precision                                   |
| 🟩 High          | `High`    | Green      | Can be auto-processed  | Model consensus is strong; prediction is likely accurate.                |
| 🟧 Medium        | `Medium`  | Orange     | Optional review        | Some confidence, but context or format may impact correctness.           |
| 🟥 Low           | `Low`     | Red        | Manual review required | Extraction is uncertain or likely incorrect. Model disagreement is high. |

This color-coding system allows product teams to **highlight uncertainty directly in the user interface**, enabling faster decisions, reducing cognitive load, and streamlining exception handling.

## Activate Confidence Scores

{% hint style="warning" icon="money-check-dollar-pen" %}
This feature is not available on all plans, check the [Plans](/account-management/plans.md#feature-comparison) section for more information.
{% endhint %}

For best results, be sure to follow the [Data Schema Best Practices](/extraction-models/data-schema-best-practices.md) before activating this feature.

If the Data Schema is not optimized, this feature may not significantly improve model accuracy.

### Activate Confidence Scores on the Platform

When setting the activation state of a feature on the Platform, this will be the default.\
All API calls will use the default state unless explicitly set otherwise during the API call.

This is useful for project managers, as it allows activating or deactivating a feature across all API calls on that model.

Anyone with write access to the model can set the option's default value.

{% @supademo/embed url="<https://app.supademo.com/demo/cmeie3irw9fe7h3pytuktflxs>" demoId="cmeie3irw9fe7h3pytuktflxs" %}

### Activate Confidence Scores via API Calls

If you need finer-grained control over when the feature is used, you can activate or deactivate it when making API calls.

This allows dynamically setting the activation state using your internal business or domain logic.

Check the [Configure the Client](/integrations/client-libraries-sdk/configure-the-client.md#optional-features-configuration) section if using our [Client Libraries / SDKs](/integrations/client-libraries-sdk.md).

{% hint style="info" %}
When the **Automation** feature is not activated, the `confidence` attribute in the response will always be `null`.
{% endhint %}

### Using Confidence Scores in Processing

You can easily add various business and/or processing logic rules in your code to handle complex workflows.

Take a look at the [Process Result Fields](/integrations/client-libraries-sdk/process-result-fields.md#confidence) section for implementation details.

## Towards 100% Automation

By combining confidence-based automation with Mindee’s **RAG-powered continuous learning loop**, you can drive your workflows toward **near 100% automation**.

Low-confidence extractions are not only flagged for human validation, but also used as feedback signals to refine models dynamically, through retrieval-augmented generation and targeted retraining.

This creates a virtuous cycle where every uncertain case contributes to future accuracy improvements, progressively reducing manual intervention and expanding the scope of trusted predictions.

## Frequently Asked Questions

### How is the confidence score computed?

The confidence score in Automation is a consensus-based reliability measure, not a simple probability. It is computed by analyzing the level of agreement between multiple models, each trained independently or with complementary strategies, on the same document field.

When these models produce matching or highly similar predictions, the confidence is high. When they disagree significantly, the confidence drops. On top of that, a dedicated arbitration and correction model acts as a referee: it takes all predictions, compares their structural and semantic coherence, and assigns a final confidence level (`low`, `medium`, `high`, or soon `certain`).

### Does Automation introduce additional latency?

Yes, Automation introduces some additional latency, but in most cases, it remains minimal. This is because the ensemble of models used for prediction is executed in parallel, which allows us to keep response times close to those of a single-model pipeline.

However, depending on the number and complexity of models involved, or the document type, the latency can occasionally be a few times longer than a standard call. The tradeoff is intentional: slightly longer processing time in exchange for higher accuracy and richer metadata, including the confidence score.

### What should I do with low confidence extractions?

We recommend routing `Medium` and lower confidence extractions to a human validation layer, or using fallback logic (e.g., default values, user input).

Lower confidence extractions are ideal candidates for feedback-driven improvement via our continuous learning loop using the RAG feature.

### Does Automation work with any type of documents or fields?

Automation is fully compatible with all document types and extracted fields supported by Mindee.\
Every extracted field, whether it's a piece of text, a number, a date, an amount, or any other data type, benefits from the same ensemble evaluation and confidence scoring logic. This consistent approach ensures a uniform and predictable developer experience, regardless of the document format or use case.

Moreover, nested objects and arrays of objects (e.g., `line_items` in invoices or tables in receipts) also receive individual confidence scores per field, enabling fine-grained control over complex data structures.


---

# Agent Instructions: Querying This Documentation

If you need additional information that is not directly available in this page, you can query the documentation dynamically by asking a question.

Perform an HTTP GET request on the current page URL with the `ask` query parameter:

```
GET https://docs.mindee.com/extraction-models/optional-features/automation-confidence-score.md?ask=<question>
```

The question should be specific, self-contained, and written in natural language.
The response will contain a direct answer to the question and relevant excerpts and sources from the documentation.

Use this mechanism when the answer is not explicitly present in the current page, you need clarification or additional context, or you want to retrieve related documentation sections.
