# Vision Models

## Overview

Our API enables you to use machine learning models for tasks that require visual capabilities. These models are referred to as *vision models*. Within our API, we offer two categories of vision models: **OCR** and **OFR**.

### OCR: Optical Character Recognition

With OCR technology, you can analyze any document and extract text as well as other characters and symbols. This allows you to detect:

* Text
* Paragraph blocks
* Handwriting
* Text inside PDF/TIFF files

{% content-ref url="vision-models/ocr-optical-character-recognition" %}
[ocr-optical-character-recognition](https://docs.aimlapi.com/api-references/vision-models/ocr-optical-character-recognition)
{% endcontent-ref %}

### OFR: Optical Feature Recognition

In contrast to OCR, OFR allows you to analyze not just documents but also images. You can filter exactly what you want to find in the image by the features they include:

* Crop hints
* Faces
* Image properties
* Labels
* Landmarks
* Logos
* Multiple objects
* Explicit content
* Web entities and pages
* And many more

{% content-ref url="vision-models/ofr-optical-feature-recognition" %}
[ofr-optical-feature-recognition](https://docs.aimlapi.com/api-references/vision-models/ofr-optical-feature-recognition)
{% endcontent-ref %}


---

# Agent Instructions: Querying This Documentation

If you need additional information that is not directly available in this page, you can query the documentation dynamically by asking a question.

Perform an HTTP GET request on the current page URL with the `ask` query parameter:

```
GET https://docs.aimlapi.com/api-references/vision-models.md?ask=<question>
```

The question should be specific, self-contained, and written in natural language.
The response will contain a direct answer to the question and relevant excerpts and sources from the documentation.

Use this mechanism when the answer is not explicitly present in the current page, you need clarification or additional context, or you want to retrieve related documentation sections.
