Baidu

paddleocr-vl-1.5

PaddleOCR-VL-1.5

OCR-specialized visual model for forms, receipts, documents, and extraction-heavy pipelines.

FREEPublic model detailMultimodal Transformer

Params

1B OCR

Context

Document OCR

Max Output

Structured OCR

License

Open

TTFT

N/A

No 5m benchmark sample

5m RPM

N/A

Why pick it

  • Good fit for document AI
  • Free public route

Pricing

TierStandardCachedSiliconFlowSavings
Realtime$0.00 / $0.00N/AN/AN/A
Batch$0.000 / $0.000N/AN/AN/A
Free model.

Production pricing proof

How this route settles on a real request

When the model executes live, response headers can expose X-BatchIn-Provider, X-BatchIn-Route-Reason, X-BatchIn-Effective-Cost-Cents, and X-BatchIn-Uncached-Cost-Cents.
The cached price here means prompt-cache discount on input tokens. Durable response-cache hits are proven separately through X-BatchIn-Response-Cache-Mode and request lookup.
Streaming calls start with X-Request-Id, then resolve final cost, cache mode, and route truth through lookup after completion.

Route source of truth

See pricing, request proof, and the upgrade path on one page

Standard, prompt-cache, batch, and SiliconFlow comparison stay visible without leaving the route.
Real requests return X-Request-Id, and buffered calls can expose route reason, billed cost, and uncached cost directly.
BatchIn supports Playground validation first, then batch, white-label, or dedicated capacity conversations.

Quick start

OpenAI-compatible surface. Swap the base URL and ship.

Python
from openai import OpenAI

client = OpenAI(
    base_url="https://api.luminapath.tech/v1",
    api_key="BATCHIN_API_KEY"
)

resp = client.chat.completions.create(
    model="paddleocr-vl-1.5",
    messages=[{"role": "user", "content": "Summarize why this model is a fit for my workload."}]
)

print(resp.choices[0].message.content)
JavaScript
import OpenAI from "openai";

const client = new OpenAI({
  baseURL: "https://api.luminapath.tech/v1",
  apiKey: process.env.BATCHIN_API_KEY,
});

const resp = await client.chat.completions.create({
  model: "paddleocr-vl-1.5",
  messages: [{ role: "user", content: "Summarize why this model is a fit for my workload." }],
});

console.log(resp.choices[0]?.message?.content);
cURL
curl https://api.luminapath.tech/v1/chat/completions \
  -H "Authorization: Bearer $BATCHIN_API_KEY" \
  -H "Content-Type: application/json" \
  -d '{
    "model": "paddleocr-vl-1.5",
    "messages": [{"role":"user","content":"Summarize why this model is a fit for my workload."}]
  }'

Specs

Architecture

Multimodal Transformer

Vendor group

Baidu

Context window

Document OCR

Max output

Structured OCR

Best for

ocr
documents

Related models

Back to model center

MiniMax

minimax-m2.5

MiniMax M2.5

Multimodal flagship for text, images, and agent-style reasoning with very good coding scores.

View detail

Z.ai

glm-5.1

GLM-5.1

Open-source coding flagship built for long-horizon autonomous engineering and deep reasoning.

View detail

Z.ai

glm-5

GLM-5

Lower-cost GLM route for production reasoning, agents, and long-context workflows.

View detail

Z.ai

glm-4.7

GLM-4.7

Mid-tier GLM reasoning route for engineering teams that need quality below flagship spend.

View detail
AI Assistant