Most APIs serving AI agents have docs written for humans. That is backwards.
Agents do not read documentation pages. They need machine-readable specs to discover endpoints, understand parameters, and generate client code automatically.
The Problem
When an AI agent needs to integrate with your API, it either:
- Has a human developer manually write the integration
- Reads your docs page and tries to parse natural language into API calls
- Uses a structured spec (OpenAPI, MCP) to auto-discover everything
Option 3 is the only one that scales.
What We Shipped
We just added a full OpenAPI 3.1 spec to GateSolve (our CAPTCHA solving API for agents). It covers every endpoint:
- POST /api/solve -- Submit a CAPTCHA solve request
- GET /api/solve?id=X -- Poll for results
- GET /api/detect?url=X -- Classify URL access blocks
- POST /api/v1/solve/dry-run -- Validate requests without consuming credits
- GET /api/v1/usage -- Check API key quota
Each endpoint has typed request/response schemas, auth requirements, and error codes.
Why OpenAPI Over Custom Docs
- Code generation: Any OpenAPI-compatible tool can generate typed clients
- MCP compatibility: MCP servers can wrap OpenAPI specs directly
- Directory indexing: API directories can automatically catalog your endpoints
- IDE support: Developers get autocomplete and type checking for free
The spec IS the documentation for machines.
Try it: gatesolve.dev/openapi.json
Free API key: POST to gatesolve.dev/api/waitlist with your email. 100 free solves.
Top comments (0)