Skip to content

Deploy via OpenAPI Import

Thousands of APIs already publish an OpenAPI or Swagger specification — payment processors, CRMs, project management tools, cloud infrastructure, monitoring platforms, communication services. Every major SaaS product has one.

Vinkius Cloud turns any of them into a fully functional MCP server in under 30 seconds. Paste the spec URL, and the platform auto-discovers every endpoint, generates typed MCP tools with semantic annotations, configures authentication, and deploys to the global edge. No code to write, no SDK to install, no infrastructure to manage.

That means every API that already has a spec is one click away from being callable by AI models.

83%
API SPEC ADOPTION RATE
#1
API STANDARD WORLDWIDE
< 30s
SPEC TO MCP SERVER

Why this changes everything

Before Vinkius Cloud, connecting an AI model to a REST API required writing a custom MCP server — defining tools by hand, mapping parameters, handling authentication, implementing error handling, and deploying infrastructure. For a 20-endpoint API, that could mean days of work.

With OpenAPI Import, the entire surface area of an existing API becomes available to AI in seconds. The spec already describes every endpoint, every parameter type, every response shape. Vinkius Cloud reads that machine-readable contract and transforms it into MCP tools automatically — complete with:

  • Typed input schemas derived from your spec's parameter definitions
  • Semantic annotations (read-only, destructive, idempotent) inferred from HTTP verbs
  • Authentication injection configured once, applied to every upstream request
  • DLP and FinOps governance active from the first tool call

The result: any API that publishes a spec is instantly AI-ready. No glue code. No custom server. No deployment pipeline.


THE WIZARD
One URL in,
every tool out.
Paste your spec URL and we do the rest — endpoints discovered, tools generated, annotations mapped. All on the fly.

The deployment wizard is a split-screen experience. On the left, a contextual sidebar explains what's happening at each step. On the right, a minimal form collects only what's needed. Three steps — or two, if your API is public.

Step 1 — Connection

02 — CONNECTION
One URL.
Every tool.
Paste your spec URL and we do the rest — endpoints discovered, tools generated, annotations mapped. All on the fly.
OpenAPI 2.0, 3.0, 3.1. JSON or YAML. We handle it.
VINKIUS CLOUD SETUPOpenAPI
OPENAPI / SWAGGER URL
https://api.company.com/openapi.json
24 endpoints ready for AI
SERVER NAME
Production ERP
Auto-filled from spec
BASE URL
Optional
https://api.company.com/v1
Auto-detected from the OpenAPI spec if left empty.

Paste your OpenAPI / Swagger spec URL. The platform fetches it live and detects:

  • Endpoint count — displayed instantly as "24 endpoints ready for AI"
  • Server name — auto-filled from info.title
  • Base URL — auto-filled from servers[0].url, override if needed
  • Auth type — detected from the security block and suggested in the next step

Supported formats

OpenAPI 2.0, 3.0, 3.1. JSON or YAML. External $ref references are resolved recursively.

Step 2 — Security

03 — SECURITY
Credentials never
leave the vault.
AI agents call your API through us. Keys are AES-256 encrypted at rest, injected at runtime, invisible to the LLM.
Public API? Skip this — no auth required.
🔒 AES-256 VAULT
AUTHENTICATION VAULT
AUTH TYPE
Bearer Token ▾
TOKEN
••••••••••••••••••••
AES-256 encrypted at rest. Credentials never leave your vault.

If the spec declares a security block, the wizard detects and suggests the right authentication type. You choose:

Auth typeWhat to provide
None — Public APINothing. Step is skipped entirely.
Bearer TokenYour API token or access key
Basic AuthUsername and password
Custom HeaderHeader name (e.g., X-API-Key) and value

Credentials are AES-256 encrypted at rest, decrypted in-memory only when an AI model invokes a tool, and immediately discarded after the upstream call. They never appear in logs, dashboards, or API responses — not even to team admins.

Public APIs

If no security block is present in the spec and you leave auth as "None", this entire step is auto-skipped. The wizard jumps directly to Governance.

Step 3 — Governance

04 — GOVERNANCE
Invisible
armor.
PII is masked before it ever touches the model. Large payloads shrink automatically — fewer tokens, same intelligence.
Both optional. Enable now or toggle later from settings.
RAW
{ "email": "user@org.com" }
REDACTED
{ "email": "u***@***.com" }
GOVERNANCE CONTROLS
Zero-Trust PII Redaction
Masks emails, SSNs, and credit cards in RAM before returning data to the AI. GDPR-compliant by default.
Smart Array Truncation
Large arrays shrink automatically. Same data, fewer tokens — your context window stays focused.

Both controls are enabled by default and toggle on/off with a single switch:

  • Zero-Trust PII Redaction — Scans every API response for emails, SSNs, credit cards, and personal identifiers. Masks them in-memory before the data reaches any AI model. GDPR-compliant by default.
  • Smart Array Truncation — When an API returns arrays with hundreds or thousands of items, the truncation engine shrinks them automatically. The AI gets the same intelligence with fewer tokens. Cost control without losing context.

DEPLOY
Hit Deploy.
Watch it go live.

What happens when you deploy

Click Deploy to Edge. In real-time, the wizard shows each phase:

DEPLOYING...
Parsing OpenAPI specs…
Sealing V8 Isolate Sandbox…
Injecting Zero-Trust DLP…
Distributing to Global Edge…

Each phase runs sequentially in the background:

  1. Parsing OpenAPI specs — The platform parses your spec, extracts every path/verb combination, and builds typed input schemas from parameter definitions.
  2. Sealing V8 Isolate Sandbox — Your server is sealed inside a V8 Isolate — no filesystem, no network escape, no process injection. The sandbox enforces strict memory and execution limits.
  3. Injecting Zero-Trust DLP — The PII redaction engine and array truncation rules are compiled into the request pipeline. Every tool call passes through governance controls before data reaches the AI.
  4. Distributing to Global Edge — The sealed bundle is distributed to edge nodes worldwide. Your MCP server is live and reachable from any AI client within seconds.

After deployment: you're live

When deployment completes, the wizard transitions to the success screen. You receive three outputs:

MCP URL
https://edge.vinkius.com/tok_abc123/mcp
The SSE endpoint. Paste it into any MCP client.
CONFIG JSON
{ "mcpServers": { ... } }
Download or copy — ready for Claude, Cursor, VS Code.
CONNECTION TOKEN
vk_••••••••••••
HMAC-SHA256 authenticated. Shown once — save it now.

Connection tokens

Every deployed server is protected by connection tokens — HMAC-SHA256 authenticated credentials that AI clients use to establish SSE connections.

  • Tokens are generated on deploy and shown exactly once. The plaintext is never stored on the server side — only the HMAC hash is persisted.
  • You can generate additional tokens from the server's Connection Tokens page, each labeled by client (e.g., "Cursor — Production", "Claude — Staging").
  • Every token can be revoked individually without affecting other connected clients.

Kill switch: 40ms global revocation

When an AI model has access to your production APIs, the ability to cut that access instantly is a security requirement. Vinkius Cloud revokes any connection token globally in 40 milliseconds.

!
One click. Any token. Globally killed in 40 milliseconds.
No blast radius, no lingering access, no exceptions.
When you revoke a token, it's invalidated across every edge node within 40ms. Active SSE connections using that token are terminated immediately — the AI model loses access mid-sentence if needed. No grace period, no retry window, no cache to expire.

Why 40ms matters

An AI model with a compromised token can make dozens of tool calls per second — accessing production data, creating records, or triggering destructive actions. In enterprise environments, the window between compromise and revocation determines the blast radius. At 40ms, that window is effectively closed before the next tool call completes.

Post-deploy security architecture

Your deployed MCP server runs inside a multi-layered security perimeter:

V8 ISOLATE
Sandbox isolation
No filesystem, no network escape, no process injection. Each server runs in its own V8 Isolate with strict memory and CPU limits.
HMAC LOCKFILE
Signed deployments
Every deploy is HMAC-signed with a lockfile. If the API surface changes unexpectedly, the deploy stops automatically — no tampered code reaches production.
EGRESS FIREWALL
SSRF protection
Outbound requests are locked to the declared base URL. The server cannot reach internal networks, metadata services, or any host not in the allowlist.
AUDIT LOG
Full observability
Every request, every response, every tool call — logged immutably. You don't wonder what AI did. You know. Available from the server's Audit Log tab.

AFTER YOU DEPLOY
What you get.
And what you can change.

What you get after deploy

Your OpenAPI spec is transformed into a fully functional MCP server. Every endpoint becomes a tool that AI models can discover, understand, and call. But the platform doesn't just convert — it solves three problems that every team deploying AI on top of APIs will face.

Problem 1 — Tool fatigue

When you expose a large API (50, 100, 200+ endpoints) to an AI model, every tool definition consumes tokens from the model's context window. Research shows that AI performance degrades significantly after just 10–15 tools — the model starts choosing wrong tools, hallucinating parameters, or entering retry loops.

How Vinkius Cloud solves this:

AUTO-GROUPING
Related endpoints share one tool
When your API exceeds the threshold, related endpoints are automatically grouped by OpenAPI tag. Instead of 50 individual tools flooding the context window, the AI sees 8 grouped tools with an action discriminator.
TOON COMPRESSION
30–50% fewer tokens per tool
TOON compresses tool descriptions into a compact tabular format. The AI model reads the same information in fewer tokens. Works best with GPT-4, Claude 3.5+, and Gemini 2.0+.

Problem 2 — Context window overflow

When an API returns arrays with hundreds or thousands of items — a list of all customers, all orders, all products — the raw response can consume the AI's entire context window in a single tool call. The model loses the ability to process the data, and the conversation stalls.

How Vinkius Cloud solves this:

Smart Array Truncation shrinks oversized arrays automatically. A /customers endpoint that returns 2,000 records is truncated to 50 items. The AI gets the same intelligence in a fraction of the tokens. You control the threshold — globally or per server.

Problem 3 — Safe vs. destructive operations

AI models don't inherently know which API calls are safe and which aren't. A GET /orders is read-only; a DELETE /orders/{id} is destructive. Without annotations, the model treats both the same — which can lead to unintended data changes.

How Vinkius Cloud solves this:

Every generated tool receives MCP annotations inferred from the HTTP verb:

GET
Read-only
AI calls freely without confirmation
POST
Non-idempotent
May prompt for confirmation
PUT
Idempotent
Safe to retry automatically
DEL
Destructive
Requires explicit user approval

What you can customize after deploy

Your server is live, but it's not locked. Everything generated automatically can be refined from the dashboard:

TOOL EDITOR
Edit any generated tool
Override descriptions, change display titles, adjust annotations. Useful when a POST is actually idempotent (e.g., a search endpoint), or when a GET has side effects.
SYSTEM RULES
Guide how AI interprets responses
Inject instructions into tool responses — "Always respond in Portuguese", "Use tables for lists", "Amounts in cents, display as currency". The AI follows these rules when presenting data.
TOGGLE ON/OFF
Disable individual tools
Don't want AI to access certain endpoints? Toggle them off. The tool is removed from the MCP discovery response without deleting it — re-enable it anytime.
CACHE CONTROL
Tell AI when data is stale
Set cache hints per tool — "no-store" for volatile data the AI should always re-fetch, or "immutable" for reference data safe to cache forever.

Real-world use cases

PAYMENTS
Payment processors
Stripe, Adyen, PayPal — import the OpenAPI spec and let AI query charges, refunds, and subscriptions directly. DLP redacts credit card numbers automatically.
CRM / ERP
CRMs and ERPs
Salesforce, HubSpot, SAP — large APIs with 200+ endpoints. Auto-grouping reduces tool count from 200+ to manageable groups. AI navigates the entire CRM naturally.
INFRASTRUCTURE
Cloud infrastructure
AWS, Cloudflare, DigitalOcean — import the spec, disable destructive tools (DELETE), and let AI query infrastructure status, logs, and metrics safely.

Next steps


Frequently Asked Questions

Which OpenAPI versions does Vinkius Cloud support?

Vinkius Cloud supports OpenAPI 2.0 (Swagger), OpenAPI 3.0, and OpenAPI 3.1 specifications in both JSON and YAML formats. You can paste a public URL to the spec or upload the file directly.

How does the platform convert OpenAPI endpoints into MCP tools?

The platform parses every path and operation in your spec, generates typed MCP tools with names, descriptions, and input schemas derived from your OpenAPI parameters and request bodies, and applies semantic annotations (read-only, idempotent, or destructive) based on the HTTP verb.

Can I control which endpoints become MCP tools?

Yes. After import, all tools are enabled by default, but you can toggle individual tools on or off from the server detail page. You can also edit their descriptions, rename them, or adjust input schemas — without modifying the original spec.

What authentication methods are supported for the upstream API?

Vinkius Cloud supports Bearer Token, Basic Auth, Custom Header, and None (public APIs). Credentials are AES-256 encrypted at rest and decrypted only in-memory at the moment the AI calls the tool.

Do I need to host the OpenAPI spec myself?

No. You can use any publicly accessible URL — GitHub raw file links, Swagger UI endpoints, API documentation pages that serve the spec. The platform fetches the spec once at import time.

What happens if I update my API specification later?

You can re-import the spec at any time from the server detail page. Vinkius Cloud will detect new, modified, and removed endpoints and update your MCP tools accordingly — no need to create a new server.