Integrations
Integrate Provisional into your own application. These examples show how to connect the MCP server to your system so your LLM can discover and call all 18 legal research tools automatically.
Setup
- 1
Install an LLM SDK that supports remote MCP servers.
- 2
Create an API key in your Provisional dashboard.
- 3
Include Provisional in your tools and start calling.
Usage
npm install @anthropic-ai/sdkimport Anthropic from "@anthropic-ai/sdk";
const client = new Anthropic();
const response = await client.beta.messages.create({
model: "claude-sonnet-4-6",
max_tokens: 4096,
mcp_servers: [
{
type: "url",
url: "https://mcp.provision.al/mcp",
name: "provisional",
authorization_token: "YOUR_PROVISIONAL_API_KEY",
},
],
tools: [{ type: "mcp_toolset", mcp_server_name: "provisional" }],
messages: [
{ role: "user", content: "Find EU regulations about the right to erasure" },
],
betas: ["mcp-client-2025-11-20"],
});
console.log(response);Authentication
| Header | Format | Notes |
|---|---|---|
Authorization | ApiKey sk-... | Recommended for SDK usage |
x-api-key | sk-... | Alternative header |
Authorization | Bearer eyJ... | OAuth tokens (for browser-based flows) |
Do not send API keys as Authorization: Bearer sk-... — the server rejects that format.
Good to know
Sessions are stored in memory. A deploy or restart invalidates every session — the SDK will need to reconnect. The server returns a 400 with invalid_session_id when this happens.
Tool responses are plain text formatted for LLM consumption, not JSON. Search results include pagination hints embedded in the text (e.g., "Use offset: 20 to see more").
The MCP SDKs also support Java, C#, Kotlin, and Swift. The connection pattern is the same — create a transport with the URL and API key header, then connect.