Skip to main content

Check out Port for yourselfย 

Port AI API Interaction

Closed Beta

Port's AI offerings are currently in closed beta and will be gradually rolled out to users by the end of 2025.

Port AI can be accessed programmatically through Port's API, enabling integration into custom applications and workflows. This provides the most flexible way to incorporate Port AI capabilities into your existing tools and processes.

API Endpointsโ€‹

Port AI provides streaming API endpoints for real-time interaction:

  • Port AI Assistant: /v1/ai/invoke - General-purpose AI interactions.
  • AI Agents: /v1/agent/<AGENT_IDENTIFIER>/invoke - Domain-specific agent interactions.

All interactions use streaming responses as Server-Sent Events (SSE) to provide real-time updates during execution. The response will be in text/event-stream format.

Interaction Processโ€‹

  1. Invoke Port AI
  2. The API will start sending Server-Sent Events
  3. Your client should process these events as they arrive, with each event providing information about the AI's progress or final response

Basic API Examplesโ€‹

Port AI Assistant:

curl 'https://api.port.io/v1/ai/invoke' \
-H 'Authorization: Bearer <YOUR_API_TOKEN>' \
-H 'Content-Type: application/json' \
--data-raw '{"prompt":"What services are failing health checks?"}'

AI Agents:

curl 'https://api.port.io/v1/agent/<AGENT_IDENTIFIER>/invoke' \
-H 'Authorization: Bearer <YOUR_API_TOKEN>' \
-H 'Content-Type: application/json' \
--data-raw '{"prompt":"Analyze the health of our production services"}'

With metadata labels:

curl 'https://api.port.io/v1/ai/invoke' \
-H 'Authorization: Bearer <YOUR_API_TOKEN>' \
-H 'Content-Type: application/json' \
--data-raw '{
"prompt":"What services are failing health checks?",
"tools": ["^(list|get|search)_.*"],
"labels": {
"source": "monitoring_system",
"environment": "production",
"triggered_by": "automated_check"
}
}

Streaming Response Formatโ€‹

The API responds with Content-Type: text/event-stream; charset=utf-8.

Each event in the stream has the following format:

event: <event_name>
data: <json_payload_or_string>

Note the blank line after data: ... which separates events.

Example Event Sequenceโ€‹

event: tool_call
data: { "id": "call_0", "name": "get_entities", "arguments": "{\"blueprint\":\"service\"}" }

event: tool_result
data: { "id": "call_0", "content": "Found 15 services in your catalog..." }

event: tool_call
data: { "id": "call_1", "name": "run_action", "arguments": "{\"actionIdentifier\":\"create_incident\"}" }

event: tool_result
data: { "id": "call_1", "content": "Action run created successfully with ID: run_12345" }

event: execution
data: I found 15 services in your catalog and created an incident report as requested.

event: done
data: {
"rateLimitUsage": {
"maxRequests": 200,
"remainingRequests": 193,
"maxTokens": 200000,
"remainingTokens": 179910,
"remainingTimeMs": 903
},
"monthlyQuotaUsage": {
"monthlyLimit": 20,
"remainingQuota": 19,
"month": "2025-09",
"remainingTimeMs": 1766899073
}
}

Event Typesโ€‹

tool_call (Click to expand)

Indicates that Port AI is about to execute a tool. This event provides details about the tool being called and its arguments. For large arguments, the data may be sent in multiple chunks.

{
"id": "call_0",
"name": "get_entities",
"arguments": "{\"blueprint\":\"service\",\"limit\":10}",
"lastChunk": true
}

Fields:

  • id: Unique identifier for this tool call.
  • name: Name of the tool being executed (only included in the first chunk).
  • arguments: JSON string containing the tool arguments (may be chunked for large payloads).
  • lastChunk: Boolean indicating if this is the final chunk for this tool call (optional, only present on the last chunk).
tool_result (Click to expand)

Contains the result of a tool execution. For large results, the data may be sent in multiple chunks.

{
"id": "call_0",
"content": "Found 15 services in your catalog: api-gateway, user-service, payment-service...",
"lastChunk": true
}

Fields:

  • id: Unique identifier matching the corresponding tool call.
  • content: The result content from the tool execution (may be chunked for large responses).
  • lastChunk: Boolean indicating if this is the final chunk for this tool result (optional, only present on the last chunk).
execution (Click to expand)

The final textual answer or a chunk of the answer from Port AI. For longer responses, multiple execution events might be sent.

done (Click to expand)

Signals that Port AI has finished processing and the response stream is complete. This event also includes quota usage information for managing your API limits.

{
"rateLimitUsage": {
"maxRequests": 200,
"remainingRequests": 193,
"maxTokens": 200000,
"remainingTokens": 179910,
"remainingTimeMs": 903
},
"monthlyQuotaUsage": {
"monthlyLimit": 20,
"remainingQuota": 19,
"month": "2025-09",
"remainingTimeMs": 1766899073
}
}

Quota Usage Fields:

  • maxRequests: Maximum number of requests allowed in the current rolling window.
  • remainingRequests: Number of requests remaining in the current window.
  • maxTokens: Maximum number of tokens allowed in the current rolling window.
  • remainingTokens: Number of tokens remaining in the current window.
  • remainingTimeMs: Time in milliseconds until the rolling window resets.

Processing Quota Informationโ€‹

Managing quota usage

Use the quota information in the done event to implement client-side rate limiting and avoid hitting API limits. When remainingRequests or remainingTokens are low, consider adding delays between requests or queuing them for later execution.

JavaScript Example: Processing Quota Information (Click to expand)

When processing the streaming response, you'll receive quota usage information in the final done event. Here's a JavaScript example of how to handle this:

const eventSource = new EventSource(apiUrl);

eventSource.addEventListener('done', (event) => {
const data = JSON.parse(event.data);

if (data.rateLimitUsage) {
const { remainingRequests, remainingTokens, remainingTimeMs } = data.rateLimitUsage;

// Check if quota is running low
if (remainingRequests < 10 || remainingTokens < 10000) {
console.warn('Quota running low, consider rate limiting');
// Implement rate limiting logic
}

// Schedule next request after quota reset if needed
if (remainingRequests === 0) {
setTimeout(() => {
// Safe to make next request
}, remainingTimeMs);
}
}

eventSource.close();
});

Rate Limits and Quotasโ€‹

Port AI operates with specific limits to ensure optimal performance for all users:

LLM Provider Limits

These limits apply when using Port's managed AI infrastructure. When you configure your own LLM provider, these Port-specific limits no longer apply, and usage will be governed by your provider's own limits and pricing.

Port acts as a bridge to leading LLM providers and doesn't host LLM models internally.

Rate Limits (Hourly)โ€‹

  • Request limit: 1,000 requests per hour.
  • Token usage limit: 800,000 tokens per hour.
  • These limits reset hourly.

Monthly Quotaโ€‹

  • Default quota: 20 AI invocations per month.
  • Each interaction with Port AI counts as one request against your quota.
  • Quota resets monthly.
Usage limits

Usage limits may change without prior notice. Once a limit is reached, you will need to wait until it resets.
If you attempt to interact with Port AI after reaching a limit, you will receive an error message indicating that the limit has been exceeded. The query limit is estimated and depends on the actual token usage.

Monitor your usageโ€‹

You can monitor your current usage in several ways:

Rate limitsโ€‹

  • Check the final done event in streaming responses for remaining requests, tokens, and reset time.

Monthly quotaโ€‹

You can monitor your current monthly quota usage using the Get monthly AI invocations quota usage API endpoint

Proactive quota monitoring

Check your monthly quota before making multiple Port AI requests to avoid hitting limits. When remainingQuota is low, consider implementing rate limiting or queuing requests until the monthly quota resets. Note that you may also encounter hourly rate limits, which are separate from this monthly quota.

Selecting Modelโ€‹

Port AI allows you to specify which LLM provider and model to use for specific API requests, giving you fine-grained control over AI processing on a per-request basis.

How LLM Providers Workโ€‹

Port AI supports multiple LLM providers and models. You can either use Port's managed AI infrastructure (default) or configure your own LLM providers for additional control over data privacy, costs, and compliance.

Learn more about LLM Provider Management and see the supported models and providers.

Specifying Provider and Modelโ€‹

When making API requests, you can include provider and model parameters (if none specified, your organization's default will be used). See the Invoke an agent API reference for detailed example.

Default Behaviorโ€‹

If no provider is specified in your API request, the system uses your organization's configured defaults, or falls back to Port's system defaults if none are configured.

Tool Selectionโ€‹

Port AI allows you to control which specific tools from the Port MCP server are available for each API interaction. This provides fine-grained control over what actions Port AI can perform, enabling you to create secure, purpose-specific AI interactions.

Permission-Based Tool Filtering

Selected tools will be available based on your regex patterns but won't include tools that are not within your permission scope. This means:

  • If you request an action to create a Jira ticket but this action is not available to you as a user, it won't be available to Port AI.
  • Members trying to use builder tools like "Create_blueprint" will not have access to these tools through Port AI if they lack the necessary permissions.
  • Tool availability is determined by the intersection of your regex selection AND your user permissions.

Port AI respects your individual user permissions and cannot access tools or perform actions that you don't have permission to use.

How Tool Selection Worksโ€‹

Include a tools parameter in your API request with an array of regex patterns. Port AI will only use tools whose names match at least one of these patterns.

Basic format:

{
"prompt": "Your question or request",
"tools": ["regex_pattern_1", "regex_pattern_2"]
}

Common Tool Selection Patternsโ€‹

Read-only Operations (Click to expand)

Perfect for monitoring dashboards and reporting systems where no modifications should be made.

["^(list|get|search|track|describe)_.*"]

What this matches:

  • get_entities, get_blueprint, get_scorecard.
  • list_entities, search_entities.
  • describe_user_details.
  • search_port_docs_sources, ask_port_docs.
Action Execution Only (Click to expand)

Allows only action execution tools while blocking data query operations.

["^run_.*"]

What this matches:

  • run_create_service, run_deploy_to_production.
  • run_github_create_issue, run_jira_create_ticket.
  • run_slack_notify_team.
Specific Integration Actions (Click to expand)

Target specific third-party service integrations.

["run_.*github.*", "run_.*jira.*", "run_.*zendesk.*"]

What this matches:

  • run_github_create_issue, run_github_merge_pr.
  • run_jira_create_ticket, run_jira_update_status.
  • run_zendesk_create_ticket.
Safe Entity Operations (Click to expand)

Enables entity operations while preventing accidental deletions.

["(?!delete_)\\w+_entity$", "get_.*", "list_.*"]

What this matches:

  • get_entity, list_entities, create_entity, update_entity.
  • Excludes: delete_entity.
Documentation and Help Tools (Click to expand)

Focus on documentation search and help functionality.

[".*docs.*", "search_.*", "ask_.*", "describe_.*"]

What this matches:

  • search_port_docs_sources, ask_port_docs.
  • describe_user_details.
Blueprint and Scorecard Analysis (Click to expand)

Focus on catalog structure and quality metrics without action execution.

[".*blueprint.*", ".*scorecard.*", "^(get|list)_.*"]

What this matches:

  • get_blueprints, get_blueprint.
  • get_scorecards, get_scorecard.
  • All get/list operations.

Interactive Tool Matcherโ€‹

Test your regex patterns to see which MCP tools would be available to Port AI. Enter your patterns in JSON array format (e.g., ["^(list|get)_.*", "run_.*github.*"]) and see the matching tools in real-time.

Enter an array of regex patterns in JSON format. Patterns automatically match from the beginning of tool names (^ is added automatically).

Tools (0 of 27 matched)

create_blueprintlist_blueprintsget_blueprintupdate_blueprintdelete_blueprintcreate_entityget_entitiesget_entityupdate_entitydelete_entitycreate_scorecardget_scorecardsget_scorecardupdate_scorecarddelete_scorecardcreate_actionlist_actionsget_actionupdate_actiondelete_actiontrack_action_runrun_*get_action_permissionsupdate_action_permissionsask_port_docssearch_port_docs_sourcesdescribe_user_details
Action Tools Note

Action tools (starting with run_*) depend on your Port configuration. The examples shown represent common action patterns, but your actual available actions may differ based on the self-service actions configured in your Port instance.

Best Practicesโ€‹

Security and Control
  • Principle of least privilege: Only include tools necessary for the specific use case.
  • Test patterns: Use the interactive matcher above to verify your regex patterns.
  • Automated systems: Use highly restrictive patterns for automated workflows.
  • User-facing interfaces: Consider broader patterns for interactive use cases.

Integration Patternsโ€‹

Direct API Calls

Integrate Port AI directly into your applications using HTTP requests:

# Basic Port AI request
curl 'https://api.port.io/v1/ai/invoke' \
-H 'Authorization: Bearer <YOUR_API_TOKEN>' \
-H 'Content-Type: application/json' \
--data-raw '{
"prompt": "What services are failing health checks?",
"tools": ["^(list|get|search)_.*"],
"labels": {
"source": "monitoring_system",
"check_type": "health_analysis"
}'

# AI Agent request
curl 'https://api.port.io/v1/agent/<AGENT_IDENTIFIER>/invoke' \
-H 'Authorization: Bearer <YOUR_API_TOKEN>' \
-H 'Content-Type: application/json' \
--data-raw '{
"prompt": "Analyze the health of our production services",
"labels": {
"source": "monitoring_dashboard",
"environment": "production"
}'

Application Integration Example

// Example: Monitoring dashboard integration
async function checkServiceHealth(serviceName) {
const response = await fetch('/api/port-ai/check-service', {
method: 'POST',
headers: { 'Content-Type': 'application/json' },
body: JSON.stringify({
prompt: `Analyze the health of service ${serviceName}`,
tools: ['^(list|get|search)_.*'],
labels: {
source: 'monitoring_dashboard',
service: serviceName,
check_type: 'health_analysis'
}
})
});

// Process streaming response
const reader = response.body.getReader();
// Handle SSE parsing...
}

Error Handlingโ€‹

Common error scenarios and handling strategies:

Rate Limit Exceededโ€‹

{
"error": "Rate limit exceeded",
"type": "RATE_LIMIT_ERROR",
"retryAfter": 3600
}

Quota Exceededโ€‹

{
"error": "Monthly quota exceeded",
"type": "QUOTA_ERROR",
"resetDate": "2025-10-01T00:00:00Z"
}
Implementation Example: Error Handling (Click to expand)
async function handlePortAIRequest(prompt) {
try {
const response = await invokePortAI(prompt);
return response;
} catch (error) {
if (error.type === 'RATE_LIMIT_ERROR') {
// Wait and retry
await new Promise(resolve => setTimeout(resolve, error.retryAfter * 1000));
return handlePortAIRequest(prompt);
} else if (error.type === 'QUOTA_ERROR') {
// Queue for next month or upgrade plan
console.log('Monthly quota exceeded, queuing request');
return null;
}
throw error;
}
}

Security Considerationsโ€‹

When integrating Port AI via API:

  • Authentication: Always use secure API token storage and rotation.
  • Data privacy: Port AI respects your organization's RBAC and data access policies.
  • Audit trail: All API interactions are logged and trackable.
  • Rate limiting: Implement client-side rate limiting to avoid hitting API limits.

For comprehensive security information, see AI Security and Data Controls.