Deploy MCP Server
AI/ML Platform Bearer Token

IOX Cloud REST API

Enterprise AI infrastructure and model orchestration API

IOX Cloud provides a unified REST API for deploying, managing, and orchestrating AI models across multi-cloud infrastructure. Developers use it to abstract away infrastructure complexity, scale ML workloads dynamically, and integrate multiple model providers through a single interface. The platform handles model versioning, A/B testing, traffic routing, and observability for production AI applications.

Base URL https://api.iox.cloud/v1

API Endpoints

MethodEndpointDescription
GET/modelsList all deployed models with metadata, status, and performance metrics
POST/modelsDeploy a new model to the IOX Cloud infrastructure
GET/models/{modelId}Retrieve detailed information about a specific model deployment
PATCH/models/{modelId}Update model configuration, scaling rules, or deployment settings
DELETE/models/{modelId}Undeploy and remove a model from the infrastructure
POST/models/{modelId}/inferenceExecute inference requests against a deployed model
POST/models/{modelId}/batchSubmit batch inference jobs for asynchronous processing
GET/models/{modelId}/metricsRetrieve real-time and historical performance metrics for a model
POST/experimentsCreate A/B testing experiments to compare model versions
GET/experiments/{experimentId}Get experiment details, traffic distribution, and performance results
GET/deploymentsList all active deployments across regions and environments
POST/deployments/{deploymentId}/scaleManually scale deployment resources or update autoscaling policies
GET/infrastructure/healthCheck infrastructure health status and resource availability
GET/logsQuery model inference logs and system events with filtering
POST/webhooksConfigure webhooks for deployment events, alerts, and metric thresholds

Code Examples

curl -X POST https://api.iox.cloud/v1/models/gpt-4-turbo/inference \
  -H 'Authorization: Bearer iox_sk_abc123xyz' \
  -H 'Content-Type: application/json' \
  -d '{
    "input": "Explain quantum computing",
    "max_tokens": 500,
    "temperature": 0.7
  }'

Connect IOX Cloud to AI

Deploy a IOX Cloud MCP server on IOX Cloud and connect it to Claude, ChatGPT, Cursor, or any AI client. Your AI assistant gets direct access to IOX Cloud through these tools:

deploy_model Deploy a new AI model to IOX Cloud infrastructure with specified configuration and scaling parameters
run_inference Execute inference requests against deployed models and return predictions or generated content
create_experiment Set up A/B testing experiments to compare different model versions or configurations
monitor_metrics Retrieve and analyze model performance metrics including latency, throughput, and accuracy
manage_scaling Adjust model deployment scaling policies and resource allocation based on traffic patterns

Deploy in 60 seconds

Describe what you need, AI generates the code, and IOX deploys it globally.

Deploy IOX Cloud MCP Server →

Related APIs