Developer Protocol
Access our high-throughput Sovereign Intelligence network via a unified API gateway. Designed for enterprise-scale inference and hyper-minimal latency.
Quick Start Initialization
The VORTEX AI API follows an OpenAI-compatible architecture. You can drop our Intelligence Layer into any existing codebase with just two lines of configuration change.
Python SDK
import openai
client = openai.OpenAI(
api_key="VORTEX_ACCESS_KEY",
base_url="https://api.vortexaillm.com/v1"
)
response = client.chat.completions.create(
model="vortex-intelligence-pro",
messages=[{"role": "user", "content": "Analyze node health."}]
)
Intelligence Endpoints
All traffic is routed through our sovereign edge network. We support streaming responses and function calling across all primary model nodes.
POST
/v1/chat/completions
GET
/v1/models
POST
/v1/embeddings
Security & Rate Protocols
Free Tier access is restricted to 1,000 tokens per minute. For Enterprise throughput levels, please review your active contracts in the settings panel or contact support.