Docs
cortecs.aiDedicated ModelsServerless ModelsLogin
  • Introduction
  • DEDICATED INFERENCE
    • Quickstart
    • Provisioning API
      • Authentication
      • User
      • Instances
      • Models
      • Hardware Types
    • Python client
      • Objects
      • Langchain integration
    • Examples
      • Batch jobs
      • Realtime streams
  • SERVERLESS INFERENCE
    • Quickstart
    • About Serverless Routing
    • API
      • Chat Completions
      • Models
  • Discord
Powered by GitBook
On this page
  1. SERVERLESS INFERENCE

API

Cortecs offers a simple, OpenAI-compatible API for serverless inference across multiple providers. It supports two endpoints:

🔁 POST /v1/chat/completions

Submit chat requests using any available model. Supports standard OpenAI parameters like messages, temperature, and max_tokens. Use preference to optimize for speed, cost, or balanced.

📦 GET /v1/models

List all available models and their capabilities.

PreviousAbout Serverless RoutingNextChat Completions

Last updated 11 days ago