Docs
cortecs.aiDedicated ModelsServerless ModelsLogin
  • Introduction
  • DEDICATED INFERENCE
    • Quickstart
    • Provisioning API
      • Authentication
      • User
      • Instances
      • Models
      • Hardware Types
    • Python client
      • Objects
      • Langchain integration
    • Examples
      • Batch jobs
      • Realtime streams
  • SERVERLESS INFERENCE
    • Quickstart
    • About Serverless Routing
    • API
      • Chat Completions
      • Models
  • Discord
Powered by GitBook
On this page
  • 1. Register
  • 2. Start a model
  • 3. Query the model
  1. DEDICATED INFERENCE

Quickstart

Get up and running in just a few clicks

PreviousIntroductionNextProvisioning API

Last updated 11 days ago

1. Register

Register at and follow these steps to set up your account:

  • Fill out your billing address in the and press Save.

  • Enter your credit card details.

  • Increase your account balance. Press Top up to increase your account balance.

If your balance reaches zero, your instances will be discontinued. To avoid this, use Auto top-up to set an amount that is automatically transferred when your balance falls below a specified threshold.

2. Start a model

To start a model, follow these steps:

  • Start the model and wait until the status indicates it is running. This setup process can take a few minutes to complete.

3. Query the model

Our endpoints are compatible with OpenAI API by default. We assume you have either Python or Node.js setup. This example is based on cortecs/phi-4-FP8-Dynamic but works for all models supported by cortecs.

export OPENAI_API_KEY="<YOUR_CORTECS_API_KEY>"
export OPENAI_BASE_URL="<YOUR_MODEL_URL>"

OpenAI Client

You can use popular libraries provided by OpenAI for Python or Node.js. First, install the library:

pip install openai
npm install openai

Query the model by calling the completion endpoint. Don't forget to pass your API key and the model URL if you didn't set them as environment variable already.

from openai import OpenAI

openai_api_key = "<OPENAI_API_KEY>"
openai_api_base = "<MODEL_URL>"

client = OpenAI(
    api_key=openai_api_key,
    base_url=openai_api_base,
)

completion = client.chat.completions.create(
    model="cortecs/phi-4-FP8-Dynamic",
    messages=[
        {
            "role": "user",
            "content": "Tell me a joke."
        }
    ]
)

print(completion.choices[0].message)
import OpenAI from "openai";

const openai = new OpenAI({
    apiKey: '<OPENAI_API_KEY>',
    baseURL: '<MODEL_URL>'
});

async function main() {
  const completion = await openai.chat.completions.create({
    messages: [
    {
        role: "user",
        content: "Tell me a joke.",
    }],
    model: "cortecs/phi-4-FP8-Dynamic"
  });

  console.log(completion.choices[0].message);
}

main();

LangChain Client

pip install langchain-openai

Query the model by calling the completion endpoint. Don't forget to pass your API key and the model URL if you didn't set them as environment variable already.

from langchain_openai import ChatOpenAI

llm = ChatOpenAI(model_name='cortecs/phi-4-FP8-Dynamic', base_url='<MODEL_URL>')

res = llm.invoke('Tell me a joke.')
print(res.content)

Select a model from our .

Accessing your model requires an API key, which you get on your . Once we have a key we'll want to set our environment variables by running:

is another powerful Python library for building LLM-based applications. It is popular for more complex use cases.

Optionally follow the to:

For more advanced use cases and easier instance management, check out our client library .

catalog
profile page
LangChain
langchain docs
Build an Extraction Chain
Generate synthetic data
Classify text into labels
Summarize text
cortecs-py
cortecs.ai
profile page
Profile page after successfully adding 100€ to the account balance
Model is up and running.