# Tooling

Enterprise developers increasingly rely on agent orchestration frameworks to build, deploy, and manage complex LLM-powered applications. These frameworks let you:

* Chain LLM calls for multi-step reasoning,
* Connect to private data sources like databases and vector stores,
* Monitor performance and output quality.

These frameworks are flexible and model-agnostic, allowing connections to any **OpenAI-compatible** endpoint. This lets developers switch between open-source or commercial LLMs to optimize cost, performance, or capabilities.

You can connect any frameworks that supports OpenAI-compatible endpoints. Here are quickstart guides to help you get started with [LangChain](https://docs.cortecs.ai/integration-examples/tooling/langchain) or [Langfuse](https://docs.cortecs.ai/integration-examples/tooling/langfuse).
