Skip to main content

LiteLLM

LiteLLM provides a unified interface to call 100+ LLMs. Apertis is supported as a native provider.

Installation

pip install litellm

Environment Setup

Set your Apertis API key:

export APERTIS_API_KEY="sk-your-api-key"

Or in Python:

import os
os.environ["APERTIS_API_KEY"] = "sk-your-api-key"

Get your API key from apertis.ai/token.

Basic Usage

Completion

import os
from litellm import completion

os.environ["APERTIS_API_KEY"] = "sk-your-api-key"

messages = [{"role": "user", "content": "What is the capital of France?"}]

response = completion(
model="apertis/gpt-5.2",
messages=messages
)
print(response.choices[0].message.content)

Streaming

from litellm import completion

messages = [{"role": "user", "content": "Write a short poem about AI"}]

response = completion(
model="apertis/gpt-5.2",
messages=messages,
stream=True
)

for chunk in response:
if chunk.choices[0].delta.content:
print(chunk.choices[0].delta.content, end="")

Using Different Models

Access 400+ models through the apertis/ prefix:

# OpenAI GPT
response = completion(model="apertis/gpt-5.2", messages=messages)

# Anthropic Claude
response = completion(model="apertis/claude-sonnet-4.5", messages=messages)

# Google Gemini
response = completion(model="apertis/gemini-3-flash-preview", messages=messages)

LiteLLM Proxy Configuration

1. Export API Key

export APERTIS_API_KEY="sk-your-api-key"

2. Configure config.yaml

model_list:
- model_name: gpt-5.2
litellm_params:
model: apertis/gpt-5.2
api_key: os.environ/APERTIS_API_KEY

- model_name: claude-sonnet
litellm_params:
model: apertis/claude-sonnet-4.5
api_key: os.environ/APERTIS_API_KEY

- model_name: gemini-flash
litellm_params:
model: apertis/gemini-3-flash-preview
api_key: os.environ/APERTIS_API_KEY

3. Start Proxy

litellm --config config.yaml

Supported Parameters

All standard OpenAI-compatible parameters are supported:

ParameterDescription
messagesChat messages array
modelModel ID with apertis/ prefix
streamEnable streaming responses
temperatureSampling temperature (0-2)
top_pNucleus sampling
max_tokensMaximum response tokens
frequency_penaltyFrequency penalty (-2 to 2)
presence_penaltyPresence penalty (-2 to 2)
stopStop sequences
toolsFunction/tool definitions
tool_choiceTool selection mode
ProviderModel ID
OpenAIapertis/gpt-5.2, apertis/gpt-4.1-mini
Anthropicapertis/claude-sonnet-4.5, apertis/claude-haiku-4.5
Googleapertis/gemini-3-pro-preview, apertis/gemini-3-flash-preview

For the full list of models, visit Apertis Models.