How to create a LiteLLM Proxy API key to use in Ozeki AI Gateway

This comprehensive guide demonstrates how to create a LiteLLM Proxy API key and configure it in Ozeki AI Gateway. You'll learn how to access the LiteLLM admin panel, generate a new API key with model access, and set up LiteLLM Proxy as a provider in your gateway. By following these steps, you can connect your Ozeki AI Gateway to LiteLLM Proxy and access multiple AI models through a unified interface.

API URL: http://localhost:4000/v1 (or your LiteLLM proxy URL)
Example API Key: sk-1234567890abcdef...
Default model: gpt-5.2

What is a LiteLLM Proxy API key?

A LiteLLM Proxy API key is a unique authentication credential that allows applications to access multiple AI models through LiteLLM's unified proxy server. LiteLLM Proxy provides a single interface to access models from OpenAI, Anthropic, Google, and other providers. When configuring LiteLLM Proxy as a provider in Ozeki AI Gateway, you need this API key to establish the connection between your gateway and the LiteLLM Proxy instance.

Overview
Overview

Steps to follow

  1. Open LiteLLM Proxy
  2. Go to admin panel
  3. Login to LiteLLM
  4. Begin key creation
  5. Enter key name and choose AI models
  6. Create the key
  7. Copy API key
  8. Open Providers page in Ozeki AI Gateway
  9. Create provider

How to create a LiteLLM Proxy API key video

The following video shows how to create a LiteLLM Proxy API key and configure it in Ozeki AI Gateway step-by-step. The video covers accessing the LiteLLM admin panel, generating a new API key, and setting up LiteLLM Proxy as a provider in your gateway.

Step 0 - Install Ozeki AI Gateway

Before configuring LiteLLM Proxy as a provider, you need to have Ozeki AI Gateway installed and running on your system. If you haven't installed Ozeki AI Gateway yet, follow our Installation on Linux guide or Installation on Windows guide to complete the initial setup.

Step 1 - Open LiteLLM Proxy

Open your web browser and navigate to your LiteLLM Proxy URL, typically http://localhost:4000 (or http://127.0.0.1:4000) if running locally. The LiteLLM Proxy interface provides access to the admin panel where you can manage API keys and model configurations (Figure 1).

Open LiteLLM Proxy
Figure 1 - Open LiteLLM Proxy

Step 2 - Go to admin panel

From the LiteLLM Proxy home page, locate and click on the link or button to access the admin panel. The admin panel provides administrative controls for managing users, API keys, and proxy configuration (Figure 2).

Go to admin panel
Figure 2 - Navigate to admin panel

Step 3 - Login to LiteLLM

Enter your LiteLLM admin credentials to log in to the admin panel. If this is your first time accessing the admin panel, use the default credentials configured during LiteLLM setup. After logging in, you'll have access to the key management interface (Figure 3).

Login to LiteLLM
Figure 3 - Login to LiteLLM admin panel

Step 4 - Begin key creation

In the admin panel, in the "Virtual keys" menu, locate and click the button to create a new API key. This initiates the key generation workflow where you'll configure the key's permissions and model access (Figure 4).

Begin key creation
Figure 4 - Begin key creation

Step 5 - Enter key name and choose AI models

Enter a descriptive name for your API key, such as "Ozeki AI Gateway". Select which AI models this key should have access to from the available models configured in your LiteLLM Proxy. You can grant access to all models or restrict access to specific models based on your requirements (Figure 5).

Enter key name and choose models
Figure 5 - Enter key name and choose AI models

Step 6 - Create the key

After entering the key details and selecting model access permissions, click the "Create Key" button to create your new LiteLLM Proxy API key. For testing purposes, the "Default" key type is sufficient. The system will generate a unique authentication credential for accessing the proxy (Figure 6).

Create the key
Figure 6 - Click Create to generate key

Step 7 - Copy API key

After the key is generated, LiteLLM displays the complete key value. Click the copy button to copy the key to your clipboard. Store this key securely as it provides access to your LiteLLM Proxy and the configured AI models (Figure 7).

The API key is shown only once during creation. Make sure to copy and securely store the key before closing this window.

Copy API key
Figure 7 - Copy API key

Step 8 - Open Providers page in Ozeki AI Gateway

Open the Ozeki AI Gateway web interface and navigate to the Providers page. This is where you'll configure LiteLLM Proxy as a new provider using the API key you just created (Figure 8).

Open Providers page
Figure 8 - Open Providers page in Ozeki AI Gateway

Step 9 - Create provider

Click the "New" button to begin the provider creation process. This opens a form where you'll enter the connection details for LiteLLM Proxy (Figure 9).

Add new provider
Figure 9 - Click add new provider button

Fill in the provider configuration form with your LiteLLM Proxy details. Enter the provider name, select "OpenAI compatible" as the provider type, paste your LiteLLM Proxy URL, and paste your API key. Select your preferred model from the dropdown list, then click "Create" to save the configuration (Figure 10).

http://localhost:4000/v1

Enter details and create provider
Figure 10 - Enter LiteLLM Proxy details and create

After creation, the LiteLLM Proxy provider is added to Ozeki AI Gateway. The provider now appears in the providers list, confirming that your gateway can route requests to LiteLLM Proxy. You can now create routes that allow users to access multiple AI models through your gateway (Figure 11).

Provider created
Figure 11 - LiteLLM Proxy provider created successfully

Final thoughts

You have successfully created a LiteLLM Proxy API key and configured it as a provider in Ozeki AI Gateway. Your gateway can now communicate with LiteLLM Proxy and access multiple AI models from various providers through a unified interface. This setup allows you to manage LiteLLM Proxy access for multiple users and applications from a single point.

More information