How to set up Continue VS Code agent with Ozeki AI Gateway

This guide walks you through installing and configuring the Continue extension in Visual Studio Code to work with Ozeki AI Gateway. By following this tutorial, you'll learn how to install the Continue extension, open its configuration file, and add Ozeki AI Gateway as a model provider.

What is Continue?

Continue is an open-source VS Code extension that brings AI-powered coding assistance directly into your editor. It supports code completion, inline edits, and chat-based interactions with LLMs. Because it supports OpenAI-compatible APIs, it can be connected to Ozeki AI Gateway, allowing you to route all requests through your gateway and use any configured models.

Overview
Overview

Continue VS Code agent models config

models:
  - name: Ozeki AI Gateway
    provider: openai
    model: your-llm-model
    apiBase: http://localhost/v1
    apiKey: ozkey-abc123

Steps to follow

We assume Ozeki AI Gateway is already installed on your system. You can install it on Linux, Windows or Mac.

You will also need Visual Studio Code installed on your system.

  1. Install and open Continue extension
  2. Open the local config file
  3. Add Ozeki AI Gateway to the config
  4. Send a test prompt

How to set up Continue VS Code agent with Ozeki AI Gateway video

The following video shows how to install and configure the Continue extension to work with Ozeki AI Gateway step-by-step. The video covers installing the extension, editing the configuration file, and testing the connection with a sample prompt.

Step 1 - Install and open Continue extension

Open Visual Studio Code and navigate to the Extensions panel by clicking the Extensions icon in the left sidebar or pressing Ctrl+Shift+X. Search for Continue and click the Install button on the extension (Figure 1).

Install Continue extension in VS Code
Figure 1 - Install the Continue extension in VS Code

Once installed, open the extension's panel by clicking the Continue icon that appears in the left sidebar (Figure 2).

Open Continue from the extensions
Figure 2 - Open Continue from the extensions panel

Step 2 - Open the local config file

Inside the Continue panel, locate and click the settings icon in the top right corner to open the settings menu (Figure 3).

Navigate to Continue settings
Figure 3 - Navigate to Continue settings

In the settings menu, select Configs from the left sidebar and click the gear icon next to Local Config to open it in the editor (Figure 4).

Select configs and open local config
Figure 4 - Select Configs and open the local config file

Step 3 - Add Ozeki AI Gateway to the config

In the config file, locate the models: section and add the following block. Replace your-llm-model with the name of the model you want to use, and ozkey-abc123 with the API key you created in Ozeki AI Gateway (Figure 5).

models:
  - name: Ozeki AI Gateway
    provider: openai
    model: your-llm-model
    apiBase: http://localhost/v1
    apiKey: ozkey-abc123

Add Ozeki AI Gateway to config models
Figure 5 - Add Ozeki AI Gateway to the config models

After saving the config file, click the back button to exit the settings view and return to the main chat panel (Figure 6).

Exit from settings using back button
Figure 6 - Return to the Continue chat panel using the back button

Step 4 - Send a test prompt

In the Continue chat panel, open the model selector and choose "Ozeki AI Gateway" from the list, then type a test prompt and press Enter (Figure 7).

Ensure Ozeki AI Gateway is selected and send a prompt
Figure 7 - Select Ozeki AI Gateway and send a test prompt

You should see a response appear in the chat panel, confirming that Continue is correctly connected to Ozeki AI Gateway (Figure 8).

LLM response
Figure 8 - Continue displays the AI response

Conclusion

You have successfully installed and configured the Continue extension to work with Ozeki AI Gateway. You can now use Continue's chat and code assistance features inside VS Code, with all requests routed through your gateway and handled by the AI models available there.


More information