Skip to content

Set your AI model for the Autonomous Debugger

Note

Available on demand from version 1.55 for Java, JavaScript, and TypeScript applications in JetBrains IDE. Support for additional languages and IDEs will be coming soon.

The Lightrun Autonomous Debugger uses an AI model that analyzes data, recognizes patterns, and makes predictions or decisions to automate and enhance tasks. By default, it is set to use Lightrun’s OpenAI account. No setup is required unless you want to use your own provider. This guide outlines how administrators can configure a custom AI model, such as a company OpenAI key or a Private Azure OpenAI deployment.

Supported AI model providers

  • OpenAI: By default, the Autonomous Debugger uses Lightrun's OpenAI as the default AI model provider. No configuration is required.
  • Customer OpenAI: Requires adding your API key.
  • Private Azure OpenAI Service: Requires copying a few values from your Private Azure deployment.

Set your Open AI account as the AI model provider

To use your company’s Open AI account, you will need to obtain your company’s API Key.

  1. Log in to the Lightrun plugin in the IDE.
  2. Click the AI tab.
  3. To access the Autonomous Debugger settings, click the More options icon.
  4. From the list, select Settings.
  5. In the Model Configuration tab, select Customer OpenAI service.

    Customer OpenAI model

  6. In the API Key field, paste your company’s API Key.

  7. Click Save.

Set Microsoft Azure OpenAI as your AI model provider

You can choose to set up your own OpenAI account using your Private Azure OpenAI service. To set up the model, you are required to enter a number of settings that are required to be obtained in Azure.

Prerequisites

Stage 1: [Azure AI Foundry] Obtain the Azure API parameters

  1. Log in to https://ai.azure.com/.
  2. In the left navigation bar, click Models + endpoints.

    The Manage deployments of your models and services opens.

  3. Click Get endpoint.

    Get endpoints

  4. Copy the key.

    Copy Key

  5. Copy the Completion and Embedding deployment names.

    a. In the left navigation bar, click the Models+endpoints page.

    A list of available deployed models are displayed.

    List of deployed models

    b. Copy the gpt-4o model deployment name to be pasted in the Completion deployment name field.

    c. Copy the text-embedding-3-model deployment name, for the Embedding deployment name field. Note that under certain circumstances, the names could be gpt-4o and text-embedding-3-small respectively.

  6. Copy the Endpoint. Do not copy the endpoint from the Get endpoint dialog. Instead, follow these steps:

    1. Go to the Azure portal in https://portal.azure.com/.

    2. Click the Azure AI services and then the Azure OpenAI links.

    3. Select the same deployment, selected in the previous steps.

    4. In the left navigation pane, navigate to Resource management, and select Keys and Endpoint.

    The Keys and Endpoint dialog opens.

    Keys and endpoints

    5. Copy the Endpoint which should match the following syntax: https://azure-cognitive-XXXXX.openai.azure.com/.

Stage 2: [Lightrun Plugin] Configure Azure Open AI in the plugin

  1. Log in to the Lightrun plugin in the IDE.
  2. Click the AI tab
  3. To access the Autonomous Debugger settings, click the More options icon.
  4. From the list, select Settings.
  5. Select the Model Configuration tab and select Private Azure OpenAI service.

    Private Azure model config

  6. Paste the parameters you copied in Stage 1 from Azure AI Foundry.

    Parameter Description
    API Key Your Azure Key.
    Endpoint The URL for the Azure Endpoint, which is used for sending API requests.
    Completion deployment name The deployment name for the Azure gpt-4o model, which is used for generating text completions.
    Embedding deployment name The deployment name for Azure text-embedding-3-model, which is used for generating vector embeddings.
  7. Click Save.


Last update: April 17, 2025