Skip to main content

Configuration

Accessing Code GPT's Settings

1.To set up the response, first pick the model. 2. In Visual Studio Code, click on button. 3. Then Configuration. This will open the Settings window.
4. On this windows you can fit the Max Tokens, Window Memory and Temperature

Available Settings

Provider

Select your AI provider from the dropdown menu, then enter the API Key for the selected provider.

Max Token

Tokens can be thought of as pieces of words. Before the API processes the prompts, the input is broken down into tokens.

Each model has a maximum number of tokens. Select the number of tokens depending on the length of the response you want to get and the selected model.

Learn more: Tokens by OpenAI

Temperature

This is a parameter that can be adjusted. It determines the level of randomness or "creativity" in the generated text. A higher temperature will result in more varied and creative output, while a lower temperature will produce output that is more similar to the training data and less likely to contain unexpected or surprising content.

It is a value between 0 and 1. 0 being the most deterministic and 1 being the most random and creative. The default temperature is 0.3

Learn more: Temperature by Cohere

Window Memory

It stores a list of past conversations, but only the last K (a specific number) conversations are used from the saved list. This is done to prevent the buffer from becoming too large and exceeding the token limit.

  • default: 4
  • minimum: 1
  • maximum: 50

Also, you can set the configuration directly from ⚙

CodeGPT Settings

image