Skip to content

LLM Connection Configuration

While there are circumstances in which AIDR can be operated using an open, freely accessible LLM, typically organizations are using LLMs protected by API keys, running on dedicated company-hosted instances, and otherwise requiring container-level configuration to interact seamlessly with AIDR for GenAI. It’s worth noting that operating the AIDR proxy in forward-proxy (“enriched”) mode requires the LLM to be configured in the container settings, and not via API key. Additionally, configuring the connection in the container spares the necessity of sending the key with every request.

The container environment variables in this section are used to configure a backend connection to LLMs on the container level. Note that many of them do not have defaults, as they are only accessed when connecting to a specific LLM instance.

AWS Configuration

AIDR can be configured to route traffic to a single AWS account or multiple AWS accounts.

  • If there is only one AWS account configured, this account is used as the default for routing all traffic.
  • If multiple AWS accounts are registered, while there is a default account, it is not used for routing all traffic. Requests with incorrect AWS account information will result in an error message.

    • There must be a default registered account. A single account configuration will be set as the default account.
    • For any additional named credential sets, the suffix X in the environment variables should be the actual AWS Access Key ID. See the Default AWS Account and Additional AWS Accounts examples below.
Environment Key Required Example Value Description
HL_LLM_PROXY_AWS_ACCESS_KEY_ID_DEFAULT True Default AWS Access Key ID
HL_LLM_PROXY_AWS_SECRET_ACCESS_KEY_DEFAULT True Default AWS Secret Access Key
HL_LLM_PROXY_AWS_SESSION_TOKEN_DEFAULT False Default AWS Session Token (required if using temporary credentials)
HL_LLM_PROXY_AWS_REGION_DEFAULT False us-east-1 Default AWS Region used if none is specified for another set
HL_LLM_PROXY_AWS_BEDROCK_BASE_URL False https://bedrock-runtime.{region}.amazonaws.com Default Bedrock base URL used if none is specified for another set
HL_LLM_PROXY_AWS_SAGEMAKER_BASE_URL False https://runtime.sagemaker.{region}.amazonaws.com Default Sagemaker base URL used if none is specified for another set
HL_LLM_PROXY_AWS_ACCESS_KEY_ID_X False AWS Secret Access Key for the credential set
HL_LLM_PROXY_AWS_SESSION_TOKEN_X False AWS Session Token for the credential set
HL_LLM_PROXY_AWS_REGION_X False Uses default region AWS Region for the credential set
HL_LLM_PROXY_AWS_BEDROCK_BASE_URL_X False Uses default Bedrock Bedrock base URL for the credential set
HL_LLM_PROXY_AWS_SAGEMAKER_BASE_URL_X False Uses default Sagemaker Sagemaker base URL for the credential set

Default AWS Account

If only one AWS account is registered with AIDR GenAI, this account is used by default to route traffic.

Example - Default AWS Account

This will create a default credential set. The application can then use this credential set to configure AWS clients.

export HL_LLM_PROXY_AWS_ACCESS_KEY_ID_DEFAULT=<your-secret-access-key>
export HL_LLM_PROXY_AWS_SECRET_ACCESS_KEY_DEFAULT=<your-session-token>
export HL_LLM_PROXY_AWS_REGION_DEFAULT=us-east-1

Additional AWS Accounts

AIDR can be configured to route traffic to multiple AWS accounts. If there are multiple keys registered, there is no default account for routing traffic. Requests with incorrect AWS account information will result in an error message.

When configuring multiple AWS accounts, you must create a default account and additional accounts.

Example - Default AWS Account

This will create a default credential set.

export HL_LLM_PROXY_AWS_ACCESS_KEY_ID_DEFAULT=<your-secret-access-key>
export HL_LLM_PROXY_AWS_SECRET_ACCESS_KEY_DEFAULT=<your-session-token>
export HL_LLM_PROXY_AWS_REGION_DEFAULT=us-east-1

Example - Additional AWS Accounts

If your AWS Access Key ID is AKIAXXXXXXXXFOO, this will create a credential set named AKIAXXXXXXXXFOO.

export HL_LLM_PROXY_AWS_ACCESS_KEY_ID_AKIAXXXXXXXXFOO=<your-secret-access-key>
export HL_LLM_PROXY_AWS_SESSION_TOKEN_AKIAXXXXXXXXFOO=<your-session-token>
export HL_LLM_PROXY_AWS_REGION_AKIAXXXXXXXXFOO=us-east-1

Azure Configuration

AIDR can be configured to route traffic to an Azure OpenAI account.

Azure

Environment Key Required Example Value Description
HL_LLM_PROXY_AZURE_TENANT_ID False A unique identifier for your Azure tenant
HL_LLM_PROXY_AZURE_CLIENT_ID False A unique identiifer assigned to your application registration
HL_LLM_PROXY_AZURE_CLIENT_SECRET False A unique secret assigned to your application registration
HL_LLM_PROXY_AZURE_BASE_URL False Default Azure application base URL used if none is specified for another set
HL_LLM_PROXY_AZURE_REGION False eastus Default Azure region

Azure OpenAI

Environment Key Required Example Value Description
HL_LLM_PROXY_OPENAI_AZURE_API_KEY False The API key to access Azure OpenAI
HL_LLM_PROXY_OPENAI_AZURE_SCHEME False The schema for Azure OpenAI output
HL_LLM_PROXY_OPENAI_AZURE_HOST False The name of the Azure OpenAI host

OpenAI Configuration

AIDR can be configured to route traffic to an OpenAI account.

Environment Key Required Example Value Description
HL_LLM_PROXY_OPENAI_DEFAULT_MODEL False Enforces the use of only the model specified in this setting. For example, gpt-4o.
HL_LLM_PROXY_OPENAI_BASE_URL False https://api.openai.com Default OpenAI base URL used if none is specified for another set
HL_LLM_PROXY_OPENAI_API_KEY False The API key to access OpenAI

Hugging Face Configuration

AIDR can be configured to route traffic to a Hugging Face model.

Environment Key Required Example Value Description
HL_LLM_PROXY_HUGGINGFACE_TGI_PROVIDER False huggingface-tgi Identifying Hugging Face as the model provider
HL_LLM_PROXY_HUGGINGFACE_TGI_NAMES False Qwen/Qwen3-0.6B The Hugging Face model name
HL_LLM_PROXY_HUGGINGFACE_TGI_BASE_URLS False Default Hugging Face base URL used if none is specified for another set

Custom Configuration

AIDR can be configured to route traffic to a custom model, like Ollama.

Environment Key Required Example Value Description
HL_LLM_PROXY_CUSTOM_{{MODELNAME}} False HL_LLM_PROXY_CUSTOM_LLAMA=tinyllama Used to provide an appropriate name for the model
HL_LLM_PROXY_CUSTOM_{{MODELNAME}}_PROVIDER False HL_LLM_PROXY_CUSTOM_LLAMA_PROVIDER=ollama Provider of the custom model
HL_LLM_PROXY_CUSTOM_{{MODELNAME}}_BASE_URL False HL_LLM_PROXY_CUSTOM_LLAMA_BASE_URL=http://ollama:11434 Base URL of the endpoint where the model can be queried
HL_LLM_PROXY_CUSTOM_{{MODELNAME}}_API_KEY False The API key for the custom model.