AI Detection & Response Release Notes¶
Overview¶
We are excited to announce the latest release of HiddenLayer’s AI Detection & Response (AIDR), a testament to our commitment to providing the industry's most advanced AI security solutions. This latest update introduces new features and enhancements that offer an even more robust, intuitive, and effective cybersecurity for AI experience.
AIDR for GenAI version 25.6.0¶
Release Notes for AIDR (LLM Proxy) v25.6.0 released on June 26, 2025.
What’s New¶
Prompt Injection Model version 5¶
-
The AIDR model version 5
- Provides protection against TokenBreak attacks. TokenBreak attacks can bypass safety features and guardrails for large language models (LLMs).
- Significantly lower false positives across the board.
- Improved true positive rate for non-English languages.
Projects and Custom Rulesets¶
- Customers can now create projects to represent unique AI applications or use cases.
- Projects can use a custom ruleset to configure AIDR detectors, rather than using one universal ruleset for all use cases.
Enriched Sandbox¶
- The Sandbox in the Console now provides historical chat context and more sample prompts.
AIDR version 25.5.2¶
Release Notes for AIDR (LLM Proxy) v25.5.2 released on May 29, 2025.
What’s New¶
Remote Ruleset Configuration¶
-
Remotely manage detection configuration from the console, rather than requiring configuration via environment variables or headers.
-
Detection configuration for Prompt Injection, DDOS, PII, and Guardrails are supported.
- Headers sent in the request can override remote rules if the local configuration is set to allow header overrides.
- Once a default ruleset is established, it will override any conflicting local configurations.
- Rulesets are polled for changes once a minute, so changes should propagate to your deployment in less than five minutes.
Prompt Injection Model v4.5¶
- Releasing the V4.5 Prompt Injection model, with significantly higher performance on English and other supported languages (German, Spanish, Korean, Japanese, French, and Italian).
Previous Releases¶
AIDR version 25.3.1¶
Release Notes for AIDR (LLM Proxy) v25.3.1 released on April 3, 2025.
What’s New¶
Mistral - Amazon Bedrock¶
- Mistral models are now supported in Amazon Bedrock.
What’s Improved¶
Quick Scan Updates¶
- Quick Scan now scans the full input, where previously it scanned just the first 512 tokens. It doesn't do a second scan with non-alphanumeric characters stripped. As a consequence, long prompts will have higher latency with Quick Scan.
- Full Scan does the same but also does a second scan with non-alphanumeric characters stripped.
AIDR version 25.3.0¶
Release Notes for AIDR (LLM Proxy) v25.3.0 released on March 13, 2025.
What’s New¶
OWASP 2025 Update¶
- For the LLM Sandbox in the Console, 2025 has been added to the OWASP Scenario. Example: LLM01: 2025 Prompt Injection.
- For AI Detection & Response, the prompt analyzer returns owasp_2025 in the response body.
What's Improved¶
Language improvement PI model version 4¶
A new version of the prompt injection detection classifier with an emphasis on language translation support.
AIDR version 25.2.0¶
Release Notes for AIDR (LLM Proxy) v25.2.0 released on February 19, 2025.
What’s New¶
GPU Utilization¶
AI Detections & Response can utilize a GPU to improve throughput and performance.
Refusals Detection¶
Refusals can now be detected based on the language in outputs. Turning on this feature (configurable) allows customers to detect refusals caused by custom system prompts, rather than the generic set of safety guardrails established by upstream LLM providers.
Error Passthrough¶
For reverse proxy requests (i.e., passthrough requests), Proxy will now propagate upstream errors back to the requester instead of responding with a HTTP 502 or HTTP 500. SageMaker and Bedrock will continue always responding with HTTP 502 for upstream errors.
AIDR version 25.1.0¶
Release Notes for AIDR (LLM Proxy) v25.1.0 released on January 21, 2025.
What’s Improved¶
Support for Multiple AWS Bedrock Accounts¶
A single AIDR instance can now route traffic to multiple AWS Bedrock instances with different credentials.
AIDR version 24.12.0¶
Release Notes for AIDR (LLM Proxy) v24.12.0 released on December 17, 2024.
What’s Improved¶
Increased Prompt Efficacy¶
Increased efficacy on Japanese and Korean prompts.
AI Detection & Response v24.10.3¶
Release Notes for the AIDR (LLM Proxy) v24.10.3 released on November 26, 2024.
What’s New¶
Prompt Injection Detection¶
A new prompt injection detection for AIDR, Control tokens, identifies when a prompt contains specific control tokens known to confuse LLMs and allow a malicious user to subvert their instructions.
Guardrail Detection for Gemini¶
AIDR now supports Guardrail Detection for Gemini, as well as OpenAI and models hosted on Azure. Guardrail detection works by deterministically passing through the field set by the LLM provider, allowing a guardrail event to be added into security workflows. Anthropic does not offer a guardrail field today, so AIDR does not support guardrail detection for Anthropic models.
AWS Instance Profiles Authentication¶
AIDR supports AWS instance profiles for authentication to Sagemaker/Bedrock.
AI Detection & Response v24.10.0¶
Release Notes for the AIDR (LLM Proxy) v24.10.0 released on October 15, 2024.
What’s New¶
New Model Version¶
Released a new version of the model with higher accuracy.
AI Detection & Response v24.9.1¶
Release Notes for the AIDR (LLM Proxy) v24.9.1 released on September 23, 2024.
What’s Improved¶
AIDR Logging¶
he AIDR SIEM logs now provide a persistent event_id
.
AI Detection & Response v24.9.0¶
Release Notes for the AIDR (LLM Proxy) v24.9.0 released on September 17, 2024.
What’s Improved¶
AIDR API¶
The Prompt Analyzer now offers full AIDR functionality, allowing customers running AIDR to check the input and output safety flexibly throughout their application.
Improved Efficacy for Code Detections¶
Enhancements to AIDR code detections to improve security and performance.
Reduction in Compute Usage¶
This release of AIDR reduces compute usage.
AI Detection & Response v24.8.1¶
Release Notes for the AIDR (LLM Proxy) v24.8.1 released on September 3, 2024.
What’s Improved¶
Added a non-root user to the runtime-image¶
This allows users who require that their pods cannot run as root to install AIDR.
What’s Resolved¶
This release also includes performance improvements and bug fixes.
AI Detection & Response v24.8.0¶
Release Notes for the AIDR (LLM Proxy) v24.8.0 released on August 20, 2024.
What’s New¶
Route Passthrough¶
AIDR now supports unmapped route passthrough, sending your request to the generative AI provider, regardless of whether the endpoint is directly supported. This new, optional global policy setting in the LLM Proxy provides a seamless user experience.
AI Detection & Response v24.7.0¶
Release Notes for the AIDR (LLM Proxy) v24.7.0 released on July 23, 2024.
What’s Resolved¶
This release also includes performance improvements and bug fixes.
AI Detection & Response v24.6.0¶
Release Notes for the AIDR (LLM Proxy) v24.6.0 released on June 20, 2024.
What’s New¶
New LLM Proxy Configurations¶
This release includes three new global policy settings to improve the user experience.
- Include Block Message Reasons - When enabled, includes the block message reasons in the response.
- Proxy Enabled Passthrough Streaming - When enabled, the proxy will immediately start streaming the response back to the requester. Currently available for OpenAI and Azure OpenAI.
- Max Request Size - The maximum size for a request or a response, in bytes.
What’s Updated¶
Improved Detections¶
Modified and improved code detections in the analyzer.
MITRE Tags for Detections¶
When a modality restriction detection occurs, the LLM Proxy includes the MITRE ATLAS ID.
What’s Resolved¶
The following items have been resolved for this release.
- Proper handling of PII redaction when a context window is set to LAST.
- Added support for OpenAI chat completion functions.
- Resolved an issue with reverse-proxy routes for Sagemaker and Bedrock. Updated the LLM Proxy Deployment Guide with examples for Sagemaker and Bedrock.