Changelog

AI Gateway Helm chart v0.1.1

March 31, 2025

🆕 Features

  • Adds readiness configuration and endpoint for AI Gateway Core image.

  • Updates AI Gateway Core image to v1.0.2.

  • Updates AI Gateway Processors image to v1.0.1.

✅ Compatibility:

  • Kubernetes version 1.25.0-0 or higher

📦 Helm registry:

  • oci://private-registry.f5.com/aigw/aigw:0.1.1

AI Gateway Core v1.0.2

March 31, 2025

🐛 Bug fixes:

  • Fix the core not rejecting request/response for parallel stages: if a stage was parallel, the core would not reject request/response, even if a processor rejected it.

✅ Compatibility:

  • Kubernetes version 1.25.0-0 or higher

📦 Container images:

  • private-registry.nginx.com/aigw/aigw:v1.0.2

AI Gateway Core v1.0.1

March 28, 2025

✨ Highlights:

  • The release includes a number of bug fixes.

🐛 Bug fixes:

  • Improve error messages in logs: make the messages more descriptive and user-friendly.

  • Fix the readiness probe health check: previously, the probe could have succeeded even if the core wasn’t ready to receive the traffic.

  • Return 502 Bad Gateway status when the core doesn’t select any service for a request. Previously, the core would return 404.

  • Ensure the core adds an error to an OTel trace for a span that corresponds to the profile sequential stage.

  • Ensure the core validates the processor parameters (the params field of the processor configuration).

  • Ensure the core returns a 500 Internal Server Error when a processor executed in a parallel stage modifies the client request or LLM response. Such modifications are not allowed. Previously, the core would disregard any modifications, but incorrectly proceed further with any processing.

  • Ensure the core fails to start if aigw start is invoked without the config file path.

  • Do not print usage if the core terminates with an error.

  • Ensure the core returns JSON errors when it cannot match a request against any route. Previously, it would return plain text errors.

  • When a service returns an error, ensure the core propagates the HTTP status code to the client. Previously, the core would always return the 500 status code.

  • Ensure the core returns errors to clients in JSON using the OpenAI-like error format. Previously, for some cases, the core would return errors in different format.

  • Ensure when the core is being shutdown, it exports any pending unexpected transaction. Note: if transactions are not exported in 30s (default terminationGracePeriodSeconds pod parameter), Kubernetes will kill the pod, and, as a result, the unexported transactions will get lost.

  • Improve validation of the path field in routes.

  • Change the minimal acceptable TLS version for connections to processors and services from v1.3 to v1.2.

✅ Compatibility:

  • Kubernetes version 1.25.0-0 or higher

📦 Container images:

  • private-registry.nginx.com/aigw/aigw:v1.0.1

AI Gateway Processors v1.0.1

March 28, 2025

✨ Highlights:

  • Minor optimizations for the language-id and prompt-injection processors

  • Improved validation of model pipeline parameters

🐛 Bug fixes:

  • Fixed an issue with processors server not utilizing worker processors when WEB_CONCURRENCY is set

✅ Compatibility:

  • Kubernetes version 1.25.0-0 or higher

📦 Container images:

  • private-registry.nginx.com/aigw/aigw-processors-f5:v1.0.1

AI Gateway v1.0.0

February 25, 2025

AI Gateway Core

✨ Highlights:

  • Export OpenTelemetry Metrics

  • Simplify Route schema configuration.

  • Entitlement and Visibility tracking enabled (for usage-based pricing compliance)

🐛 Bug fixes:

  • Fixed an issue with OpenTelemetry tracing that sometimes resulted in duplicate spans.

  • Improve OpenTelemetry Tracing configuration to avoid performance degradation under load

  • Exit on TLS misconfiguration

  • Fix various Selector evaluation and configuration bugs.

  • Support path style addressing when connecting to AWS S3 for transaction exporting.

AI Gateway Processors

✨ Highlights:

  • F5 processor set revised to:

    • prompt-injection

    • system-prompt

    • language-id

    • repetition-detect

  • prompt-injection and language-id processors now process the full conversation history

  • system-prompt processor adds the ability to strip existing system prompts. Changes made to ordering of inserted prompts to fit model expectations

  • language-id processor now allows rejection

  • Processors now accept common parameters reject, annotate, modify for more predictability in processor actions

  • Support for enabling GPU support for model-based processors added to helm chart.

✅ Compatibility:

  • Kubernetes version 1.25.0-0 or higher

📦 Container images:

  • private-registry.nginx.com/aigw/aigw:v1.0.0

  • private-registry.nginx.com/aigw/aigw-processors-f5:v1.0.0

AI Gateway Core v0.2.0

January 23, 2025

AI Gateway Core

🆕 Features

  • Secure communications between core, processors, and upstream LLMs using private TLS certificates.

  • Authenticate processors using mTLS; private certificates are supported here as well.

  • Configure core application loging level, verbosity, and contextual output.

  • Model-based LLM request routing.

  • List models API for OpenAI

🐛 Bug fixes:

  • Transaction exporter “busy looping” causing unnecessary CPU utilization

  • OTEL Exporter fails to shut down cleanly when receiving SIGINT

✅ Compatibility:

  • Kubernetes version 1.25.0-0 or higher

📦 Container images:

  • private-registry.nginx.com/aigw/aigw:v0.2.0

v0.1.0

November 13, 2024

✨ Highlights:

F5 AI Gateway routes generative AI traffic to an appropriate Large Language Model (LLM) backend and protects the traffic against common threats, which includes:

  • Inspecting and filtering client requests and LLM responses

  • Preventing malicious inputs from reaching an LLM backend

  • Ensuring that LLM responses are safe to send to clients

  • Protecting against leaking sensitive information

AI Gateway is a standalone product that runs on Kubernetes.

AI Gateway Core

🆕 Features:

  • Support the following LLM APIs: OpenAI, Microsoft Azure OpenAI Service, Anthropic, Ollama (only between AI Gateway and Ollama)

  • Authentication of client requests using JWT and TLS

  • Authorization of client requests using JWT claims and HTTP headers

  • Executing a chain of processors for client requests and LLM responses.

  • Routing client requests based on request attributes and metadata generated by processors.

  • Conversion between LLM API schemas when the request schema and the LLM backend schema do not match.

  • Recording every request/response on disk with an ability to export to AWS S3 or S3-compatible storage.

  • Generating and exporting observability data via OpenTelemetry.

  • Providing a configuration interface via an HTTP API and a config file.

✅ Compatibility:

  • Kubernetes >= 1.25.0-0

📦 Container images:

  • private-registry.nginx.com/aigw/aigw:v0.1.0

AI Gateway Processors

🆕 Features:

  • Added the following processors:

    • Language ID

    • Prompt enforcement

    • Prompt injection detection (PI Processor)

    • Repetition detection

  • Efficacy tests for above processors

  • Added GPU support for model-based processors

  • TLS support added to the processor server

  • Added support for system messages in the processor with new API revision

✅ Compatibility:

  • Kubernetes >= 1.25.0-0

📦 Container images:

  • private-registry.nginx.com/aigw/aigw-processors-f5:v0.1.0