Last modified December 5, 2025
Installing Gateway API Inference Extension
Overview
The Kubernetes Gateway API Inference Extension enables AI/ML workloads to be exposed and managed through the Gateway API. This extension provides a standardized way to configure routing and load balancing for inference endpoints, making it easier to deploy and scale AI services in your Kubernetes clusters.
This guide walks you through installing the Gateway API Inference Extension on Giant Swarm workload clusters. You enable the inference pool Custom Resource Definitions (CRDs) in your Gateway API bundle configuration.
Installation
Our Gateway API CRDs app already supports the inference extensions. In your Gateway API bundle configmap, add the following block:
apiVersion: v1
kind: ConfigMap
metadata:
name: <CLUSTER_NAME>-gateway-api-bundle
namespace: org-<ORGANIZATION>
data:
values: |
clusterID: <CLUSTER_NAME>
organization: <ORGANIZATION>
apps:
gatewayApiCrds:
userConfig:
configMap:
values: |
install:
inferencepools: "standard"
Run the kubectl apply -f <configmap-file.yaml> command on your management cluster to apply the updated bundle configuration, then wait until the new CRDs are deployed.
Further reading
Need help, got feedback?
We listen to your Slack support channel. You can also reach us at support@giantswarm.io. And of course, we welcome your pull requests!