Configuring Ingress
Learn how you can configure your Ingress setup to meet your workload needs.
Preserving the source IP address
To preserve source IP addresses, you can enable the PROXY protocol for VPC clusters. This option is available for clusters that run version 4.13 or later.
The PROXY protocol provides a convenient way to transport connection information, such as a client's address, across multiple layers of NAT or TCP proxies. For more information on the PROXY protocol, see the HAProxy specification.
By default, the Ingress controller receives connections that contain only the source address associated with the load balancer. You can enable the PROXY protocol in VPC clusters to configure the load balancer to preserve the original client address for connections that the Ingress controller receives.
Enabling the PROXY protocol
-
Edit the Ingress Controller resource.
oc -n openshift-ingress-operator edit ingresscontroller/default -
In the Ingress controller resource, find the
spec.endpointPublishingStrategy.loadBalancersection and define the followingproviderParametersvalues.endpointPublishingStrategy: loadBalancer: providerParameters: type: IBM ibm: protocol: PROXY scope: External type: LoadBalancerService -
Save and apply the resource.
Disabling PROXY protocol
-
Edit the Ingress Controller resource.
oc -n openshift-ingress-operator edit ingresscontroller/default -
In the Ingress controller resource, find the
spec.endpointPublishingStrategy.loadBalancersection and define the followingproviderParametersvalues.endpointPublishingStrategy: loadBalancer: providerParameters: type: IBM ibm: protocol: TCP scope: External type: LoadBalancerService -
Save and apply the resource.
Customizing Ingress routing with annotations
If you want to customize routing rules for your app, you can use route-specific HAProxy annotations in the Ingress resources that you define.
These supported annotations are in the format haproxy.router.openshift.io/<annotation> or router.openshift.io/<annotation>.IBM Cloud Kubernetes Service annotations (ingress.bluemix.net/<annotation>)
and NGINX annotations (nginx.ingress.kubernetes.io/<annotation>) are not supported for the Ingress controller or the Ingress resource in Red Hat OpenShift version 4.
Enabling access logging
HTTP access logs contain information of incoming HTTP requests. Having access logs is useful when debugging complex problems, but the OpenShift router doesn't support access logging by default. For OpenShift clusters, you can configure access
logging for routers, which results in having a sidecar container in the router pod that runs a syslog server and writes access logs to its stdout.
-
Edit the
IngressControllerconfiguration and add the following configuration to the.spec.logging: access: destination: type: Container httpCaptureHeaders: request: - maxLength: 256 name: Host httpLogFormat: '{"time_date":"%t","client":"%ci","host":"%[capture.req.hdr(0)]","ssl_version":"%sslv","request_method":"%HM", "request_uri":"%HU","status":%ST,"upstream_addr":"%si:%sp","request_time":%Tt,"upstream_connect_time":%Tc, "upstream_header_time":%Tr,"termination_state":"%ts"}'Edit command
kubectl edit ingresscontroller -n openshift-ingress-operator default ingresscontroller.operator.openshift.io/default edited -
Check if the router pods are running two containers.
kubectl get pod -n openshift-ingress -wExample output
NAME READY STATUS RESTARTS AGE router-default-66945cc7c4-4xlnh 2/2 Running 0 36s router-default-66945cc7c4-5cxpn 2/2 Running 0 36s -
Check access logs.
kubectl logs -n openshift-ingress router-default-66945cc7c4-4xlnh -c logsExample output
... 2025-01-28T12:29:07.038592+00:00 router-default-7fc484bbb8-qfm5d router-default-7fc484bbb8-qfm5d haproxy[41]:{"time_date": "28/Jan/2025:12:29:06.879","client":"10.5.207.79","host":"alb-autoscale-example-service-default.pvg-classic-z9g5zltmgb1ir -1e7743ca80a399c9cff4eaf617434c72-0000.us-south.stg.kube.appdomain.cloud","ssl_version":"TLSv1.3","request_method": "GET","request_uri":"/","status":200,"upstream_addr":"172.30.210.252:8080","request_time":159,"upstream_connect_time":1, "upstream_header_time":2,"termination_state":"--"} 2025-01-28T12:29:09.572129+00:00 router-default-7fc484bbb8-qfm5d router-default-7fc484bbb8-qfm5d haproxy[41]: {"time_date": "28/Jan/2025:12:29:09.405","client":"10.5.207.79","host":"alb-autoscale-example-service-default.pvg-classic-z9g5zltmgb1ir -1e7743ca80a399c9cff4eaf617434c72-0000.us-south.stg.kube.appdomain.cloud","ssl_version":"TLSv1.3","request_method":"GET", "request_uri":"/","status":200,"upstream_addr":"172.30.210.252:8080","request_time":166,"upstream_connect_time":0, "upstream_header_time":3,"termination_state":"--"} ...
Fine-tuning connection handling
The clientTimeout and serverTimeout parameters are crucial configurations that dictate how long connections remain active between clients, the Ingress controller, and the backend servers. These timeouts play a significant
role in optimizing request handling, particularly when dealing with long-lasting client connections, delayed responses from backend servers, and safeguarding valuable resources from being unnecessarily occupied.
If you anticipate that clients will keep connections open for a longer period of time, it is advisable to increase the clientTimeout setting to cater to these scenarios. Conversely, if your backend servers have high latency due
to high traffic volumes or processing loads, adjusting the serverTimeout can provide the necessary allowance for the servers to complete their request processing before the Ingress controller terminates the connection.
You can change the above parameters in your IngressController resource:
apiVersion: operator.openshift.io/v1
kind: IngressController
...
spec:
tuningOptions:
clientTimeout: 5s
serverTimeout: 5s
For more information about tuning options see the OpenShift documentation.
Adjusting Timeouts
If your clusters are exposed with IBM Cloud Cloud Internet Services (CIS) / Cloudflare and use Web Application Firewall (WAF) or global load balancing you should set clientTimeout and serverTimeout to values exceeding
900 seconds to prevent premature connection terminations. For more information, see the Cloudflare documentation.
-
Identify your Ingress Controller: Begin by listing your
IngressControllerresources. This can be achieved with the command:oc get ingresscontrollers -n openshift-ingress-operator -
Update the timeout parameters: To modify the
clientTimeoutandserverTimeoutparameters for a specificIngressController, you can execute a patch command. For example the following command updates the timeout setting for thedefaultIngress Controllers to 905 seconds:oc patch ingresscontrollers default --patch '{"spec": {"tuningOptions":{"clientTimeout": "905s", "serverTimeout": "905s"}}}' --type=merge -n openshift-ingress-operator -
VPC clusters only: In cases where you are operating within a Virtual Private Cloud (VPC), it is necessary to tune the idle connection timeout of the VPC load balancer along with your Ingress Controller settings. We recommend choosing a greater idle connection timeout than the timeout settings of your Ingress Controller. The command below demonstrates how to update the idle connection timeout for the
router-defaultLoadBalancer service to 910 seconds:oc annotate svc -n openshift-ingress service.kubernetes.io/ibm-load-balancer-cloud-provider-vpc-idle-connection-timeout="910" router-default