We've just released a new Ingress Controller (docs here), which supersedes the operator.
Pomerium Ingress Controller addresses shortcomings in the operator and allows Pomerium to directly handle Ingress
resources without the need for an external/third-party ingress controller. Additionally, the ingress controller supports Pomerium's new policy language and other features introduced in the last year or so.
As such, pomerium-operator will no longer be receiving updates. Most practically, the operator will not be supported on Kubernetes v1.22+ due to the deprecation of the v1beta1/Ingress
API.
While it is possible to deploy the ingress controller in an "operator compatible" manner, the new project is meant to function as a first class ingress controller and we strongly recommend migrating to the native functionality. This provides higher performance, stronger security guarantees, lower complexity, and reduced error opportunities compared to using a third party ingress integration via forward-auth.
See https://github.com/pomerium/pomerium-helm/tree/master/charts/pomerium#2500-1 for upgrade steps if you'd like to continue using forward-auth and a separate proxy.
Note: Beginning in Helm chart v25.0.0
, the operator deployment has been replaced with Pomerium Ingress Controller.
An operator for running Pomerium on a Kubernetes cluster.
pomerium-operator intends to be the way to automatically configure pomerium based on the state of Ingress, Service and CRD resources in the Kubernetes API Server. It has aspects of both an Operator and a Controller and in many ways functions as an add-on Ingress Controller.
The pomerium operator should be installed with the pomerium helm chart at https://helm.pomerium.io.
The operator may be run from outside the cluster for development or testing. In this case, it will use the default configuration at ~/.kube/config
, or you may specify a kubeconfig via the KUBECONFIG
env var. Your current context from the config will be used in either case.
Due to current capabilities, the pomerium-operator is most useful when utilizing forward auth. At this time, you must provide the appropriate annotations for your ingress controller to have pomerium protect your endpoint. Examples can be found in the pomerium documentation.
With the operator installed on your cluster (typically via helm chart), it will begin watching Ingress
and Service
resources in all namespaces or the
namespace specified by the namespace
flag. Following standard ingress controller behavior, pomerium-operator will respond only to resources that match
the configured kubernetes.io/ingress.class
and kubernetes.io/service.class
annotations, or resources without any annotation at all.
For a given matching resource, pomerium-operator will process all ingress.pomerium.io/*
annotations and create a policy based on ingress host
rules (from
in pomerium policy) and backend
service names (to
in pomerium policy).
Annotations will apply to all rules defined by an ingress resource.
Services must have an ingress.pomerium.io/from
annotation or they will be ignored as invalid.
pomerium-operator uses a similar syntax for proxying to endpoints based on both Ingress and Service resources.
Policy is set by annotation, as are typical Ingress Controller semantics.
Key | Description |
---|---|
kubernetes.io/ingress.class | standard kubernetes ingress class |
kubernetes.io/service.class | class for service control. effectively signals pomerium-operator to watch/configure this resource |
pomerium.ingress.kubernetes.io/backend-protocol | set backend protocol to http or https. similar to nginx |
ingress.pomerium.io/[policy_config_key] | policy_config_key is mapped to a policy configuration of the same name in yaml form. eg, ingress.pomerium.io/allowed_groups is mapped to allowed_groups in the policy block for all service targets in this Ingress. This value should be JSON format. |
apiVersion: extensions/v1beta1
kind: Ingress
metadata:
annotations:
ingress.pomerium.io/allowed_domains: '["pomerium.io"]'
nginx.ingress.kubernetes.io/auth-signin: https://forwardauth.pomerium.io/?uri=$scheme://$host$request_uri
nginx.ingress.kubernetes.io/auth-url: https://forwardauth.pomerium.io/verify?uri=$scheme://$host$request_uri
labels:
app: grafana
chart: grafana-4.3.2
heritage: Tiller
release: prometheus
name: prometheus-grafana
spec:
rules:
- host: grafana.pomerium.io
http:
paths:
- backend:
serviceName: prometheus-grafana
servicePort: 80
path: /
This ingress:
- Sets up external auth for nginx-ingress via the
nginx.ingress.kubernetes.io
annotations - Maps
grafana.pomerium.io
to the service atprometheus-grafana
- Permits all users from domain
pomerium.io
to access this endpoint
The appropriate policy entry will be generated and injected into the pomerium config Secret
:
apiVersion: v1
stringData:
config.yaml: |
policy:
- from: https://grafana.pomerium.io
to: http://grafana.default.svc.cluster.local:80
allowed_domains:
- pomerium.io
pomerium-operator utilizes go-task for development related tasks:
task build
-
Basic CM update functionality. Provide enough functionality to implement the Forward Auth deployment model. Basically this is just policy updates being automated and compatible with the current helm chart.
-
Introduce a mutating webhook that speaks the 3 forward auth dialects and annotates your Ingress for you. Maybe introduce this configuration via CRD.
-
Get "table stakes" Ingress features into pomerium. Target model is Inverted Double Ingress or Simple Ingress. We need cert handling up to snuff, but load balancing and path based routing can be offloaded to a next-hop ingress controller or kube-proxy via Service. CRD maps which "next-hop" service to use for the IDI model from the ingress class.
-
Introduce backend load balancing via Endpoint discovery to allow for skipping a second ingress for most configurations.
-
Allow non-Ingress/Service based policy via CRD. Helm chart does conversion on the backend.
-
Pomerium deployment itself is managed by CRD. The helm chart becomes a wrapper to this CRD. Move the templating and resource generation logic into pomerium-operator.