Publish to a Kubernetes cluster

So far we’ve only tested the controller code locally. Now we’d like to deploy it to a cluster and test it working.

Run the controller locally

Start a minikube cluster.

If you already have a Kubernetes cluster running, you can skip this step. The cluster must be 1.11+ if you enabled the status subresource earlier.

minikube start

Run make install to install the generated CRDs into the cluster.

make install

Run make run to run the controller process locally, talking to the kubectl-configured cluster (this will be minikube if you used minikube start earlier).

make run

In the reference project, this produces log output like this:

{"level":"info","ts":1546896846.3001022,"logger":"kubebuilder.controller","msg":"Starting Controller","controller":"samplesource-controller"}
{"level":"info","ts":1546896846.4004664,"logger":"kubebuilder.controller","msg":"Starting workers","controller":"samplesource-controller","worker count":1}

Create a sample source

In a different terminal, use kubectl apply to create a source.

kubectl apply -f config/samples

The controller should log an error saying that the sink reference is nil. In the reference project, that error looks like this:

Stacktraces in log messages have been elided for clarity.

  "level": "error",
  "ts": 1546896989.0428371,
  "logger": "kubebuilder.controller",
  "msg": "Reconciler error",
  "controller": "samplesource-controller",
  "request": "default/samplesource-sample",
  "error": "Failed to get sink URI: sink reference is nil",
  "stacktrace": "..."

Create a TestSink CRD to use as an Addressable.

If you already have an Addressable resource in your cluster, you can skip this step.

echo "apiVersion:
kind: CustomResourceDefinition
    kind: TestSink
    plural: testsinks
  scope: Namespaced
  version: v1alpha1" | kubectl apply -f -

Create a TestSink object with an Addressable status.

echo "apiVersion:
kind: TestSink
  name: foosink
    hostname:" | kubectl apply -f -

Update the source to include a reference to the sink. In the reference project, that command looks like this.

echo "apiVersion:
kind: SampleSource
  name: samplesource-sample
    kind: TestSink
    name: foosink
    namespace: default" | kubectl apply -f -

Check the controller logs in the first terminal. You should see an Updated Status log line. In the reference project, that line looks like this:

  "level": "info",
  "ts": 1546898070.4645903,
  "logger": "controller",
  "msg": "Updating Status",
  "request": { "namespace": "default", "name": "samplesource-sample" }

Verify that the source’s SinkURI was updated by the controller. In the reference project, that command looks like this.

kubectl get samplesources samplesource-sample -oyaml

We expect to see this in the output:

    kind: TestSink
    name: foosink
    namespace: default

Run the controller in cluster

Normally controllers run inside the Kubernetes cluster. This requires publishing a container image and creating several Kubernetes objects:

  • Namespace to run the controller pod in
  • StatefulSet or Deployment to manage the controller pod
  • RBAC rules granting permissions to manipulate Kubernetes resources

Export the IMG environment variable with a value equal to the desired container image URL. This URL will be different depending on your container image registry. The reference project uses Docker Hub.

export IMG=grantrodgers/samplesource-manager:latest

Run make docker-build to build the container image.

make docker-build

Notice that after running this command, the config/default/manager_image_patch.yaml file is updated with the URL of the built image.

Run make docker-push to publish the container image.

make docker-push

Run make deploy to create the resources to run the controller in the Kubernetes cluster. This will use the cluster referenced in the current kubectl context.

make deploy

Verify that the controller is running by checking for a Running pod in the correct namespace. This will be the project name suffixed with -system. In the reference project, the namespace is sample-source-system.

kubectl get pods -n sample-source-system

Now you can use the verification procedure outlined above in Create a sample source.

Next: Dispatching Events

Last modified 01.01.0001