Version v1.15 of the documentation is no longer actively maintained. The site that you are currently viewing is an archived snapshot. For up-to-date documentation, see the latest version.
v1.0.0to use a Kubebuilder-style.
The motivations for the new layout are related to bringing more flexibility to users and part of the process to Integrating Kubebuilder and Operator SDK. Because of this integration you may be referred to the Kubebuilder documentation https://book.kubebuilder.io/ for more information about certain topics. When using this document just remember to replace
$ kubebuilder <command> with
$ operator-sdk <command>.
Note: It is recommended that you have your project upgraded to the latest SDK v1.y release version before following the steps in this guide to migrate to the new layout. However, the steps might work from previous versions as well. In this case, if you find an issue which is not covered here then check the previous Migration Guides which might help out.
What was changed
deploydirectory was replaced with the
configdirectory including a new layout of Kubernetes manifests files:
- CRD manifests in
deploy/crds/are now in
- CR manifests in
deploy/crds/are now in
- Controller manifest
deploy/operator.yamlis now in
- RBAC manifests in
deployare now in
- CRD manifests in
build/Dockerfileis moved to
Dockerfilein the project root directory
What is new
Scaffolded projects now use:
- kustomize to manage Kubernetes resources needed to deploy your operator
Makefilewith helpful targets for build, test, and deployment, and to give you flexibility to tailor things to your project’s needs
- Updated metrics configuration using kube-auth-proxy, a
--metrics-bind-addressflag, and kustomize-based deployment of a Kubernetes
Serviceand prometheus operator
- Preliminary support for CLI plugins. For more info see the [plugins design document][plugins-phase1-design-doc]
PROJECTconfiguration file to store information about GVKs, plugins, and help the CLI make decisions.
Generated files with the default API versions:
apiextensions/v1for generated CRDs (
apiextensions/v1beta1was deprecated in Kubernetes
1.16and will be removed in
admissionregistration.k8s.io/v1for webhooks (
admissionregistration.k8s.io/v1beta1was deprecated in Kubernetes
1.16and will be removed in
How to migrate
The easy migration path is to initialize a new project, re-recreate APIs, then copy pre-v1.0.0 configuration files into the new project.
- Go through the installation guide.
- User authorized with
- An accessible image registry for various operator images (ex. hub.docker.com,
quay.io) and be logged in in your command line environment.
example.comis used as the registry Docker Hub namespace in these examples. Replace it with another value if using a different registry or namespace.
- Authentication and certificates if the registry is private or uses a custom CA.
Creating a new project
In Kubebuilder-style projects, CRD groups are defined using two different flags
When we initialize a new project, we need to specify the domain that all APIs in our project will share, so before creating the new project, we need to determine which domain we’re using for the APIs in our existing project.
To determine the domain, look at the
spec.group field in your CRDs in the
The domain is everything after the first DNS segment. Using
demo.example.com as an
--domain would be
So let’s create a new project with the same domain (
mkdir nginx-operator cd nginx-operator operator-sdk init --plugins=helm --domain=example.com
Now that we have our new project initialized, we need to re-create each of our APIs.
Using our API example from earlier (
demo.example.com), we’ll use
demo for the
--kind, we use
For each API in the existing project, run:
operator-sdk create api \ --group=demo \ --version=<version> \ --kind=<Kind> \ --helm-chart=<path_to_existing_project>/helm-charts/<chart>
Migrating your Custom Resource samples
Update the CR manifests in
config/samples with the values of the CRs in your existing project which are in
Check if you have custom options in the
watches.yaml file of your existing project. If so, update the new
watches.yaml file to match. In our example, it will look like:
# Use the 'create api' subcommand to add watches to this file. - group: example.com version: v1alpha1 kind: Nginx chart: helm-charts/nginx #+kubebuilder:scaffold:watch
NOTE: Do not remove the
+kubebuilder:scaffold:watch marker. It allows the tool to update the watches file when new APIs are created.
Checking RBAC Permissions
In your new project, roles are automatically generated in
If you modified these permissions manually in
deploy/role.yaml in your existing
project, you need to re-apply them in
New projects are configured to watch all namespaces by default, so they need a
ClusterRole to have the necessary permissions. Ensure that
config/rbac/role.yaml remains a
ClusterRole if you want to retain the default behavior of the new project conventions.
The following rules were used in earlier versions of helm-operator to automatically create and manage services and servicemonitors for metrics collection. If your operator’s charts don’t require these rules, they can safely be left out of the new
- apiGroups: - monitoring.coreos.com resources: - servicemonitors verbs: - get - create - apiGroups: - apps resourceNames: - nginx-operator resources: - deployments/finalizers verbs: - update
Updating your ServiceAccount
New Helm projects come with a ServiceAccount
Your project’s RoleBinding and ClusterRoleBinding subjects, and Deployment’s
that reference a ServiceAccount already refer to this new name. When you run
your project’s name will be prepended to
controller-manager, making it unique within a namespace,
much like your old
deploy/service_account.yaml. If you wish to use the old ServiceAccount,
make sure to update all RBAC bindings and your manager Deployment.
Configuring your Operator
If your existing project has customizations in
deploy/operator.yaml, they need to be ported to
config/manager/manager.yaml. If you are passing custom arguments in your deployment, make sure to also update
Note that the following environment variables are no longer used.
OPERATOR_NAMEis deprecated. It is used to define the name for a leader election config map. Operator authors should begin using
POD_NAMEwas used to enable a particular pod to hold the leader election lock when the Helm operator used the leader for life mechanism. Helm operator now uses controller-runtime’s leader with lease mechanism, and
POD_NAMEis no longer necessary.
If you are using metrics and would like to keep them exported you will need to configure
it in the
config/default/kustomization.yaml. Please see the metrics doc to know how you can perform this setup.
The default port used by the metric endpoint binds to was changed from
:8080. To continue using port
--metrics-bind-address=:8383 when you start the operator.
Verify the migration
The project can now be deployed on cluster by running the command:
make deploy IMG=example.com/nginx-operator:v0.0.1
You can troubleshoot your deployment by checking container logs:
kubectl logs deployment.apps/nginx-operator-controller-manager -n nginx-operator-system -c manager
For further steps regarding the deployment of the operator, creation of custom resources, and cleaning up of resources, see the tutorial.