# cockroach-operator **Repository Path**: mirrors_cockroachdb/cockroach-operator ## Basic Information - **Project Name**: cockroach-operator - **Description**: k8s operator for CRDB - **Primary Language**: Unknown - **License**: Apache-2.0 - **Default Branch**: master - **Homepage**: None - **GVP Project**: No ## Statistics - **Stars**: 1 - **Forks**: 0 - **Created**: 2020-09-24 - **Last Updated**: 2025-06-23 ## Categories & Tags **Categories**: Uncategorized **Tags**: None ## README # CockroachDB Kubernetes Operator The CockroachDB Kubernetes Operator deploys CockroachDB on a Kubernetes cluster. You can use the Operator to manage the configuration of a running CockroachDB cluster, including: - Authenticating certificates - Configuring resource requests and limits - Scaling the cluster - Performing a rolling upgrade ## Build Status GKE Nightly: [![GKE Nightly](https://teamcity.cockroachdb.com/guestAuth/app/rest/builds/buildType:Cockroach_CockroachOperator_Nightlies_GkeNightly/statusIcon)](https://teamcity.cockroachdb.com/viewType.html?buildTypeId=Cockroach_CockroachOperator_Nightlies_GkeNightly) OpenShift Nightly: [![OpenShift Nightly](https://teamcity.cockroachdb.com/guestAuth/app/rest/builds/buildType:Cockroach_CockroachOperator_Nightlies_OpenShiftNightly/statusIcon)](https://teamcity.cockroachdb.com/viewType.html?buildTypeId=Cockroach_CockroachOperator_Nightlies_OpenShiftNightly) ## Limitations - The Operator currently runs on GKE. VMware Tanzu, EKS, and AKS have not been tested. - The Operator does not yet support [multiple Kubernetes clusters for multi-region deployments](https://www.cockroachlabs.com/docs/stable/orchestrate-cockroachdb-with-kubernetes-multi-cluster.html#eks). - [Migrating from a deployment using the Helm Chart to the Operator](https://github.com/cockroachdb/cockroach-operator/issues/140) has not been defined or tested. - The Operator does not yet [automatically create an ingress object](https://github.com/cockroachdb/cockroach-operator/issues/76). - The Operator has not been tested with [Istio](https://istio.io/). ## Prerequisites - Kubernetes 1.18 or higher - [kubectl](https://kubernetes.io/docs/tasks/tools/install-kubectl/) - A GKE cluster (`n2-standard-4` is the minimum requirement for testing) ## Install the Operator Apply the custom resource definition (CRD) for the Operator: ``` kubectl apply -f https://raw.githubusercontent.com/cockroachdb/cockroach-operator/master/install/crds.yaml ``` Apply the Operator manifest. By default, the Operator is configured to install in the `cockroach-operator-system` namespace. To use the Operator in a custom namespace, download the [Operator manifest](https://raw.githubusercontent.com/cockroachdb/cockroach-operator/master/install/operator.yaml) and edit all instances of `namespace: cockroach-operator-system` to specify your custom namespace. Then apply this version of the manifest to the cluster with `kubectl apply -f {local-file-path}` instead of using the command below. ``` kubectl apply -f https://raw.githubusercontent.com/cockroachdb/cockroach-operator/master/install/operator.yaml ``` Validate that the Operator is running: ``` kubectl get pods ``` ``` NAME READY STATUS RESTARTS AGE cockroach-operator-6f7b86ffc4-9ppkv 1/1 Running 0 54s ``` ## Start CockroachDB Download the [`example.yaml`](https://github.com/cockroachdb/cockroach-operator/blob/master/examples/example.yaml) custom resource. > **Note:** The latest stable CockroachDB release is specified by default in `image.name`. ### Resource requests and limits By default, the Operator allocates 2 CPUs and 8Gi memory to CockroachDB in the Kubernetes pods. These resources are appropriate for `n2-standard-4` (GCP) and `m5.xlarge` (AWS) machines. On a production deployment, you should modify the `resources.requests` object in the custom resource with values appropriate for your workload. For details, see the [CockroachDB documentation](https://www.cockroachlabs.com/docs/stable/operate-cockroachdb-kubernetes.html#allocate-resources). ### Certificate signing The Operator generates and approves 1 root and 1 node certificate for the cluster. ### Apply the custom resource Apply `example.yaml`: ``` kubectl create -f example.yaml ``` Check that the pods were created: ``` kubectl get pods ``` ``` NAME READY STATUS RESTARTS AGE cockroach-operator-6f7b86ffc4-9t9zb 1/1 Running 0 3m22s cockroachdb-0 1/1 Running 0 2m31s cockroachdb-1 1/1 Running 0 102s cockroachdb-2 1/1 Running 0 46s ``` Each pod should have `READY` status soon after being created. ## Access the SQL shell To use the CockroachDB SQL client, first launch a secure pod running the `cockroach` binary. ``` kubectl create -f https://raw.githubusercontent.com/cockroachdb/cockroach-operator/master/examples/client-secure-operator.yaml ``` Get a shell into the client pod: ``` kubectl exec -it cockroachdb-client-secure -- ./cockroach sql --certs-dir=/cockroach/cockroach-certs --host=cockroachdb-public ``` If you want to [access the DB Console](#access-the-db-console), create a SQL user with a password while you're here: ``` CREATE USER roach WITH PASSWORD 'Q7gc8rEdS'; ``` Then assign `roach` to the `admin` role to enable access to [secure DB Console pages](https://www.cockroachlabs.com/docs/stable/ui-overview.html#db-console-security): ``` GRANT admin TO roach; ``` ``` \q ``` ## Access the DB Console To access the cluster's [DB Console](https://www.cockroachlabs.com/docs/stable/ui-overview.html), port-forward from your local machine to the `cockroachdb-public` service: ``` kubectl port-forward service/cockroachdb-public 8080 ``` Access the DB Console at `https://localhost:8080`. ### Scale the CockroachDB cluster > **Note:** Due to a [known issue](https://github.com/cockroachdb/cockroach-operator/issues/542), automatic pruning of PVCs is currently disabled by default. This means that after decommissioning and removing a node, the Operator will not remove the persistent volume that was mounted to its pod. If you plan to eventually scale up the cluster after scaling down, you will need to manually delete any PVCs that were orphaned by node removal before scaling up. For more information, see the [documentation](https://www.cockroachlabs.com/docs/stable/operate-cockroachdb-kubernetes.html#remove-nodes). To scale the cluster up and down, modify `nodes` in the custom resource. For details, see the [CockroachDB documentation](https://www.cockroachlabs.com/docs/stable/operate-cockroachdb-kubernetes#scale-the-cluster). Do **not** scale down to fewer than 3 nodes. This is considered an anti-pattern on CockroachDB and will cause errors. > **Note:** You must scale by updating the `nodes` value in the Operator configuration. Using `kubectl scale statefulset --replicas=4` will result in new pods immediately being terminated. ### Upgrade the CockroachDB cluster Perform a rolling upgrade by changing `image.name` in the custom resource. For details, see the [CockroachDB documentation](https://www.cockroachlabs.com/docs/stable/operate-cockroachdb-kubernetes#upgrade-the-cluster). ## Stop the CockroachDB cluster Delete the custom resource: ``` kubectl delete -f example.yaml ``` Remove the Operator: ``` kubectl delete -f https://raw.githubusercontent.com/cockroachdb/cockroach-operator/master/install/operator.yaml ``` > **Note:** If you want to delete the persistent volumes and free up the storage used by CockroachDB, be sure you have a backup copy of your data. Data **cannot** be recovered once the persistent volumes are deleted. For more information, see the [Kubernetes documentation](https://kubernetes.io/docs/tasks/run-application/delete-stateful-set/#persistent-volumes). # Releases We have a few phases to our releases. The first involves creating a new branch, updating the version, and then getting a PR merged into master with all of the generated files. Subsequent steps will need to be carried out in TeamCity and RedHat Connect. ## Creating a new release PR From a clean, up-to-date master (seriously...check), run the following where `` is the desired new version (e.g. `2.2.0`). ``` $ make release/new VERSION= ... ... $ git push origin release-$(cat version.txt) ``` This will do the following for you: - Create a new branch named `release-` - Update version.txt - Generate the manifest, bundles, etc. - Commit the changes with the message `Bump version to `. - Push to a new branch on origin (that wasn't automated) ## Tag the release After the PR is merged run the following to create the tag (you'll need to be a member of CRL to do this). git tag v$(cat version.txt) git push upstream v$(cat version.txt) ## Run Release Automation From here, the rest of the release process is done with TeamCity. A CRL team member will need to perform some manual steps in RedHat Connect as well. Ping one of us in Slack for info.