How to rotate all secrets / credentials ¶
There are two types of credentials managed by kOps:
-
"secrets" are symmetric credentials.
-
"keypairs" are pairs of X.509 certificates and their corresponding private keys. The exceptions are "service-account" keypairs, which are stored as certificate and private key pairs, but do not use any part of the certificates other than the public keys.
Keypairs are grouped into named "keysets", according to their use. For example, the "kubernetes-ca" keyset is used for the cluster's Kubernetes general CA. Each keyset has a single primary keypair, which is the one whose private key is used. The remaining, secondary keypairs are either trusted or distrusted. The trusted keypairs, including the primary keypair, have their certificates included in relevant trust stores.
Rotating keypairs ¶
Introduced |
---|
kOps 1.22 |
You may gracefully rotate keypairs of keysets that are either Certificate Authorities
or are "service-account" by performing the following procedure. Other keypairs will be
automatically reissued by a non-dryrun kops update cluster
when their issuing
CA is rotated.
1. Create and stage new keypair ¶
Create a new keypair for each keyset that you are going to rotate. Then update the cluster and perform a rolling update. To stage all rotatable keysets, run:
kops create keypair all
kops update cluster --yes
kops rolling-update cluster --yes
Rollback procedure ¶
A failure at this stage is unlikely. To roll back this change:
- Use
kops get keypairs
to get the IDs of the newly created keysets. - Then use
kops distrust keypair
to distrust each of them by keyset and ID. - Then use
kops update cluster --yes
- Then use
kops rolling-update cluster --yes
2. Export and distribute new kubeconfig certificate-authority-data ¶
If you are rotating the Kubernetes general CA ("kubernetes-ca" or "all") and
you are not using a load balancer for the Kubernetes API with its own separate
certificate, export a new kubeconfig with the new CA certificate
included in the certificate-authority-data
field for the cluster:
kops export kubecfg
Distribute the new certificate-authority-data
to all clients of that cluster's
Kubernetes API.
Rollback procedure ¶
To roll back this change, distribute the previous kubeconfig certificate-authority-data
.
3. Promote the new keypairs ¶
Promote the new keypairs to primary with:
kops promote keypair all
kops update cluster --yes
kops rolling-update cluster --yes
On cloud providers, such as AWS, that use kops-controller to bootstrap worker nodes, after
the kops update cluster --yes
step there is a temporary impediment to node scale-up.
Instances using the new launch template will not be able to bootstrap off of old kops-controllers.
Similarly, instances using the old launch template and which have not yet bootstrapped will not
be able to bootstrap off of new kops-controllers. The subsequent rolling update will eventually
replace all instances using the old launch template.
Rollback procedure ¶
The most likely failure at this stage would be a client of the Kubernetes API that
did not get the new certificate-authority-data
and thus do not trust the
new TLS server certificate.
To roll back this change:
- Use
kops get keypairs
to get the IDs of the previous primary keysets, most likely by identifying the issue dates. - Then use
kops promote keypair
to promote each of them by keyset and ID. - Then use
kops update cluster --yes
- Then use
kops rolling-update cluster --yes
4. Export and distribute new kubeconfig admin credentials ¶
If you are rotating the Kubernetes general CA ("kubernetes-ca" or "all") and have kubeconfigs with cluster admin credentials, export new kubeconfigs with new admin credentials for the cluster:
kops export kubecfg --admin=DURATION
where DURATION
is the desired lifetime of the admin credential.
Distribute the new credentials to all clients that require them.
Rollback procedure ¶
To roll back this change, distribute the previous kubeconfig admin credentials.
5. Distrust the previous keypairs ¶
Remove trust in the previous keypairs with:
kops distrust keypair all
kops update cluster --yes
kops rolling-update cluster --yes
Rollback procedure ¶
The most likely failure at this stage would be a client of the Kubernetes API that is still using a credential issued by the previous keypair.
To roll back this change:
- Use
kops get keypairs --distrusted
to get the IDs of the previously trusted keysets, most likely by identifying the distrust dates. - Then use
kops trust keypair
to trust each of them by keyset and ID. - Then use
kops update cluster --yes
- Then use
kops rolling-update cluster --force --yes
6. Export and distribute new kubeconfig certificate-authority-data ¶
If you are rotating the Kubernetes general CA ("kubernetes-ca" or "all") and
you are not using a load balancer for the Kubernetes API with its own separate
certificate, export a new kubeconfig with the previous CA certificate
removed from the certificate-authority-data
field for the cluster:
kops export kubecfg
Distribute the new certificate-authority-data
to all clients of that cluster's
Kubernetes API.
Rollback procedure ¶
To roll back this change, distribute the previous kubeconfig certificate-authority-data
.
Rotating the API Server encryptionconfig ¶
See the Kubernetes documentation for information on how to gracefully rotate keys in the encryptionconfig.
Use kops create secret encryptionconfig --force
to update the encryptionconfig secret.
Following that, use kops update cluster --yes
and kops rolling-update cluster --yes
.
Rotating the Cilium IPSec keys ¶
See the Cilium documentation for information on how to gracefully rotate the Cilium IPSec keys.
Use kops create secret ciliumpassword --force
to update the cilium-ipsec-keys secret.
Following that, use kops update cluster --yes
and kops rolling-update cluster --yes
.
Rotating the Docker secret ¶
[TODO]
Use kops create secret dockerconfig --force
to update the Docker secret.
Following that, use kops update cluster --yes
and kops rolling-update cluster --yes
.
Legacy procedure ¶
The following is the procedure to rotate secrets and keypairs in kOps versions prior to 1.22.
This is a disruptive procedure.
1. Delete all secrets ¶
Delete all secrets & keypairs that kOps is holding:
kops get secrets | grep '^Secret' | awk '{print $2}' | xargs -I {} kops delete secret secret {}
kops get secrets | grep '^Keypair' | awk '{print $2}' | xargs -I {} kops delete secret keypair {}
2. Recreate all secrets ¶
Now run kops update
to regenerate the secrets & keypairs.
kops update cluster
kops update cluster --yes
kOps may fail to recreate all the keys on first try. If you get errors about ca key for 'ca' not being found, run kops update cluster --yes
once more.
3. Force cluster to use new secrets ¶
Now you will have to remove the etcd certificates from every master.
Find all the master IPs. One easy way of doing that is running
kops toolbox dump
Then SSH into each node and run
sudo find /mnt/ -name server.* | xargs -I {} sudo rm {}
sudo find /mnt/ -name me.* | xargs -I {} sudo rm {}
You need to reboot every node (using a rolling-update). You have to use --cloudonly
because the keypair no longer matches.
kops rolling-update cluster --cloudonly --force --yes
Re-export kubecfg with new settings:
kops export kubecfg
4. Recreate all service accounts ¶
Now the service account tokens will need to be regenerated inside the cluster:
kops toolbox dump
and find a master IP
Then ssh admin@${IP}
and run this to delete all the service account tokens:
# Delete all service account tokens in all namespaces
NS=`kubectl get namespaces -o 'jsonpath={.items[*].metadata.name}'`
for i in ${NS}; do kubectl get secrets --namespace=${i} --no-headers | grep "kubernetes.io/service-account-token" | awk '{print $1}' | xargs -I {} kubectl delete secret --namespace=$i {}; done
# Allow for new secrets to be created
sleep 60
# Bounce all pods to make use of the new service tokens
pkill -f kube-controller-manager
kubectl delete pods --all --all-namespaces
5. Verify the cluster is back up ¶
The last command from the previous section will take some time. Meanwhile you can check validation to see the cluster gradually coming back online.
kops validate cluster --wait 10m