Upgrade notes
This page is intended to deal with specific, special circumstances you may encounter when upgrading between versions of Charmed Kubernetes. The notes are organised according to the upgrade path below, but also be aware that any upgrade that spans more than one minor version may need to beware of notes in any of the intervening steps.
Upgrades to all versions deployed to Juju's localhost
LXD based cloud
There is a known issue (https://bugs.launchpad.net/juju/+bug/1904619) with container profiles not surviving an upgrade in clouds running on LXD. If your container-based applications fail to work properly after an upgrade, please see this topic on the troubleshooting page
Upgrading to 1.29
There are several important changes starting in 1.29 that will affect all users:
kubeapi-load-balancer
,kubernetes-control-plane
, andkubernetes-worker
charms can be observed using the COS rather than LMA.- Dropped specific relations and features which are outsourced to other charms
Observability Relations
These represent relations which were removed in favour of observability with the Canonical Observability Stack(COS).
LMA Relations:
nrpe-external-master
(provides:nrpe-external-master
on KCP and KW)prometheus
(provides:prometheus-manual
on KCP)scrape
(provides:prometheus
on KW)grafana
(provides:grafana-dashboard
)
In order to prepare for observability, see the Integration with COS Lite docs which can be performed following an upgrade of the charms but prior to an upgrade of the Kubernetes cluster.
kube-api-endpoint relation dropped
The kubernetes-control-plane:kube-api-endpoint
and
kubernetes-worker:kube-api-endpoint
relations have been removed since these
APIs are provided by the kube-control
relation. Ensure these two apps are
linked by kube-control
relation before removing this relation.
juju integrate kubernetes-control-plane:kube-control kubernetes-worker:kube-control
juju remove-relation kubernetes-control-plane:kube-api-endpoint kubernetes-worker:kube-api-endpoint
loadbalancer relation dropped
The kubernetes-control-plane:loadbalancer
relation has been removed in favour
of using the loadbalancer-internal
and loadbalancer-external
relations.
juju integrate kubernetes-control-plane:loadbalancer-internal kubeapi-loadbalancer
juju integrate kubernetes-control-plane:loadbalancer-external kubeapi-loadbalancer
juju remove-relation kubernetes-control-plane:loadbalancer kubeapi-loadbalancer
openstack relation deprecated
The kubernetes-control-plane:openstack
relation is being deprecated.
Integration with openstack is still important with Charmed Kubernetes, but
continues that integration through the charms openstack-cloud-controller
and
cinder-csi
. These two charms better manage versions of those deployment
integrations with Kubernetes. See openstack-integration for more details on
using these charms.
After upgrading the kubernetes-control-plane
charm, the unit may enter
blocked
status with the message: openstack relation is no longer managed
.
If you see this message, you can resolve it by removing the openstack
relation:
juju deploy openstack-cloud-controller --channel=1.29/stable
juju deploy cinder-csi --channel=1.29/stable
juju integrate openstack-cloud-controller openstack-integrator
juju integrate cinder-csi openstack-integrator
juju integrate openstack-cloud-controller:kube-control kubernetes-control-plane:kube-control
juju integrate openstack-cloud-controller:external-cloud-provider kubernetes-control-plane:external-cloud-provider
juju integrate openstack-cloud-controller:openstack openstack-integrator:clients
juju integrate kubernetes-control-plane:kube-control cinder-csi:kube-control
juju integrate openstack-integrator:clients cinder-csi:openstack
# The following could be vault:certificates instead of easyrsa:client
# Check what supplies the certificates for the kubernetes-control-plane
juju status kubernetes-control-plane --relations | grep ':certificates'
juju integrate openstack-cloud-controller:certificates easyrsa:client
juju integrate cinder-csi:certificates easyrsa:client
# Wait for the units to be active/idle, then
juju remove-relation kubernetes-control-plane:openstack openstack:clients
nvidia gpu operator deprecated
The kubernetes-control-plane
has allowed the configuration of
enable-nvidia-plugin=auto
where it would automatically detect a worker node
ready for GPU workloads and deploy the nvidia-plugin operator into the cluster.
After upgrading the kubernetes-control-plane
charm, the unit may enter
blocked
status with the message: nvidia-plugin is no longer managed
.
If you see this message, you can resolve it by following the
nvidia-gpu-operator docs to deploy a new charm. Once deployed, correcting
the config enable-nvidia-plugin
juju config kubernetes-control-plane enable-nvidia-plugin=false
ceph-client relation deprecated
The kubernetes-control-plane:ceph-client
relation is being deprecated
starting in the 1.29 release of charmed-kubernetes
Ceph integration is still a priority, but continues with the ceph-csi
charm
which integrates Ceph with Kubernetes.
After upgrading the kubernetes-control-plane
charm, the unit may enter
blocked
status with the message: ceph-client relation deprecated, use
ceph-csi charm instead
.
If you see this message, you can resolve it by removing the ceph-client
relation and deploying the ceph-csi charm to mimic the previous behaviour.
juju deploy ceph-csi --channel=1.29/stable --config release="v3.8.1"
juju integrate ceph-csi:ceph-client ceph-mon
juju remove-relation kubernetes-control-plane:ceph-client ceph-mon
juju integrate ceph-csi:kubernetes kubernetes-control-plane
Keystone/K8s Authentication management
Charmed Kubernetes was installing and managing an older version of keystone-auth which manages authentication and authorisation through Keystone.
This service is better suited to be managed externally from the
kubernetes-control-plane
charm. However, the charm provides the following
upgrade method to maintain the deployment of this service beyond 1.28.
One can determine if Keystone management is applicable with:
juju status --relations | grep kubernetes-control-plane:keystone-credentials
If this is empty, no steps regarding Keystone management are required.
If this states:
keystone:identity-credentials kubernetes-control-plane:keystone-credentials keystone-credentials regular
...then you'll need to prepare a bit before the upgrade.
Resources
The upstream Keystone docs cover keystone-auth in detail and should be the main reference for implementation details.
Keystone has two "Auth" options:
- Authentication of users only called keystone-authentication
- Authentication and authorisation of users, called keystone-authorization
Both options require the deployment and management of the
k8s-keystone-auth webhook service, a deployment which
provides a service endpoint for the kubernetes-api-server
to use as an
intermediate to interact with an external Keystone service.
Preparation
Starting from version 1.29, the kubernetes-control-plane
charm will drop the
following:
kubernetes-control-plane:keystone-credentials
relationkeystone-policy
configenable-keystone-authorization
configkeystone-ssl-ca
config
Before upgrading, it is important to capture the state of these config options:
mkdir keystone-upgrade
juju config kubernetes-control-plane keystone-policy > keystone-upgrade/keystone-policy.yaml
juju config kubernetes-control-plane enable-keystone-authorization > keystone-upgrade/keystone-authorization
juju config kubernetes-control-plane keystone-ssl-ca | base64 -d > keystone-upgrade/keystone-webhook-ca.crt
juju exec -u kubernetes-control-plane/leader -- 'cat /root/cdk/keystone/webhook.yaml' > keystone-upgrade/webhook.yaml
Migration
After upgrading, the charm will enter a blocked
state with the status
message: Keystone credential relation is no longer managed
. This indicates
that the k8s-keystone-auth
webhook service is still running, but is no longer
managed.
If keystone-upgrade/keystone-authorization
contains true
, then the webhook
should be enabled. This command adds the Keystone authorisation webhook config
and the Webhook
authorisation mode:
juju config kubernetes-control-plane \
authorization-webhook-config-file="$(cat keystone-upgrade/webhook.yaml)" \
authorization-mode="Node,RBAC,Webhook"
Finally, acknowledge the charm no longer manages Keystone by removing the relation:
juju remove-relation kubernetes-control-plane:keystone-credentials keystone
Day 2 Operations Manually
After migration, the deployment, service, secrets, and policies associated with
keystone-auth
are no longer handled by the kubernetes-control-plane
charm.
The following components remain in the cluster, unmanaged by the charm, and should be considered managed by the cluster administrators.
Deployment/kube-system/k8s-keystone-auth
Service/kube-system/k8s-keystone-auth-service
Secret/kube-system/keystone-auth-certs
ConfigMap/kube-system/k8s-auth-policy
ClusterRole/k8s-keystone-auth
Day 2 Operations via Charm
The keystone-k8s-auth
charm also provides management of these above resources.
The charm can be installed after the 1.29 upgrade, and used to manage these resource.
See keystone-k8s-auth for more details
Administrative Actions missing
The kubernetes-control-plane
and kubernetes-worker
actions list was
substantially reduced during development of 1.29. The following are no longer
present, but are slated to be reintroduced:
restart
namespace-list
namespace-create
namespace-delete
user-create
user-delete
user-list
apply-manifest
CIS-Benchmark Action missing
The kubernetes-control-plane
and kubernetes-worker
action for cis-benchmark
were removed during the development of the 1.29 charms and an engineering
decision to reintroduce these actions are on-going, but development and testing
incomplete. Details in LP#2044219
Automatic labelling of GPU nodes
While current worker nodes would remain unaffected as they would already be
labelled, the worker charm in 1.29 no longer labels the nodes with gpu=true
and cuda=true
.
Parity with this feature has been attained by using the nvidia-gpu-operator
Upgrading to 1.24
There are several important changes to 1.24 that will affect all users:
- Charms have migrated to the Charmhub.io store.
- Control-plane units will switch to a new charm named
kubernetes-control-plane
.The application in the juju model and all relations to it will remain under the same namekubernetes-master
, only the charm supporting the application will switch. See inclusive-naming for an explanation about this.
Due to these and other changes, it is recommended to follow the specific upgrade procedure described in the upgrading to 1.24 docs.
ceph-storage relation deprecated
The kubernetes-control-plane:ceph-storage
relation has been deprecated. After
upgrading the kubernetes-control-plane charm, the charm may enter blocked
status with the message:
ceph-storage relation deprecated, use ceph-client instead
.
If you see this message, you can resolve it by removing the ceph-storage relation:
juju remove-relation kubernetes-control-plane:ceph-storage ceph-mon
Upgrading to 1.19
New in 1.19, master units rely on Kubernetes secrets for authentication. Entries in the previously used "basic_auth.csv" and "known_tokens.csv" will be migrated to secrets and new kubeconfig files will be created during the upgrade. Administrators should update any existing kubeconfig files that are used outside of the cluster.
There is a known issue with etcd 3.2 and Kubernetes 1.19. If you are still running etcd 3.2, upgrade to etcd 3.4 prior to upgrading Kubernetes.
As of Kubernetes 1.19, kube-proxy's userspace proxier no longer works. Before you upgrade, check the proxy-extra-args configs to make sure that the userspace proxy mode is not being used in your cluster:
juju config kubernetes-master proxy-extra-args
juju config kubernetes-worker proxy-extra-args
If you see proxy-mode=userspace
in the charm configs, remove it, then proceed
with the upgrade.
Please follow the upgrade instructions for 1.19.
Upgrading to 1.18
CDK Addons
As stated in the release notes, from 1.18, the cluster-monitoring
addons (Heapster, InfluxDB, and Grafana)
have been removed from the Kubernetes source tree and therefore removed from the cdk-addons
snap as well.
Customers relying on these addons should migrate to a metrics-server
solution prior to upgrading.
Note: these removals do not affect the Kubernetes Dashboard nor the methods described in Monitoring Charmed Kubernetes.
Upgrading to 1.16
Docker Registry with Containerd
Prior to 1.16, some fixes were required to support using the Docker Registry charm with Containerd.
This charm, if used, is now supported through standard relations. Before upgrading,
remove any reference of the registry in the custom_registries
containerd charm configuration.
After upgrading, see the docker registry instructions for details of how to configure a registry.
Admission control plugins
In Charmed Kubernetes 1.16, the API server parameter by which additional,
non-default admission control plugins is specified has changed. The old
parameter was --admission-control
; the new parameter is --enable-admission-plugins
.
For example, prior to 1.16, The 'PodSecurityPolicy' admission plugin could be applied like this:
juju config kubernetes-master api-extra-args="admission-control=PodSecurityPolicy"
As of 1.16, this changes to:
juju config kubernetes-master api-extra-args="enable-admission-plugins=PodSecurityPolicy"
If using non-default admission plugins, be sure to upgrade your charm config accordingly after upgrading to 1.16.
Upgrading to 1.15
This upgrade switches the container runtime to make use of containerd, rather
than Docker. You have the option of keeping Docker as the container runtime,
but even in that case you must perform the upgrade steps. To facilitate
different container runtimes, the architecture of Charmed Kubernetes has
changed slightly, and the container runtime is now part of a separate,
subordinate charm rather than being included in the kubernetes-master
and
kubernetes-worker
charms.
To keep Docker as the container runtime
Docker is currently installed on your kubernetes-worker units. The Docker subordinate charm includes clean-up code to manage the transition to the new pluggable architecture. To upgrade whilst retaining Docker as the runtime, you need to additionally deploy the new charm and add relations to the master and worker components:
juju deploy docker
juju add-relation docker kubernetes-master
juju add-relation docker kubernetes-worker
The upgrade is complete, and your worker nodes will continue to use the Docker runtime. For information on configuring the Docker charm, see the Docker configuration page.
To migrate to containerd
If you intend to switch to containerd, it's recommended that you first add some temporary extra worker units. While not strictly necessary, skipping this step will result in some cluster down time. Adding temporary additional workers provides a place for keeping pods running while new workers are brought online. The temporary workers can then be removed as the pods are migrated to the new workers. The rest of these instructions assume that you have deployed temporary workers.
Deploy temporary workers
CURRENT_WORKER_REV=$(juju status | grep '^kubernetes-worker\s' | awk '{print $7}')
CURRENT_WORKER_COUNT=$(juju status | grep '^kubernetes-worker/' | wc -l | sed -e 's/^[ \t]*//')
juju deploy kubernetes-worker-$CURRENT_WORKER_REV kubernetes-worker-temp -n $CURRENT_WORKER_COUNT
Add necessary relations
juju status --relations | grep worker: | awk '{print $1,$2}' | sed 's/worker:/worker-temp:/g' | xargs -n2 juju relate
Wait for the temporary workers to become active before continuing. Upgrade the master and worker charms:
juju upgrade-charm kubernetes-master
juju upgrade-charm kubernetes-worker
The kubernetes-worker units will enter a blocked state, with status message "Connect a container runtime."
Deploy and relate the new Docker charm
This step is needed even if you do not intend to use Docker following the
upgrade. Docker is already installed on your kubernetes-worker
units, and
the Docker subordinate includes clean-up code to uninstall Docker when the
Docker charm is replaced with the containerd charm.
juju deploy docker
juju add-relation docker kubernetes-master
juju add-relation docker kubernetes-worker
Switching to Containerd
Now, pause the existing workers, which will move pods to the temporary workers.
juju run-action [unit] pause
For example:
juju run-action kubernetes-worker/0 pause --wait
juju run-action kubernetes-worker/1 pause --wait
juju run-action kubernetes-worker/2 pause --wait
One-liner:
juju status | grep ^kubernetes-worker/ | awk '{print $1}' | tr -d '*' | xargs -n1 -I '{}' juju run-action {} pause --wait
Remove Docker
This will uninstall Docker from the workers.
juju remove-relation docker kubernetes-master
juju remove-relation docker kubernetes-worker
juju remove-application docker
Deploy Containerd
juju deploy containerd
juju add-relation containerd kubernetes-master
juju add-relation containerd kubernetes-worker
Resume workers
Now we can allow pods to be rescheduled to our original workers.
juju run-action [unit] resume --wait
E.g.
juju run-action kubernetes-worker/0 resume --wait
juju run-action kubernetes-worker/1 resume --wait
juju run-action kubernetes-worker/2 resume --wait
One-liner:
juju status | grep ^kubernetes-worker/ | awk '{print $1}' | tr -d '*' | xargs -n1 -I '{}' juju run-action {} resume --wait
Cleanup
You can now pause the temporary workers to force all pods to migrate back to your "real" workers, then remove the temporary workers.
juju status | grep ^kubernetes-worker-temp/ | awk '{print $1}' | tr -d '*' | xargs -n1 -I '{}' juju run-action {} pause --wait
juju remove-application kubernetes-worker-temp
Mixing Containerd and Docker
Once you have a Containerd backed Charmed Kubernetes running, you can add Docker backed workers like so:
juju deploy kubernetes-worker kubernetes-worker-docker
juju deploy docker
juju relate docker kubernetes-worker-docker
Clusters which previously ran etcd 2.3
If your cluster has run etcd 2.3 at any point in the past, then it is strongly recommended not to upgrade the etcd charm to revision 449. Doing so will cause etcd to lose all of its data. For details, see https://bugs.launchpad.net/charm-etcd/+bug/1843497
We recommend upgrading etcd directly to charm revision 460 instead.
Upgrading to 1.14
This upgrade includes support for CoreDNS 1.4.0. All new deployments of Charmed Kubernetes will install CoreDNS by default instead of KubeDNS.
Existing deployments which are upgraded to Charmed Kubernetes 1.14 will continue to use KubeDNS until the operator chooses to upgrade to CoreDNS. To upgrade, set the new dns-provider config:
juju config kubernetes-master dns-provider=core-dns
Please be aware that changing DNS providers will momentarily interrupt DNS availability within the cluster. It is not necessary to recreate pods after the upgrade completes.
The enable-kube-dns
option has been removed to avoid confusion. The new
dns-provider
option allows you to enable or disable KubeDNS as needed.
For more information on the new dns-provider config
, see the
dns-provider config description.
Upgrading from 1.9.x to 1.10.x
This upgrade includes a transistion between major versions of etcd, from 2.3 to 3.x. Between these releases, etcd changed the way it accesses storage, so it is necessary to reconfigure this. There is more detailed information on the change and the upgrade proceedure in the upstream etcd documentation.
To make this upgrade more convenient for users of Charmed Kubernetes, a script has been prepared to manage the transition. The script can be examined here.
To use the script to update etcd, follow these steps:
1. Download the script with the command:
curl -O https://raw.githubusercontent.com/juju-solutions/cdk-etcd-2to3/master/migrate
2. Make the script executable:
chmod +x migrate
3. Execute the script:
./migrate
4. etcd OutputSed
The script should update etcd and you will see output similar to the following:
· Waiting for etcd units to be active and idle...
· Acquiring configured version of etcd...
· Upgrading etcd to version 3.0/stable from 2.3/stable.
· Waiting for etcd upgrade to 3.0/stable............................................................
· Waiting for etcd units to be active and idle....................................................
· Waiting for etcd cluster convergence...
· Stopping all Kubernetes api servers...
· Waiting for etcd cluster convergence...
· Stopping etcd...
· Migrating etcd/0...
· Migrating etcd/1...
· Migrating etcd/2...
· Starting etcd...
· Configuring storage-backend to etcd3...
· Waiting for all Kubernetes api servers to restart.......
· Done.
You can now proceed with the rest of the upgrade.
See the guide to contributing or discuss these docs in our public Mattermost channel.