Your submission was sent successfully! Close

Thank you for contacting us. A member of our team will be in touch shortly. Close

You have successfully unsubscribed! Close

Thank you for signing up for our newsletter!
In these regular emails you will find the latest updates about Ubuntu and upcoming events where you can meet our team.Close

Release notes

1.18+ck2 Bugfix release

August 12, 2020 - charmed-kubernetes-485

Fixes

Bug fixes included in this release can be found at https://launchpad.net/charmed-kubernetes/+milestone/1.18+ck2.

1.18+ck1 Bugfix release

June 11, 2020 - charmed-kubernetes-464

Before upgrading from 1.17 or earlier, please read the upgrade notes.

What's new

  • New options for custom TLS data in container runtime charms

All container runtime subordinate charms now support a custom-registry-ca option that can be used to specify a base64 encoded Certificate Authority (CA) certificate. The value set here will be installed as a system-wide trusted CA. See the related issue for more details.

For users that require custom TLS configuration per registry, the containerd subordinate charm has expanded the custom_registries config option to support ca_file, cert_file, and cert_key. These can be set for each custom registry to enable TLS without altering the system-wide trusted CAs. See the related issue for more details.

Both of the above options allow the container runtime located on kubernetes-worker units to pull containers from a registry that utilizes custom TLS certificates.

  • New memory constraint for kubeapi-load-balancer

Deploying Charmed Kubernetes now requires a minimum of 4GB of RAM for the kubeapi-load-balancer. This addresses OOM errors reported in the related issue.

  • Updated profile when deploying to LXD

An updated LXD profile has been included in kubernetes-master and kubernetes-worker charms. This resolves an issue where containers would fail to start in a LXD environment.

Fixes

Bug fixes included in this release can be found at https://launchpad.net/charmed-kubernetes/+milestone/1.18+ck1.

1.18

April 13, 2020 - charmed-kubernetes-430

Before upgrading, please read the upgrade notes.

What's new

  • New SSL options for nginx-ingress-controller

New configuration options on the kubernetes-worker charm, ingress-default-ssl-certificate and ingress-default-ssl-key, allow you to configure nginx-ingress-controller with your own SSL certificate for serving Kubernetes ingress traffic.

  • Multus support

This release of Charmed Kubernetes introduces support for Multus, a CNI provider that makes it possible to attach multiple network interfaces to your pods.

Along the way, we've also updated existing charms to make it possible for multiple CNI providers to be deployed together in the same cluster.

For more details on Multus support in Charmed Kubernetes and how to get started, please refer to the Multus documentation page.

  • CIS Benchmark 1.5.0

The cis-benchmark action now supports version 1.5.0 of the CIS Kubernetes Benchmark. See the CIS compliance page for information on running this action on Charmed Kubernetes components.

  • Containerd version hold

The version of containerd will now be held. This means that the version of containerd will not be upgraded along with the charm. To update containerd to the latest stable, currently 1.3.3, you can call the upgrade-containerd action:

juju run-action --wait containerd/0 upgrade-containerd

This will perform the upgrade and return any output. If you have more than one unit of containerd, you should run this on each unit. The upgrades can be staggered to avoid downtime.

Component Upgrades

Many of the components in Charmed Kubernetes 1.18 have been upgraded. The following list highlights some of the more notable version changes:

  • containerd 1.3.3 (see above)
  • coredns 1.6.7
  • dashboard 2.0.0-rc5
  • etcd 3.3.15
  • openstack-provider 1.17

Fixes

A list of bug fixes and other minor feature updates in this release can be found at https://launchpad.net/charmed-kubernetes/+milestone/1.18.

Notes / Known Issues

  • Heapster, InfluxDB, Grafana addons have been removed from cdk-addons

Heapster was initially [deprecated][heapster-deprecation] in 1.11; users were encouraged to move to the metrics-server for similar functionality. With 1.18, the cluster-monitoring addons (Heapster, InfluxDB, and Grafana) have been removed from the Kubernetes source tree and therefore removed from the cdk-addons snap as well. Customers relying on these addons should migrate to a metrics-server solution prior to upgrading. Note: these removals do not affect the Kubernetes Dashboard nor the methods described in Monitoring Charmed Kubernetes.

  • Containerd cannot pull images from a registry with TLS mutual authentication

An issue with the containerd charm prevents pulling images from a private container registry when TLS mutual authentication is enabled. Where possible, users can workaround this issue by disabling mutual authentication on the registry. More details can be found in the following bug:

https://bugs.launchpad.net/charm-containerd/+bug/1853653

  • New provisioner value for Cinder storage classes

The new version of the openstack-provisioner includes an upstream change to the provisioner field for storage classes using Cinder. The cdk-cinder storage class will be automatically updated, but any manually created storage classes will need to be edited and the provisioner field changed to cinder.csi.openstack.org. Existing volumes will be unaffected, but new PVCs using those storage classes will hang until the storage class is updated.

1.17+ck2 Bugfix release

March 2, 2020 - charmed-kubernetes-410

Fixes

CephFS is now supported in Charmed Kubernetes. This allows for ReadWriteMany volumes which can be attached to multiple pods. More information can be found in the storage documentation.

Additional bug fixes included in this release can be found at https://launchpad.net/charmed-kubernetes/+milestone/1.17+ck2.

1.17+ck1 Bugfix release

January 15, 2020 - charmed-kubernetes-372

Fixes

We fixed an issue where pod-to-pod network traffic was being unnecessarily masqueraded when using Flannel or Canal. More details can be found at https://launchpad.net/charmed-kubernetes/+milestone/1.17+ck1.

1.17

December 17, 2019 - charmed-kubernetes-335

Before upgrading, please read the upgrade notes.

What's new

  • CIS Benchmark

The Center for Internet Security (CIS) maintains a [Kubernetes benchmark][cis-benchmark] that is helpful to ensure clusters are deployed in accordance with security best practices. See the CIS Compliance documentation for instructions on how to run this compliance benchmark.

  • Snap Coherence

Beginning with Charmed Kubernetes 1.17, updates to Kubernetes snap packages used by kubernetes-master and kubernetes-worker charms will be applied in a controlled fashion. Known as Snap Coherence, this feature ensures snap updates are first applied to individual master units, followed by workers. If an update fails, the process is aborted before affecting the entire cluster. This feature also allows snap revisions to be controlled by a snap store proxy. See snap coherence documentation for details.

  • Nagios checks

Additional Nagios checks have been added for the kubernetes-master and kubernetes-worker charms. These checks enhance the monitoring and reporting available via Nagios by collecting data on node registration and API server connectivity.

  • Improved metrics

kube-state-metrics is now added by default to the cluster when monitoring is enabled. New default dashboards are also included to highlight these metrics with Prometheus/Grafana.

  • Storage Classes created by default

Storage classes will now be created if the kubernetes-master charm is related to an integrator charm. These classes are for AWS, GCE, OpenStack, and Azure and are named cdk-ebs, cdk-gce-pd, cdk-cinder, and cdk-azure-disk, respectively.

  • Support for etcd 3.3 and 3.4

Whilst Charmed Kubernetes 1.17 ships with etcd 3.3 by default, it also brings support for running etcd 3.4. To do so, you can simply run the following Juju command:

juju config etcd channel=3.4/stable

Component Upgrades

Many of the components in Charmed Kubernetes 1.17 have been upgraded. The following list highlights some of the more notable version changes:

  • calico 3.10.1
  • coredns 1.6.5
  • etcd 3.3
  • nfs-provisioner 3.1.0
  • nginx-ingress-controller 0.26.1

Fixes

A list of bug fixes and other minor feature updates in this release can be found at https://launchpad.net/charmed-kubernetes/+milestone/1.17.

Notes / Known Issues

  • The registry action for the kubernetes-worker charm has been deprecated and will be removed in a future release. To enable a custom container registry, please see the registry documentation.

1.16+ck2 Bugfix release

November 22, 2019 - charmed-kubernetes-316

Fixes

A list of bug fixes and other minor feature updates in this release can be found at https://launchpad.net/charmed-kubernetes/+milestone/1.16+ck2.

1.16+ck1 Bugfix release

October 4, 2019 - charmed-kubernetes-270

Fixes

A list of bug fixes and other minor feature updates in this release can be found at https://launchpad.net/charmed-kubernetes/+milestone/1.16+ck1.

1.16

September 27, 2019 - charmed-kubernetes-252

Before upgrading, please read the upgrade notes.

What's new

  • Kata Containers support

Beginning with Charmed Kubernetes 1.16, the Kata Containers runtime can be used with containerd to safely run insecure or untrusted pods. When enabled, Kata provides hypervisor isolation for pods that request it, while trusted pods can continue to run on a shared kernel via runc. For details on using Kata Containers with Charmed Kubernetes, consult the documentation.

  • AWS IAM support

Amazon AWS IAM authentication and authorisation is now supported via a subordinate charm. See AWS-IAM documentation for details on how to use AWS credentials to log in to your Charmed Kubernetes cluster.

  • SSL passthrough support

A new configuration parameter was added to the kubernetes-worker charm to enable ssl passthrough. This allows TLS termination to happen on the workload. Refer to the upstream documentation for more information.

  • Improved LXD support

LXD containers used for hosting Kubernetes components require some specific profile settings. These profiles are now embedded in the charms themselves and applied when deployed, dramatically simplifying the process of installing Charmed Kubernetes on a single machine. See the Local install documentation for the updated instructions.

  • Improved Prometheus/Grafana integration

The setup and configuration of Prometheus and Grafana has been significantly streamlined with new relations to allow the charms to manage the scraper job and dashboards. This means that monitoring can now be added by specifying a single overlay when deploying Charmed Kubernetes. Refer to the updated documentation for more information.

  • Improved OpenStack integration

The OpenStack Integrator charm can now replace the Kube API Load Balancer by providing a native OpenStack load balancer (Octavia or Neutron) to provide HA load balancing for the Kubernetes control plane. Refer to the updated documentation for more information.

  • Docker Registry with Containerd

The Docker registry charm can now be related directly to the Containerd runtime charm. Refer to the documentation for instructions on how to deploy the charm.

  • Renamed default container registry

The Canonical container image registry has a new, firewall-friendly name: image-registry.canonical.com:5000 is now rocks.canonical.com. The old URL is an alias for rocks and will continue to work. However, the default configuration for current charms has changed to the new URL.

Fixes

A list of bug fixes and other minor feature updates in this release can be found at https://launchpad.net/charmed-kubernetes/+milestone/1.16.

Special thanks to pierrop for contributing a fix to issue 1841965!

Known Issues

The Kubernetes Dashboard shipped with Charmed Kubernetes 1.16 is version 2.0.0-beta4. While unusual to ship a beta component with a stable release, in this case it was necessary, since the latest stable dashboard (v1.10.1) does not work with Kubernetes 1.16.

1.15+ck1 Bugfix release

August 15, 2019 - charmed-kubernetes-209

Fixes

A list of bug fixes and other minor feature updates in this release can be found at https://launchpad.net/charmed-kubernetes/+milestone/1.15+ck1.

1.15

June 28, 2019 - charmed-kubernetes-142

What's new

  • Containerd support

Although Docker is still supported, containerd is now the default container runtime in Charmed Kubernetes. Containerd brings significant peformance improvements and prepares the way for Charmed Kubernetes integration with Kata in the future.

Container runtime code has been moved out of the kubernetes-worker charm, and into subordinate charms (one for Docker and one for containerd). This allows the operator to swap the container runtime as desired, and even mix container runtimes within a cluster. It also allows for additional container runtimes to be supported in the future. Because this is a significant change, you are advised to read the upgrade notes before upgrading from a previous version.

  • Calico 3.x support

The Calico and Canal charms have been updated to install Calico 3.6.1 by default. For users currently running Calico 2.x, the next time you upgrade your Calico or Canal charm, the charm will automatically upgrade to Calico 3.6.1 with no user intervention required.

The Calico charm's ipip config option has been changed from a boolean to a string to allow for the addition of a new mode. This change is illustrated in the table below:

New value Old value Description
"Never" false Never use IPIP encapsulation. (The default)
"Always" true Always use IPIP encapsulation.
"CrossSubnet" \ Only use IPIP encapsulation for cross-subnet traffic.
  • Calico BGP support

Several new config options have been added to the Calico charm to support BGP functionality within Calico. These additions make it possible to configure external BGP peers, route reflectors, and multiple IP pools. For instructions on how to use the new config options, see the [CNI with Calico documentation][cni-calico].

  • Custom load balancer addresses

Support has been added to specify the IP address of an external load balancer. This support is in the kubeapi-load-balancer and the kubernetes-master charms. This allows a virtual IP address on the kubeapi-load-balancer charm or the IP address of an external load balancer. See the custom load balancer page for more information.

  • Container image registry

By default, all container images required by the deployment come from the Canonical image registry. This includes images used by the cdk-addons snap, ingress, dns, storage providers, etc. The registry can be configured with the new image-registry config option on the kubernetes-master charm.

The addons-registry config option is now deprecated. If set, this will take precedence over the new image-registry option when deploying images from the cdk-addons snap. However, the addons-registry option will be removed in 1.17. Users are encouraged to migrate to the new image-registry option as soon as possible.

Fixes

A list of bug fixes and other minor feature updates in this release can be found at https://launchpad.net/charmed-kubernetes/+milestone/1.15.

Known Issues

  • Docker-registry interface does not support containerd (bug 1833579)

When a docker-registry charm is related, kubernetes-worker units will attempt to configure the Docker daemon.json configuration file and may also attempt to use docker login to authenticate with the connected registry. This will not work in a containerd environment, as there is no daemon.json file nor docker command available to invoke.

Users relying on docker-registry to serve container images to Kubernetes deployments should continue to use the Docker subordinate runtime as outlined in the upgrade notes, under the heading "To keep Docker as the container runtime".

We intend to fix this shortly after release. For now, if you want to deploy Charmed Kubernetes on LXD, we recommend using the Docker subordinate charm instead. Instructions for this can be found in the [Container runtimes][container-runtime] section of our documentation.

  • New provisioner value for Cinder storage classes

The switch to the external cloud provider for OpenStack includes an upstream change to the provisioner field for storage classes using Cinder. A cdk-cinder storage class will be automatically created with the correct value, but any manually created storage classes will need to be edited and the provisioner field changed to csi-cinderplugin. Existing volumes will be unaffected, but new PVCs using those storage classes will hang until the storage class is updated.

1.14 Bugfix release

June 19th, 2019 - charmed-kubernetes-124

Fixes

  • Fixed leader_set being called by kubernetes-master followers (Issue)

1.14 Bugfix release

June 6th, 2019 - charmed-kubernetes-96

Fixes

  • Fixed leader_get import error in .reactive/kubernetes_master_worker_base.py (Issue)
  • Fixed kernel network tunables need better defaults and to be configurable (Issue)
  • Fixed proxy-extra-args config missing from kubernetes-master (Issue)

1.14 Bugfix release

May 23rd, 2019 - charmed-kubernetes-74

Fixes

  • Fixed missing core snap resource for etcd, kubernetes-master, kubernetes-worker, and kubernetes-e2e charms (Issue)
  • Fixed kubernetes-master charm resetting user changes to basic_auth.csv (Issue)
  • Fixed charm upgrades removing /srv/kubernetes directory (Issue)
  • Fixed docker-opts charm config being ignored on kubernetes-worker (Issue)
  • Fixed master services constantly restarting due to cert change (Issue)
  • Fixed kubernetes-worker tag error on GCP (Issue)

1.14 Bugfix release

April 23rd, 2019 - charmed-kubernetes-31

Fixes

  • Added automatic and manual cleanup for subnet tags (
  • Added action apply-manifest (Issue)
  • Added label to inform Juju of cloud (Issue)
  • Added support for loadbalancer-ips (Issue)
  • Fixed handling "not found" error message
  • Fixed snapd_refresh smashed by subordinate charm (Issue)
  • Fixed making sure cert has proper IP as well as DNS (Issue)
  • Fixed etcd charm stuck on "Requesting tls certificates" (Issue)
  • Fixed cert relation thrashing due to random SAN order (Issue)
  • Fixed contact point for keystone to be public address (Issue)
  • Fixed cluster tag not being sent to new worker applications (Issue)
  • Fixed removal of ceph relations causing trouble (Issue)
  • Fixed pause/resume actions (Issue)
  • Fixed ingress address selection to avoid fan IPs (Issue)
  • Fixed snapd_refresh handler (Issue)
  • Fixed credentials fields to allow for fallback and override

1.14 Bugfix release

April 4th, 2019 - canonical-kubernetes-471

Fixes

  • Fixed Ceph PV fails to mount in pod (Issue)
  • Fixed Problems switching from kube-dns to CoreDNS (Issue)
  • Fixed defaultbackend-s390x image
  • Fixed keystone-ssl-ca config description
  • Partial fix for using custom CA with Keystone

1.14

March 27, 2019 - canonical-kubernetes-466

What's new

  • Tigera Secure EE support

CDK extends its support for CNI solutions by adding the option of using [Tigera Secure EE][tigera-home], the enterprise-ready alternative to Calico. Users are now able to deploy CDK with Tigera Secure EE installed and subsequently configure additional features such as ElasticSearch and the CNX secure connectivity manager. For further details, please see the [CDK CNI documentation][tigera-docs]

  • Additional options for High Availability

Version 1.13 of CDK introduced support for keepalived to provide HA for the api-loadbalancer. This new release adds support for both HAcluster and MetalLB. See the relevant [HAcluster][hacluster-docs] and [MetalLB][metallb-docs] pages in the documentation, as well as the [HA overview][haoverview] for more information.

  • Added CoreDNS support

All new deployments of CDK 1.14 will install CoreDNS 1.4.0 by default instead of KubeDNS.

Existing deployments that are upgraded to CDK 1.14 will continue to use KubeDNS until the operator chooses to upgrade to CoreDNS. See the [upgrade notes][upgrade-notes] for details.

  • Docker upgrades: Docker 18.09.2 is the new default in Ubuntu. CDK now includes a charm action to simplify [upgrading Docker across a set of worker nodes][upgrading-docker].
  • Registry enhancements: Read-only mode, frontend support, and additional TLS configuration options have been added to the Docker registry charm.

  • Cloud integrations: New configuration options have been added to the vSphere (folder and respool_path) and OpenStack (ignore-volume-az, bs-version, trust-device-path) integrator charms.

Fixes

  • Added an action to upgrade Docker
  • Added better multi-client support to EasyRSA
  • Added block storage options for OpenStack
  • Added dashboard-auth config option to master
  • Added docker registry handling to master
  • Added more TLS options to Docker registry (Issue)
  • Added new folder/respool_path config for vSphere
  • Added proxy support to Docker registry (Issue)
  • Added read-only mode for Docker registry (Issue)
  • Fixed allow-privileged not enabled when Ceph relation joins
  • Fixed apt install source for VaultLocker
  • Fixed Ceph relation join not creating necessary pools
  • Fixed Ceph volume provisioning fails with "No such file or directory"
  • Fixed detecting of changed AppKV values
  • Fixed docker-ce-version config not working for non-NVIDIA configuration
  • Fixed Docker registry behavior with multiple frontends (Issue)
  • Fixed Docker registry not cleaning up old relation data (Issue)
  • Fixed Docker registry to correctly handle frontend removal (Issue)
  • Fixed Docker registry to work behind a TLS-terminating frontend (Issue)
  • Fixed error: snap "etcd" is not compatible with --classic
  • Fixed file descriptor limit on api server
  • Fixed GCP NetworkUnavailable hack when only some pods pending
  • Fixed handle_requests being called when no clients are related
  • Fixed handling of nameless and SANless server certificates
  • Fixed inconsistent cert flags
  • Fixed ingress=false not allowing custom ingress to be used
  • Fixed installing from outdated docker APT respository
  • Fixed IPv6 disabled on kubeapi-loadbalancer machines leads to error during installation
  • Fixed Keystone not working with multiple masters
  • Fixed kubeconfig should contain the VIP when keepalived used with kubeapi-load-balancer
  • Fixed metrics server for k8s 1.11
  • Fixed proxy var to apply when adding an apt-key
  • Fixed RBAC enabled results in error: unable to upgrade connection
  • Fixed registry action creating configmap in the wrong namespace
  • Fixed rules for metrics-server
  • Fixed status when writing kubeconfig file
  • Fixed "subnet not found" to be non-fatal
  • Fixed vSphere integrator charm not updating cloud-config when setting new charm defaults
  • Removed deprecated allow-privileged config from worker
  • Removed use of global / shared client certificate
  • Updated default nginx-ingress controller to 0.22.0 for amd64 and arm64

1.13 Bugfix Release

February 21, 2019 - canonical-kubernetes-435

Fixes

  • Fixed docker does not start when docker_runtime is set to nvidia
  • Fixed snapd_refresh charm option conflict

CVE-2018-18264

January 10, 2019

What happened

  • A security vulnerability was found in the Kubernetes dashboard that affected all versions of the dashboard.

A new dashboard version, v1.10.1, was released to address this vulnerability. This includes an important change to logging in to the dashboard. The Skip button is now missing from the login page and a user and password is now required. The easiest way to log in to the dashboard is to select your ~/.kube/config file and use credentials from there.

1.13 Release Notes

December 10, 2018

What's new

  • LDAP and Keystone support

Added support for LDAP-based authentication and authorisation via Keystone. Please read the documentation for details on how to enable this.

  • Vault PKI support

Added support for using Vault for PKI in place of EasyRSA. Vault is more secure and robust than EasyRSA and supports more advanced features for certificate management. See the documentation for details of how to add Vault to Charmed Kubernetes and configure it as a root or intermediary CA.

  • Encryption-at-rest support using Vault

Added support for encryption-at-rest for cluster secrets, leveraging Vault for data protection. This ensures that even the keys used to encrypt the data are protected at rest, unlike many configurations of encryption-at-rest for Kubernetes. Please see the documentation for further details.

  • Private Docker registry support

Added support for the Docker Registry charm to provide Docker images to cluster components without requiring access to public registries. Full instructions on using this feature are in the documentation.

  • Keepalived support

The keepalived charm can be used to run multiple kube-api-loadbalancers behind a virtual IP. For more details, please see the documentation.

  • Nginx update

Nginx was updated to v0.21.0, which brings a few changes of which to be aware. The first is that nginx is now in a namespace by itself, which is derived from the application name. By default this will be ingress-nginx-kubernetes-worker. The second change relates to custom configmaps. The name has changed to nginx-configuration and the configmap needs to reside in the same namespace as the nginx deployment.

Fixes

  • Added post deployment script for jaas/jujushell
  • Added support for load-balancer failover
  • Added always restart for etcd
  • Added Xenial support to Azure integrator
  • Added Bionic support to OpenStack integrator
  • Added support for ELB service-linked role
  • Added ability to configure Docker install source
  • Fixed EasyRSA does not run as an LXD container on 18.04
  • Fixed ceph volumes cannot be attached to the pods after 1.12
  • Fixed ceph volumes fail to attach with "node has no NodeID annotation"
  • Fixed ceph-xfs volumes failing to format due to "executable file not found in $PATH"
  • Fixed ceph volumes not detaching properly
  • Fixed ceph-csi addons not getting cleaned up properly
  • Fixed Calico/Canal not working with kube-proxy on master
  • Fixed issue with Canal charm not populating the kubeconfig option in 10-canal.conflist
  • Fixed cannot access logs after enabling RBAC
  • Fixed RBAC breaking prometheus/grafana metric collection
  • Fixed upstream Docker charm config option using wrong package source
  • Fixed a timing issue where ceph can appear broken when it's not
  • Fixed status when cni is not ready
  • Fixed an issue with calico-node service failures not surfacing
  • Fixed empty configuration due to timing issue with cni.
  • Fixed an issue where the calico-node service failed to start
  • Fixed updating policy definitions during upgrade-charm on AWS integrator
  • Fixed parsing credentials config value
  • Fixed pvc stuck in pending (azure-integrator)
  • Fixed updating properties of the openstack integrator charm do not propagate automatically (openstack-integrator)
  • Fixed flannel error during install hook due to incorrect resource (flannel)
  • Updated master and worker to handle upstream changes from OpenStack Integrator
  • Updated to CNI 0.7.4
  • Updated to Flannel v0.10.0
  • Updated Calico and Canal charms to Calico v2.6.12
  • Updated to latest CUDA and removed version pins of nvidia-docker stack
  • Updated to nginx-ingress-controller v0.21.0
  • Removed portmap from Calico resource
  • Removed CNI bins from flannel resource

Known issues

  • A current bug in Kubernetes could prevent the upgrade from properly deleting old pods. kubectl delete pod <pod_name> --force --grace-period=0 can be used to clean them up.

1.12 Release Notes

  • Added support for Ubuntu 18.04 (Bionic)

New deployments will get Ubuntu 18.04 machines by default. We will also continue to support Charmed Kubernetes on Ubuntu 16.04 (Xenial) machines for existing deployments.

  • Added kube-proxy to kubernetes-master

The kubernetes-master charm now installs and runs kube-proxy along with the other master services. This makes it possible for the master services to reach Service IPs within the cluster, making it easier to enable certain integrations that depend on this functionality (e.g. Keystone).

For operators of offline deployments, please note that this change may require you to attach a kube-proxy resource to kubernetes-master.

  • New kubernetes-worker charm config: kubelet-extra-config

In Kubernetes 1.10, a new KubeletConfiguration file was introduced, and many of Kubelet's command line options were moved there and marked as deprecated. In order to accomodate this change, we've introduced a new charm config to kubernetes-worker: kubelet-extra-config.

This config can be used to override KubeletConfiguration values provided by the charm, and is usable on any Canonical cluster running Kubernetes 1.10+.

The value for this config must be a YAML mapping that can be safely merged with a KubeletConfiguration file. For example:

juju config kubernetes-worker kubelet-extra-config="{evictionHard: {memory.available: 200Mi}}"

For more information about KubeletConfiguration, see upstream docs: https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/

  • Added support for Dynamic Kubelet Configuration

While we recommend kubelet-extra-config as a more robust and approachable way to configure Kubelet, we've also made it possible to configure kubelet using the Dynamic Kubelet Configuration feature that comes with Kubernetes 1.11+.

  • New etcd charm config: bind_to_all_interfaces (PR)

Default true, which retains the old behavior of binding to 0.0.0.0. Setting this to false makes etcd bind only to the addresses it expects traffic on, as determined by the configuration of Juju endpoint bindings.

Special thanks to @rmescandon for this contribution!

  • Updated proxy configuration

For operators who currently use the http-proxy, https-proxy and no-proxy Juju model configs, we recommend using the newer juju-http-proxy, juju-https-proxy and juju-no-proxy model configs instead. See the Proxy configuration page for details.

Fixes

  • Fixed kube-dns constantly restarting on 18.04
  • Fixed LXD machines not working on 18.04
  • Fixed kubernetes-worker unable to restart services after kubernetes-master leader is removed
  • Fixed kubeapi-load-balancer default timeout might be too low
  • Fixed unable to deploy on NVidia hardware

We appreciate your feedback on the documentation. You can edit this page or file a bug here.

See the guide to contributing or discuss these docs in our public Mattermost channel.