text
stringlengths
1
1k
id
int64
0
8.58k
node is initialized by cloud- controller-manager In cloud provider scenarios, kube-proxy can end up being scheduled on new worker nodes before the cloud-controller-manager has initialized the node addresses. This causes kube-proxy to fail to pick up the node's IP address properly and has knock-on effects to the proxy function managing load balancers. The following error can be seen in kube-proxy Pods: server.go:610] Failed to retrieve node IP: host IP unknown; known addresses: [] proxier.go:340] invalid nodeIP, initializing kube-proxy with 127.0.0.1 as nodeIPβ€’
8,100
A known solution is to patch the kube-proxy DaemonSet to allow scheduling it on control- plane nodes regardless of their conditions, keeping it off of other nodes until their initial guarding conditions abate: kubectl -n kube-system patch ds kube-proxy -p='{ "spec": { "template": { "spec": { "tolerations": [ { "key": "CriticalAddonsOnly", "operator": "Exists" }, { "effect": "NoSchedule", "key": "node-role.kubernetes.io/control-plane" } ] } } } }' The tracking issue for this problem is here. /usr is mounted read-only on nodes On Linux distributions such as Fedora CoreOS or Flatcar Container Linux, the directory /usr is mounted as a read-only filesystem. For flex-volume support , Kubernetes components like the kubelet and kube-controller-manager use the default path of /usr/libexec/kubernetes/kubelet- plugins/volume/exec/ , yet the flex-volume directory must be
8,101
writeable for the feature to work. Note: FlexVolume was deprecated in the Kubernetes v1.23 release. To workaround this issue, you can configure the flex-volume directory using the kubeadm configuration file . On the primary control-plane Node (created using kubeadm init ), pass the following file using --config : apiVersion : kubeadm.k8s.io/v1beta3 kind: InitConfiguration nodeRegistration : kubeletExtraArgs : volume-plugin-dir : "/opt/libexec/kubernetes/kubelet-plugins/volume/exec/" --- apiVersion : kubeadm.k8s.io/v1beta3 kind: ClusterConfiguration controllerManager : extraArgs : flex-volume-plugin-dir : "/opt/libexec/kubernetes/kubelet-plugins/volume/exec/" On joining Nodes
8,102
apiVersion : kubeadm.k8s.io/v1beta3 kind: JoinConfiguration nodeRegistration : kubeletExtraArgs : volume-plugin-dir : "/opt/libexec/kubernetes/kubelet-plugins/volume/exec/" Alternatively, you can modify /etc/fstab to make the /usr mount writeable, but please be advised that this is modifying a design principle of the Linux distribution. kubeadm upgrade plan prints out context deadline exceeded error message This error message is shown when upgrading a Kubernetes cluster with kubeadm in the case of running an external etcd. This is not a critical bug and happens because older versions of kubeadm perform a version check on the external etcd cluster. You can proceed with kubeadm upgrade apply ... . This issue is fixed as of version 1.19. kubeadm reset unmounts /var/lib/kubelet If /var/lib/kubelet is being mounted, performing a kubeadm reset will effectively unmount it. To workaround the issue, re-mount the /var/lib/kubelet directory after performing the kubeadm reset ope
8,103
ration. This is a regression introduced in kubeadm 1.15. The issue is fixed in 1.20. Cannot use the metrics-server securely in a kubeadm cluster In a kubeadm cluster, the metrics-server can be used insecurely by passing the --kubelet- insecure-tls to it. This is not recommended for production clusters. If you want to use TLS between the metrics-server and the kubelet there is a problem, since kubeadm deploys a self-signed serving certificate for the kubelet. This can cause the following errors on the side of the metrics-server: x509: certificate signed by unknown authority x509: certificate is valid for IP-foo not IP-bar See Enabling signed kubelet serving certificates to understand how to configure the kubelets in a kubeadm cluster to have properly signed serving certificates. Also see How to run the metrics-server securely . Upgrade fails due to etcd hash not changing Only applicable to upgrading a control plane node with a kubeadm binary v1.28.3 or later, where the node is curren
8,104
tly managed by kubeadm versions v1.28.0, v1.28.1 or v1.28.2
8,105
Here is the error message you may encounter: [upgrade/etcd] Failed to upgrade etcd: couldn't upgrade control plane. kubeadm has tried to recover everything into the earlier state. Errors faced: static Pod hash for component etcd on Node kinder-upgrade-control-plane-1 did not change after 5m0s: timed out waiting for the condition [upgrade/etcd] Waiting for previous etcd to become available I0907 10:10:09.109104 3704 etcd.go:588] [etcd] attempting to see if all cluster endpoints ([https://172.17.0.6:2379/ https://172.17.0.4:2379/ https://172.17.0.3:2379/]) are available 1/10 [upgrade/etcd] Etcd was rolled back and is now available static Pod hash for component etcd on Node kinder-upgrade-control-plane-1 did not change after 5m0s: timed out waiting for the condition couldn't upgrade control plane. kubeadm has tried to recover everything into the earlier state. Errors faced k8s.io/kubernetes/cmd/kubeadm/app/phases/upgrade.rollbackOldManifests cmd/kubeadm/app/phases/upgrade/staticp
8,106
ods.go:525 k8s.io/kubernetes/cmd/kubeadm/app/phases/upgrade.upgradeComponent cmd/kubeadm/app/phases/upgrade/staticpods.go:254 k8s.io/kubernetes/cmd/kubeadm/app/phases/upgrade.performEtcdStaticPodUpgrade cmd/kubeadm/app/phases/upgrade/staticpods.go:338 ... The reason for this failure is that the affected versions generate an etcd manifest file with unwanted defaults in the PodSpec. This will result in a diff from the manifest comparison, and kubeadm will expect a change in the Pod hash, but the kubelet will never update the hash. There are two way to workaround this issue if you see it in your cluster: The etcd upgrade can be skipped between the affected versions and v1.28.3 (or later) by using: kubeadm upgrade {apply|node } [version ] --etcd-upgrade =false This is not recommended in case a new etcd version was introduced by a later v1.28 patch version. Before upgrade, patch the manifest for the etcd static pod, to remove the problematic defaulted attributes: diff --git a/etc/kubernetes
8,107
/manifests/etcd_defaults.yaml b/etc/kubernetes/manifests/ etcd_origin.yaml index d807ccbe0aa..46b35f00e15 100644 --- a/etc/kubernetes/manifests/etcd_defaults.yaml +++ b/etc/kubernetes/manifests/etcd_origin.yaml @@ -43,7 +43,6 @@ spec: scheme: HTTP initialDelaySeconds: 10 periodSeconds: 10 - successThreshold: 1 timeoutSeconds: 15 name: etcd resources: @@ -59,26 +58,18 @@ spec: scheme: HTTPβ€’
8,108
initialDelaySeconds: 10 periodSeconds: 10 - successThreshold: 1 timeoutSeconds: 15 - terminationMessagePath: /dev/termination-log - terminationMessagePolicy: File volumeMounts: - mountPath: /var/lib/etcd name: etcd-data - mountPath: /etc/kubernetes/pki/etcd name: etcd-certs - dnsPolicy: ClusterFirst - enableServiceLinks: true hostNetwork: true priority: 2000001000 priorityClassName: system-node-critical - restartPolicy: Always - schedulerName: default-scheduler securityContext: seccompProfile: type: RuntimeDefault - terminationGracePeriodSeconds: 30 volumes: - hostPath: path: /etc/kubernetes/pki/etcd More information can be found in the tracking issue for this bug. Creating a cluster with kubeadm Using kubeadm , you can create a minimum viable Kubernetes cluster that conforms to best practices. In fact, you can use kubeadm to set up a cluster that will pass the Kubernetes Conformance tests . kubeadm also
8,109
supports other cluster lifecycle functions, such as bootstrap tokens and cluster upgrades. The kubeadm tool is good if you need: A simple way for you to try out Kubernetes, possibly for the first time. A way for existing users to automate setting up a cluster and test their application. A building block in other ecosystem and/or installer tools with a larger scope. You can install and use kubeadm on various machines: your laptop, a set of cloud servers, a Raspberry Pi, and more. Whether you're deploying into the cloud or on-premises, you can integrate kubeadm into provisioning systems such as Ansible or Terraform. Before you begin To follow this guide, you need: One or more machines running a deb/rpm-compatible Linux OS; for example: Ubuntu or CentOS.β€’ β€’ β€’
8,110
2 GiB or more of RAM per machine--any less leaves little room for your apps. At least 2 CPUs on the machine that you use as a control-plane node. Full network connectivity among all machines in the cluster. You can use either a public or a private network. You also need to use a version of kubeadm that can deploy the version of Kubernetes that you want to use in your new cluster. Kubernetes' version and version skew support policy applies to kubeadm as well as to Kubernetes overall. Check that policy to learn about what versions of Kubernetes and kubeadm are supported. This page is written for Kubernetes v1.29. The kubeadm tool's overall feature state is General Availability (GA). Some sub-features are still under active development. The implementation of creating the cluster may change slightly as the tool evolves, but the overall implementation should be pretty stable. Note: Any commands under kubeadm alpha are, by definition, supported on an alpha level. Objectives Install a s
8,111
ingle control-plane Kubernetes cluster Install a Pod network on the cluster so that your Pods can talk to each other Instructions Preparing the hosts Component installation Install a container runtime and kubeadm on all the hosts. For detailed instructions and other prerequisites, see Installing kubeadm . Note: If you have already installed kubeadm, see the first two steps of the Upgrading Linux nodes document for instructions on how to upgrade kubeadm. When you upgrade, the kubelet restarts every few seconds as it waits in a crashloop for kubeadm to tell it what to do. This crashloop is expected and normal. After you initialize your control-plane, the kubelet runs normally. Network setup kubeadm similarly to other Kubernetes components tries to find a usable IP on the network interfaces associated with a default gateway on a host. Such an IP is then used for the advertising and/or listening performed by a component. To find out what this IP is on a Linux host you can use: ip route sh
8,112
ow # Look for a line starting with "default via"β€’ β€’ β€’ β€’
8,113
Note: If two or more default gateways are present on the host, a Kubernetes component will try to use the first one it encounters that has a suitable global unicast IP address. While making this choice, the exact ordering of gateways might vary between different operating systems and kernel versions. Kubernetes components do not accept custom network interface as an option, therefore a custom IP address must be passed as a flag to all components instances that need such a custom configuration. Note: If the host does not have a default gateway and if a custom IP address is not passed to a Kubernetes component, the component may exit with an error. To configure the API server advertise address for control plane nodes created with both init and join, the flag --apiserver-advertise-address can be used. Preferably, this option can be set in the kubeadm API as InitConfiguration.localAPIEndpoint and JoinConfiguration.controlPlane.localAPIEndpoint . For kubelets on all nodes, the --nod
8,114
e-ip option can be passed in .nodeRegistration.kubeletExtraArgs inside a kubeadm configuration file ( InitConfiguration or JoinConfiguration ). For dual-stack see Dual-stack support with kubeadm . The IP addresses that you assign to control plane components become part of their X.509 certificates' subject alternative name fields. Changing these IP addresses would require signing new certificates and restarting the affected components, so that the change in certificate files is reflected. See Manual certificate renewal for more details on this topic. Warning: The Kubernetes project recommends against this approach (configuring all component instances with custom IP addresses). Instead, the Kubernetes maintainers recommend to setup the host network, so that the default gateway IP is the one that Kubernetes components auto-detect and use. On Linux nodes, you can use commands such as ip route to configure networking; your operating system might also provide higher level network mana
8,115
gement tools. If your node's default gateway is a public IP address, you should configure packet filtering or other security measures that protect the nodes and your cluster. Preparing the required container images This step is optional and only applies in case you wish kubeadm init and kubeadm join to not download the default container images which are hosted at registry.k8s.io . Kubeadm has commands that can help you pre-pull the required images when creating a cluster without an internet connection on its nodes. See Running kubeadm without an internet connection for more details. Kubeadm allows you to use a custom image repository for the required images. See Using custom images for more details
8,116
Initializing your control-plane node The control-plane node is the machine where the control plane components run, including etcd (the cluster database) and the API Server (which the kubectl command line tool communicates with). (Recommended) If you have plans to upgrade this single control-plane kubeadm cluster to high availability you should specify the --control-plane-endpoint to set the shared endpoint for all control-plane nodes. Such an endpoint can be either a DNS name or an IP address of a load-balancer. Choose a Pod network add-on, and verify whether it requires any arguments to be passed to kubeadm init . Depending on which third-party provider you choose, you might need to set the --pod-network-cidr to a provider-specific value. See Installing a Pod network add-on . (Optional) kubeadm tries to detect the container runtime by using a list of well known endpoints. To use different container runtime or if there are more than one installed on the provisioned node, specify
8,117
the --cri-socket argument to kubeadm . See Installing a runtime . To initialize the control-plane node run: kubeadm init <args> Considerations about apiserver-advertise-address and ControlPlaneEndpoint While --apiserver-advertise-address can be used to set the advertise address for this particular control-plane node's API server, --control-plane-endpoint can be used to set the shared endpoint for all control-plane nodes. --control-plane-endpoint allows both IP addresses and DNS names that can map to IP addresses. Please contact your network administrator to evaluate possible solutions with respect to such mapping. Here is an example mapping: 192.168.0.102 cluster-endpoint Where 192.168.0.102 is the IP address of this node and cluster-endpoint is a custom DNS name that maps to this IP. This will allow you to pass --control-plane-endpoint=cluster-endpoint to kubeadm init and pass the same DNS name to kubeadm join . Later you can modify cluster- endpoint to point to the address
8,118
of your load-balancer in an high availability scenario. Turning a single control plane cluster created without --control-plane-endpoint into a highly available cluster is not supported by kubeadm. More information For more information about kubeadm init arguments, see the kubeadm reference guide . To configure kubeadm init with a configuration file see Using kubeadm init with a configuration file .1. 2. 3
8,119
To customize control plane components, including optional IPv6 assignment to liveness probe for control plane components and etcd server, provide extra arguments to each component as documented in custom arguments . To reconfigure a cluster that has already been created see Reconfiguring a kubeadm cluster . To run kubeadm init again, you must first tear down the cluster . If you join a node with a different architecture to your cluster, make sure that your deployed DaemonSets have container image support for this architecture. kubeadm init first runs a series of prechecks to ensure that the machine is ready to run Kubernetes. These prechecks expose warnings and exit on errors. kubeadm init then downloads and installs the cluster control plane components. This may take several minutes. After it finishes you should see: Your Kubernetes control-plane has initialized successfully! To start using your cluster, you need to run the following as a regular user: mkdir -p $HOME/.kube sudo
8,120
cp -i /etc/kubernetes/admin.conf $HOME/.kube/config sudo chown $(id -u):$(id -g) $HOME/.kube/config You should now deploy a Pod network to the cluster. Run "kubectl apply -f [podnetwork].yaml" with one of the options listed at: /docs/concepts/cluster-administration/addons/ You can now join any number of machines by running the following on each node as root: kubeadm join <control-plane-host>:<control-plane-port> --token <token> --discovery-token- ca-cert-hash sha256:<hash> To make kubectl work for your non-root user, run these commands, which are also part of the kubeadm init output: mkdir -p $HOME /.kube sudo cp -i /etc/kubernetes/admin.conf $HOME /.kube/config sudo chown $(id -u ):$(id -g) $HOME /.kube/config Alternatively, if you are the root user, you can run: export KUBECONFIG =/etc/kubernetes/admin.conf Warning: The kubeconfig file admin.conf that kubeadm init generates contains a certificate with Subject: O = kubeadm:cluster-admins, CN = kubernetes-admin . The group
8,121
kubeadm:cluster-admins is bound to the built-in cluster-admin ClusterRole. Do not share the admin.conf file with anyone. kubeadm init generates another kubeconfig file super-admin.conf that contains a certificate with Subject: O = system:masters, CN = kubernetes-super-admin . system:masters is a break
8,122
glass, super user group that bypasses the authorization layer (for example RBAC). Do not share the super-admin.conf file with anyone. It is recommended to move the file to a safe location. See Generating kubeconfig files for additional users on how to use kubeadm kubeconfig user to generate kubeconfig files for additional users. Make a record of the kubeadm join command that kubeadm init outputs. You need this command to join nodes to your cluster . The token is used for mutual authentication between the control-plane node and the joining nodes. The token included here is secret. Keep it safe, because anyone with this token can add authenticated nodes to your cluster. These tokens can be listed, created, and deleted with the kubeadm token command. See the kubeadm reference guide . Installing a Pod network add-on Caution: This section contains important information about networking setup and deployment order. Read all of this advice carefully before proceeding. You must deploy a
8,123
Container Network Interface (CNI) based Pod network add-on so that your Pods can communicate with each other. Cluster DNS (CoreDNS) will not start up before a network is installed. Take care that your Pod network must not overlap with any of the host networks: you are likely to see problems if there is any overlap. (If you find a collision between your network plugin's preferred Pod network and some of your host networks, you should think of a suitable CIDR block to use instead, then use that during kubeadm init with -- pod-network-cidr and as a replacement in your network plugin's YAML). By default, kubeadm sets up your cluster to use and enforce use of RBAC (role based access control). Make sure that your Pod network plugin supports RBAC, and so do any manifests that you use to deploy it. If you want to use IPv6--either dual-stack, or single-stack IPv6 only networking--for your cluster, make sure that your Pod network plugin supports IPv6. IPv6 support was added to CNI in v0.6.0
8,124
. Note: Kubeadm should be CNI agnostic and the validation of CNI providers is out of the scope of our current e2e testing. If you find an issue related to a CNI plugin you should log a ticket in its respective issue tracker instead of the kubeadm or kubernetes issue trackers. Several external projects provide Kubernetes Pod networks using CNI, some of which also support Network Policy . See a list of add-ons that implement the Kubernetes networking model . Please refer to the Installing Addons page for a non-exhaustive list of networking addons supported by Kubernetes. You can install a Pod network add-on with the following command on the control-plane node or a node that has the kubeconfig credentials: kubectl apply -f <add-on.yaml>β€’ β€’
8,125
You can install only one Pod network per cluster. Once a Pod network has been installed, you can confirm that it is working by checking that the CoreDNS Pod is Running in the output of kubectl get pods --all-namespaces . And once the CoreDNS Pod is up and running, you can continue by joining your nodes. If your network is not working or CoreDNS is not in the Running state, check out the troubleshooting guide for kubeadm . Managed node labels By default, kubeadm enables the NodeRestriction admission controller that restricts what labels can be self-applied by kubelets on node registration. The admission controller documentation covers what labels are permitted to be used with the kubelet --node-labels option. The node- role.kubernetes.io/control-plane label is such a restricted label and kubeadm manually applies it using a privileged client after a node has been created. To do that manually you can do the same by using kubectl label and ensure it is using a privileged kubeconfig
8,126
such as the kubeadm managed /etc/kubernetes/admin.conf . Control plane node isolation By default, your cluster will not schedule Pods on the control plane nodes for security reasons. If you want to be able to schedule Pods on the control plane nodes, for example for a single machine Kubernetes cluster, run: kubectl taint nodes --all node-role.kubernetes.io/control-plane- The output will look something like: node "test-01" untainted ... This will remove the node-role.kubernetes.io/control-plane:NoSchedule taint from any nodes that have it, including the control plane nodes, meaning that the scheduler will then be able to schedule Pods everywhere. Additionally, you can execute the following command to remove the node.kubernetes.io/ exclude-from-external-load-balancers label from the control plane node, which excludes it from the list of backend servers: kubectl label nodes --all node.kubernetes.io/exclude-from-external-load-balancers- Joining your nodes The nodes are where your worklo
8,127
ads (containers and Pods, etc) run. To add new nodes to your cluster do the following for each machine: SSH to the machine Become root (e.g. sudo su - ) Install a runtime if needed Run the command that was output by kubeadm init . For example:β€’ β€’ β€’
8,128
kubeadm join --token <token> <control-plane-host>:<control-plane-port> --discovery- token-ca-cert-hash sha256:<hash> If you do not have the token, you can get it by running the following command on the control- plane node: kubeadm token list The output is similar to this: TOKEN TTL EXPIRES USAGES DESCRIPTION EXTRA GROUPS 8ewj1p.9r9hcjoqgajrj4gi 23h 2018-06-12T02:51:28Z authentication, The default bootstrap system: signing token generated by bootstrappers: 'kubeadm init'. kubeadm: default-node-token By default, tokens expire after 24 hours. If you are joining a node to the cluster after the current token has expired, you can create a new token by running the following command on the control-plane node:
8,129
kubeadm token create The output is similar to this: 5didvk.d09sbcov8ph2amjw If you don't have the value of --discovery-token-ca-cert-hash , you can get it by running the following command chain on the control-plane node: openssl x509 -pubkey -in /etc/kubernetes/pki/ca.crt | openssl rsa -pubin -outform der 2>/dev/ null | \ openssl dgst -sha256 -hex | sed 's/^.* //' The output is similar to: 8cb2de97839780a412b93877f8507ad6c94f73add17d5d7058e91741c9d5ec78 Note: To specify an IPv6 tuple for <control-plane-host>:<control-plane-port> , IPv6 address must be enclosed in square brackets, for example: [2001:db8::101]:2073 . The output should look something like: [preflight] Running pre-flight checks ... (log output of join workflow) ... Node join complete: * Certificate signing request sent to control-plane and response received. * Kubelet informed of new secure connection details. Run 'kubectl get nodes' on control-plane to see this machine join
8,130
A few seconds later, you should notice this node in the output from kubectl get nodes when run on the control-plane node. Note: As the cluster nodes are usually initialized sequentially, the CoreDNS Pods are likely to all run on the first control-plane node. To provide higher availability, please rebalance the CoreDNS Pods with kubectl -n kube-system rollout restart deployment coredns after at least one new node is joined. (Optional) Controlling your cluster from machines other than the control-plane node In order to get a kubectl on some other computer (e.g. laptop) to talk to your cluster, you need to copy the administrator kubeconfig file from your control-plane node to your workstation like this: scp root@<control-plane-host>:/etc/kubernetes/admin.conf . kubectl --kubeconfig ./admin.conf get nodes Note: The example above assumes SSH access is enabled for root. If that is not the case, you can copy the admin.conf file to be accessible by some other user and scp using that other
8,131
user instead. The admin.conf file gives the user superuser privileges over the cluster. This file should be used sparingly. For normal users, it's recommended to generate an unique credential to which you grant privileges. You can do this with the kubeadm kubeconfig user --client-name <CN> command. That command will print out a KubeConfig file to STDOUT which you should save to a file and distribute to your user. After that, grant privileges by using kubectl create (cluster)rolebinding . (Optional) Proxying API Server to localhost If you want to connect to the API Server from outside the cluster you can use kubectl proxy : scp root@<control-plane-host>:/etc/kubernetes/admin.conf . kubectl --kubeconfig ./admin.conf proxy You can now access the API Server locally at http://localhost:8001/api/v1 Clean up If you used disposable servers for your cluster, for testing, you can switch those off and do no further clean up. You can use kubectl config delete-cluster to delete your local refer
8,132
ences to the cluster. However, if you want to deprovision your cluster more cleanly, you should first drain the node and make sure that the node is empty, then deconfigure the node. Remove the node Talking to the control-plane node with the appropriate credentials, run
8,133
kubectl drain <node name> --delete-emptydir-data --force --ignore-daemonsets Before removing the node, reset the state installed by kubeadm : kubeadm reset The reset process does not reset or clean up iptables rules or IPVS tables. If you wish to reset iptables, you must do so manually: iptables -F && iptables -t nat -F && iptables -t mangle -F && iptables -X If you want to reset the IPVS tables, you must run the following command: ipvsadm -C Now remove the node: kubectl delete node <node name> If you wish to start over, run kubeadm init or kubeadm join with the appropriate arguments. Clean up the control plane You can use kubeadm reset on the control plane host to trigger a best-effort clean up. See the kubeadm reset reference documentation for more information about this subcommand and its options. Version skew policy While kubeadm allows version skew against some components that it manages, it is recommended that you match the kubeadm version with the versions of the control pla
8,134
ne components, kube-proxy and kubelet. kubeadm's skew against the Kubernetes version kubeadm can be used with Kubernetes components that are the same version as kubeadm or one version older. The Kubernetes version can be specified to kubeadm by using the -- kubernetes-version flag of kubeadm init or the ClusterConfiguration.kubernetesVersion field when using --config . This option will control the versions of kube-apiserver, kube-controller- manager, kube-scheduler and kube-proxy. Example: kubeadm is at 1.29 kubernetesVersion must be at 1.29 or 1.28 kubeadm's skew against the kubelet Similarly to the Kubernetes version, kubeadm can be used with a kubelet version that is the same version as kubeadm or three versions older.β€’
8,135
Example: kubeadm is at 1.29 kubelet on the host must be at 1.29, 1.28, 1.27 or 1.26 kubeadm's skew against kubeadm There are certain limitations on how kubeadm commands can operate on existing nodes or whole clusters managed by kubeadm. If new nodes are joined to the cluster, the kubeadm binary used for kubeadm join must match the last version of kubeadm used to either create the cluster with kubeadm init or to upgrade the same node with kubeadm upgrade . Similar rules apply to the rest of the kubeadm commands with the exception of kubeadm upgrade . Example for kubeadm join : kubeadm version 1.29 was used to create a cluster with kubeadm init Joining nodes must use a kubeadm binary that is at version 1.29 Nodes that are being upgraded must use a version of kubeadm that is the same MINOR version or one MINOR version newer than the version of kubeadm used for managing the node. Example for kubeadm upgrade : kubeadm version 1.28 was used to create or upgrade the node The version of kube
8,136
adm used for upgrading the node must be at 1.28 or 1.29 To learn more about the version skew between the different Kubernetes component see the Version Skew Policy . Limitations Cluster resilience The cluster created here has a single control-plane node, with a single etcd database running on it. This means that if the control-plane node fails, your cluster may lose data and may need to be recreated from scratch. Workarounds: Regularly back up etcd . The etcd data directory configured by kubeadm is at /var/lib/etcd on the control-plane node. Use multiple control-plane nodes. You can read Options for Highly Available topology to pick a cluster topology that provides high-availability . Platform compatibility kubeadm deb/rpm packages and binaries are built for amd64, arm (32-bit), arm64, ppc64le, and s390x following the multi-platform proposal . Multiplatform container images for the control plane and addons are also supported since v1.12.β€’ β€’ β€’ β€’ β€’ β€’ β€’
8,137
Only some of the network providers offer solutions for all platforms. Please consult the list of network providers above or the documentation from each provider to figure out whether the provider supports your chosen platform. Troubleshooting If you are running into difficulties with kubeadm, please consult our troubleshooting docs . What's next Verify that your cluster is running properly with Sonobuoy See Upgrading kubeadm clusters for details about upgrading your cluster using kubeadm . Learn about advanced kubeadm usage in the kubeadm reference documentation Learn more about Kubernetes concepts and kubectl . See the Cluster Networking page for a bigger list of Pod network add-ons. See the list of add-ons to explore other add-ons, including tools for logging, monitoring, network policy, visualization & control of your Kubernetes cluster. Configure how your cluster handles logs for cluster events and from applications running in Pods. See Logging Architecture for an overview of
8,138
what is involved. Feedback For bugs, visit the kubeadm GitHub issue tracker For support, visit the #kubeadm Slack channel General SIG Cluster Lifecycle development Slack channel: #sig-cluster-lifecycle SIG Cluster Lifecycle SIG information SIG Cluster Lifecycle mailing list: kubernetes-sig-cluster-lifecycle Customizing components with the kubeadm API This page covers how to customize the components that kubeadm deploys. For control plane components you can use flags in the ClusterConfiguration structure or patches per-node. For the kubelet and kube-proxy you can use KubeletConfiguration and KubeProxyConfiguration , accordingly. All of these options are possible via the kubeadm configuration API. For more details on each field in the configuration you can navigate to our API reference pages . Note: Customizing the CoreDNS deployment of kubeadm is currently not supported. You must manually patch the kube-system/coredns ConfigMap and recreate the CoreDNS Pods after that. Alternati
8,139
vely, you can skip the default CoreDNS deployment and deploy your own variant. For more details on that see Using init phases with kubeadm . Note: To reconfigure a cluster that has already been created see Reconfiguring a kubeadm cluster .β€’ β€’ β€’ β€’ β€’ β€’ β€’ β€’ β€’ β€’ β€’
8,140
Customizing the control plane with flags in ClusterConfiguration The kubeadm ClusterConfiguration object exposes a way for users to override the default flags passed to control plane components such as the APIServer, ControllerManager, Scheduler and Etcd. The components are defined using the following structures: apiServer controllerManager scheduler etcd These structures contain a common extraArgs field, that consists of key: value pairs. To override a flag for a control plane component: Add the appropriate extraArgs to your configuration. Add flags to the extraArgs field. Run kubeadm init with --config <YOUR CONFIG YAML> . Note: You can generate a ClusterConfiguration object with default values by running kubeadm config print init-defaults and saving the output to a file of your choice. Note: The ClusterConfiguration object is currently global in kubeadm clusters. This means that any flags that you add, will apply to all instances of the same component on different nodes
8,141
. To apply individual configuration per component on different nodes you can use patches . Note: Duplicate flags (keys), or passing the same flag --foo multiple times, is currently not supported. To workaround that you must use patches . APIServer flags For details, see the reference documentation for kube-apiserver . Example usage: apiVersion : kubeadm.k8s.io/v1beta3 kind: ClusterConfiguration kubernetesVersion : v1.16.0 apiServer : extraArgs : anonymous-auth : "false" enable-admission-plugins : AlwaysPullImages,DefaultStorageClass audit-log-path : /home/johndoe/audit.log ControllerManager flags For details, see the reference documentation for kube-controller-manager . Example usage: apiVersion : kubeadm.k8s.io/v1beta3 kind: ClusterConfiguration kubernetesVersion : v1.16.0 controllerManager :β€’ β€’ β€’ β€’ 1. 2. 3
8,142
extraArgs : cluster-signing-key-file : /home/johndoe/keys/ca.key deployment-controller-sync-period : "50" Scheduler flags For details, see the reference documentation for kube-scheduler . Example usage: apiVersion : kubeadm.k8s.io/v1beta3 kind: ClusterConfiguration kubernetesVersion : v1.16.0 scheduler : extraArgs : config : /etc/kubernetes/scheduler-config.yaml extraVolumes : - name : schedulerconfig hostPath : /home/johndoe/schedconfig.yaml mountPath : /etc/kubernetes/scheduler-config.yaml readOnly : true pathType : "File" Etcd flags For details, see the etcd server documentation . Example usage: apiVersion : kubeadm.k8s.io/v1beta3 kind: ClusterConfiguration etcd: local : extraArgs : election-timeout : 1000 Customizing with patches FEATURE STATE: Kubernetes v1.22 [beta] Kubeadm allows you to pass a directory with patch files to InitConfiguration and JoinConfiguration on individual nodes. These patches can be used as the last
8,143
customization step before component configuration is written to disk. You can pass this file to kubeadm init with --config <YOUR CONFIG YAML> : apiVersion : kubeadm.k8s.io/v1beta3 kind: InitConfiguration patches : directory : /home/user/somedir Note: For kubeadm init you can pass a file containing both a ClusterConfiguration and InitConfiguration separated by ---
8,144
You can pass this file to kubeadm join with --config <YOUR CONFIG YAML> : apiVersion : kubeadm.k8s.io/v1beta3 kind: JoinConfiguration patches : directory : /home/user/somedir The directory must contain files named target[suffix][+patchtype].extension . For example, kube-apiserver0+merge.yaml or just etcd.json . target can be one of kube-apiserver , kube-controller-manager , kube-scheduler , etcd and kubeletconfiguration . patchtype can be one of strategic , merge or json and these must match the patching formats supported by kubectl . The default patchtype is strategic . extension must be either json or yaml . suffix is an optional string that can be used to determine which patches are applied first alpha-numerically. Note: If you are using kubeadm upgrade to upgrade your kubeadm nodes you must again provide the same patches, so that the customization is preserved after upgrade. To do that you can use the --patches flag, which must point to the same directory. kubeadm up
8,145
grade currently does not support a configuration API structure that can be used for the same purpose. Customizing the kubelet To customize the kubelet you can add a KubeletConfiguration next to the ClusterConfiguration or InitConfiguration separated by --- within the same configuration file. This file can then be passed to kubeadm init and kubeadm will apply the same base KubeletConfiguration to all nodes in the cluster. For applying instance-specific configuration over the base KubeletConfiguration you can use the kubeletconfiguration patch target . Alternatively, you can use kubelet flags as overrides by passing them in the nodeRegistration.kubeletExtraArgs field supported by both InitConfiguration and JoinConfiguration . Some kubelet flags are deprecated, so check their status in the kubelet reference documentation before using them. For additional details see Configuring each kubelet in your cluster using kubeadm Customizing kube-proxy To customize kube-proxy you can pa
8,146
ss a KubeProxyConfiguration next your ClusterConfiguration or InitConfiguration to kubeadm init separated by ---. For more details you can navigate to our API reference pages . Note: kubeadm deploys kube-proxy as a DaemonSet , which means that the KubeProxyConfiguration would apply to all instances of kube-proxy in the cluster.β€’ β€’ β€’
8,147
Options for Highly Available Topology This page explains the two options for configuring the topology of your highly available (HA) Kubernetes clusters. You can set up an HA cluster: With stacked control plane nodes, where etcd nodes are colocated with control plane nodes With external etcd nodes, where etcd runs on separate nodes from the control plane You should carefully consider the advantages and disadvantages of each topology before setting up an HA cluster. Note: kubeadm bootstraps the etcd cluster statically. Read the etcd Clustering Guide for more details. Stacked etcd topology A stacked HA cluster is a topology where the distributed data storage cluster provided by etcd is stacked on top of the cluster formed by the nodes managed by kubeadm that run control plane components. Each control plane node runs an instance of the kube-apiserver , kube-scheduler , and kube- controller-manager . The kube-apiserver is exposed to worker nodes using a load balancer. Each control plane
8,148
node creates a local etcd member and this etcd member communicates only with the kube-apiserver of this node. The same applies to the local kube-controller-manager and kube-scheduler instances. This topology couples the control planes and etcd members on the same nodes. It is simpler to set up than a cluster with external etcd nodes, and simpler to manage for replication. However, a stacked cluster runs the risk of failed coupling. If one node goes down, both an etcd member and a control plane instance are lost, and redundancy is compromised. You can mitigate this risk by adding more control plane nodes. You should therefore run a minimum of three stacked control plane nodes for an HA cluster. This is the default topology in kubeadm. A local etcd member is created automatically on control plane nodes when using kubeadm init and kubeadm join --control-plane . Stacked etcd topology External etcd topology An HA cluster with external etcd is a topology where the distributed data stora
8,149
ge cluster provided by etcd is external to the cluster formed by the nodes that run control plane components. Like the stacked etcd topology, each control plane node in an external etcd topology runs an instance of the kube-apiserver , kube-scheduler , and kube-controller-manager . And the kube-β€’
8,150
apiserver is exposed to worker nodes using a load balancer. However, etcd members run on separate hosts, and each etcd host communicates with the kube-apiserver of each control plane node. This topology decouples the control plane and etcd member. It therefore provides an HA setup where losing a control plane instance or an etcd member has less impact and does not affect the cluster redundancy as much as the stacked HA topology. However, this topology requires twice the number of hosts as the stacked HA topology. A minimum of three hosts for control plane nodes and three hosts for etcd nodes are required for an HA cluster with this topology. External etcd topology What's next Set up a highly available cluster with kubeadm Creating Highly Available Clusters with kubeadm This page explains two different approaches to setting up a highly available Kubernetes cluster using kubeadm: With stacked control plane nodes. This approach requires less infrastructure. The etcd members and control
8,151
plane nodes are co-located. With an external etcd cluster. This approach requires more infrastructure. The control plane nodes and etcd members are separated. Before proceeding, you should carefully consider which approach best meets the needs of your applications and environment. Options for Highly Available topology outlines the advantages and disadvantages of each. If you encounter issues with setting up the HA cluster, please report these in the kubeadm issue tracker . See also the upgrade documentation . Caution: This page does not address running your cluster on a cloud provider. In a cloud environment, neither approach documented here works with Service objects of type LoadBalancer, or with dynamic PersistentVolumes. Before you begin The prerequisites depend on which topology you have selected for your cluster's control plane: Stacked etcd External etcdβ€’ β€’ β€’ β€’
8,152
You need: Three or more machines that meet kubeadm's minimum requirements for the control- plane nodes. Having an odd number of control plane nodes can help with leader selection in the case of machine or zone failure. including a container runtime , already set up and working Three or more machines that meet kubeadm's minimum requirements for the workers including a container runtime, already set up and working Full network connectivity between all machines in the cluster (public or private network) Superuser privileges on all machines using sudo You can use a different tool; this guide uses sudo in the examples. SSH access from one device to all nodes in the system kubeadm and kubelet already installed on all machines. See Stacked etcd topology for context. You need: Three or more machines that meet kubeadm's minimum requirements for the control- plane nodes. Having an odd number of control plane nodes can help with leader selection in the case of machine or zone failure. incl
8,153
uding a container runtime , already set up and working Three or more machines that meet kubeadm's minimum requirements for the workers including a container runtime, already set up and working Full network connectivity between all machines in the cluster (public or private network) Superuser privileges on all machines using sudo You can use a different tool; this guide uses sudo in the examples. SSH access from one device to all nodes in the system kubeadm and kubelet already installed on all machines. And you also need: Three or more additional machines, that will become etcd cluster members. Having an odd number of members in the etcd cluster is a requirement for achieving optimal voting quorum. These machines again need to have kubeadm and kubelet installed. These machines also require a container runtime, that is already set up and working. See External etcd topology for context. Container images Each host should have access read and fetch images from the Kubernetes containe
8,154
r image registry, registry.k8s.io . If you want to deploy a highly-available cluster where the hosts do not have access to pull images, this is possible. You must ensure by some other means that the correct container images are already available on the relevant hosts. Command line interface To manage Kubernetes once your cluster is set up, you should install kubectl on your PC. It is also useful to install the kubectl tool on each control plane node, as this can be helpful for troubleshooting.β€’ β—¦ β€’ β—¦ β€’ β€’ β—¦ β€’ β€’ β€’ β—¦ β€’ β—¦ β€’ β€’ β—¦ β€’ β€’ β€’ β—¦
8,155
First steps for both methods Create load balancer for kube-apiserver Note: There are many configurations for load balancers. The following example is only one option. Your cluster requirements may need a different configuration. Create a kube-apiserver load balancer with a name that resolves to DNS. In a cloud environment you should place your control plane nodes behind a TCP forwarding load balancer. This load balancer distributes traffic to all healthy control plane nodes in its target list. The health check for an apiserver is a TCP check on the port the kube-apiserver listens on (default value :6443 ). It is not recommended to use an IP address directly in a cloud environment. The load balancer must be able to communicate with all control plane nodes on the apiserver port. It must also allow incoming traffic on its listening port. Make sure the address of the load balancer always matches the address of kubeadm's ControlPlaneEndpoint . Read the Options for Software Load Balancing
8,156
guide for more details. Add the first control plane node to the load balancer, and test the connection: nc -v <LOAD_BALANCER_IP> <PORT> A connection refused error is expected because the API server is not yet running. A timeout, however, means the load balancer cannot communicate with the control plane node. If a timeout occurs, reconfigure the load balancer to communicate with the control plane node. Add the remaining control plane nodes to the load balancer target group. Stacked control plane and etcd nodes Steps for the first control plane node Initialize the control plane: sudo kubeadm init --control-plane-endpoint "LOAD_BALANCER_DNS:LOAD_BALANC ER_PORT" --upload-certs You can use the --kubernetes-version flag to set the Kubernetes version to use. It is recommended that the versions of kubeadm, kubelet, kubectl and Kubernetes match. The --control-plane-endpoint flag should be set to the address or DNS and port of the load balancer. The --upload-certs flag is used to upload the
8,157
certificates that should be shared across all the control-plane instances to the cluster. If instead, you prefer to copy1. β—¦ β—¦ β—¦ β—¦ β—¦ 2. 3. 1. β—¦ β—¦
8,158
certs across control-plane nodes manually or using automation tools, please remove this flag and refer to Manual certificate distribution section below. Note: The kubeadm init flags --config and --certificate-key cannot be mixed, therefore if you want to use the kubeadm configuration you must add the certificateKey field in the appropriate config locations (under InitConfiguration and JoinConfiguration: controlPlane ). Note: Some CNI network plugins require additional configuration, for example specifying the pod IP CIDR, while others do not. See the CNI network documentation . To add a pod CIDR pass the flag --pod-network-cidr , or if you are using a kubeadm configuration file set the podSubnet field under the networking object of ClusterConfiguration . The output looks similar to: ... You can now join any number of control-plane node by running the following command on each as a root: kubeadm join 192.168.0.200:6443 --token 9vr73a.a8uxyaju799qwdjv --discovery-token
8,159
- ca-cert-hash sha256:7c2e69131a36ae2a042a339b33381c6d0d43887e2de83720eff5359e26aec866 --control- plane --certificate-key f8902e114ef118304e561c3ecd4d0b543adc226b7a07f675f56564185ffe0c07 Please note that the certificate-key gives access to cluster sensitive data, keep it secret! As a safeguard, uploaded-certs will be deleted in two hours; If necessary, you can use kubeadm init phase upload-certs to reload certs afterward. Then you can join any number of worker nodes by running the following on each as root: kubeadm join 192.168.0.200:6443 --token 9vr73a.a8uxyaju799qwdjv --discovery-token- ca-cert-hash sha256:7c2e69131a36ae2a042a339b33381c6d0d43887e2de83720eff5359e26aec866 Copy this output to a text file. You will need it later to join control plane and worker nodes to the cluster. When --upload-certs is used with kubeadm init , the certificates of the primary control plane are encrypted and uploaded in the kubeadm-certs Secret. To re-upload the certificates and generate a new
8,160
decryption key, use the following command on a control plane node that is already joined to the cluster: sudo kubeadm init phase upload-certs --upload-certs You can also specify a custom --certificate-key during init that can later be used by join. To generate such a key you can use the following command: kubeadm certs certificate-key The certificate key is a hex encoded string that is an AES key of size 32 bytes. Note: The kubeadm-certs Secret and the decryption key expire after two hours. Caution: As stated in the command output, the certificate key gives access to cluster sensitive data, keep it secret!β—¦ β—¦ β—¦
8,161
Apply the CNI plugin of your choice: Follow these instructions to install the CNI provider. Make sure the configuration corresponds to the Pod CIDR specified in the kubeadm configuration file (if applicable). Note: You must pick a network plugin that suits your use case and deploy it before you move on to next step. If you don't do this, you will not be able to launch your cluster properly. Type the following and watch the pods of the control plane components get started: kubectl get pod -n kube-system -w Steps for the rest of the control plane nodes For each additional control plane node you should: Execute the join command that was previously given to you by the kubeadm init output on the first node. It should look something like this: sudo kubeadm join 192.168.0.200:6443 --token 9vr73a.a8uxyaju799qwdjv --discovery- token-ca-cert-hash sha256:7c2e69131a36ae2a042a339b33381c6d0d43887e2de83720eff5359e26aec866 --control- plane --certificate-key f8902e114ef118304e561c3ecd4d0b543adc226
8,162
b7a07f675f56564185ffe0c07 The --control-plane flag tells kubeadm join to create a new control plane. The --certificate-key ... will cause the control plane certificates to be downloaded from the kubeadm-certs Secret in the cluster and be decrypted using the given key. You can join multiple control-plane nodes in parallel. External etcd nodes Setting up a cluster with external etcd nodes is similar to the procedure used for stacked etcd with the exception that you should setup etcd first, and you should pass the etcd information in the kubeadm config file. Set up the etcd cluster Follow these instructions to set up the etcd cluster. Set up SSH as described here. Copy the following files from any etcd node in the cluster to the first control plane node: export CONTROL_PLANE ="[email protected]" scp /etc/kubernetes/pki/etcd/ca.crt "${CONTROL_PLANE }": scp /etc/kubernetes/pki/apiserver-etcd-client.crt "${CONTROL_PLANE }": scp /etc/kubernetes/pki/apiserver-etcd-client.key "${CONTROL_PLA
8,163
NE }": Replace the value of CONTROL_PLANE with the user@host of the first control- plane node.2. 3. 1. β—¦ β—¦ 1. 2. 3.
8,164
Set up the first control plane node Create a file called kubeadm-config.yaml with the following contents: --- apiVersion : kubeadm.k8s.io/v1beta3 kind: ClusterConfiguration kubernetesVersion : stable controlPlaneEndpoint : "LOAD_BALANCER_DNS:LOAD_BALANCER_PORT" # change this (see below) etcd: external : endpoints : - https://ETCD_0_IP:2379 # change ETCD_0_IP appropriately - https://ETCD_1_IP:2379 # change ETCD_1_IP appropriately - https://ETCD_2_IP:2379 # change ETCD_2_IP appropriately caFile : /etc/kubernetes/pki/etcd/ca.crt certFile : /etc/kubernetes/pki/apiserver-etcd-client.crt keyFile : /etc/kubernetes/pki/apiserver-etcd-client.key Note: The difference between stacked etcd and external etcd here is that the external etcd setup requires a configuration file with the etcd endpoints under the external object for etcd. In the case of the stacked etcd topology, this is managed automatically. Replace the following variables in the config temp
8,165
late with the appropriate values for your cluster: LOAD_BALANCER_DNS LOAD_BALANCER_PORT ETCD_0_IP ETCD_1_IP ETCD_2_IP The following steps are similar to the stacked etcd setup: Run sudo kubeadm init --config kubeadm-config.yaml --upload-certs on this node. Write the output join commands that are returned to a text file for later use. Apply the CNI plugin of your choice. Note: You must pick a network plugin that suits your use case and deploy it before you move on to next step. If you don't do this, you will not be able to launch your cluster properly. Steps for the rest of the control plane nodes The steps are the same as for the stacked etcd setup: Make sure the first control plane node is fully initialized. Join each control plane node with the join command you saved to a text file. It's recommended to join the control plane nodes one at a time. Don't forget that the decryption key from --certificate-key expires after two hours, by default.1. β—¦ β–ͺ β–ͺ β–ͺ β–ͺ β–ͺ 1. 2. 3. β€’ β€’
8,166
Common tasks after bootstrapping control plane Install workers Worker nodes can be joined to the cluster with the command you stored previously as the output from the kubeadm init command: sudo kubeadm join 192.168.0.200:6443 --token 9vr73a.a8uxyaju799qwdjv --discovery-token-ca- cert-hash sha256:7c2e69131a36ae2a042a339b33381c6d0d43887e2de83720eff5359e26aec866 Manual certificate distribution If you choose to not use kubeadm init with the --upload-certs flag this means that you are going to have to manually copy the certificates from the primary control plane node to the joining control plane nodes. There are many ways to do this. The following example uses ssh and scp: SSH is required if you want to control all nodes from a single machine. Enable ssh-agent on your main device that has access to all other nodes in the system: eval $(ssh-agent) Add your SSH identity to the session: ssh-add ~/.ssh/path_to_private_key SSH between nodes to check that the connection is working correctly. W
8,167
hen you SSH to any node, add the -A flag. This flag allows the node that you have logged into via SSH to access the SSH agent on your PC. Consider alternative methods if you do not fully trust the security of your user session on the node. ssh -A 10.0.0.7 When using sudo on any node, make sure to preserve the environment so SSH forwarding works: sudo -E -s After configuring SSH on all the nodes you should run the following script on the first control plane node after running kubeadm init . This script will copy the certificates from the first control plane node to the other control plane nodes: In the following example, replace CONTROL_PLANE_IPS with the IP addresses of the other control plane nodes. USER =ubuntu # customizable CONTROL_PLANE_IPS ="10.0.0.7 10.0.0.8" for host in ${CONTROL_PLANE_IPS }; do scp /etc/kubernetes/pki/ca.crt "${USER }"@$host : scp /etc/kubernetes/pki/ca.key "${USER }"@$host :1. 2. 3. β—¦ β—¦ 4
8,168
scp /etc/kubernetes/pki/sa.key "${USER }"@$host : scp /etc/kubernetes/pki/sa.pub "${USER }"@$host : scp /etc/kubernetes/pki/front-proxy-ca.crt "${USER }"@$host : scp /etc/kubernetes/pki/front-proxy-ca.key "${USER }"@$host : scp /etc/kubernetes/pki/etcd/ca.crt "${USER }"@$host :etcd-ca.crt # Skip the next line if you are using external etcd scp /etc/kubernetes/pki/etcd/ca.key "${USER }"@$host :etcd-ca.key done Caution: Copy only the certificates in the above list. kubeadm will take care of generating the rest of the certificates with the required SANs for the joining control- plane instances. If you copy all the certificates by mistake, the creation of additional nodes could fail due to a lack of required SANs. Then on each joining control plane node you have to run the following script before running kubeadm join . This script will move the previously copied certificates from the home directory to /etc/kubernetes/pki : USER =ubuntu # customizable mkdir -p /etc/
8,169
kubernetes/pki/etcd mv /home/ ${USER }/ca.crt /etc/kubernetes/pki/ mv /home/ ${USER }/ca.key /etc/kubernetes/pki/ mv /home/ ${USER }/sa.pub /etc/kubernetes/pki/ mv /home/ ${USER }/sa.key /etc/kubernetes/pki/ mv /home/ ${USER }/front-proxy-ca.crt /etc/kubernetes/pki/ mv /home/ ${USER }/front-proxy-ca.key /etc/kubernetes/pki/ mv /home/ ${USER }/etcd-ca.crt /etc/kubernetes/pki/etcd/ca.crt # Skip the next line if you are using external etcd mv /home/ ${USER }/etcd-ca.key /etc/kubernetes/pki/etcd/ca.key Set up a High Availability etcd Cluster with kubeadm Note: While kubeadm is being used as the management tool for external etcd nodes in this guide, please note that kubeadm does not plan to support certificate rotation or upgrades for such nodes. The long-term plan is to empower the tool etcdadm to manage these aspects. By default, kubeadm runs a local etcd instance on each control plane node. It is also possible to treat the etcd cluster as external and provision etcd instances on separa
8,170
te hosts. The differences between the two approaches are covered in the Options for Highly Available topology page. This task walks through the process of creating a high availability external etcd cluster of three members that can be used by kubeadm during cluster creation. Before you begin Three hosts that can talk to each other over TCP ports 2379 and 2380. This document assumes these default ports. However, they are configurable through the kubeadm config file. Each host must have systemd and a bash compatible shell installed.5. β€’
8,171
Each host must have a container runtime, kubelet, and kubeadm installed . Each host should have access to the Kubernetes container image registry ( registry.k8s.io ) or list/pull the required etcd image using kubeadm config images list/pull . This guide will set up etcd instances as static pods managed by a kubelet. Some infrastructure to copy files between hosts. For example ssh and scp can satisfy this requirement. Setting up the cluster The general approach is to generate all certs on one node and only distribute the necessary files to the other nodes. Note: kubeadm contains all the necessary cryptographic machinery to generate the certificates described below; no other cryptographic tooling is required for this example. Note: The examples below use IPv4 addresses but you can also configure kubeadm, the kubelet and etcd to use IPv6 addresses. Dual-stack is supported by some Kubernetes options, but not by etcd. For more details on Kubernetes dual-stack support see Dual-stack supp
8,172
ort with kubeadm . Configure the kubelet to be a service manager for etcd. Note: You must do this on every host where etcd should be running. Since etcd was created first, you must override the service priority by creating a new unit file that has higher precedence than the kubeadm-provided kubelet unit file. cat << EOF > /etc/systemd/system/kubelet.service.d/kubelet.conf # Replace "systemd" with the cgroup driver of your container runtime. The default value in the kubelet is "cgroupfs". # Replace the value of "containerRuntimeEndpoint" for a different container runtime if needed. # apiVersion: kubelet.config.k8s.io/v1beta1 kind: KubeletConfiguration authentication: anonymous: enabled: false webhook: enabled: false authorization: mode: AlwaysAllow cgroupDriver: systemd address: 127.0.0.1 containerRuntimeEndpoint: unix:///var/run/containerd/containerd.sock staticPodPath: /etc/kubernetes/manifests EOF cat << EOF > /etc/systemd/system/kubelet.service.d/20-etcd-service-man
8,173
ager.conf [Service] ExecStart= ExecStart=/usr/bin/kubelet --config=/etc/systemd/system/kubelet.service.d/kubelet.conf Restart=always EOFβ€’ β€’ β€’ 1
8,174
systemctl daemon-reload systemctl restart kubelet Check the kubelet status to ensure it is running. systemctl status kubelet Create configuration files for kubeadm. Generate one kubeadm configuration file for each host that will have an etcd member running on it using the following script. # Update HOST0, HOST1 and HOST2 with the IPs of your hosts export HOST0 =10.0.0.6 export HOST1 =10.0.0.7 export HOST2 =10.0.0.8 # Update NAME0, NAME1 and NAME2 with the hostnames of your hosts export NAME0 ="infra0" export NAME1 ="infra1" export NAME2 ="infra2" # Create temp directories to store files that will end up on other hosts mkdir -p /tmp/ ${HOST0 }/ /tmp/ ${HOST1 }/ /tmp/ ${HOST2 }/ HOSTS =(${HOST0 } ${HOST1 } ${HOST2 }) NAMES =(${NAME0 } ${NAME1 } ${NAME2 }) for i in "${!HOSTS[@] }"; do HOST =${HOSTS [$i]} NAME =${NAMES [$i]} cat << EOF > /tmp/${HOST}/kubeadmcfg.yaml --- apiVersion: "kubeadm.k8s.io/v1beta3" kind: InitConfiguration nodeRegistration: name: ${NAME} localAPIEndpoint:
8,175
advertiseAddress: ${HOST} --- apiVersion: "kubeadm.k8s.io/v1beta3" kind: ClusterConfiguration etcd: local: serverCertSANs: - "${HOST}" peerCertSANs: - "${HOST}" extraArgs: initial-cluster: ${NAMES[0]}=https://${HOSTS[0]}:2380,${NAMES[1]}=https://$ {HOSTS[1]}:2380,${NAMES[2]}=https://${HOSTS[2]}:2380 initial-cluster-state: new name: ${NAME}2
8,176
listen-peer-urls: https://${HOST}:2380 listen-client-urls: https://${HOST}:2379 advertise-client-urls: https://${HOST}:2379 initial-advertise-peer-urls: https://${HOST}:2380 EOF done Generate the certificate authority. If you already have a CA then the only action that is copying the CA's crt and key file to / etc/kubernetes/pki/etcd/ca.crt and /etc/kubernetes/pki/etcd/ca.key . After those files have been copied, proceed to the next step, "Create certificates for each member". If you do not already have a CA then run this command on $HOST0 (where you generated the configuration files for kubeadm). kubeadm init phase certs etcd-ca This creates two files: /etc/kubernetes/pki/etcd/ca.crt /etc/kubernetes/pki/etcd/ca.key Create certificates for each member. kubeadm init phase certs etcd-server --config =/tmp/ ${HOST2 }/kubeadmcfg.yaml kubeadm init phase certs etcd-peer --config =/tmp/ ${HOST2 }/kubeadmcfg.yaml kubeadm init phase certs etcd-healthcheck-c
8,177
lient --config =/tmp/ ${HOST2 }/ kubeadmcfg.yaml kubeadm init phase certs apiserver-etcd-client --config =/tmp/ ${HOST2 }/kubeadmcfg.yaml cp -R /etc/kubernetes/pki /tmp/ ${HOST2 }/ # cleanup non-reusable certificates find /etc/kubernetes/pki -not -name ca.crt -not -name ca.key -type f -delete kubeadm init phase certs etcd-server --config =/tmp/ ${HOST1 }/kubeadmcfg.yaml kubeadm init phase certs etcd-peer --config =/tmp/ ${HOST1 }/kubeadmcfg.yaml kubeadm init phase certs etcd-healthcheck-client --config =/tmp/ ${HOST1 }/ kubeadmcfg.yaml kubeadm init phase certs apiserver-etcd-client --config =/tmp/ ${HOST1 }/kubeadmcfg.yaml cp -R /etc/kubernetes/pki /tmp/ ${HOST1 }/ find /etc/kubernetes/pki -not -name ca.crt -not -name ca.key -type f -delete kubeadm init phase certs etcd-server --config =/tmp/ ${HOST0 }/kubeadmcfg.yaml kubeadm init phase certs etcd-peer --config =/tmp/ ${HOST0 }/kubeadmcfg.yaml kubeadm init phase certs etcd-healthcheck-client --config =/tmp/ ${HOST0 }/ kubeadmcfg.yaml k
8,178
ubeadm init phase certs apiserver-etcd-client --config =/tmp/ ${HOST0 }/kubeadmcfg.yaml # No need to move the certs because they are for HOST0 # clean up certs that should not be copied off this host find /tmp/ ${HOST2 } -name ca.key -type f -delete find /tmp/ ${HOST1 } -name ca.key -type f -delete Copy certificates and kubeadm configs.3. β—¦ β—¦ 4. 5
8,179
The certificates have been generated and now they must be moved to their respective hosts. USER =ubuntu HOST =${HOST1 } scp -r /tmp/ ${HOST }/* ${USER }@${HOST }: ssh ${USER }@${HOST } USER@HOST $ sudo -Es root@HOST $ chown -R root:root pki root@HOST $ mv pki /etc/kubernetes/ Ensure all expected files exist. The complete list of required files on $HOST0 is: /tmp/${HOST0} └── kubeadmcfg.yaml --- /etc/kubernetes/pki β”œβ”€β”€ apiserver-etcd-client.crt β”œβ”€β”€ apiserver-etcd-client.key └── etcd β”œβ”€β”€ ca.crt β”œβ”€β”€ ca.key β”œβ”€β”€ healthcheck-client.crt β”œβ”€β”€ healthcheck-client.key β”œβ”€β”€ peer.crt β”œβ”€β”€ peer.key β”œβ”€β”€ server.crt └── server.key On $HOST1 : $HOME └── kubeadmcfg.yaml --- /etc/kubernetes/pki β”œβ”€β”€ apiserver-etcd-client.crt β”œβ”€β”€ apiserver-etcd-client.key └── etcd β”œβ”€β”€ ca.crt β”œβ”€β”€ healthcheck-client.crt β”œβ”€β”€ healthcheck-client.key β”œβ”€β”€ peer.crt β”œβ”€β”€ peer.key β”œβ”€β”€ server.crt └── server.key On $HOST2 : $HOME └── kubeadmcfg.yaml --- /etc/kubernetes/pki6.
8,180
β”œβ”€β”€ apiserver-etcd-client.crt β”œβ”€β”€ apiserver-etcd-client.key └── etcd β”œβ”€β”€ ca.crt β”œβ”€β”€ healthcheck-client.crt β”œβ”€β”€ healthcheck-client.key β”œβ”€β”€ peer.crt β”œβ”€β”€ peer.key β”œβ”€β”€ server.crt └── server.key Create the static pod manifests. Now that the certificates and configs are in place it's time to create the manifests. On each host run the kubeadm command to generate a static manifest for etcd. root@HOST0 $ kubeadm init phase etcd local --config =/tmp/ ${HOST0 }/kubeadmcfg.yaml root@HOST1 $ kubeadm init phase etcd local --config =$HOME /kubeadmcfg.yaml root@HOST2 $ kubeadm init phase etcd local --config =$HOME /kubeadmcfg.yaml Optional: Check the cluster health. If etcdctl isn't available, you can run this tool inside a container image. You would do that directly with your container runtime using a tool such as crictl run and not through Kubernetes ETCDCTL_API =3 etcdctl \ --cert /etc/kubernetes/pki/etcd/peer.crt \ --key /etc/kubernetes/pki/etcd/peer.key \ --cacer
8,181
t /etc/kubernetes/pki/etcd/ca.crt \ --endpoints https:// ${HOST0 }:2379 endpoint health ... https:// [HOST0 IP ]:2379 is healthy: successfully committed proposal: took = 16.283339ms https:// [HOST1 IP ]:2379 is healthy: successfully committed proposal: took = 19.44402ms https:// [HOST2 IP ]:2379 is healthy: successfully committed proposal: took = 35.926451ms Set ${HOST0} to the IP address of the host you are testing. What's next Once you have an etcd cluster with 3 working members, you can continue setting up a highly available control plane using the external etcd method with kubeadm . Configuring each kubelet in your cluster using kubeadm Note: Dockershim has been removed from the Kubernetes project as of release 1.24. Read the Dockershim Removal FAQ for further details. FEATURE STATE: Kubernetes v1.11 [stable]7. 8.
8,182
The lifecycle of the kubeadm CLI tool is decoupled from the kubelet , which is a daemon that runs on each node within the Kubernetes cluster. The kubeadm CLI tool is executed by the user when Kubernetes is initialized or upgraded, whereas the kubelet is always running in the background. Since the kubelet is a daemon, it needs to be maintained by some kind of an init system or service manager. When the kubelet is installed using DEBs or RPMs, systemd is configured to manage the kubelet. You can use a different service manager instead, but you need to configure it manually. Some kubelet configuration details need to be the same across all kubelets involved in the cluster, while other configuration aspects need to be set on a per-kubelet basis to accommodate the different characteristics of a given machine (such as OS, storage, and networking). You can manage the configuration of your kubelets manually, but kubeadm now provides a KubeletConfiguration API type for managing your kubelet c
8,183
onfigurations centrally . Kubelet configuration patterns The following sections describe patterns to kubelet configuration that are simplified by using kubeadm, rather than managing the kubelet configuration for each Node manually. Propagating cluster-level configuration to each kubelet You can provide the kubelet with default values to be used by kubeadm init and kubeadm join commands. Interesting examples include using a different container runtime or setting the default subnet used by services. If you want your services to use the subnet 10.96.0.0/12 as the default for services, you can pass the --service-cidr parameter to kubeadm: kubeadm init --service-cidr 10.96.0.0/12 Virtual IPs for services are now allocated from this subnet. You also need to set the DNS address used by the kubelet, using the --cluster-dns flag. This setting needs to be the same for every kubelet on every manager and Node in the cluster. The kubelet provides a versioned, structured API object that can conf
8,184
igure most parameters in the kubelet and push out this configuration to each running kubelet in the cluster. This object is called KubeletConfiguration . The KubeletConfiguration allows the user to specify flags such as the cluster DNS IP addresses expressed as a list of values to a camelCased key, illustrated by the following example: apiVersion : kubelet.config.k8s.io/v1beta1 kind: KubeletConfiguration clusterDNS : - 10.96.0.10 For more details on the KubeletConfiguration have a look at this section
8,185
Providing instance-specific configuration details Some hosts require specific kubelet configurations due to differences in hardware, operating system, networking, or other host-specific parameters. The following list provides a few examples. The path to the DNS resolution file, as specified by the --resolv-conf kubelet configuration flag, may differ among operating systems, or depending on whether you are using systemd-resolved . If this path is wrong, DNS resolution will fail on the Node whose kubelet is configured incorrectly. The Node API object .metadata.name is set to the machine's hostname by default, unless you are using a cloud provider. You can use the --hostname-override flag to override the default behavior if you need to specify a Node name different from the machine's hostname. Currently, the kubelet cannot automatically detect the cgroup driver used by the container runtime, but the value of --cgroup-driver must match the cgroup driver used by the container runtime to
8,186
ensure the health of the kubelet. To specify the container runtime you must set its endpoint with the --container-runtime- endpoint=<path> flag. The recommended way of applying such instance-specific configuration is by using KubeletConfiguration patches . Configure kubelets using kubeadm It is possible to configure the kubelet that kubeadm will start if a custom KubeletConfiguration API object is passed with a configuration file like so kubeadm ... --config some-config-file.yaml . By calling kubeadm config print init-defaults --component-configs KubeletConfiguration you can see all the default values for this structure. It is also possible to apply instance-specific patches over the base KubeletConfiguration . Have a look at Customizing the kubelet for more details. Workflow when using kubeadm init When you call kubeadm init , the kubelet configuration is marshalled to disk at /var/lib/kubelet/ config.yaml , and also uploaded to a kubelet-config ConfigMap in the kube-system na
8,187
mespace of the cluster. A kubelet configuration file is also written to /etc/kubernetes/kubelet.conf with the baseline cluster-wide configuration for all kubelets in the cluster. This configuration file points to the client certificates that allow the kubelet to communicate with the API server. This addresses the need to propagate cluster-level configuration to each kubelet . To address the second pattern of providing instance-specific configuration details , kubeadm writes an environment file to /var/lib/kubelet/kubeadm-flags.env , which contains a list of flags to pass to the kubelet when it starts. The flags are presented in the file like this: KUBELET_KUBEADM_ARGS ="--flag1=value1 --flag2=value2 ..."β€’ β€’ β€’
8,188
In addition to the flags used when starting the kubelet, the file also contains dynamic parameters such as the cgroup driver and whether to use a different container runtime socket (--cri-socket ). After marshalling these two files to disk, kubeadm attempts to run the following two commands, if you are using systemd: systemctl daemon-reload && systemctl restart kubelet If the reload and restart are successful, the normal kubeadm init workflow continues. Workflow when using kubeadm join When you run kubeadm join , kubeadm uses the Bootstrap Token credential to perform a TLS bootstrap, which fetches the credential needed to download the kubelet-config ConfigMap and writes it to /var/lib/kubelet/config.yaml . The dynamic environment file is generated in exactly the same way as kubeadm init . Next, kubeadm runs the following two commands to load the new configuration into the kubelet: systemctl daemon-reload && systemctl restart kubelet After the kubelet loads the new configuration, kub
8,189
eadm writes the /etc/kubernetes/bootstrap- kubelet.conf KubeConfig file, which contains a CA certificate and Bootstrap Token. These are used by the kubelet to perform the TLS Bootstrap and obtain a unique credential, which is stored in /etc/kubernetes/kubelet.conf . When the /etc/kubernetes/kubelet.conf file is written, the kubelet has finished performing the TLS Bootstrap. Kubeadm deletes the /etc/kubernetes/bootstrap-kubelet.conf file after completing the TLS Bootstrap. The kubelet drop-in file for systemd kubeadm ships with configuration for how systemd should run the kubelet. Note that the kubeadm CLI command never touches this drop-in file. This configuration file installed by the kubeadm package is written to /etc/systemd/system/ kubelet.service.d/10-kubeadm.conf and is used by systemd. It augments the basic kubelet.service : Note: The contents below are just an example. If you don't want to use a package manager follow the guide outlined in the ( Without a package manag
8,190
er ) section. [Service] Environment="KUBELET_KUBECONFIG_ARGS=--bootstrap-kubeconfig=/etc/kubernetes/ bootstrap-kubelet.conf --kubeconfig=/etc/kubernetes/kubelet.conf" Environment="KUBELET_CONFIG_ARGS=--config=/var/lib/kubelet/config.yaml" # This is a file that "kubeadm init" and "kubeadm join" generate at runtime, populating # the KUBELET_KUBEADM_ARGS variable dynamically EnvironmentFile=-/var/lib/kubelet/kubeadm-flags.env # This is a file that the user can use for overrides of the kubelet args as a last resort. Preferably, # the user should use the .NodeRegistration.KubeletExtraArgs object in the configuration file
8,191
instead. # KUBELET_EXTRA_ARGS should be sourced from this file. EnvironmentFile=-/etc/default/kubelet ExecStart= ExecStart=/usr/bin/kubelet $KUBELET_KUBECONFIG_ARGS $KUBELET_CONFIG_ARGS $KUBELET_KUBEADM_ARGS $KUBELET_EXTRA_ARGS This file specifies the default locations for all of the files managed by kubeadm for the kubelet. The KubeConfig file to use for the TLS Bootstrap is /etc/kubernetes/bootstrap- kubelet.conf , but it is only used if /etc/kubernetes/kubelet.conf does not exist. The KubeConfig file with the unique kubelet identity is /etc/kubernetes/kubelet.conf . The file containing the kubelet's ComponentConfig is /var/lib/kubelet/config.yaml . The dynamic environment file that contains KUBELET_KUBEADM_ARGS is sourced from /var/lib/kubelet/kubeadm-flags.env . The file that can contain user-specified flag overrides with KUBELET_EXTRA_ARGS is sourced from /etc/default/kubelet (for DEBs), or /etc/sysconfig/kubelet (for RPMs). KUBELET_EXTRA_ARGS is last in the flag chain and
8,192
has the highest priority in the event of conflicting settings. Kubernetes binaries and package contents The DEB and RPM packages shipped with the Kubernetes releases are: Package nameDescription kubeadmInstalls the /usr/bin/kubeadm CLI tool and the kubelet drop-in file for the kubelet. kubelet Installs the /usr/bin/kubelet binary. kubectl Installs the /usr/bin/kubectl binary. cri-tools Installs the /usr/bin/crictl binary from the cri-tools git repository . kubernetes-cni Installs the /opt/cni/bin binaries from the plugins git repository . Dual-stack support with kubeadm FEATURE STATE: Kubernetes v1.23 [stable] Your Kubernetes cluster includes dual-stack networking, which means that cluster networking lets you use either address family. In a cluster, the control plane can assign both an IPv4 address and an IPv6 address to a single Pod or a Service . Before you begin You need to have installed the kubeadm tool, following the steps from Installing kubeadm . For each server that
8,193
you want to use as a node , make sure it allows IPv6 forwarding. On Linux, you can set this by running run sysctl -w net.ipv6.conf.all.forwarding=1 as the root user on each server.β€’ β€’ β€’ β€’
8,194
You need to have an IPv4 and and IPv6 address range to use. Cluster operators typically use private address ranges for IPv4. For IPv6, a cluster operator typically chooses a global unicast address block from within 2000::/3 , using a range that is assigned to the operator. You don't have to route the cluster's IP address ranges to the public internet. The size of the IP address allocations should be suitable for the number of Pods and Services that you are planning to run. Note: If you are upgrading an existing cluster with the kubeadm upgrade command, kubeadm does not support making modifications to the pod IP address range (β€œcluster CIDR”) nor to the cluster's Service address range (β€œService CIDR”). Create a dual-stack cluster To create a dual-stack cluster with kubeadm init you can pass command line arguments similar to the following example: # These address ranges are examples kubeadm init --pod-network-cidr =10.244.0.0/16,2001:db8:42:0::/56 --service-cidr =10.96.0.0/16,200 1:db
8,195
8:42:1::/112 To make things clearer, here is an example kubeadm configuration file kubeadm-config.yaml for the primary dual-stack control plane node. --- apiVersion : kubeadm.k8s.io/v1beta3 kind: ClusterConfiguration networking : podSubnet : 10.244.0.0 /16,2001:db8:42:0::/56 serviceSubnet : 10.96.0.0 /16,2001:db8:42:1::/112 --- apiVersion : kubeadm.k8s.io/v1beta3 kind: InitConfiguration localAPIEndpoint : advertiseAddress : "10.100.0.1" bindPort : 6443 nodeRegistration : kubeletExtraArgs : node-ip : 10.100.0.2 ,fd00:1:2:3::2 advertiseAddress in InitConfiguration specifies the IP address that the API Server will advertise it is listening on. The value of advertiseAddress equals the --apiserver-advertise-address flag of kubeadm init . Run kubeadm to initiate the dual-stack control plane node: kubeadm init --config =kubeadm-config.yaml The kube-controller-manager flags --node-cidr-mask-size-ipv4|--node-cidr-mask-size-ipv6 are set with default values. See configure IPv
8,196
4/IPv6 dual stack . Note: The --apiserver-advertise-address flag does not support dual-stack
8,197
Join a node to dual-stack cluster Before joining a node, make sure that the node has IPv6 routable network interface and allows IPv6 forwarding. Here is an example kubeadm configuration file kubeadm-config.yaml for joining a worker node to the cluster. apiVersion : kubeadm.k8s.io/v1beta3 kind: JoinConfiguration discovery : bootstrapToken : apiServerEndpoint : 10.100.0.1 :6443 token : "clvldh.vjjwg16ucnhp94qr" caCertHashes : - "sha256:a4863cde706cfc580a439f842cc65d5ef112b7b2be31628513a9881cf0d9fe0e" # change auth info above to match the actual token and CA certificate hash for your cluster nodeRegistration : kubeletExtraArgs : node-ip : 10.100.0.3 ,fd00:1:2:3::3 Also, here is an example kubeadm configuration file kubeadm-config.yaml for joining another control plane node to the cluster. apiVersion : kubeadm.k8s.io/v1beta3 kind: JoinConfiguration controlPlane : localAPIEndpoint : advertiseAddress : "10.100.0.2" bindPort : 6443 discovery : boots
8,198
trapToken : apiServerEndpoint : 10.100.0.1 :6443 token : "clvldh.vjjwg16ucnhp94qr" caCertHashes : - "sha256:a4863cde706cfc580a439f842cc65d5ef112b7b2be31628513a9881cf0d9fe0e" # change auth info above to match the actual token and CA certificate hash for your cluster nodeRegistration : kubeletExtraArgs : node-ip : 10.100.0.4 ,fd00:1:2:3::4 advertiseAddress in JoinConfiguration.controlPlane specifies the IP address that the API Server will advertise it is listening on. The value of advertiseAddress equals the --apiserver-advertise- address flag of kubeadm join . kubeadm join --config =kubeadm-config.yaml Create a single-stack cluster Note: Dual-stack support doesn't mean that you need to use dual-stack addressing. You can deploy a single-stack cluster that has the dual-stack networking feature enabled
8,199