When you create a Kubernetes cluster by using TKE, you must select models from various configuration options. This document describes and compares available feature models and gives suggestions to help you select models that are most applicable to your services.
Kubernetes Versions
Kubernetes versions are iterated quickly. New versions usually include many bug fixes and new features. Meanwhile, earlier versions will be phased out. We recommend that you select the latest version that is supported by the current TKE when creating a cluster. Subsequently, you can upgrade existing master components and nodes to the latest versions generated during iteration.
Container Network Plugins: GlobalRouter and VPC-CNI
Network modes
TKE supports the following two network modes. For more information, see How to Choose TKE Network Mode.
GlobalRouter mode:
- In this mode, container network capabilities are implemented based on container networking interfaces (CNIs) and network bridges, whereas container routing is implemented based on the underlying VPC layer.
- Containers are located on the same network plane as nodes. IP ranges of containers cover abundant IP addresses and do not overlap those of VPC instances.
VPC-CNI mode:
- In this mode, container network capabilities are implemented based on CNIs and VPC ENIs, whereas container routing is implemented based on ENIs. The performance of this mode is approximately 10% higher than that of the GlobalRouter mode.
- Containers are located on the same network plane as nodes. The IP ranges of containers fall within those of VPC instances.
- Pods can use static IP addresses.
How to use
TKE allows you to specify network modes in the following ways:
- Specify the GlobalRouter mode when creating a cluster.
- Specify the VPC-CNI mode when creating a cluster. Subsequently, all pods must be created in VPC-CNI mode.
- Specify the GlobalRouter mode when creating a cluster. You can enable the VPC-CNI mode for the cluster when needed. In this case, the two modes are mixed.
Recommendations
- In general cases, we recommend that you select the GlobalRouter mode, because IP ranges of containers cover abundant IP addresses, allow high scalability, and support large-scale services.
- If a subsequent service needs to run in VPC-CNI mode, you can enable the VPC-CNI mode for the GlobalRouter cluster. In this case, the GlobalRouter mode is mixed with the VPC-CNI mode, but only some services run in VPC-CNI mode.
- If you fully understand and accept the use limits of the VPC-CNI mode and all pods in the cluster need to run in VPC-CNI mode, we recommend that you select the VPC-CNI mode when creating the cluster.
Runtime Components: Docker and Containerd (Under Beta Testing)
Runtime components
TKE supports two types of runtime components: Docker and containerd. For more information, see How to Choose Between containerd and Docker.
Comparison between the two runtimes
- The call chain of the containerd runtime bypasses dockerd and therefore is shorter. Accordingly, the containerd solution requires fewer components, occupies fewer node resources, and bypasses dockerd bugs. However, containerd has some bugs that need to be fixed. Currently, containerd is under beta testing and has fixed some bugs.
- Having been used for a long time, the Docker solution is more mature, supports the Docker API, and provides abundant features. This solution is friendly to most users.
Recommendations
- The Docker solution is more mature than the containerd solution. If you require high stability, we recommend that you select the Docker solution.
- In the following scenarios, you can select the Docker solution only:
- You need to run a Docker host inside of another Docker host (Docker-in-Docker). This requirement usually occurs during continuous integration (CI).
- You need to use Docker commands on a node.
- You need to call the Docker API.
In other scenarios, we recommend that you select the containerd solution.
Service Forwarding Modes: iptables and IPVS
The following figure shows how a Service is forwarded.
- The kube-proxy component on the node watches API Server to obtain the Service and the Endpoint. Then, the kube-proxy component converts the Service to an iptables or IPVS rule based on the forwarding mode and writes the rule to the node.
- The client in the cluster gains access to the Service through the cluster IP address. Then, according to the iptable or IPVS rule, the client is load-balanced to the backend pod corresponding to the Service.
Comparison between the two forwarding modes
- The IPVS mode provides higher performance but has some outstanding bugs.
- The iptables mode is more mature and stable.
Recommendations
If you require extremely high stability with less than 2,000 Services running in the cluster, we recommend that you select iptables. In other scenarios, we recommend that you preferably select IPVS.
Cluster Types: Managed Cluster and Self-Deployed Cluster
TKE supports the following types of clusters:
Managed clusters:
Master components are invisible to you but are managed by Tencent Cloud.
Clusters with most new features are preferably managed.
Computing resources of master components are automatically scaled up based on the cluster scale.
You do not need to pay for master components.
Self-deployed clusters:
Master components are fully under your control.
You need to purchase master components.
Recommendations
In regular cases, we recommend that you select managed clusters. If you need to fully control master components, such as specifying custom features to implement advanced features, you can select self-deployed clusters.
Node Operating System
TKE supports three distributions of operating systems, Tencent Linux, Ubuntu and CentOS. The Tencent Linux operating system uses TencentOS-kernel that is a customized kernel maintained by Tencent Cloud team. Other operating systems use the open-source kernel provided by the official Linux community. The following shows the supported operating systems.
Note:
TKE-Optimized series images is once used for improving the stability of the image and providing more features, but it is not available for the clusters in TKE console after the Tencent Linux public image is launched. For more information, see TKE-Optimized Series Images.
Recommendations
We recommend that you use the Tencent Linux operating system. It is a public image of Tencent Cloud that contains the TencentOS-kernel. TKE now supports this image and uses it as the default image.
Node Pool
The node pool is mainly used to batch manage nodes with the following items:
- Label and Taint properties of nodes
- Startup parameters of node components
- Custom launch script for nodes
For more information, see Node Pool Overview.
Application scenarios
- Manage heterogeneous nodes by group to reduce management costs.
- Use the Label and Taint properties to enable a cluster to support complex scheduling rules.
- Frequently scale out and in nodes to reduce operation costs.
- Routinely maintain nodes, such as upgrade node versions.
Examples
Some I/O-intensive services require models with high I/O throughput. You can create a node pool for a service of these kinds, configure a model, centrally specify Label and Taint properties for the nodes, and configure affinity with I/O-intensive services. You can select Labels to schedule the service to a node with a high I/O model. To avoid other service pods from being scheduled to the node, you can select specific Taints.
When the service traffic increases, the I/O-intensive service needs more computing resources. During peak hours, the HPA feature automatically scales out pods for the service, and the computing resources of the node become insufficient. In this case, the auto scaling feature of the node pool automatically scales out nodes to withstand the traffic spike.
Launch Script
Custom parameters for components
Note:
To use this feature, submit a ticket to apply for it.
- When creating a cluster, you can customize some startup parameters of master components in Advanced Settings under Cluster Information.
- In Select Model step, you can customize some startup parameters of kubelet in Advanced Settings under Worker Configurations.
Node launch configuration
- When creating a cluster, in Advanced Settings under CVM Configuration, you can specify custom data to configure the node launch script. In the script, you can modify component startup parameters and kernel parameters, as shown in the following figure:
- When adding a node, in Advanced Settings under CVM Configuration, you can specify custom data to configure the node launch script. In the script, you can modify component startup parameters and kernel parameters, as shown in the following figure:
Apakah halaman ini membantu?