Previous articles in this series:
- TiDB Operator Source Code Reading (I): Overview
- TiDB Operator Source Code Reading (II): Operator Pattern
- TiDB Operator Source Code Reading (III): The Component Control Loop
In our last article, we introduced how TiDB Operator orchestrates control loop events to manage the lifecycles of TiDB components. The TidbCluster
controller manages TiDB components’ lifecycles, and the member manager of each TiDB component encapsulates that component’s specific management logic.
In this post, I’ll explain in detail how we implement a component control loop by taking PD as an example. You’ll learn about the PD member manager and its lifecycle management operations. I’ll also compare other components with PD and show their differences.
The PD control loop
Placement Driver (PD) manages a TiDB cluster and schedules Regions in the cluster. The PD member manager maintains the PD lifecycle management logic. Most of its code resides in pkg/manager/member/pd_member_manager.go
, and other code related to scaling, upgrade, and failover is located in pd_scaler.go
, pd_upgrader.go
, and pd_failover.go
respectively.
As is illustrated in Part III, it takes the following tasks to manage a component’s lifecycle:
- Sync Service
- Start syncing StatefulSet
- Sync status
- Sync ConfigMap
- Rolling update
- Scaling in and scaling out
- Failover
- Finish syncing StatefulSet
Syncing StatefulSet is the major logic. Other tasks, like syncing status and syncing ConfigMap, are defined as sub-functions and called by the PD member manager when it syncs StatefulSet.
Sync StatefulSet
To sync StatefulSet, the PD member manager:
- Obtains the current PD StatefulSet using
StatefulSetLister
.oldPDSetTmp, err := m.deps.StatefulSetLister.StatefulSets(ns).Get(controller.PDMemberName(tcName))
if err != nil && !errors.IsNotFound(err) {
return fmt.Errorf("syncPDStatefulSetForTidbCluster: fail to get sts %s for cluster %s/%s, error: %s", controller.PDMemberName(tcName), ns, tcName, err)
}
setNotExist := errors.IsNotFound(err)
oldPDSet := oldPDSetTmp.DeepCopy() - Gets the latest status using
m.syncTidbClusterStatus(tc, oldPDSet).
if err := m.syncTidbClusterStatus(tc, oldPDSet); err != nil {
klog.Errorf("failed to sync TidbCluster: [%s/%s]'s status, error: %v", ns, tcName, err)
}
- Checks whether the TiDB cluster pauses the synchronization. If so, terminate the following reconciliation.
if tc.Spec.Paused {
klog.V(4).Infof("tidb cluster %s/%s is paused, skip syncing for pd statefulset", tc.GetNamespace(), tc.GetName())
return nil
}
Syncs ConfigMap according to the latest tc.Spec.
cm, err := m.syncPDConfigMap(tc, oldPDSet)
- Generates the latest StatefulSet template according to the latest
tc.Spec,
tc.Status
, and ConfigMap obtained in the last step.newPDSet, err := getNewPDSetForTidbCluster(tc, cm)
If the new PD StatefulSet hasn't been created yet, first create the StatefulSet.
if setNotExist {
if err := SetStatefulSetLastAppliedConfigAnnotation(newPDSet); err != nil {
return err
}
if err := m.deps.StatefulSetControl.CreateStatefulSet(tc, newPDSet); err != nil {
return err
}
tc.Status.PD.StatefulSet = &apps.StatefulSetStatus{}
return controller.RequeueErrorf("TidbCluster: [%s/%s], waiting for PD cluster running", ns, tcName)
}
- If a user configures a forced upgrade using annotations, this step configures the StatefulSet to perform a rolling upgrade directly. This is to avoid an upgrade failure if the reconciliation loop is blocked.
if !tc.Status.PD.Synced && NeedForceUpgrade(tc.Annotations) {
tc.Status.PD.Phase = v1alpha1.UpgradePhase
setUpgradePartition(newPDSet, 0)
errSTS := UpdateStatefulSet(m.deps.StatefulSetControl, tc, newPDSet, oldPDSet)
return controller.RequeueErrorf("tidbcluster: [%s/%s]'s pd needs force upgrade, %v", ns, tcName, errSTS)
}
Calls the scaling logic implemented in pd_scaler.go.
if err := m.scaler.Scale(tc, oldPDSet, newPDSet); err != nil {
return err
}
- Calls the failover logic implemented in pd_failover.go. The function first checks if the cluster needs to recover, then checks if all Pods are started and all members are healthy, and finally determines whether it needs to begin failover.
if m.deps.CLIConfig.AutoFailover {
if m.shouldRecover(tc) {
m.failover.Recover(tc)
} else if tc.PDAllPodsStarted() && !tc.PDAllMembersReady() || tc.PDAutoFailovering() {
if err := m.failover.Failover(tc); err != nil {
return err
}
}
}
Calls the upgrade logic implemented in pd_upgrader.go. When the newly generated PD StatefulSet is inconsistent with the existing one or when the two StatefulSets are consistent but tc.Status.PD.Phase is upgrade, the PD Member Manager enters upgrader to process the rolling upgrade logic.
if !templateEqual(newPDSet, oldPDSet) || tc.Status.PD.Phase == v1alpha1.UpgradePhase {
if err := m.upgrader.Upgrade(tc, oldPDSet, newPDSet); err != nil {
return err
}
}
The PD StatefulSet is synced, and the new StatefulSet is updated to the Kubernetes cluster.
Sync Service
PD uses both Services and headless Services, managed by syncPDServiceForTidbCluster
and syncPDHeadlessServiceForTidbCluster
.
The Service address is used in TiDB, TiKV, and TiFlash to configure the PD endpoint. For example, TiDB uses --path=${CLUSTER_NAME}-pd:2379
as its PD service address in the startup parameter:
ARGS="--store=tikv \
--advertise-address=${POD_NAME}.${HEADLESS_SERVICE_NAME}.${NAMESPACE}.svc \
--path=${CLUSTER_NAME}-pd:2379 \
The headless Service provides a unique identifier for each Pod. When PD starts up, the PD Pod registers its endpoint in PD members as "${POD_NAME}.${PEER_SERVICE_NAME}.${NAMESPACE}.svc"
:
domain="${POD_NAME}.${PEER_SERVICE_NAME}.${NAMESPACE}.svc"
ARGS="--data-dir=/var/lib/pd \
--name=${POD_NAME} \
--peer-urls=http://0.0.0.0:2380 \
--advertise-peer-urls=http://${domain}:2380 \
--client-urls=http://0.0.0.0:2379 \
--advertise-client-urls=http://${domain}:2379 \
--config=/etc/pd/pd.toml \
"
Sync ConfigMap
PD uses ConfigMap to manage the configurations and the start script. The syncPDConfigMap
function calls getPDConfigMap
to get the latest ConfigMap and apply it to the Kubernetes cluster. ConfigMap handles the following tasks:
- Get
PD.Config
for the follow-up sync. To remain compatible with earlier versions that use Helm, when the config object is empty, ConfigMap is not synced.config := tc.Spec.PD.Config
if config == nil {
return nil, nil
}
// override CA if tls enabled
if tc.IsTLSClusterEnabled() {
config.Set("security.cacert-path", path.Join(pdClusterCertPath, tlsSecretRootCAKey))
config.Set("security.cert-path", path.Join(pdClusterCertPath, corev1.TLSCertKey))
config.Set("security.key-path", path.Join(pdClusterCertPath, corev1.TLSPrivateKeyKey))
}
// Versions below v4.0 do not support Dashboard
if tc.Spec.TiDB != nil && tc.Spec.TiDB.IsTLSClientEnabled() && !tc.SkipTLSWhenConnectTiDB() && clusterVersionGE4 {
config.Set("dashboard.tidb-cacert-path", path.Join(tidbClientCertPath, tlsSecretRootCAKey))
config.Set("dashboard.tidb-cert-path", path.Join(tidbClientCertPath, corev1.TLSCertKey))
config.Set("dashboard.tidb-key-path", path.Join(tidbClientCertPath, corev1.TLSPrivateKeyKey))
}
- Modify TLS-related configuration. Because TiDB 4.0 and earlier versions don’t support TiDB Dashboard, PD member manager skips configuring Dashboard certificates for PD in earlier versions.
- Transform the configuration to TOML format so that PD can read it.
confText, err := config.MarshalTOML()
- Generate the PD start script using
RenderPDStartScript
. The script template is stored in thepdStartScriptTpl
variable inpkg/manager/member/template.go
. The PD start script is a Bash script.RenderPDStartScript
inserts some variables and annotations configured by theTidbCluster
object into the start script. These variables and annotations are used for PD startup and debugging.
- Assemble the PD configurations and the start script as the Kubernetes ConfigMap object, and return the ConfigMap to the
syncPDConfigMap
function.
cm := &corev1.ConfigMap{
ObjectMeta: metav1.ObjectMeta{
Name: controller.PDMemberName(tc.Name),
Namespace: tc.Namespace,
Labels: pdLabel,
OwnerReferences: []metav1.OwnerReference{controller.GetOwnerRef(tc)},
},
Data: map[string]string{
"config-file": string(confText),
"startup-script": startScript,
},
}
Scaling
The scaling logic is implemented in pkg/manager/member/pd_scaler.go
. The StatefulSet calls the Scale
function to scale out or scale in the PD cluster.
Before scaling, TiDB Operator needs to perform some preliminary operations. For example, to scale in the PD cluster, TiDB Operator needs to transfer Leaders, take members offline, and add annotations to PersistentVolumeClaims (PVCs) for deferred deletion. To scale out the cluster, TiDB Operator needs to delete the previously retained PVCs.
After the preliminary operations are completed, TiDB Operator adjusts the number of StatefulSet replicas and begins the scaling process. The preliminary operations minimize the impact of the scaling operation.
The Scale
function acts like a router. Based on the scaling direction, step, and length, it determines which scaling plan to use. If the scaling
variable is a positive number, PD is scaled out; otherwise, PD is scaled in. In each scaling operation, PD is only scaled by one step. The specific plan is implemented by the ScaleIn
and ScaleOut
functions.
func (s *pdScaler) Scale(meta metav1.Object, oldSet *apps.StatefulSet, newSet *apps.StatefulSet) error {
scaling, _, _, _ := scaleOne(oldSet, newSet)
if scaling > 0 {
return s.ScaleOut(meta, oldSet, newSet)
} else if scaling < 0 {
return s.ScaleIn(meta, oldSet, newSet)
}
return s.SyncAutoScalerAnn(meta, oldSet)
}
Scaling in
Before TiDB Operator takes a PD member offline, it must transfer the Leader. Otherwise, when the Leader member is taken offline, the rest of the PD members are forced into a Leader election, which affects cluster performance. Therefore, TiDB Operator needs to transfer the Leader to the PD member with the smallest ID. This makes sure that the PD Leader is only transferred once.
To transfer the PD Leader, first get the PD client and the PD Leader:
pdClient := controller.GetPDClient(s.deps.PDControl, tc)
leader, err := pdClient.GetPDLeader()
When the Leader name equals the member name, TiDB Operator transfers the Leader. If there is only one member left, no other member is available for transfer, so TiDB Operator skips transferring the Leader.
After the Leader is transferred, the ScaleIn
function calls PD’s DeleteMember
API and deletes the member from PD members. The PD member is then taken offline. Finally, the function calls setReplicasAndDeleteSlots
to adjust the number of StatefulSet replicas, and the scaling-in process is completed.
Scaling out
Before TiDB Operator scales out the PD cluster, it needs to delete deferred PVCs by calling deleteDeferDeletingPVC
. When TiDB Operator previously scaled in PD, these PVCs were retained for deferred deletion to achieve data reliability; now, TiDB Operator must delete them to avoid the cluster using old data. After these PVCs are deleted, TiDB Operator only has to adjust the number of StatefulSet replicas and scale out the PD cluster.
Whether you scale in or scale out the PD cluster, the operation is performed by configuring the number of StatefulSet replicas. Therefore, when TiDB Operator supports the advanced StatefulSet, it must take into account the empty slots when counting the replicas.
Rolling upgrade
The upgrade logic is implemented in pkg/manager/member/pd_upgrader.go
. The UpdateStrategy
in the StatefulSet performs the rolling upgrade on PD. To mitigate the impact of upgrade on the PD cluster, pd_upgrader.go
performs some preliminary operations when it updates UpdateStrategy
in the StatefulSet. For details on how to control UpdateStrategy
, refer to our last article.
Before upgrade, TiDB Operator completes the following status check:
-
It checks whether there are other ongoing operations. In particular, it checks whether TiCDC and TiFlash is being upgraded and whether PD is being scaled out or scaled in.
if tc.Status.TiCDC.Phase == v1alpha1.UpgradePhase || tc.Status.TiFlash.Phase == v1alpha1.UpgradePhase || tc.PDScaling()
- The PD member manager determines to enter
Upgrader
on two circumstances:-
When the newly generated PD StatefulSet is inconsistent with the existing one, the status check returns
nil
, and TiDB Operator updates the StatefulSet. There is no need to perform the following checks in steps 3 and 4.if !templateEqual(newSet, oldSet) { return nil }
-
When the two StatefulSets are consistent but
tc.Status.PD.Phase
isupgrade
,newSet
andoldSet
have the same template spec, and TiDB Operator continues the following two checks.
-
When the newly generated PD StatefulSet is inconsistent with the existing one, the status check returns
-
It compares
tc.Status.PD.StatefulSet.UpdateRevision
andtc.Status.PD.StatefulSet.CurrentRevision
. If the two variables are equal, the rolling upgrade operation is already completed. Then, TiDB Operator ends the upgrade process.if tc.Status.PD.StatefulSet.UpdateRevision == tc.Status.PD.StatefulSet.CurrentRevision
-
It checks if the
UpdateStrategy
in the StatefulSet is manually modified. If yes, follow the modified strategy.if oldSet.Spec.UpdateStrategy.Type == apps.OnDeleteStatefulSetStrategyType || oldSet.Spec.UpdateStrategy.RollingUpdate == nil { newSet.Spec.UpdateStrategy = oldSet.Spec.UpdateStrategy klog.Warningf("tidbcluster: [%s/%s] pd statefulset %s UpdateStrategy has been modified manually", ns, tcName, oldSet.GetName()) return nil }
-
It checks whether the PD Pod is updated. By comparing the
controller-revision-hash
value in the Pod Label with theUpdateRevision
value of the StatefulSet, TiDB Operator can tell whether this Pod is updated or still unprocessed. If the Pod is updated, TiDB Operator then checks whether the PD member is healthy. If the PD member is healthy, TiDB Operator moves on to the next Pod. If not, TiDB Operator returns an error and checks the health state in the next sync.revision, exist := pod.Labels[apps.ControllerRevisionHashLabelKey] if !exist { return controller.RequeueErrorf("tidbcluster: [%s/%s]'s pd pod: [%s] has no label: %s", ns, tcName, podName, apps.ControllerRevisionHashLabelKey) } if revision == tc.Status.PD.StatefulSet.UpdateRevision { if member, exist := tc.Status.PD.Members[PdName(tc.Name, i, tc.Namespace, tc.Spec.ClusterDomain)]; !exist || !member.Health { return controller.RequeueErrorf("tidbcluster: [%s/%s]'s pd upgraded pod: [%s] is not ready", ns, tcName, podName) } continue }
-
If
Pod revision != tc.Status.PD.StatefulSet.UpdateRevision
, the Pod is not updated yet. TiDB Operator calls theupgradePDPod
function to process this Pod. When TiDB Operator processes the PD Leader Pod, it transfers the Leader before it continues the upgrade process.if tc.Status.PD.Leader.Name == upgradePdName || tc.Status.PD.Leader.Name == upgradePodName { var targetName string targetOrdinal := helper.GetMaxPodOrdinal(*newSet.Spec.Replicas, newSet) if ordinal == targetOrdinal { targetOrdinal = helper.GetMinPodOrdinal(*newSet.Spec.Replicas, newSet) } targetName = PdName(tcName, targetOrdinal, tc.Namespace, tc.Spec.ClusterDomain) if _, exist := tc.Status.PD.Members[targetName]; !exist { targetName = PdPodName(tcName, targetOrdinal) } if len(targetName) > 0 { err := u.transferPDLeaderTo(tc, targetName) if err != nil { klog.Errorf("pd upgrader: failed to transfer pd leader to: %s, %v", targetName, err) return err } klog.Infof("pd upgrader: transfer pd leader to: %s successfully", targetName) return controller.RequeueErrorf("tidbcluster: [%s/%s]'s pd member: [%s] is transferring leader to pd member: [%s]", ns, tcName, upgradePdName, targetName) } } setUpgradePartition(newSet, ordinal)
Failover
The PD failover logic is implemented in pd_failover.go
. PD deletes failed Pods to fix the failure, which is different from other component failover logic.
Before performing failover, TiDB Operator also carries out health checks. If the PD cluster is unavailable, which means half of the PD members are unhealthy, recreating PD members cannot recover the cluster. Under such circumstances, TiDB Operator doesn’t perform failover.
- TiDB Operator traverses the PD member health status obtained from the PD client. When the PD member is unhealthy and the duration between its last transition time and now is longer than
failoverDeadline
, the member is marked as unhealthy. Its Pod information and PVC information are recorded intc.Status.PD.FailureMembers
.
for pdName, pdMember := range tc.Status.PD.Members {
podName := strings.Split(pdName, ".")[0]
failoverDeadline := pdMember.LastTransitionTime.Add(f.deps.CLIConfig.PDFailoverPeriod)
_, exist := tc.Status.PD.FailureMembers[pdName]
if pdMember.Health || time.Now().Before(failoverDeadline) || exist {
continue
}
pod, _ := f.deps.PodLister.Pods(ns).Get(podName)
pvcs, _ := util.ResolvePVCFromPod(pod, f.deps.PVCLister)
f.deps.Recorder.Eventf(tc, apiv1.EventTypeWarning, "PDMemberUnhealthy", "%s/%s(%s) is unhealthy", ns, podName, pdMember.ID)
pvcUIDSet := make(map[types.UID]struct{})
for _, pvc := range pvcs {
pvcUIDSet[pvc.UID] = struct{}{}
}
tc.Status.PD.FailureMembers[pdName] = v1alpha1.PDFailureMember{
PodName: podName,
MemberID: pdMember.ID,
PVCUIDSet: pvcUIDSet,
MemberDeleted: false,
CreatedAt: metav1.Now(),
}
return controller.RequeueErrorf("marking Pod: %s/%s pd member: %s as failure", ns, podName, pdMember.Name)
}
- TiDB Operator calls the
tryToDeleteAFailureMember
function to process theFailureMembers
. It traverses all failure members, and when it encounters a member whoseMemberDeleted
isFalse
, it calls the PD client to delete the member and tries to recover the Pod.
func (f *pdFailover) tryToDeleteAFailureMember(tc *v1alpha1.TidbCluster) error {
ns := tc.GetNamespace()
tcName := tc.GetName()
var failureMember *v1alpha1.PDFailureMember
var failurePodName string
var failurePDName string
for pdName, pdMember := range tc.Status.PD.FailureMembers {
if !pdMember.MemberDeleted {
failureMember = &pdMember
failurePodName = strings.Split(pdName, ".")[0]
failurePDName = pdName
break
}
}
if failureMember == nil {
klog.Infof("No PD FailureMembers to delete for tc %s/%s", ns, tcName)
return nil
}
memberID, err := strconv.ParseUint(failureMember.MemberID, 10, 64)
if err != nil {
return err
}
if err := controller.GetPDClient(f.deps.PDControl, tc).DeleteMemberByID(memberID); err != nil {
klog.Errorf("pd failover[tryToDeleteAFailureMember]: failed to delete member %s/%s(%d), error: %v", ns, failurePodName, memberID, err)
return err
}
klog.Infof("pd failover[tryToDeleteAFailureMember]: delete member %s/%s(%d) successfully", ns, failurePodName, memberID)
...
Delete the failed Pod.
pod, err := f.deps.PodLister.Pods(ns).Get(failurePodName)
if err != nil && !errors.IsNotFound(err) {
return fmt.Errorf("pd failover[tryToDeleteAFailureMember]: failed to get pod %s/%s for tc %s/%s, error: %s", ns, failurePodName, ns, tcName, err)
}
if pod != nil {
if pod.DeletionTimestamp == nil {
if err := f.deps.PodControl.DeletePod(tc, pod); err != nil {
return err
}
}
} else {
klog.Infof("pd failover[tryToDeleteAFailureMember]: failure pod %s/%s not found, skip", ns, failurePodName)
}
Delete the PVC.
for _, pvc := range pvcs {
_, pvcUIDExist := failureMember.PVCUIDSet[pvc.GetUID()]
// for backward compatibility, if there exists failureMembers and user upgrades operator to newer version
// there will be failure member structures with PVCUID set from api server, we should handle this as pvcUIDExist == true
if pvc.GetUID() == failureMember.PVCUID {
pvcUIDExist = true
}
if pvc.DeletionTimestamp == nil && pvcUIDExist {
if err := f.deps.PVCControl.DeletePVC(tc, pvc); err != nil {
klog.Errorf("pd failover[tryToDeleteAFailureMember]: failed to delete PVC: %s/%s, error: %s", ns, pvc.Name, err)
return err
}
klog.Infof("pd failover[tryToDeleteAFailureMember]: delete PVC %s/%s successfully", ns, pvc.Name)
}
}
Mark the status of tc.Status.PD.FailureMembers
as Deleted
.
setMemberDeleted(tc, failurePDName)
tc.PDStsDesiredReplicas()
obtains the number of PD StatefulSet replicas, which equals the number of StatefulSet replicas plus the deleted failure members. When syncing the StatefulSet, TiDB Operator calls the scaling-out logic to add a PD Pod for failover.
func (tc *TidbCluster) GetPDDeletedFailureReplicas() int32 {
var deletedReplicas int32 = 0
for _, failureMember := range tc.Status.PD.FailureMembers {
if failureMember.MemberDeleted {
deletedReplicas++
}
}
return deletedReplicas
}
func (tc *TidbCluster) PDStsDesiredReplicas() int32 {
return tc.Spec.PD.Replicas + tc.GetPDDeletedFailureReplicas()
}
Other component’s control loop
Now that you know how TiDB Operator manages PD’s lifecycle, I’ll move on to the differences between PD and other components.
TiKV and TiFlash
TiKV and TiFlash are two storage engines of TiDB and have a similar lifecycle management mechanism. In this section, I’ll take TiKV as an example.
- When syncing the StatefulSet, the TiKV member manager needs to set TiKV store labels using the
setStoreLabelsForTiKV
function. This function gets the node labels and sets labels for TiKV stores via the PD client’sSetStoreLables
function.
for _, store := range storesInfo.Stores {
nodeName := pod.Spec.NodeName
ls, _ := getNodeLabels(m.deps.NodeLister, nodeName, storeLabels)
if !m.storeLabelsEqualNodeLabels(store.Store.Labels, ls) {
set, err := pdCli.SetStoreLabels(store.Store.Id, ls)
if err != nil {
continue
}
if set {
setCount++
klog.Infof("pod: [%s/%s] set labels: %v successfully", ns, podName, ls)
}
}
}
- When syncing Status, the TiKV member manager calls the PD client’s
GetStores
function and obtains the TiKV store information from PD. It then categorizes the store information for later synchronization. This operation is similar to PD’s syncing status, in which the PD member manager calls theGetMembers
function and records PD member information. - When syncing Service, the TiKV member manager only creates headless Services for parsing TiKV Pod DNS.
- When syncing ConfigMap, the TiKV member manager uses a template in the
templates.go
file, calls theRenderTiKVStartScript
function to generate the TiKV start script, and calls thetransformTiKVConfigMap
function to get the TiKV configuration file. - When performing scaling operations, the TiKV member manager must safely take TiKV stores offline. It calls the PD client’s
DeleteStore
function to delete the corresponding store from the TiKV Pod. - In terms of rolling upgrade, the TiKV member manager must ensure that the Region Leader is not on the Pod to be upgraded. Before the rolling upgrade, TiKV Upgrader checks whether the Pod annotation contains
EvictLeaderBeginTime
and determines whether it has performedEvictLeader
on this Pod. If not, it calls theBeginEvictLeader
function from the PD client and adds an evict leader scheduler to the TiKV store. The Region Leader is then evicted from the TiKV store.
_, evicting := upgradePod.Annotations[EvictLeaderBeginTime]
if !evicting {
return u.beginEvictLeader(tc, storeID, upgradePod)
}
In the readyToUpgrade
function, when the Region Leader is zero, or when the evict Leader duration exceeds tc.Spec.TiKV.EvictLeaderTimeout
, the partition configuration in the StatefulSet UpdateStrategy
is updated, which triggers Pod upgrade. When PD finishes the upgrade, it calls the endEvictLeaderbyStoreID
function to end the operation.
When performing failover, the TiKV member manager records the timestamp of the last store state change.
status.LastTransitionTime = metav1.Now()
if exist && status.State == oldStore.State {
status.LastTransitionTime = oldStore.LastTransitionTime
}
When the store state is v1alpha1.TiKVStateDown
and the downtime exceeds the maximum failover timeout, the TiKV Pod is added to FailureStores
.
if store.State == v1alpha1.TiKVStateDown && time.Now().After(deadline) && !exist {
if tc.Status.TiKV.FailureStores == nil {
tc.Status.TiKV.FailureStores = map[string]v1alpha1.TiKVFailureStore{}
}
if tc.Spec.TiKV.MaxFailoverCount != nil && *tc.Spec.TiKV.MaxFailoverCount > 0 {
maxFailoverCount := *tc.Spec.TiKV.MaxFailoverCount
if len(tc.Status.TiKV.FailureStores) >= int(maxFailoverCount) {
klog.Warningf("%s/%s failure stores count reached the limit: %d", ns, tcName, tc.Spec.TiKV.MaxFailoverCount)
return nil
}
tc.Status.TiKV.FailureStores[storeID] = v1alpha1.TiKVFailureStore{
PodName: podName,
StoreID: store.ID,
CreatedAt: metav1.Now(),
}
msg := fmt.Sprintf("store[%s] is Down", store.ID)
f.deps.Recorder.Event(tc, corev1.EventTypeWarning, unHealthEventReason, fmt.Sprintf(unHealthEventMsgPattern, "tikv", podName, msg))
}
}
- When syncing TiKV StatefulSet, the number of replicas includes the number of
FailureStores
, which triggers the scaling-out logic and completes the failover operation.
TiDB, TiCDC, and Pump
TiDB, TiCDC, and Pump have a similar lifecycle management mechanism. Compared to other components, their rolling upgrade is only carried out when the member is healthy.
During scaling, TiDB Operator needs to pay extra attention to PVC usage. When scaling in the cluster, TiDB Operator needs to add deferDeleting
to ensure data safety; when scaling out the cluster, TiDB Operator also needs to remove the PVC.
In addition, TiDB Operator only supports failover logic for TiDB, but not for TiCDC or Pump.
Summary
In this post, taking PD as an example, we have learned the implementation of PD’s control loop and compared PD with other components. By now you should be familiar with the member manager design of major TiDB components and how TiDB Operator implements TiDB lifecycle management.
In the next article, we’ll discuss how TiDB Operator implements backup and restore for TiDB clusters. Stay tuned!
If you are interested in learning more about TiDB Operator, feel free to join our Slack channel or join our discussions at pingcap/tidb-operator.
If you’ve missed an article in our series, you can read them here:
Experience modern data infrastructure firsthand.
TiDB Cloud Dedicated
A fully-managed cloud DBaaS for predictable workloads
TiDB Cloud Serverless
A fully-managed cloud DBaaS for auto-scaling workloads