@@ -252,13 +252,21 @@ var _ = ginkgo.Describe("[tidb-operator][Stability]", func() {
252
252
// local-volume-provisioner, then orphan PVs will be garbaged
253
253
// collected and will not cause problem even if the name of deleted
254
254
// node is used again in the future.
255
- // - (GKE only) delete failed pods and associated PVCs/PVs if local PVs
256
- // are using unique paths
257
- // - Since GKE 1.11, the node object will not be recreated. If local
258
- // volumes are mounted at the same paths, PD/TiKV pods will be
259
- // running soon when underlying instance is recreated and running.
260
- // Otherwise, we need to delete failed pods and associated PVCs/PVs.
261
- // PVs must be deleted because their paths does not exist now.
255
+ // - (GKE only, fixed path) nothing need to do
256
+ // - Because the node name does not change, old PVs can be used. Note
257
+ // that `setPVOwnerRef` cannot be enabled because the node object
258
+ // could get deleted if it takes too long for the instance to
259
+ // recreate.
260
+ // - Optionally, you can deleted failed pods to make them to start
261
+ // soon. This is due to exponential crash loop back off.
262
+ // - (GKE only, unique paths) delete failed pods and associated PVCs/PVs
263
+ // - This is because even if the node name does not change, old PVs
264
+ // are invalid because unique volume paths are used. We must delete
265
+ // them all and wait for Kubernetes to rcreate and run agian.
266
+ // - PVs must be deleted because the PVs are invalid and should not
267
+ // exist anymore. We can configure `setPVOwnerRef` to clean unused
268
+ // PVs when the node object is deleted, but the node object will not
269
+ // get deleted if the instance is recreated soon.
262
270
//
263
271
// Note that:
264
272
// - We assume local storage is used, otherwise PV can be re-attached
@@ -277,7 +285,8 @@ var _ = ginkgo.Describe("[tidb-operator][Stability]", func() {
277
285
// the new machine.
278
286
// - In GKE (1.11+), the node object are no longer recreated on
279
287
// upgrade/repair even though the underlying instance is recreated and
280
- // local disks are wiped.
288
+ // local disks are wiped. However , he node object could get deleted
289
+ // if it takes too long for the instance to recreate.
281
290
//
282
291
// Related issues:
283
292
// - https://github.com/pingcap/tidb-operator/issues/1546
0 commit comments