/* Copyright 2015 The Kubernetes Authors. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. */ package gcp import ( "context" "fmt" "os/exec" "path" "regexp" "strconv" "strings" "time" "github.com/onsi/ginkgo/v2" v1 "k8s.io/api/core/v1" metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" clientset "k8s.io/client-go/kubernetes" "k8s.io/kubernetes/test/e2e/common" "k8s.io/kubernetes/test/e2e/feature" "k8s.io/kubernetes/test/e2e/framework" e2enode "k8s.io/kubernetes/test/e2e/framework/node" e2epod "k8s.io/kubernetes/test/e2e/framework/pod" e2eskipper "k8s.io/kubernetes/test/e2e/framework/skipper" admissionapi "k8s.io/pod-security-admission/api" ) func addMasterReplica(zone string) error { framework.Logf(fmt.Sprintf("Adding a new master replica, zone: %s", zone)) _, _, err := framework.RunCmd(path.Join(framework.TestContext.RepoRoot, "hack/e2e-internal/e2e-grow-cluster.sh"), zone, "true", "true", "false") if err != nil { return err } return nil } func removeMasterReplica(zone string) error { framework.Logf(fmt.Sprintf("Removing an existing master replica, zone: %s", zone)) _, _, err := framework.RunCmd(path.Join(framework.TestContext.RepoRoot, "hack/e2e-internal/e2e-shrink-cluster.sh"), zone, "true", "false", "false") if err != nil { return err } return nil } func addWorkerNodes(zone string) error { framework.Logf(fmt.Sprintf("Adding worker nodes, zone: %s", zone)) _, _, err := framework.RunCmd(path.Join(framework.TestContext.RepoRoot, "hack/e2e-internal/e2e-grow-cluster.sh"), zone, "true", "false", "true") if err != nil { return err } return nil } func removeWorkerNodes(zone string) error { framework.Logf(fmt.Sprintf("Removing worker nodes, zone: %s", zone)) _, _, err := framework.RunCmd(path.Join(framework.TestContext.RepoRoot, "hack/e2e-internal/e2e-shrink-cluster.sh"), zone, "true", "true", "true") if err != nil { return err } return nil } func verifyRCs(ctx context.Context, c clientset.Interface, ns string, names []string) { for _, name := range names { framework.ExpectNoError(e2epod.VerifyPods(ctx, c, ns, name, true, 1)) } } func createNewRC(c clientset.Interface, ns string, name string) { _, err := common.NewRCByName(c, ns, name, 1, nil, nil) framework.ExpectNoError(err) } func findRegionForZone(zone string) string { region, err := exec.Command("gcloud", "compute", "zones", "list", zone, "--quiet", "--format=csv[no-heading](region)").Output() framework.ExpectNoError(err) if string(region) == "" { framework.Failf("Region not found; zone: %s", zone) } return string(region) } func findZonesForRegion(region string) []string { output, err := exec.Command("gcloud", "compute", "zones", "list", "--filter=region="+region, "--quiet", "--format=csv[no-heading](name)").Output() framework.ExpectNoError(err) zones := strings.Split(string(output), "\n") return zones } // removeZoneFromZones removes zone from zones slide. // Please note that entries in zones can be repeated. In such situation only one replica is removed. func removeZoneFromZones(zones []string, zone string) []string { idx := -1 for j, z := range zones { if z == zone { idx = j break } } if idx >= 0 { return zones[:idx+copy(zones[idx:], zones[idx+1:])] } return zones } // generateMasterRegexp returns a regex for matching master node name. func generateMasterRegexp(prefix string) string { return prefix + "(-...)?" } // waitForMasters waits until the cluster has the desired number of ready masters in it. func waitForMasters(ctx context.Context, masterPrefix string, c clientset.Interface, size int, timeout time.Duration) error { for start := time.Now(); time.Since(start) < timeout; time.Sleep(20 * time.Second) { nodes, err := c.CoreV1().Nodes().List(ctx, metav1.ListOptions{}) if err != nil { framework.Logf("Failed to list nodes: %v", err) continue } // Filter out nodes that are not master replicas e2enode.Filter(nodes, func(node v1.Node) bool { res, err := regexp.Match(generateMasterRegexp(masterPrefix), ([]byte)(node.Name)) if err != nil { framework.Logf("Failed to match regexp to node name: %v", err) return false } return res }) numNodes := len(nodes.Items) // Filter out not-ready nodes. e2enode.Filter(nodes, func(node v1.Node) bool { return e2enode.IsConditionSetAsExpected(&node, v1.NodeReady, true) }) numReady := len(nodes.Items) if numNodes == size && numReady == size { framework.Logf("Cluster has reached the desired number of masters %d", size) return nil } framework.Logf("Waiting for the number of masters %d, current %d, not ready master nodes %d", size, numNodes, numNodes-numReady) } return fmt.Errorf("timeout waiting %v for the number of masters to be %d", timeout, size) } var _ = SIGDescribe("HA-master", feature.HAMaster, func() { f := framework.NewDefaultFramework("ha-master") f.NamespacePodSecurityLevel = admissionapi.LevelPrivileged var c clientset.Interface var ns string var additionalReplicaZones []string var additionalNodesZones []string var existingRCs []string ginkgo.BeforeEach(func(ctx context.Context) { e2eskipper.SkipUnlessProviderIs("gce") c = f.ClientSet ns = f.Namespace.Name framework.ExpectNoError(waitForMasters(ctx, framework.TestContext.CloudConfig.MasterName, c, 1, 10*time.Minute)) additionalReplicaZones = make([]string, 0) existingRCs = make([]string, 0) }) ginkgo.AfterEach(func(ctx context.Context) { // Clean-up additional worker nodes if the test execution was broken. for _, zone := range additionalNodesZones { removeWorkerNodes(zone) } framework.ExpectNoError(e2enode.AllNodesReady(ctx, c, 5*time.Minute)) // Clean-up additional master replicas if the test execution was broken. for _, zone := range additionalReplicaZones { removeMasterReplica(zone) } framework.ExpectNoError(waitForMasters(ctx, framework.TestContext.CloudConfig.MasterName, c, 1, 10*time.Minute)) }) type Action int const ( None Action = iota AddReplica RemoveReplica AddNodes RemoveNodes ) step := func(ctx context.Context, action Action, zone string) { switch action { case None: case AddReplica: framework.ExpectNoError(addMasterReplica(zone)) additionalReplicaZones = append(additionalReplicaZones, zone) case RemoveReplica: framework.ExpectNoError(removeMasterReplica(zone)) additionalReplicaZones = removeZoneFromZones(additionalReplicaZones, zone) case AddNodes: framework.ExpectNoError(addWorkerNodes(zone)) additionalNodesZones = append(additionalNodesZones, zone) case RemoveNodes: framework.ExpectNoError(removeWorkerNodes(zone)) additionalNodesZones = removeZoneFromZones(additionalNodesZones, zone) } framework.ExpectNoError(waitForMasters(ctx, framework.TestContext.CloudConfig.MasterName, c, len(additionalReplicaZones)+1, 10*time.Minute)) framework.ExpectNoError(e2enode.AllNodesReady(ctx, c, 5*time.Minute)) // Verify that API server works correctly with HA master. rcName := "ha-master-" + strconv.Itoa(len(existingRCs)) createNewRC(c, ns, rcName) existingRCs = append(existingRCs, rcName) verifyRCs(ctx, c, ns, existingRCs) } f.It("survive addition/removal replicas same zone", f.WithSerial(), f.WithDisruptive(), func(ctx context.Context) { zone := framework.TestContext.CloudConfig.Zone step(ctx, None, "") numAdditionalReplicas := 2 for i := 0; i < numAdditionalReplicas; i++ { step(ctx, AddReplica, zone) } for i := 0; i < numAdditionalReplicas; i++ { step(ctx, RemoveReplica, zone) } }) f.It("survive addition/removal replicas different zones", f.WithSerial(), f.WithDisruptive(), func(ctx context.Context) { zone := framework.TestContext.CloudConfig.Zone region := findRegionForZone(zone) zones := findZonesForRegion(region) zones = removeZoneFromZones(zones, zone) step(ctx, None, "") // If numAdditionalReplicas is larger then the number of remaining zones in the region, // we create a few masters in the same zone and zone entry is repeated in additionalReplicaZones. numAdditionalReplicas := 2 for i := 0; i < numAdditionalReplicas; i++ { step(ctx, AddReplica, zones[i%len(zones)]) } for i := 0; i < numAdditionalReplicas; i++ { step(ctx, RemoveReplica, zones[i%len(zones)]) } }) f.It("survive addition/removal replicas multizone workers", f.WithSerial(), f.WithDisruptive(), func(ctx context.Context) { zone := framework.TestContext.CloudConfig.Zone region := findRegionForZone(zone) zones := findZonesForRegion(region) zones = removeZoneFromZones(zones, zone) step(ctx, None, "") numAdditionalReplicas := 2 // Add worker nodes. for i := 0; i < numAdditionalReplicas && i < len(zones); i++ { step(ctx, AddNodes, zones[i]) } // Add master repilcas. // // If numAdditionalReplicas is larger then the number of remaining zones in the region, // we create a few masters in the same zone and zone entry is repeated in additionalReplicaZones. for i := 0; i < numAdditionalReplicas; i++ { step(ctx, AddReplica, zones[i%len(zones)]) } // Remove master repilcas. for i := 0; i < numAdditionalReplicas; i++ { step(ctx, RemoveReplica, zones[i%len(zones)]) } // Remove worker nodes. for i := 0; i < numAdditionalReplicas && i < len(zones); i++ { step(ctx, RemoveNodes, zones[i]) } }) })