// Blocks outgoing network traffic on 'node'. Then runs testFunc and returns its status. // At the end (even in case of errors), the network traffic is brought back to normal. // This function executes commands on a node so it will work only for some // environments. func testUnderTemporaryNetworkFailure(c clientset.Interface, ns string, node *api.Node, testFunc func()) { host := framework.GetNodeExternalIP(node) master := framework.GetMasterAddress(c) By(fmt.Sprintf("block network traffic from node %s to the master", node.Name)) defer func() { // This code will execute even if setting the iptables rule failed. // It is on purpose because we may have an error even if the new rule // had been inserted. (yes, we could look at the error code and ssh error // separately, but I prefer to stay on the safe side). By(fmt.Sprintf("Unblock network traffic from node %s to the master", node.Name)) framework.UnblockNetwork(host, master) }() framework.Logf("Waiting %v to ensure node %s is ready before beginning test...", resizeNodeReadyTimeout, node.Name) if !framework.WaitForNodeToBe(c, node.Name, api.NodeReady, true, resizeNodeReadyTimeout) { framework.Failf("Node %s did not become ready within %v", node.Name, resizeNodeReadyTimeout) } framework.BlockNetwork(host, master) framework.Logf("Waiting %v for node %s to be not ready after simulated network failure", resizeNodeNotReadyTimeout, node.Name) if !framework.WaitForNodeToBe(c, node.Name, api.NodeReady, false, resizeNodeNotReadyTimeout) { framework.Failf("Node %s did not become not-ready within %v", node.Name, resizeNodeNotReadyTimeout) } testFunc() // network traffic is unblocked in a deferred function }
MountPath: etcLocaltime, }, { Name: configVolume, MountPath: configDir, }, }, }, }, }, }) Expect(err).NotTo(HaveOccurred()) By("Wait for node problem detector running") Expect(f.WaitForPodRunning(name)).To(Succeed()) // Get the node time nodeIP := framework.GetNodeExternalIP(node) result, err := framework.SSH("date '+%FT%T.%N%:z'", nodeIP, framework.TestContext.Provider) Expect(err).ShouldNot(HaveOccurred()) Expect(result.Code).Should(BeZero()) nodeTime, err = time.Parse(time.RFC3339, strings.TrimSpace(result.Stdout)) Expect(err).ShouldNot(HaveOccurred()) }) It("should generate node condition and events for corresponding errors", func() { for _, test := range []struct { description string timestamp time.Time message string messageNum int events int conditionReason string
Expect(ok).To(Equal(true)) newNode <- n }, }, ) defer func() { // Will not explicitly close newNode channel here due to // race condition where stopCh and newNode are closed but informer onUpdate still executes. close(stopCh) }() go controller.Run(stopCh) By(fmt.Sprintf("Block traffic from node %s to the master", node.Name)) host := framework.GetNodeExternalIP(&node) master := framework.GetMasterAddress(c) defer func() { By(fmt.Sprintf("Unblock traffic from node %s to the master", node.Name)) framework.UnblockNetwork(host, master) if CurrentGinkgoTestDescription().Failed { return } By("Expect to observe node and pod status change from NotReady to Ready after network connectivity recovers") expectNodeReadiness(true, newNode) if err = framework.WaitForMatchPodsCondition(c, podOpts, "Running and Ready", podReadyTimeout, testutils.PodRunningReady); err != nil { framework.Failf("Pods on node %s did not become ready and running within %v: %v", node.Name, podReadyTimeout, err) } }()
NFS: &v1.NFSVolumeSource{ Server: nfsServerIP, Path: "/exports", ReadOnly: false, }, }, } } // Get the first ready node IP that is not hosting the NFS pod. if clientNodeIP == "" { framework.Logf("Designating test node") nodes := framework.GetReadySchedulableNodesOrDie(c) for _, node := range nodes.Items { if node.Name != nfsServerPod.Spec.NodeName { clientNode = &node clientNodeIP = framework.GetNodeExternalIP(clientNode) Expect(clientNodeIP).NotTo(BeEmpty()) break } } } }) AddCleanupAction(func() { if nfsServerPod != nil && c != nil { By("Deleting NFS server pod") nfsServerPodCleanup(c, nfsServerConfig) nfsServerPod = nil } })