Skip to content

Commit

Permalink
e2e-tests:
Browse files Browse the repository at this point in the history
move more log. calls to t.Log to make e2e output more readable

Signed-off-by: Magnus Kulke <magnuskulke@microsoft.com>
  • Loading branch information
mkulke committed Jul 30, 2024
1 parent f968c29 commit d83123a
Show file tree
Hide file tree
Showing 2 changed files with 27 additions and 28 deletions.
16 changes: 8 additions & 8 deletions src/cloud-api-adaptor/test/e2e/assessment_runner.go
Original file line number Diff line number Diff line change
Expand Up @@ -276,15 +276,15 @@ func (tc *TestCase) Run() {
t.Logf("Expected Pod State: %v", tc.podState)
yamlData, err := yaml.Marshal(pod.Status)
if err != nil {
fmt.Println("Error marshaling pod.Status to YAML: ", err.Error())
t.Logf("Error marshaling pod.Status to YAML: ", err.Error())
} else {
t.Logf("Current Pod State: %v", string(yamlData))
}
if pod.Status.Phase == v1.PodRunning {
fmt.Printf("Log of the pod %.v \n===================\n", pod.Name)
t.Logf("Log of the pod %.v \n===================\n", pod.Name)
podLogString, _ := GetPodLog(ctx, client, *pod)
fmt.Println(podLogString)
fmt.Printf("===================\n")
t.Log(podLogString)
t.Logf("===================\n")
}
t.Fatal(err)
}
Expand All @@ -295,11 +295,11 @@ func (tc *TestCase) Run() {
t.Fatal(err)
}
clusterIP := WaitForClusterIP(t, client, tc.service)
log.Printf("webserver service is available on cluster IP: %s", clusterIP)
t.Logf("webserver service is available on cluster IP: %s", clusterIP)
}
if tc.extraPods != nil {
for _, extraPod := range tc.extraPods {
fmt.Printf("Provision extra pod %s", extraPod.pod.Name)
t.Logf("Provision extra pod %s", extraPod.pod.Name)
err := ProvisionPod(ctx, client, t, extraPod.pod, extraPod.podState, extraPod.testCommands)
if err != nil {
t.Fatal(err)
Expand Down Expand Up @@ -329,7 +329,7 @@ func (tc *TestCase) Run() {
}
if podLogString != "" {
if strings.Contains(podLogString, tc.expectedPodLogString) {
log.Printf("Output Log from Pod: %s", podLogString)
t.Logf("Output Log from Pod: %s", podLogString)
} else {
t.Errorf("Job Created pod with Invalid log")
}
Expand Down Expand Up @@ -435,7 +435,7 @@ func (tc *TestCase) Run() {
t.Fatal(error)
}
} else if profile != "" {
fmt.Printf("PodVM Created with Instance Type %v", profile)
t.Logf("PodVM Created with Instance Type %v", profile)
if tc.FailReason != "" {
var podlist v1.PodList
var podLogString string
Expand Down
39 changes: 19 additions & 20 deletions src/cloud-api-adaptor/test/e2e/rolling_update.go
Original file line number Diff line number Diff line change
Expand Up @@ -12,7 +12,6 @@ import (
"testing"
"time"

log "github.com/sirupsen/logrus"
appsv1 "k8s.io/api/apps/v1"
v1 "k8s.io/api/core/v1"
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
Expand Down Expand Up @@ -143,22 +142,22 @@ func DoTestCaaDaemonsetRollingUpdate(t *testing.T, testEnv env.Environment, asse
t.Fatal(err)
}

log.Info("Creating webserver deployment...")
t.Logf("Creating webserver deployment...")
if err = client.Resources().Create(ctx, deployment); err != nil {
t.Fatal(err)
}
waitForDeploymentAvailable(t, client, deployment, rc)
log.Info("webserver deployment is available now")
t.Logf("webserver deployment is available now")

// Cache Pod VM instance IDs before upgrade
assert.CachePodVmIDs(t, deploymentName)

log.Info("Creating webserver Service")
t.Logf("Creating webserver Service")
if err = client.Resources().Create(ctx, svc); err != nil {
t.Fatal(err)
}
clusterIP := WaitForClusterIP(t, client, svc)
log.Printf("webserver service is available on cluster IP: %s", clusterIP)
t.Logf("webserver service is available on cluster IP: %s", clusterIP)

// Update verify command
verifyPod.Spec.Containers[0].Command = append(
Expand Down Expand Up @@ -196,7 +195,7 @@ func DoTestCaaDaemonsetRollingUpdate(t *testing.T, testEnv env.Environment, asse
if err = client.Resources().Get(ctx, caaDaemonSetName, caaNamespace, ds); err != nil {
t.Fatal(err)
}
log.Info("Force to update CAA pods by increasing StartupProbe.FailureThreshold")
t.Logf("Force to update CAA pods by increasing StartupProbe.FailureThreshold")
ds.Spec.Template.Spec.Containers[0].StartupProbe.FailureThreshold += 1
if err = client.Resources().Update(ctx, ds); err != nil {
t.Fatal(err)
Expand All @@ -210,33 +209,33 @@ func DoTestCaaDaemonsetRollingUpdate(t *testing.T, testEnv env.Environment, asse
if err = client.Resources().Get(ctx, verifyPodName, E2eNamespace, verifyPod); err != nil {
t.Fatal(err)
}
log.Printf("verify pod status: %s", verifyPod.Status.Phase)
t.Logf("verify pod status: %s", verifyPod.Status.Phase)
if verifyPod.Status.Phase != v1.PodRunning {
clientset, err := kubernetes.NewForConfig(client.RESTConfig())
if err != nil {
log.Printf("Failed to new client set: %v", err)
t.Logf("Failed to new client set: %v", err)
} else {
req := clientset.CoreV1().Pods(E2eNamespace).GetLogs(verifyPodName, &v1.PodLogOptions{})
podLogs, err := req.Stream(ctx)
if err != nil {
log.Printf("Failed to get pod logs: %v", err)
t.Logf("Failed to get pod logs: %v", err)
} else {
defer podLogs.Close()
buf := new(bytes.Buffer)
_, err = io.Copy(buf, podLogs)
if err != nil {
log.Printf("Failed to copy pod logs: %v", err)
t.Logf("Failed to copy pod logs: %v", err)
} else {
podLogString := strings.TrimSpace(buf.String())
log.Printf("verify pod logs: \n%s", podLogString)
t.Logf("verify pod logs: \n%s", podLogString)
}
}
}
t.Fatal(fmt.Errorf("verify pod is not running"))
}

time.Sleep(OLD_VM_DELETION_TIMEOUT)
log.Info("Verify old VM instances have been deleted:")
t.Logf("Verify old VM instances have been deleted:")
assert.VerifyOldVmDeleted(t)

return ctx
Expand All @@ -247,17 +246,17 @@ func DoTestCaaDaemonsetRollingUpdate(t *testing.T, testEnv env.Environment, asse
t.Fatal(err)
}

log.Info("Deleting verify pod...")
t.Log("Deleting verify pod...")
if err = client.Resources().Delete(ctx, verifyPod); err != nil {
t.Fatal(err)
}

log.Info("Deleting webserver service...")
t.Log("Deleting webserver service...")
if err = client.Resources().Delete(ctx, svc); err != nil {
t.Fatal(err)
}

log.Info("Deleting webserver deployment...")
t.Log("Deleting webserver deployment...")
if err = client.Resources().Delete(ctx, deployment); err != nil {
t.Fatal(err)
}
Expand All @@ -272,12 +271,12 @@ func waitForCaaDaemonSetUpdated(t *testing.T, client klient.Client, ds *appsv1.D
if err := wait.For(conditions.New(client.Resources()).ResourceMatch(ds, func(object k8s.Object) bool {
dsObj, ok := object.(*appsv1.DaemonSet)
if !ok {
log.Printf("Not a DaemonSet object: %v", object)
t.Logf("Not a DaemonSet object: %v", object)
return false
}

log.Printf("Current CAA DaemonSet UpdatedNumberScheduled: %d", dsObj.Status.UpdatedNumberScheduled)
log.Printf("Current CAA DaemonSet NumberAvailable: %d", dsObj.Status.NumberAvailable)
t.Logf("Current CAA DaemonSet UpdatedNumberScheduled: %d", dsObj.Status.UpdatedNumberScheduled)
t.Logf("Current CAA DaemonSet NumberAvailable: %d", dsObj.Status.NumberAvailable)
return dsObj.Status.UpdatedNumberScheduled == rc && dsObj.Status.NumberAvailable == rc
}), wait.WithTimeout(WAIT_DEPLOYMENT_AVAILABLE_TIMEOUT)); err != nil {
t.Fatal(err)
Expand All @@ -288,11 +287,11 @@ func waitForDeploymentAvailable(t *testing.T, client klient.Client, deployment *
if err := wait.For(conditions.New(client.Resources()).ResourceMatch(deployment, func(object k8s.Object) bool {
deployObj, ok := object.(*appsv1.Deployment)
if !ok {
log.Printf("Not a Deployment object: %v", object)
t.Logf("Not a Deployment object: %v", object)
return false
}

log.Printf("Current deployment available replicas: %d", deployObj.Status.AvailableReplicas)
t.Logf("Current deployment available replicas: %d", deployObj.Status.AvailableReplicas)
return deployObj.Status.AvailableReplicas == rc
}), wait.WithTimeout(WAIT_DEPLOYMENT_AVAILABLE_TIMEOUT)); err != nil {
t.Fatal(err)
Expand Down

0 comments on commit d83123a

Please sign in to comment.