gitpod/test/tests/workspace/ephemeral_storage_limit_test.go
Wouter Verlaek 757e3467ff
[test] Ephemeral storage limit, fix preview workspace classes (#18088)
* [test] Ephemeral storage limit, fix preview workspace classes

* Check for workspace failure message

* Link WKS-215
2023-07-03 16:44:20 +08:00

110 lines
3.4 KiB
Go

// Copyright (c) 2020 Gitpod GmbH. All rights reserved.
// Licensed under the GNU Affero General Public License (AGPL).
// See License.AGPL.txt in the project root for license information.
package workspace
import (
"context"
"strings"
"testing"
"time"
agent "github.com/gitpod-io/gitpod/test/pkg/agent/workspace/api"
"github.com/gitpod-io/gitpod/test/pkg/integration"
"sigs.k8s.io/e2e-framework/pkg/envconf"
"sigs.k8s.io/e2e-framework/pkg/features"
)
func TestEphemeralStorageLimit(t *testing.T) {
f := features.New("ephemeral storage limit").
WithLabel("component", "workspace").
WithLabel("type", "ephemeral storage limit").
Assess("it should stop a workspace that reaches the ephemeral storage limit", func(testCtx context.Context, t *testing.T, cfg *envconf.Config) context.Context {
t.Parallel()
ctx, cancel := context.WithTimeout(testCtx, time.Duration(5*time.Minute))
defer cancel()
api := integration.NewComponentAPI(ctx, cfg.Namespace(), kubeconfig, cfg.Client())
t.Cleanup(func() {
api.Done(t)
})
nfo, stopWs, err := integration.LaunchWorkspaceDirectly(t, ctx, api)
if err != nil {
t.Fatal(err)
}
t.Cleanup(func() {
sctx, scancel := context.WithTimeout(context.Background(), 5*time.Minute)
defer scancel()
sapi := integration.NewComponentAPI(sctx, cfg.Namespace(), kubeconfig, cfg.Client())
defer sapi.Done(t)
if _, err = stopWs(true, sapi); err != nil {
t.Errorf("cannot stop workspace: %q", err)
}
})
rsa, closer, err := integration.Instrument(integration.ComponentWorkspace, "workspace", cfg.Namespace(), kubeconfig, cfg.Client(), integration.WithInstanceID(nfo.Req.Id))
integration.DeferCloser(t, closer)
if err != nil {
t.Fatalf("unexpected error instrumenting workspace: %v", err)
}
defer rsa.Close()
t.Logf("allocating disk space")
var res agent.ExecResponse
err = rsa.Call("WorkspaceAgent.Exec", &agent.ExecRequest{
Dir: "/workspace",
Command: "fallocate",
Args: []string{"-l", "11G", "/var/tmp/test1"},
}, &res)
if err != nil {
t.Fatal(err)
}
if res.ExitCode != 0 {
t.Fatalf("fallocate failed (%d): %s", res.ExitCode, res.Stderr)
}
t.Logf("expecting workspace to stop")
ready := make(chan struct{}, 1)
status, err := integration.WaitForWorkspaceStop(t, ctx, ready, api, nfo.Req.Id, nfo.WorkspaceID, integration.WorkspaceCanFail)
if err != nil {
t.Fatal(err)
}
t.Logf("workspace stopped, checking for failed condition")
if status == nil || status.Conditions == nil {
t.Fatalf("workspace status is empty: %v", status)
}
if status.Conditions.Failed == "" {
t.Fatalf("expected failed condition but got none: %v", status)
}
expectedFailures := []string{
"Evicted: Pod ephemeral local storage usage exceeds the total limit of containers ",
// Until WKS-215 is fixed we can sometimes see the below error:
"container workspace completed; containers of a workspace pod are not supposed to do that",
}
foundExpectedFailure := false
for _, ef := range expectedFailures {
if strings.Contains(status.Conditions.Failed, ef) {
foundExpectedFailure = true
break
}
}
if !foundExpectedFailure {
t.Fatalf("expected failed condition to contain one of %v but got: %v", expectedFailures, status.Conditions.Failed)
}
t.Logf("workspace failed as expected: %v", status.Conditions.Failed)
return testCtx
}).
Feature()
testEnv.Test(t, f)
}