mirror of
				https://github.com/siderolabs/talos.git
				synced 2025-10-31 16:31:13 +01:00 
			
		
		
		
	Provide a trace for each step of the reset sequence taken, so if one of those fails, integration test produces a meaningful message instead of proceeding and failing somewhere else. More cleanup/refactor, should be functionally equivalent. Fixes #8635 Signed-off-by: Andrey Smirnov <andrey.smirnov@siderolabs.com>
		
			
				
	
	
		
			191 lines
		
	
	
		
			5.4 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
			
		
		
	
	
			191 lines
		
	
	
		
			5.4 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
| // This Source Code Form is subject to the terms of the Mozilla Public
 | |
| // License, v. 2.0. If a copy of the MPL was not distributed with this
 | |
| // file, You can obtain one at http://mozilla.org/MPL/2.0/.
 | |
| 
 | |
| //go:build integration_api
 | |
| 
 | |
| package api
 | |
| 
 | |
| import (
 | |
| 	"bytes"
 | |
| 	"context"
 | |
| 	"fmt"
 | |
| 	"io"
 | |
| 	"path/filepath"
 | |
| 	"testing"
 | |
| 	"time"
 | |
| 
 | |
| 	"github.com/siderolabs/go-retry/retry"
 | |
| 	"google.golang.org/grpc/codes"
 | |
| 
 | |
| 	"github.com/siderolabs/talos/internal/integration/base"
 | |
| 	machineapi "github.com/siderolabs/talos/pkg/machinery/api/machine"
 | |
| 	"github.com/siderolabs/talos/pkg/machinery/client"
 | |
| 	"github.com/siderolabs/talos/pkg/machinery/config/machine"
 | |
| 	"github.com/siderolabs/talos/pkg/machinery/constants"
 | |
| )
 | |
| 
 | |
| // EtcdRecoverSuite ...
 | |
| type EtcdRecoverSuite struct {
 | |
| 	base.K8sSuite
 | |
| 
 | |
| 	ctx       context.Context //nolint:containedctx
 | |
| 	ctxCancel context.CancelFunc
 | |
| }
 | |
| 
 | |
| // SuiteName ...
 | |
| func (suite *EtcdRecoverSuite) SuiteName() string {
 | |
| 	return "api.EtcdRecoverSuite"
 | |
| }
 | |
| 
 | |
| // SetupTest ...
 | |
| func (suite *EtcdRecoverSuite) SetupTest() {
 | |
| 	if testing.Short() {
 | |
| 		suite.T().Skip("skipping in short mode")
 | |
| 	}
 | |
| 
 | |
| 	// make sure we abort at some point in time, but give enough room for Recovers
 | |
| 	suite.ctx, suite.ctxCancel = context.WithTimeout(context.Background(), 10*time.Minute)
 | |
| }
 | |
| 
 | |
| // TearDownTest ...
 | |
| func (suite *EtcdRecoverSuite) TearDownTest() {
 | |
| 	if suite.ctxCancel != nil {
 | |
| 		suite.ctxCancel()
 | |
| 	}
 | |
| }
 | |
| 
 | |
| // TestSnapshotRecover snapshot etcd, wipes control plane nodes and recovers etcd from a snapshot.
 | |
| func (suite *EtcdRecoverSuite) TestSnapshotRecover() {
 | |
| 	if !suite.Capabilities().SupportsReboot {
 | |
| 		suite.T().Skip("cluster doesn't support reboot")
 | |
| 	}
 | |
| 
 | |
| 	if suite.Cluster == nil {
 | |
| 		suite.T().Skip("without full cluster state reset test is not reliable (can't wait for cluster readiness in between resets)")
 | |
| 	}
 | |
| 
 | |
| 	// 'init' nodes are not compatible with etcd recovery
 | |
| 	suite.Require().Empty(suite.DiscoverNodeInternalIPsByType(suite.ctx, machine.TypeInit))
 | |
| 
 | |
| 	controlPlaneNodes := suite.DiscoverNodeInternalIPsByType(suite.ctx, machine.TypeControlPlane)
 | |
| 	suite.Require().NotEmpty(controlPlaneNodes)
 | |
| 
 | |
| 	snapshotNode := suite.RandomDiscoveredNodeInternalIP(machine.TypeControlPlane)
 | |
| 	recoverNode := suite.RandomDiscoveredNodeInternalIP(machine.TypeControlPlane)
 | |
| 
 | |
| 	suite.WaitForBootDone(suite.ctx)
 | |
| 
 | |
| 	suite.T().Logf("taking etcd snapshot at node %q", snapshotNode)
 | |
| 
 | |
| 	var snapshot bytes.Buffer
 | |
| 
 | |
| 	suite.Require().NoError(suite.snapshotEtcd(snapshotNode, &snapshot))
 | |
| 
 | |
| 	// leave etcd on all nodes but one
 | |
| 	for _, node := range controlPlaneNodes[1:] {
 | |
| 		suite.T().Logf("leaving etcd on node %q", node)
 | |
| 
 | |
| 		nodeCtx := client.WithNode(suite.ctx, node)
 | |
| 
 | |
| 		_, err := suite.Client.EtcdForfeitLeadership(nodeCtx, &machineapi.EtcdForfeitLeadershipRequest{})
 | |
| 		suite.Require().NoError(err)
 | |
| 
 | |
| 		err = suite.Client.EtcdLeaveCluster(nodeCtx, &machineapi.EtcdLeaveClusterRequest{})
 | |
| 		suite.Require().NoError(err)
 | |
| 	}
 | |
| 
 | |
| 	// wipe ephemeral partition on all control plane nodes, starting with the one that still has etcd running
 | |
| 	for _, node := range controlPlaneNodes {
 | |
| 		suite.ResetNode(suite.ctx, node, &machineapi.ResetRequest{
 | |
| 			Reboot:   true,
 | |
| 			Graceful: false,
 | |
| 			SystemPartitionsToWipe: []*machineapi.ResetPartitionSpec{
 | |
| 				{
 | |
| 					Label: constants.EphemeralPartitionLabel,
 | |
| 					Wipe:  true,
 | |
| 				},
 | |
| 			},
 | |
| 		}, false)
 | |
| 	}
 | |
| 
 | |
| 	// verify that etcd data directory doesn't exist on the nodes
 | |
| 	for _, node := range controlPlaneNodes {
 | |
| 		stream, err := suite.Client.MachineClient.List(client.WithNode(suite.ctx, node), &machineapi.ListRequest{Root: filepath.Join(constants.EtcdDataPath, "member")})
 | |
| 		suite.Require().NoError(err)
 | |
| 
 | |
| 		_, err = stream.Recv()
 | |
| 		suite.Require().Error(err)
 | |
| 		suite.Require().Equal(client.StatusCode(err), codes.Unknown)
 | |
| 		suite.Require().Contains(client.Status(err).Message(), "no such file or directory")
 | |
| 	}
 | |
| 
 | |
| 	suite.T().Logf("recovering etcd snapshot at node %q", recoverNode)
 | |
| 
 | |
| 	suite.Require().NoError(suite.recoverEtcd(recoverNode, bytes.NewReader(snapshot.Bytes())))
 | |
| 
 | |
| 	suite.AssertClusterHealthy(suite.ctx)
 | |
| }
 | |
| 
 | |
| func (suite *EtcdRecoverSuite) snapshotEtcd(snapshotNode string, dest io.Writer) error {
 | |
| 	ctx := client.WithNodes(suite.ctx, snapshotNode)
 | |
| 
 | |
| 	r, err := suite.Client.EtcdSnapshot(ctx, &machineapi.EtcdSnapshotRequest{})
 | |
| 	if err != nil {
 | |
| 		return fmt.Errorf("error reading snapshot: %w", err)
 | |
| 	}
 | |
| 
 | |
| 	defer r.Close() //nolint:errcheck
 | |
| 
 | |
| 	_, err = io.Copy(dest, r)
 | |
| 
 | |
| 	return err
 | |
| }
 | |
| 
 | |
| func (suite *EtcdRecoverSuite) recoverEtcd(recoverNode string, src io.ReadSeeker) error {
 | |
| 	ctx := client.WithNodes(suite.ctx, recoverNode)
 | |
| 
 | |
| 	suite.T().Log("uploading the snapshot")
 | |
| 
 | |
| 	if err := retry.Constant(time.Minute, retry.WithUnits(time.Millisecond*200)).RetryWithContext(
 | |
| 		ctx, func(ctx context.Context) error {
 | |
| 			_, err := src.Seek(0, io.SeekStart)
 | |
| 			if err != nil {
 | |
| 				return err
 | |
| 			}
 | |
| 
 | |
| 			_, err = suite.Client.EtcdRecover(ctx, src)
 | |
| 
 | |
| 			if client.StatusCode(err) == codes.FailedPrecondition {
 | |
| 				return retry.ExpectedError(err)
 | |
| 			}
 | |
| 
 | |
| 			return err
 | |
| 		},
 | |
| 	); err != nil {
 | |
| 		return fmt.Errorf("error uploading snapshot: %w", err)
 | |
| 	}
 | |
| 
 | |
| 	suite.T().Log("bootstrapping from the snapshot")
 | |
| 
 | |
| 	return retry.Constant(time.Minute, retry.WithUnits(time.Millisecond*200)).RetryWithContext(
 | |
| 		ctx, func(ctx context.Context) error {
 | |
| 			err := suite.Client.Bootstrap(
 | |
| 				ctx, &machineapi.BootstrapRequest{
 | |
| 					RecoverEtcd: true,
 | |
| 				},
 | |
| 			)
 | |
| 
 | |
| 			if client.StatusCode(err) == codes.FailedPrecondition || client.StatusCode(err) == codes.DeadlineExceeded {
 | |
| 				return retry.ExpectedError(err)
 | |
| 			}
 | |
| 
 | |
| 			return err
 | |
| 		},
 | |
| 	)
 | |
| }
 | |
| 
 | |
| func init() {
 | |
| 	allSuites = append(allSuites, new(EtcdRecoverSuite))
 | |
| }
 |