mirror of
https://github.com/siderolabs/talos.git
synced 2025-08-08 07:37:06 +02:00
`config.Container` implements a multi-doc container which implements both `Container` interface (encoding, validation, etc.), and `Conifg` interface (accessing parts of the config). Refactor `generate` and `bundle` packages to support multi-doc, and provide backwards compatibility. Implement a first (mostly example) machine config document for SideroLink API URL. Many places don't properly support multi-doc yet (e.g. config patches). Signed-off-by: Andrey Smirnov <andrey.smirnov@talos-systems.com>
217 lines
5.4 KiB
Go
217 lines
5.4 KiB
Go
// This Source Code Form is subject to the terms of the Mozilla Public
|
|
// License, v. 2.0. If a copy of the MPL was not distributed with this
|
|
// file, You can obtain one at http://mozilla.org/MPL/2.0/.
|
|
|
|
//go:build integration_api
|
|
|
|
package api
|
|
|
|
import (
|
|
"context"
|
|
"fmt"
|
|
"sync"
|
|
"testing"
|
|
"time"
|
|
|
|
"github.com/siderolabs/go-retry/retry"
|
|
|
|
"github.com/siderolabs/talos/internal/integration/base"
|
|
"github.com/siderolabs/talos/pkg/machinery/client"
|
|
"github.com/siderolabs/talos/pkg/machinery/config/machine"
|
|
)
|
|
|
|
// RebootSuite ...
|
|
type RebootSuite struct {
|
|
base.APISuite
|
|
|
|
ctx context.Context //nolint:containedctx
|
|
ctxCancel context.CancelFunc
|
|
}
|
|
|
|
// SuiteName ...
|
|
func (suite *RebootSuite) SuiteName() string {
|
|
return "api.RebootSuite"
|
|
}
|
|
|
|
// SetupTest ...
|
|
func (suite *RebootSuite) SetupTest() {
|
|
if testing.Short() {
|
|
suite.T().Skip("skipping in short mode")
|
|
}
|
|
|
|
// make sure we abort at some point in time, but give enough room for reboots
|
|
suite.ctx, suite.ctxCancel = context.WithTimeout(context.Background(), 30*time.Minute)
|
|
}
|
|
|
|
// TearDownTest ...
|
|
func (suite *RebootSuite) TearDownTest() {
|
|
if suite.ctxCancel != nil {
|
|
suite.ctxCancel()
|
|
}
|
|
}
|
|
|
|
// TestRebootNodeByNode reboots cluster node by node, waiting for health between reboots.
|
|
func (suite *RebootSuite) TestRebootNodeByNode() {
|
|
if !suite.Capabilities().SupportsReboot {
|
|
suite.T().Skip("cluster doesn't support reboots")
|
|
}
|
|
|
|
nodes := suite.DiscoverNodeInternalIPs(suite.ctx)
|
|
suite.Require().NotEmpty(nodes)
|
|
|
|
for _, node := range nodes {
|
|
suite.T().Log("rebooting node", node)
|
|
|
|
suite.AssertRebooted(
|
|
suite.ctx, node, func(nodeCtx context.Context) error {
|
|
return base.IgnoreGRPCUnavailable(suite.Client.Reboot(nodeCtx))
|
|
}, 10*time.Minute,
|
|
)
|
|
}
|
|
}
|
|
|
|
// TestRebootMultiple reboots a node, issues consequent reboots
|
|
// reboot should cancel boot sequence, and cancel another reboot.
|
|
func (suite *RebootSuite) TestRebootMultiple() {
|
|
if !suite.Capabilities().SupportsReboot {
|
|
suite.T().Skip("cluster doesn't support reboots")
|
|
}
|
|
|
|
node := suite.RandomDiscoveredNodeInternalIP(machine.TypeWorker)
|
|
nodeCtx := client.WithNodes(suite.ctx, node)
|
|
|
|
bootID := suite.ReadBootIDWithRetry(nodeCtx, time.Minute*5)
|
|
|
|
// Issue reboot.
|
|
suite.Require().NoError(base.IgnoreGRPCUnavailable(
|
|
suite.Client.Reboot(nodeCtx),
|
|
))
|
|
|
|
// Issue reboot once again and wait for node to get a new boot id.
|
|
suite.Require().NoError(base.IgnoreGRPCUnavailable(
|
|
suite.Client.Reboot(nodeCtx),
|
|
))
|
|
|
|
suite.AssertBootIDChanged(nodeCtx, bootID, node, time.Minute*5)
|
|
|
|
bootID = suite.ReadBootIDWithRetry(nodeCtx, time.Minute*5)
|
|
|
|
suite.Require().NoError(retry.Constant(time.Second * 5).Retry(func() error {
|
|
// Issue reboot while the node is still booting.
|
|
err := suite.Client.Reboot(nodeCtx)
|
|
if err != nil {
|
|
return retry.ExpectedError(err)
|
|
}
|
|
|
|
// Reboot again and wait for cluster to become healthy.
|
|
suite.Require().NoError(base.IgnoreGRPCUnavailable(
|
|
suite.Client.Reboot(nodeCtx),
|
|
))
|
|
|
|
return nil
|
|
}))
|
|
|
|
suite.AssertBootIDChanged(nodeCtx, bootID, node, time.Minute*5)
|
|
suite.WaitForBootDone(suite.ctx)
|
|
}
|
|
|
|
// TestRebootAllNodes reboots all cluster nodes at the same time.
|
|
//
|
|
//nolint:gocyclo
|
|
func (suite *RebootSuite) TestRebootAllNodes() {
|
|
if !suite.Capabilities().SupportsReboot {
|
|
suite.T().Skip("cluster doesn't support reboots")
|
|
}
|
|
|
|
nodes := suite.DiscoverNodeInternalIPs(suite.ctx)
|
|
suite.Require().NotEmpty(nodes)
|
|
|
|
errCh := make(chan error, len(nodes))
|
|
|
|
var initialBootID sync.Map
|
|
|
|
for _, node := range nodes {
|
|
go func(node string) {
|
|
errCh <- func() error {
|
|
nodeCtx := client.WithNodes(suite.ctx, node)
|
|
|
|
// read boot_id before reboot
|
|
bootIDBefore, err := suite.ReadBootID(nodeCtx)
|
|
if err != nil {
|
|
return fmt.Errorf("error reading initial bootID (node %q): %w", node, err)
|
|
}
|
|
|
|
initialBootID.Store(node, bootIDBefore)
|
|
|
|
return nil
|
|
}()
|
|
}(node)
|
|
}
|
|
|
|
for range nodes {
|
|
suite.Require().NoError(<-errCh)
|
|
}
|
|
|
|
allNodesCtx := client.WithNodes(suite.ctx, nodes...)
|
|
|
|
err := base.IgnoreGRPCUnavailable(suite.Client.Reboot(allNodesCtx))
|
|
|
|
suite.Require().NoError(err)
|
|
|
|
for _, node := range nodes {
|
|
go func(node string) {
|
|
errCh <- func() error {
|
|
bootIDBeforeInterface, ok := initialBootID.Load(node)
|
|
if !ok {
|
|
return fmt.Errorf("bootID record not found for %q", node)
|
|
}
|
|
|
|
bootIDBefore := bootIDBeforeInterface.(string) //nolint:errcheck,forcetypeassert
|
|
|
|
nodeCtx := client.WithNodes(suite.ctx, node)
|
|
|
|
return retry.Constant(10 * time.Minute).Retry(
|
|
func() error {
|
|
requestCtx, requestCtxCancel := context.WithTimeout(nodeCtx, 5*time.Second)
|
|
defer requestCtxCancel()
|
|
|
|
bootIDAfter, err := suite.ReadBootID(requestCtx)
|
|
if err != nil {
|
|
// API might be unresponsive during reboot
|
|
return retry.ExpectedError(fmt.Errorf("error reading bootID for node %q: %w", node, err))
|
|
}
|
|
|
|
if bootIDAfter == bootIDBefore {
|
|
// bootID should be different after reboot
|
|
return retry.ExpectedError(
|
|
fmt.Errorf(
|
|
"bootID didn't change for node %q: before %s, after %s",
|
|
node,
|
|
bootIDBefore,
|
|
bootIDAfter,
|
|
),
|
|
)
|
|
}
|
|
|
|
return nil
|
|
},
|
|
)
|
|
}()
|
|
}(node)
|
|
}
|
|
|
|
for range nodes {
|
|
suite.Assert().NoError(<-errCh)
|
|
}
|
|
|
|
if suite.Cluster != nil {
|
|
// without cluster state we can't do deep checks, but basic reboot test still works
|
|
// NB: using `ctx` here to have client talking to init node by default
|
|
suite.AssertClusterHealthy(suite.ctx)
|
|
}
|
|
}
|
|
|
|
func init() {
|
|
allSuites = append(allSuites, new(RebootSuite))
|
|
}
|