mirror of
https://github.com/hashicorp/vault.git
synced 2025-08-07 07:07:05 +02:00
When creating database connections, there is a race condition when multiple goroutines try to create the connection at the same time. This happens, for example, on leadership changes in a cluster. Normally, the extra database connections are cleaned up when this is detected. However, some database implementations, notably Postgres, do not seem to clean up in a timely manner, and can leak in these scenarios. To fix this, we create a global lock when creating database connections to prevent multiple connections from being created at the same time. We also clean up the logic at the end so that if (somehow) we ended up creating an additional connection, we use the existing one rather than the new one. This by itself would solve our problem long-term, however, would still involve many transient database connections being created and immediately killed on leadership changes. It's not ideal to have a single global lock for database connection creation. Some potential alternatives: * a map of locks from the connection name to the lock. The biggest downside is the we probably will want to garbage collect this map so that we don't have an unbounded number of locks. * a small pool of locks, where we hash the connection names to pick the lock. Using such a pool generally is a good way to introduce deadlock, but since we will only use it in a specific case, and the purpose is to improve performance for concurrent connection creation, this is probably acceptable. Co-authored-by: Jason O'Donnell <2160810+jasonodonnell@users.noreply.github.com>
110 lines
2.9 KiB
Go
110 lines
2.9 KiB
Go
// Copyright (c) HashiCorp, Inc.
|
|
// SPDX-License-Identifier: BUSL-1.1
|
|
|
|
package database
|
|
|
|
import (
|
|
"context"
|
|
"sync"
|
|
"testing"
|
|
|
|
"github.com/hashicorp/vault/sdk/helper/consts"
|
|
"github.com/hashicorp/vault/sdk/helper/pluginutil"
|
|
"github.com/hashicorp/vault/sdk/logical"
|
|
"github.com/hashicorp/vault/sdk/queue"
|
|
)
|
|
|
|
func newSystemViewWrapper(view logical.SystemView) logical.SystemView {
|
|
return &systemViewWrapper{
|
|
view,
|
|
}
|
|
}
|
|
|
|
type systemViewWrapper struct {
|
|
logical.SystemView
|
|
}
|
|
|
|
var _ logical.ExtendedSystemView = (*systemViewWrapper)(nil)
|
|
|
|
func (s *systemViewWrapper) RequestWellKnownRedirect(ctx context.Context, src, dest string) error {
|
|
panic("nope")
|
|
}
|
|
|
|
func (s *systemViewWrapper) DeregisterWellKnownRedirect(ctx context.Context, src string) bool {
|
|
panic("nope")
|
|
}
|
|
|
|
func (s *systemViewWrapper) Auditor() logical.Auditor {
|
|
panic("nope")
|
|
}
|
|
|
|
func (s *systemViewWrapper) ForwardGenericRequest(ctx context.Context, request *logical.Request) (*logical.Response, error) {
|
|
panic("nope")
|
|
}
|
|
|
|
func (s *systemViewWrapper) APILockShouldBlockRequest() (bool, error) {
|
|
panic("nope")
|
|
}
|
|
|
|
func (s *systemViewWrapper) GetPinnedPluginVersion(ctx context.Context, pluginType consts.PluginType, pluginName string) (*pluginutil.PinnedVersion, error) {
|
|
return nil, pluginutil.ErrPinnedVersionNotFound
|
|
}
|
|
|
|
func (s *systemViewWrapper) LookupPluginVersion(ctx context.Context, pluginName string, pluginType consts.PluginType, version string) (*pluginutil.PluginRunner, error) {
|
|
return &pluginutil.PluginRunner{
|
|
Name: mockv5,
|
|
Type: consts.PluginTypeDatabase,
|
|
Builtin: true,
|
|
BuiltinFactory: New,
|
|
}, nil
|
|
}
|
|
|
|
func getDbBackend(t *testing.T) (*databaseBackend, logical.Storage) {
|
|
t.Helper()
|
|
config := logical.TestBackendConfig()
|
|
config.System = newSystemViewWrapper(config.System)
|
|
config.StorageView = &logical.InmemStorage{}
|
|
// Create and init the backend ourselves instead of using a Factory because
|
|
// the factory function kicks off threads that cause racy tests.
|
|
b := Backend(config)
|
|
if err := b.Setup(context.Background(), config); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
b.schedule = &TestSchedule{}
|
|
b.credRotationQueue = queue.New()
|
|
b.populateQueue(context.Background(), config.StorageView)
|
|
|
|
return b, config.StorageView
|
|
}
|
|
|
|
// TestGetConnectionRaceCondition checks that GetConnection always returns the same instance, even when asked
|
|
// by multiple goroutines in parallel.
|
|
func TestGetConnectionRaceCondition(t *testing.T) {
|
|
ctx := context.Background()
|
|
b, s := getDbBackend(t)
|
|
defer b.Cleanup(ctx)
|
|
configureDBMount(t, s)
|
|
|
|
goroutines := 16
|
|
|
|
wg := sync.WaitGroup{}
|
|
wg.Add(goroutines)
|
|
dbis := make([]*dbPluginInstance, goroutines)
|
|
errs := make([]error, goroutines)
|
|
for i := 0; i < goroutines; i++ {
|
|
go func(i int) {
|
|
defer wg.Done()
|
|
dbis[i], errs[i] = b.GetConnection(ctx, s, mockv5)
|
|
}(i)
|
|
}
|
|
wg.Wait()
|
|
for i := 0; i < goroutines; i++ {
|
|
if errs[i] != nil {
|
|
t.Fatal(errs[i])
|
|
}
|
|
if dbis[0] != dbis[i] {
|
|
t.Fatal("Error: database instances did not match")
|
|
}
|
|
}
|
|
}
|