mirror of
				https://github.com/prometheus/prometheus.git
				synced 2025-11-04 10:21:02 +01:00 
			
		
		
		
	
		
			
				
	
	
		
			254 lines
		
	
	
		
			6.4 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
			
		
		
	
	
			254 lines
		
	
	
		
			6.4 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
// Copyright 2013 The Prometheus Authors
 | 
						|
// Licensed under the Apache License, Version 2.0 (the "License");
 | 
						|
// you may not use this file except in compliance with the License.
 | 
						|
// You may obtain a copy of the License at
 | 
						|
//
 | 
						|
// http://www.apache.org/licenses/LICENSE-2.0
 | 
						|
//
 | 
						|
// Unless required by applicable law or agreed to in writing, software
 | 
						|
// distributed under the License is distributed on an "AS IS" BASIS,
 | 
						|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 | 
						|
// See the License for the specific language governing permissions and
 | 
						|
// limitations under the License.
 | 
						|
 | 
						|
package remote
 | 
						|
 | 
						|
import (
 | 
						|
	"fmt"
 | 
						|
	"sync"
 | 
						|
	"sync/atomic"
 | 
						|
	"testing"
 | 
						|
	"time"
 | 
						|
 | 
						|
	"github.com/prometheus/common/model"
 | 
						|
)
 | 
						|
 | 
						|
type TestStorageClient struct {
 | 
						|
	receivedSamples map[string]model.Samples
 | 
						|
	expectedSamples map[string]model.Samples
 | 
						|
	wg              sync.WaitGroup
 | 
						|
	mtx             sync.Mutex
 | 
						|
}
 | 
						|
 | 
						|
func NewTestStorageClient() *TestStorageClient {
 | 
						|
	return &TestStorageClient{
 | 
						|
		receivedSamples: map[string]model.Samples{},
 | 
						|
		expectedSamples: map[string]model.Samples{},
 | 
						|
	}
 | 
						|
}
 | 
						|
 | 
						|
func (c *TestStorageClient) expectSamples(ss model.Samples) {
 | 
						|
	c.mtx.Lock()
 | 
						|
	defer c.mtx.Unlock()
 | 
						|
 | 
						|
	for _, s := range ss {
 | 
						|
		ts := s.Metric.String()
 | 
						|
		c.expectedSamples[ts] = append(c.expectedSamples[ts], s)
 | 
						|
	}
 | 
						|
	c.wg.Add(len(ss))
 | 
						|
}
 | 
						|
 | 
						|
func (c *TestStorageClient) waitForExpectedSamples(t *testing.T) {
 | 
						|
	c.wg.Wait()
 | 
						|
 | 
						|
	c.mtx.Lock()
 | 
						|
	defer c.mtx.Unlock()
 | 
						|
	for ts, expectedSamples := range c.expectedSamples {
 | 
						|
		for i, expected := range expectedSamples {
 | 
						|
			if !expected.Equal(c.receivedSamples[ts][i]) {
 | 
						|
				t.Fatalf("%d. Expected %v, got %v", i, expected, c.receivedSamples[ts][i])
 | 
						|
			}
 | 
						|
		}
 | 
						|
	}
 | 
						|
}
 | 
						|
 | 
						|
func (c *TestStorageClient) Store(ss model.Samples) error {
 | 
						|
	c.mtx.Lock()
 | 
						|
	defer c.mtx.Unlock()
 | 
						|
 | 
						|
	for _, s := range ss {
 | 
						|
		ts := s.Metric.String()
 | 
						|
		c.receivedSamples[ts] = append(c.receivedSamples[ts], s)
 | 
						|
	}
 | 
						|
	c.wg.Add(-len(ss))
 | 
						|
	return nil
 | 
						|
}
 | 
						|
 | 
						|
func (c *TestStorageClient) Name() string {
 | 
						|
	return "teststorageclient"
 | 
						|
}
 | 
						|
 | 
						|
func TestSampleDelivery(t *testing.T) {
 | 
						|
	// Let's create an even number of send batches so we don't run into the
 | 
						|
	// batch timeout case.
 | 
						|
	n := defaultQueueManagerConfig.QueueCapacity * 2
 | 
						|
 | 
						|
	samples := make(model.Samples, 0, n)
 | 
						|
	for i := 0; i < n; i++ {
 | 
						|
		name := model.LabelValue(fmt.Sprintf("test_metric_%d", i))
 | 
						|
		samples = append(samples, &model.Sample{
 | 
						|
			Metric: model.Metric{
 | 
						|
				model.MetricNameLabel: name,
 | 
						|
			},
 | 
						|
			Value: model.SampleValue(i),
 | 
						|
		})
 | 
						|
	}
 | 
						|
 | 
						|
	c := NewTestStorageClient()
 | 
						|
	c.expectSamples(samples[:len(samples)/2])
 | 
						|
 | 
						|
	cfg := defaultQueueManagerConfig
 | 
						|
	cfg.MaxShards = 1
 | 
						|
	m := NewQueueManager(nil, cfg, nil, nil, c)
 | 
						|
 | 
						|
	// These should be received by the client.
 | 
						|
	for _, s := range samples[:len(samples)/2] {
 | 
						|
		m.Append(s)
 | 
						|
	}
 | 
						|
	// These will be dropped because the queue is full.
 | 
						|
	for _, s := range samples[len(samples)/2:] {
 | 
						|
		m.Append(s)
 | 
						|
	}
 | 
						|
	m.Start()
 | 
						|
	defer m.Stop()
 | 
						|
 | 
						|
	c.waitForExpectedSamples(t)
 | 
						|
}
 | 
						|
 | 
						|
func TestSampleDeliveryOrder(t *testing.T) {
 | 
						|
	ts := 10
 | 
						|
	n := defaultQueueManagerConfig.MaxSamplesPerSend * ts
 | 
						|
 | 
						|
	samples := make(model.Samples, 0, n)
 | 
						|
	for i := 0; i < n; i++ {
 | 
						|
		name := model.LabelValue(fmt.Sprintf("test_metric_%d", i%ts))
 | 
						|
		samples = append(samples, &model.Sample{
 | 
						|
			Metric: model.Metric{
 | 
						|
				model.MetricNameLabel: name,
 | 
						|
			},
 | 
						|
			Value:     model.SampleValue(i),
 | 
						|
			Timestamp: model.Time(i),
 | 
						|
		})
 | 
						|
	}
 | 
						|
 | 
						|
	c := NewTestStorageClient()
 | 
						|
	c.expectSamples(samples)
 | 
						|
	m := NewQueueManager(nil, defaultQueueManagerConfig, nil, nil, c)
 | 
						|
 | 
						|
	// These should be received by the client.
 | 
						|
	for _, s := range samples {
 | 
						|
		m.Append(s)
 | 
						|
	}
 | 
						|
	m.Start()
 | 
						|
	defer m.Stop()
 | 
						|
 | 
						|
	c.waitForExpectedSamples(t)
 | 
						|
}
 | 
						|
 | 
						|
// TestBlockingStorageClient is a queue_manager StorageClient which will block
 | 
						|
// on any calls to Store(), until the `block` channel is closed, at which point
 | 
						|
// the `numCalls` property will contain a count of how many times Store() was
 | 
						|
// called.
 | 
						|
type TestBlockingStorageClient struct {
 | 
						|
	numCalls uint64
 | 
						|
	block    chan bool
 | 
						|
}
 | 
						|
 | 
						|
func NewTestBlockedStorageClient() *TestBlockingStorageClient {
 | 
						|
	return &TestBlockingStorageClient{
 | 
						|
		block:    make(chan bool),
 | 
						|
		numCalls: 0,
 | 
						|
	}
 | 
						|
}
 | 
						|
 | 
						|
func (c *TestBlockingStorageClient) Store(s model.Samples) error {
 | 
						|
	atomic.AddUint64(&c.numCalls, 1)
 | 
						|
	<-c.block
 | 
						|
	return nil
 | 
						|
}
 | 
						|
 | 
						|
func (c *TestBlockingStorageClient) NumCalls() uint64 {
 | 
						|
	return atomic.LoadUint64(&c.numCalls)
 | 
						|
}
 | 
						|
 | 
						|
func (c *TestBlockingStorageClient) unlock() {
 | 
						|
	close(c.block)
 | 
						|
}
 | 
						|
 | 
						|
func (c *TestBlockingStorageClient) Name() string {
 | 
						|
	return "testblockingstorageclient"
 | 
						|
}
 | 
						|
 | 
						|
func (t *QueueManager) queueLen() int {
 | 
						|
	t.shardsMtx.Lock()
 | 
						|
	defer t.shardsMtx.Unlock()
 | 
						|
	queueLength := 0
 | 
						|
	for _, shard := range t.shards.queues {
 | 
						|
		queueLength += len(shard)
 | 
						|
	}
 | 
						|
	return queueLength
 | 
						|
}
 | 
						|
 | 
						|
func TestSpawnNotMoreThanMaxConcurrentSendsGoroutines(t *testing.T) {
 | 
						|
	// Our goal is to fully empty the queue:
 | 
						|
	// `MaxSamplesPerSend*Shards` samples should be consumed by the
 | 
						|
	// per-shard goroutines, and then another `MaxSamplesPerSend`
 | 
						|
	// should be left on the queue.
 | 
						|
	n := defaultQueueManagerConfig.MaxSamplesPerSend * 2
 | 
						|
 | 
						|
	samples := make(model.Samples, 0, n)
 | 
						|
	for i := 0; i < n; i++ {
 | 
						|
		name := model.LabelValue(fmt.Sprintf("test_metric_%d", i))
 | 
						|
		samples = append(samples, &model.Sample{
 | 
						|
			Metric: model.Metric{
 | 
						|
				model.MetricNameLabel: name,
 | 
						|
			},
 | 
						|
			Value: model.SampleValue(i),
 | 
						|
		})
 | 
						|
	}
 | 
						|
 | 
						|
	c := NewTestBlockedStorageClient()
 | 
						|
	cfg := defaultQueueManagerConfig
 | 
						|
	cfg.MaxShards = 1
 | 
						|
	cfg.QueueCapacity = n
 | 
						|
	m := NewQueueManager(nil, cfg, nil, nil, c)
 | 
						|
 | 
						|
	m.Start()
 | 
						|
 | 
						|
	defer func() {
 | 
						|
		c.unlock()
 | 
						|
		m.Stop()
 | 
						|
	}()
 | 
						|
 | 
						|
	for _, s := range samples {
 | 
						|
		m.Append(s)
 | 
						|
	}
 | 
						|
 | 
						|
	// Wait until the runShard() loops drain the queue.  If things went right, it
 | 
						|
	// should then immediately block in sendSamples(), but, in case of error,
 | 
						|
	// it would spawn too many goroutines, and thus we'd see more calls to
 | 
						|
	// client.Store()
 | 
						|
	//
 | 
						|
	// The timed wait is maybe non-ideal, but, in order to verify that we're
 | 
						|
	// not spawning too many concurrent goroutines, we have to wait on the
 | 
						|
	// Run() loop to consume a specific number of elements from the
 | 
						|
	// queue... and it doesn't signal that in any obvious way, except by
 | 
						|
	// draining the queue.  We cap the waiting at 1 second -- that should give
 | 
						|
	// plenty of time, and keeps the failure fairly quick if we're not draining
 | 
						|
	// the queue properly.
 | 
						|
	for i := 0; i < 100 && m.queueLen() > 0; i++ {
 | 
						|
		time.Sleep(10 * time.Millisecond)
 | 
						|
	}
 | 
						|
 | 
						|
	if m.queueLen() != defaultQueueManagerConfig.MaxSamplesPerSend {
 | 
						|
		t.Fatalf("Failed to drain QueueManager queue, %d elements left",
 | 
						|
			m.queueLen(),
 | 
						|
		)
 | 
						|
	}
 | 
						|
 | 
						|
	numCalls := c.NumCalls()
 | 
						|
	if numCalls != uint64(1) {
 | 
						|
		t.Errorf("Saw %d concurrent sends, expected 1", numCalls)
 | 
						|
	}
 | 
						|
}
 |