mirror of
				https://github.com/prometheus/prometheus.git
				synced 2025-11-04 02:11:01 +01:00 
			
		
		
		
	Patterned after `Chunk.Iterator()`: pass the old iterator in so it can be re-used to avoid allocating a new object. (This commit does not do any re-use; it is just changing all the method signatures so re-use is possible in later commits.) Signed-off-by: Bryan Boreham <bjboreham@gmail.com>
		
			
				
	
	
		
			711 lines
		
	
	
		
			24 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
			
		
		
	
	
			711 lines
		
	
	
		
			24 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
// Copyright 2017 The Prometheus Authors
 | 
						|
// Licensed under the Apache License, Version 2.0 (the "License");
 | 
						|
// you may not use this file except in compliance with the License.
 | 
						|
// You may obtain a copy of the License at
 | 
						|
//
 | 
						|
// http://www.apache.org/licenses/LICENSE-2.0
 | 
						|
//
 | 
						|
// Unless required by applicable law or agreed to in writing, software
 | 
						|
// distributed under the License is distributed on an "AS IS" BASIS,
 | 
						|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 | 
						|
// See the License for the specific language governing permissions and
 | 
						|
// limitations under the License.
 | 
						|
 | 
						|
package tsdb
 | 
						|
 | 
						|
import (
 | 
						|
	"context"
 | 
						|
	"encoding/binary"
 | 
						|
	"errors"
 | 
						|
	"fmt"
 | 
						|
	"hash/crc32"
 | 
						|
	"math/rand"
 | 
						|
	"os"
 | 
						|
	"path/filepath"
 | 
						|
	"sort"
 | 
						|
	"strconv"
 | 
						|
	"testing"
 | 
						|
 | 
						|
	"github.com/go-kit/log"
 | 
						|
	prom_testutil "github.com/prometheus/client_golang/prometheus/testutil"
 | 
						|
	"github.com/stretchr/testify/require"
 | 
						|
 | 
						|
	"github.com/prometheus/prometheus/model/histogram"
 | 
						|
	"github.com/prometheus/prometheus/model/labels"
 | 
						|
	"github.com/prometheus/prometheus/storage"
 | 
						|
	"github.com/prometheus/prometheus/tsdb/chunkenc"
 | 
						|
	"github.com/prometheus/prometheus/tsdb/chunks"
 | 
						|
	"github.com/prometheus/prometheus/tsdb/fileutil"
 | 
						|
	"github.com/prometheus/prometheus/tsdb/tsdbutil"
 | 
						|
	"github.com/prometheus/prometheus/tsdb/wlog"
 | 
						|
)
 | 
						|
 | 
						|
// In Prometheus 2.1.0 we had a bug where the meta.json version was falsely bumped
 | 
						|
// to 2. We had a migration in place resetting it to 1 but we should move immediately to
 | 
						|
// version 3 next time to avoid confusion and issues.
 | 
						|
func TestBlockMetaMustNeverBeVersion2(t *testing.T) {
 | 
						|
	dir := t.TempDir()
 | 
						|
 | 
						|
	_, err := writeMetaFile(log.NewNopLogger(), dir, &BlockMeta{})
 | 
						|
	require.NoError(t, err)
 | 
						|
 | 
						|
	meta, _, err := readMetaFile(dir)
 | 
						|
	require.NoError(t, err)
 | 
						|
	require.NotEqual(t, 2, meta.Version, "meta.json version must never be 2")
 | 
						|
}
 | 
						|
 | 
						|
func TestSetCompactionFailed(t *testing.T) {
 | 
						|
	tmpdir := t.TempDir()
 | 
						|
 | 
						|
	blockDir := createBlock(t, tmpdir, genSeries(1, 1, 0, 1))
 | 
						|
	b, err := OpenBlock(nil, blockDir, nil)
 | 
						|
	require.NoError(t, err)
 | 
						|
	require.Equal(t, false, b.meta.Compaction.Failed)
 | 
						|
	require.NoError(t, b.setCompactionFailed())
 | 
						|
	require.Equal(t, true, b.meta.Compaction.Failed)
 | 
						|
	require.NoError(t, b.Close())
 | 
						|
 | 
						|
	b, err = OpenBlock(nil, blockDir, nil)
 | 
						|
	require.NoError(t, err)
 | 
						|
	require.Equal(t, true, b.meta.Compaction.Failed)
 | 
						|
	require.NoError(t, b.Close())
 | 
						|
}
 | 
						|
 | 
						|
func TestCreateBlock(t *testing.T) {
 | 
						|
	tmpdir := t.TempDir()
 | 
						|
	b, err := OpenBlock(nil, createBlock(t, tmpdir, genSeries(1, 1, 0, 10)), nil)
 | 
						|
	require.NoError(t, err)
 | 
						|
	require.NoError(t, b.Close())
 | 
						|
}
 | 
						|
 | 
						|
func BenchmarkOpenBlock(b *testing.B) {
 | 
						|
	tmpdir := b.TempDir()
 | 
						|
	blockDir := createBlock(b, tmpdir, genSeries(1e6, 20, 0, 10))
 | 
						|
	b.Run("benchmark", func(b *testing.B) {
 | 
						|
		for i := 0; i < b.N; i++ {
 | 
						|
			block, err := OpenBlock(nil, blockDir, nil)
 | 
						|
			require.NoError(b, err)
 | 
						|
			require.NoError(b, block.Close())
 | 
						|
		}
 | 
						|
	})
 | 
						|
}
 | 
						|
 | 
						|
func TestCorruptedChunk(t *testing.T) {
 | 
						|
	for _, tc := range []struct {
 | 
						|
		name     string
 | 
						|
		corrFunc func(f *os.File) // Func that applies the corruption.
 | 
						|
		openErr  error
 | 
						|
		iterErr  error
 | 
						|
	}{
 | 
						|
		{
 | 
						|
			name: "invalid header size",
 | 
						|
			corrFunc: func(f *os.File) {
 | 
						|
				require.NoError(t, f.Truncate(1))
 | 
						|
			},
 | 
						|
			openErr: errors.New("invalid segment header in segment 0: invalid size"),
 | 
						|
		},
 | 
						|
		{
 | 
						|
			name: "invalid magic number",
 | 
						|
			corrFunc: func(f *os.File) {
 | 
						|
				magicChunksOffset := int64(0)
 | 
						|
				_, err := f.Seek(magicChunksOffset, 0)
 | 
						|
				require.NoError(t, err)
 | 
						|
 | 
						|
				// Set invalid magic number.
 | 
						|
				b := make([]byte, chunks.MagicChunksSize)
 | 
						|
				binary.BigEndian.PutUint32(b[:chunks.MagicChunksSize], 0x00000000)
 | 
						|
				n, err := f.Write(b)
 | 
						|
				require.NoError(t, err)
 | 
						|
				require.Equal(t, chunks.MagicChunksSize, n)
 | 
						|
			},
 | 
						|
			openErr: errors.New("invalid magic number 0"),
 | 
						|
		},
 | 
						|
		{
 | 
						|
			name: "invalid chunk format version",
 | 
						|
			corrFunc: func(f *os.File) {
 | 
						|
				chunksFormatVersionOffset := int64(4)
 | 
						|
				_, err := f.Seek(chunksFormatVersionOffset, 0)
 | 
						|
				require.NoError(t, err)
 | 
						|
 | 
						|
				// Set invalid chunk format version.
 | 
						|
				b := make([]byte, chunks.ChunksFormatVersionSize)
 | 
						|
				b[0] = 0
 | 
						|
				n, err := f.Write(b)
 | 
						|
				require.NoError(t, err)
 | 
						|
				require.Equal(t, chunks.ChunksFormatVersionSize, n)
 | 
						|
			},
 | 
						|
			openErr: errors.New("invalid chunk format version 0"),
 | 
						|
		},
 | 
						|
		{
 | 
						|
			name: "chunk not enough bytes to read the chunk length",
 | 
						|
			corrFunc: func(f *os.File) {
 | 
						|
				// Truncate one byte after the segment header.
 | 
						|
				require.NoError(t, f.Truncate(chunks.SegmentHeaderSize+1))
 | 
						|
			},
 | 
						|
			iterErr: errors.New("cannot populate chunk 8 from block 00000000000000000000000000: segment doesn't include enough bytes to read the chunk size data field - required:13, available:9"),
 | 
						|
		},
 | 
						|
		{
 | 
						|
			name: "chunk not enough bytes to read the data",
 | 
						|
			corrFunc: func(f *os.File) {
 | 
						|
				fi, err := f.Stat()
 | 
						|
				require.NoError(t, err)
 | 
						|
				require.NoError(t, f.Truncate(fi.Size()-1))
 | 
						|
			},
 | 
						|
			iterErr: errors.New("cannot populate chunk 8 from block 00000000000000000000000000: segment doesn't include enough bytes to read the chunk - required:26, available:25"),
 | 
						|
		},
 | 
						|
		{
 | 
						|
			name: "checksum mismatch",
 | 
						|
			corrFunc: func(f *os.File) {
 | 
						|
				fi, err := f.Stat()
 | 
						|
				require.NoError(t, err)
 | 
						|
 | 
						|
				// Get the chunk data end offset.
 | 
						|
				chkEndOffset := int(fi.Size()) - crc32.Size
 | 
						|
 | 
						|
				// Seek to the last byte of chunk data and modify it.
 | 
						|
				_, err = f.Seek(int64(chkEndOffset-1), 0)
 | 
						|
				require.NoError(t, err)
 | 
						|
				n, err := f.Write([]byte("x"))
 | 
						|
				require.NoError(t, err)
 | 
						|
				require.Equal(t, n, 1)
 | 
						|
			},
 | 
						|
			iterErr: errors.New("cannot populate chunk 8 from block 00000000000000000000000000: checksum mismatch expected:cfc0526c, actual:34815eae"),
 | 
						|
		},
 | 
						|
	} {
 | 
						|
		t.Run(tc.name, func(t *testing.T) {
 | 
						|
			tmpdir := t.TempDir()
 | 
						|
 | 
						|
			series := storage.NewListSeries(labels.FromStrings("a", "b"), []tsdbutil.Sample{sample{1, 1, nil, nil}})
 | 
						|
			blockDir := createBlock(t, tmpdir, []storage.Series{series})
 | 
						|
			files, err := sequenceFiles(chunkDir(blockDir))
 | 
						|
			require.NoError(t, err)
 | 
						|
			require.Greater(t, len(files), 0, "No chunk created.")
 | 
						|
 | 
						|
			f, err := os.OpenFile(files[0], os.O_RDWR, 0o666)
 | 
						|
			require.NoError(t, err)
 | 
						|
 | 
						|
			// Apply corruption function.
 | 
						|
			tc.corrFunc(f)
 | 
						|
			require.NoError(t, f.Close())
 | 
						|
 | 
						|
			// Check open err.
 | 
						|
			b, err := OpenBlock(nil, blockDir, nil)
 | 
						|
			if tc.openErr != nil {
 | 
						|
				require.Equal(t, tc.openErr.Error(), err.Error())
 | 
						|
				return
 | 
						|
			}
 | 
						|
			defer func() { require.NoError(t, b.Close()) }()
 | 
						|
 | 
						|
			querier, err := NewBlockQuerier(b, 0, 1)
 | 
						|
			require.NoError(t, err)
 | 
						|
			defer func() { require.NoError(t, querier.Close()) }()
 | 
						|
			set := querier.Select(false, nil, labels.MustNewMatcher(labels.MatchEqual, "a", "b"))
 | 
						|
 | 
						|
			// Check chunk errors during iter time.
 | 
						|
			require.True(t, set.Next())
 | 
						|
			it := set.At().Iterator(nil)
 | 
						|
			require.Equal(t, chunkenc.ValNone, it.Next())
 | 
						|
			require.Equal(t, tc.iterErr.Error(), it.Err().Error())
 | 
						|
		})
 | 
						|
	}
 | 
						|
}
 | 
						|
 | 
						|
func TestLabelValuesWithMatchers(t *testing.T) {
 | 
						|
	tmpdir := t.TempDir()
 | 
						|
 | 
						|
	var seriesEntries []storage.Series
 | 
						|
	for i := 0; i < 100; i++ {
 | 
						|
		seriesEntries = append(seriesEntries, storage.NewListSeries(labels.Labels{
 | 
						|
			{Name: "tens", Value: fmt.Sprintf("value%d", i/10)},
 | 
						|
			{Name: "unique", Value: fmt.Sprintf("value%d", i)},
 | 
						|
		}, []tsdbutil.Sample{sample{100, 0, nil, nil}}))
 | 
						|
	}
 | 
						|
 | 
						|
	blockDir := createBlock(t, tmpdir, seriesEntries)
 | 
						|
	files, err := sequenceFiles(chunkDir(blockDir))
 | 
						|
	require.NoError(t, err)
 | 
						|
	require.Greater(t, len(files), 0, "No chunk created.")
 | 
						|
 | 
						|
	// Check open err.
 | 
						|
	block, err := OpenBlock(nil, blockDir, nil)
 | 
						|
	require.NoError(t, err)
 | 
						|
	defer func() { require.NoError(t, block.Close()) }()
 | 
						|
 | 
						|
	indexReader, err := block.Index()
 | 
						|
	require.NoError(t, err)
 | 
						|
	defer func() { require.NoError(t, indexReader.Close()) }()
 | 
						|
 | 
						|
	testCases := []struct {
 | 
						|
		name           string
 | 
						|
		labelName      string
 | 
						|
		matchers       []*labels.Matcher
 | 
						|
		expectedValues []string
 | 
						|
	}{
 | 
						|
		{
 | 
						|
			name:           "get tens based on unique id",
 | 
						|
			labelName:      "tens",
 | 
						|
			matchers:       []*labels.Matcher{labels.MustNewMatcher(labels.MatchEqual, "unique", "value35")},
 | 
						|
			expectedValues: []string{"value3"},
 | 
						|
		}, {
 | 
						|
			name:           "get unique ids based on a ten",
 | 
						|
			labelName:      "unique",
 | 
						|
			matchers:       []*labels.Matcher{labels.MustNewMatcher(labels.MatchEqual, "tens", "value1")},
 | 
						|
			expectedValues: []string{"value10", "value11", "value12", "value13", "value14", "value15", "value16", "value17", "value18", "value19"},
 | 
						|
		}, {
 | 
						|
			name:           "get tens by pattern matching on unique id",
 | 
						|
			labelName:      "tens",
 | 
						|
			matchers:       []*labels.Matcher{labels.MustNewMatcher(labels.MatchRegexp, "unique", "value[5-7]5")},
 | 
						|
			expectedValues: []string{"value5", "value6", "value7"},
 | 
						|
		}, {
 | 
						|
			name:           "get tens by matching for absence of unique label",
 | 
						|
			labelName:      "tens",
 | 
						|
			matchers:       []*labels.Matcher{labels.MustNewMatcher(labels.MatchNotEqual, "unique", "")},
 | 
						|
			expectedValues: []string{"value0", "value1", "value2", "value3", "value4", "value5", "value6", "value7", "value8", "value9"},
 | 
						|
		},
 | 
						|
	}
 | 
						|
 | 
						|
	for _, tt := range testCases {
 | 
						|
		t.Run(tt.name, func(t *testing.T) {
 | 
						|
			actualValues, err := indexReader.SortedLabelValues(tt.labelName, tt.matchers...)
 | 
						|
			require.NoError(t, err)
 | 
						|
			require.Equal(t, tt.expectedValues, actualValues)
 | 
						|
 | 
						|
			actualValues, err = indexReader.LabelValues(tt.labelName, tt.matchers...)
 | 
						|
			sort.Strings(actualValues)
 | 
						|
			require.NoError(t, err)
 | 
						|
			require.Equal(t, tt.expectedValues, actualValues)
 | 
						|
		})
 | 
						|
	}
 | 
						|
}
 | 
						|
 | 
						|
// TestBlockSize ensures that the block size is calculated correctly.
 | 
						|
func TestBlockSize(t *testing.T) {
 | 
						|
	tmpdir := t.TempDir()
 | 
						|
 | 
						|
	var (
 | 
						|
		blockInit    *Block
 | 
						|
		expSizeInit  int64
 | 
						|
		blockDirInit string
 | 
						|
		err          error
 | 
						|
	)
 | 
						|
 | 
						|
	// Create a block and compare the reported size vs actual disk size.
 | 
						|
	{
 | 
						|
		blockDirInit = createBlock(t, tmpdir, genSeries(10, 1, 1, 100))
 | 
						|
		blockInit, err = OpenBlock(nil, blockDirInit, nil)
 | 
						|
		require.NoError(t, err)
 | 
						|
		defer func() {
 | 
						|
			require.NoError(t, blockInit.Close())
 | 
						|
		}()
 | 
						|
		expSizeInit = blockInit.Size()
 | 
						|
		actSizeInit, err := fileutil.DirSize(blockInit.Dir())
 | 
						|
		require.NoError(t, err)
 | 
						|
		require.Equal(t, expSizeInit, actSizeInit)
 | 
						|
	}
 | 
						|
 | 
						|
	// Delete some series and check the sizes again.
 | 
						|
	{
 | 
						|
		require.NoError(t, blockInit.Delete(1, 10, labels.MustNewMatcher(labels.MatchRegexp, "", ".*")))
 | 
						|
		expAfterDelete := blockInit.Size()
 | 
						|
		require.Greater(t, expAfterDelete, expSizeInit, "after a delete the block size should be bigger as the tombstone file should grow %v > %v", expAfterDelete, expSizeInit)
 | 
						|
		actAfterDelete, err := fileutil.DirSize(blockDirInit)
 | 
						|
		require.NoError(t, err)
 | 
						|
		require.Equal(t, expAfterDelete, actAfterDelete, "after a delete reported block size doesn't match actual disk size")
 | 
						|
 | 
						|
		c, err := NewLeveledCompactor(context.Background(), nil, log.NewNopLogger(), []int64{0}, nil, nil)
 | 
						|
		require.NoError(t, err)
 | 
						|
		blockDirAfterCompact, err := c.Compact(tmpdir, []string{blockInit.Dir()}, nil)
 | 
						|
		require.NoError(t, err)
 | 
						|
		blockAfterCompact, err := OpenBlock(nil, filepath.Join(tmpdir, blockDirAfterCompact.String()), nil)
 | 
						|
		require.NoError(t, err)
 | 
						|
		defer func() {
 | 
						|
			require.NoError(t, blockAfterCompact.Close())
 | 
						|
		}()
 | 
						|
		expAfterCompact := blockAfterCompact.Size()
 | 
						|
		actAfterCompact, err := fileutil.DirSize(blockAfterCompact.Dir())
 | 
						|
		require.NoError(t, err)
 | 
						|
		require.Greater(t, actAfterDelete, actAfterCompact, "after a delete and compaction the block size should be smaller %v,%v", actAfterDelete, actAfterCompact)
 | 
						|
		require.Equal(t, expAfterCompact, actAfterCompact, "after a delete and compaction reported block size doesn't match actual disk size")
 | 
						|
	}
 | 
						|
}
 | 
						|
 | 
						|
func TestReadIndexFormatV1(t *testing.T) {
 | 
						|
	/* The block here was produced at the commit
 | 
						|
	    706602daed1487f7849990678b4ece4599745905 used in 2.0.0 with:
 | 
						|
	   db, _ := Open("v1db", nil, nil, nil)
 | 
						|
	   app := db.Appender()
 | 
						|
	   app.Add(labels.FromStrings("foo", "bar"), 1, 2)
 | 
						|
	   app.Add(labels.FromStrings("foo", "baz"), 3, 4)
 | 
						|
	   app.Add(labels.FromStrings("foo", "meh"), 1000*3600*4, 4) // Not in the block.
 | 
						|
	   // Make sure we've enough values for the lack of sorting of postings offsets to show up.
 | 
						|
	   for i := 0; i < 100; i++ {
 | 
						|
	     app.Add(labels.FromStrings("bar", strconv.FormatInt(int64(i), 10)), 0, 0)
 | 
						|
	   }
 | 
						|
	   app.Commit()
 | 
						|
	   db.compact()
 | 
						|
	   db.Close()
 | 
						|
	*/
 | 
						|
 | 
						|
	blockDir := filepath.Join("testdata", "index_format_v1")
 | 
						|
	block, err := OpenBlock(nil, blockDir, nil)
 | 
						|
	require.NoError(t, err)
 | 
						|
 | 
						|
	q, err := NewBlockQuerier(block, 0, 1000)
 | 
						|
	require.NoError(t, err)
 | 
						|
	require.Equal(t, query(t, q, labels.MustNewMatcher(labels.MatchEqual, "foo", "bar")),
 | 
						|
		map[string][]tsdbutil.Sample{`{foo="bar"}`: {sample{t: 1, v: 2}}})
 | 
						|
 | 
						|
	q, err = NewBlockQuerier(block, 0, 1000)
 | 
						|
	require.NoError(t, err)
 | 
						|
	require.Equal(t, query(t, q, labels.MustNewMatcher(labels.MatchNotRegexp, "foo", "^.?$")),
 | 
						|
		map[string][]tsdbutil.Sample{
 | 
						|
			`{foo="bar"}`: {sample{t: 1, v: 2}},
 | 
						|
			`{foo="baz"}`: {sample{t: 3, v: 4}},
 | 
						|
		})
 | 
						|
}
 | 
						|
 | 
						|
func BenchmarkLabelValuesWithMatchers(b *testing.B) {
 | 
						|
	tmpdir := b.TempDir()
 | 
						|
 | 
						|
	var seriesEntries []storage.Series
 | 
						|
	metricCount := 1000000
 | 
						|
	for i := 0; i < metricCount; i++ {
 | 
						|
		// Note these series are not created in sort order: 'value2' sorts after 'value10'.
 | 
						|
		// This makes a big difference to the benchmark timing.
 | 
						|
		seriesEntries = append(seriesEntries, storage.NewListSeries(labels.Labels{
 | 
						|
			{Name: "a_unique", Value: fmt.Sprintf("value%d", i)},
 | 
						|
			{Name: "b_tens", Value: fmt.Sprintf("value%d", i/(metricCount/10))},
 | 
						|
			{Name: "c_ninety", Value: fmt.Sprintf("value%d", i/(metricCount/10)/9)}, // "0" for the first 90%, then "1"
 | 
						|
		}, []tsdbutil.Sample{sample{100, 0, nil, nil}}))
 | 
						|
	}
 | 
						|
 | 
						|
	blockDir := createBlock(b, tmpdir, seriesEntries)
 | 
						|
	files, err := sequenceFiles(chunkDir(blockDir))
 | 
						|
	require.NoError(b, err)
 | 
						|
	require.Greater(b, len(files), 0, "No chunk created.")
 | 
						|
 | 
						|
	// Check open err.
 | 
						|
	block, err := OpenBlock(nil, blockDir, nil)
 | 
						|
	require.NoError(b, err)
 | 
						|
	defer func() { require.NoError(b, block.Close()) }()
 | 
						|
 | 
						|
	indexReader, err := block.Index()
 | 
						|
	require.NoError(b, err)
 | 
						|
	defer func() { require.NoError(b, indexReader.Close()) }()
 | 
						|
 | 
						|
	matchers := []*labels.Matcher{labels.MustNewMatcher(labels.MatchEqual, "c_ninety", "value0")}
 | 
						|
 | 
						|
	b.ResetTimer()
 | 
						|
	b.ReportAllocs()
 | 
						|
 | 
						|
	for benchIdx := 0; benchIdx < b.N; benchIdx++ {
 | 
						|
		actualValues, err := indexReader.LabelValues("b_tens", matchers...)
 | 
						|
		require.NoError(b, err)
 | 
						|
		require.Equal(b, 9, len(actualValues))
 | 
						|
	}
 | 
						|
}
 | 
						|
 | 
						|
func TestLabelNamesWithMatchers(t *testing.T) {
 | 
						|
	tmpdir := t.TempDir()
 | 
						|
 | 
						|
	var seriesEntries []storage.Series
 | 
						|
	for i := 0; i < 100; i++ {
 | 
						|
		seriesEntries = append(seriesEntries, storage.NewListSeries(labels.Labels{
 | 
						|
			{Name: "unique", Value: fmt.Sprintf("value%d", i)},
 | 
						|
		}, []tsdbutil.Sample{sample{100, 0, nil, nil}}))
 | 
						|
 | 
						|
		if i%10 == 0 {
 | 
						|
			seriesEntries = append(seriesEntries, storage.NewListSeries(labels.Labels{
 | 
						|
				{Name: "tens", Value: fmt.Sprintf("value%d", i/10)},
 | 
						|
				{Name: "unique", Value: fmt.Sprintf("value%d", i)},
 | 
						|
			}, []tsdbutil.Sample{sample{100, 0, nil, nil}}))
 | 
						|
		}
 | 
						|
 | 
						|
		if i%20 == 0 {
 | 
						|
			seriesEntries = append(seriesEntries, storage.NewListSeries(labels.Labels{
 | 
						|
				{Name: "tens", Value: fmt.Sprintf("value%d", i/10)},
 | 
						|
				{Name: "twenties", Value: fmt.Sprintf("value%d", i/20)},
 | 
						|
				{Name: "unique", Value: fmt.Sprintf("value%d", i)},
 | 
						|
			}, []tsdbutil.Sample{sample{100, 0, nil, nil}}))
 | 
						|
		}
 | 
						|
 | 
						|
	}
 | 
						|
 | 
						|
	blockDir := createBlock(t, tmpdir, seriesEntries)
 | 
						|
	files, err := sequenceFiles(chunkDir(blockDir))
 | 
						|
	require.NoError(t, err)
 | 
						|
	require.Greater(t, len(files), 0, "No chunk created.")
 | 
						|
 | 
						|
	// Check open err.
 | 
						|
	block, err := OpenBlock(nil, blockDir, nil)
 | 
						|
	require.NoError(t, err)
 | 
						|
	t.Cleanup(func() { require.NoError(t, block.Close()) })
 | 
						|
 | 
						|
	indexReader, err := block.Index()
 | 
						|
	require.NoError(t, err)
 | 
						|
	t.Cleanup(func() { require.NoError(t, indexReader.Close()) })
 | 
						|
 | 
						|
	testCases := []struct {
 | 
						|
		name          string
 | 
						|
		labelName     string
 | 
						|
		matchers      []*labels.Matcher
 | 
						|
		expectedNames []string
 | 
						|
	}{
 | 
						|
		{
 | 
						|
			name:          "get with non-empty unique: all",
 | 
						|
			matchers:      []*labels.Matcher{labels.MustNewMatcher(labels.MatchNotEqual, "unique", "")},
 | 
						|
			expectedNames: []string{"tens", "twenties", "unique"},
 | 
						|
		}, {
 | 
						|
			name:          "get with unique ending in 1: only unique",
 | 
						|
			matchers:      []*labels.Matcher{labels.MustNewMatcher(labels.MatchRegexp, "unique", "value.*1")},
 | 
						|
			expectedNames: []string{"unique"},
 | 
						|
		}, {
 | 
						|
			name:          "get with unique = value20: all",
 | 
						|
			matchers:      []*labels.Matcher{labels.MustNewMatcher(labels.MatchEqual, "unique", "value20")},
 | 
						|
			expectedNames: []string{"tens", "twenties", "unique"},
 | 
						|
		}, {
 | 
						|
			name:          "get tens = 1: unique & tens",
 | 
						|
			matchers:      []*labels.Matcher{labels.MustNewMatcher(labels.MatchEqual, "tens", "value1")},
 | 
						|
			expectedNames: []string{"tens", "unique"},
 | 
						|
		},
 | 
						|
	}
 | 
						|
 | 
						|
	for _, tt := range testCases {
 | 
						|
		t.Run(tt.name, func(t *testing.T) {
 | 
						|
			actualNames, err := indexReader.LabelNames(tt.matchers...)
 | 
						|
			require.NoError(t, err)
 | 
						|
			require.Equal(t, tt.expectedNames, actualNames)
 | 
						|
		})
 | 
						|
	}
 | 
						|
}
 | 
						|
 | 
						|
// createBlock creates a block with given set of series and returns its dir.
 | 
						|
func createBlock(tb testing.TB, dir string, series []storage.Series) string {
 | 
						|
	blockDir, err := CreateBlock(series, dir, 0, log.NewNopLogger())
 | 
						|
	require.NoError(tb, err)
 | 
						|
	return blockDir
 | 
						|
}
 | 
						|
 | 
						|
func createBlockFromHead(tb testing.TB, dir string, head *Head) string {
 | 
						|
	compactor, err := NewLeveledCompactor(context.Background(), nil, log.NewNopLogger(), []int64{1000000}, nil, nil)
 | 
						|
	require.NoError(tb, err)
 | 
						|
 | 
						|
	require.NoError(tb, os.MkdirAll(dir, 0o777))
 | 
						|
 | 
						|
	// Add +1 millisecond to block maxt because block intervals are half-open: [b.MinTime, b.MaxTime).
 | 
						|
	// Because of this block intervals are always +1 than the total samples it includes.
 | 
						|
	ulid, err := compactor.Write(dir, head, head.MinTime(), head.MaxTime()+1, nil)
 | 
						|
	require.NoError(tb, err)
 | 
						|
	return filepath.Join(dir, ulid.String())
 | 
						|
}
 | 
						|
 | 
						|
func createHead(tb testing.TB, w *wlog.WL, series []storage.Series, chunkDir string) *Head {
 | 
						|
	opts := DefaultHeadOptions()
 | 
						|
	opts.ChunkDirRoot = chunkDir
 | 
						|
	head, err := NewHead(nil, nil, w, nil, opts, nil)
 | 
						|
	require.NoError(tb, err)
 | 
						|
 | 
						|
	var it chunkenc.Iterator
 | 
						|
	ctx := context.Background()
 | 
						|
	app := head.Appender(ctx)
 | 
						|
	for _, s := range series {
 | 
						|
		ref := storage.SeriesRef(0)
 | 
						|
		it = s.Iterator(it)
 | 
						|
		lset := s.Labels()
 | 
						|
		typ := it.Next()
 | 
						|
		lastTyp := typ
 | 
						|
		for ; typ != chunkenc.ValNone; typ = it.Next() {
 | 
						|
			if lastTyp != typ {
 | 
						|
				// The behaviour of appender is undefined if samples of different types
 | 
						|
				// are appended to the same series in a single Commit().
 | 
						|
				require.NoError(tb, app.Commit())
 | 
						|
				app = head.Appender(ctx)
 | 
						|
			}
 | 
						|
 | 
						|
			switch typ {
 | 
						|
			case chunkenc.ValFloat:
 | 
						|
				t, v := it.At()
 | 
						|
				ref, err = app.Append(ref, lset, t, v)
 | 
						|
			case chunkenc.ValHistogram:
 | 
						|
				t, h := it.AtHistogram()
 | 
						|
				ref, err = app.AppendHistogram(ref, lset, t, h)
 | 
						|
			default:
 | 
						|
				err = fmt.Errorf("unknown sample type %s", typ.String())
 | 
						|
			}
 | 
						|
			require.NoError(tb, err)
 | 
						|
			lastTyp = typ
 | 
						|
		}
 | 
						|
		require.NoError(tb, it.Err())
 | 
						|
	}
 | 
						|
	require.NoError(tb, app.Commit())
 | 
						|
	return head
 | 
						|
}
 | 
						|
 | 
						|
func createHeadWithOOOSamples(tb testing.TB, w *wlog.WL, series []storage.Series, chunkDir string, oooSampleFrequency int) *Head {
 | 
						|
	opts := DefaultHeadOptions()
 | 
						|
	opts.ChunkDirRoot = chunkDir
 | 
						|
	opts.OutOfOrderTimeWindow.Store(10000000000)
 | 
						|
	head, err := NewHead(nil, nil, w, nil, opts, nil)
 | 
						|
	require.NoError(tb, err)
 | 
						|
 | 
						|
	oooSampleLabels := make([]labels.Labels, 0, len(series))
 | 
						|
	oooSamples := make([]tsdbutil.SampleSlice, 0, len(series))
 | 
						|
 | 
						|
	var it chunkenc.Iterator
 | 
						|
	totalSamples := 0
 | 
						|
	app := head.Appender(context.Background())
 | 
						|
	for _, s := range series {
 | 
						|
		ref := storage.SeriesRef(0)
 | 
						|
		it = s.Iterator(it)
 | 
						|
		lset := s.Labels()
 | 
						|
		os := tsdbutil.SampleSlice{}
 | 
						|
		count := 0
 | 
						|
		for it.Next() == chunkenc.ValFloat {
 | 
						|
			totalSamples++
 | 
						|
			count++
 | 
						|
			t, v := it.At()
 | 
						|
			if count%oooSampleFrequency == 0 {
 | 
						|
				os = append(os, sample{t: t, v: v})
 | 
						|
				continue
 | 
						|
			}
 | 
						|
			ref, err = app.Append(ref, lset, t, v)
 | 
						|
			require.NoError(tb, err)
 | 
						|
		}
 | 
						|
		require.NoError(tb, it.Err())
 | 
						|
		if len(os) > 0 {
 | 
						|
			oooSampleLabels = append(oooSampleLabels, lset)
 | 
						|
			oooSamples = append(oooSamples, os)
 | 
						|
		}
 | 
						|
	}
 | 
						|
	require.NoError(tb, app.Commit())
 | 
						|
 | 
						|
	oooSamplesAppended := 0
 | 
						|
	require.Equal(tb, float64(0), prom_testutil.ToFloat64(head.metrics.outOfOrderSamplesAppended))
 | 
						|
 | 
						|
	app = head.Appender(context.Background())
 | 
						|
	for i, lset := range oooSampleLabels {
 | 
						|
		ref := storage.SeriesRef(0)
 | 
						|
		for _, sample := range oooSamples[i] {
 | 
						|
			ref, err = app.Append(ref, lset, sample.T(), sample.V())
 | 
						|
			require.NoError(tb, err)
 | 
						|
			oooSamplesAppended++
 | 
						|
		}
 | 
						|
	}
 | 
						|
	require.NoError(tb, app.Commit())
 | 
						|
 | 
						|
	actOOOAppended := prom_testutil.ToFloat64(head.metrics.outOfOrderSamplesAppended)
 | 
						|
	require.GreaterOrEqual(tb, actOOOAppended, float64(oooSamplesAppended-len(series)))
 | 
						|
	require.LessOrEqual(tb, actOOOAppended, float64(oooSamplesAppended))
 | 
						|
 | 
						|
	require.Equal(tb, float64(totalSamples), prom_testutil.ToFloat64(head.metrics.samplesAppended))
 | 
						|
 | 
						|
	return head
 | 
						|
}
 | 
						|
 | 
						|
const (
 | 
						|
	defaultLabelName  = "labelName"
 | 
						|
	defaultLabelValue = "labelValue"
 | 
						|
)
 | 
						|
 | 
						|
// genSeries generates series of float64 samples with a given number of labels and values.
 | 
						|
func genSeries(totalSeries, labelCount int, mint, maxt int64) []storage.Series {
 | 
						|
	return genSeriesFromSampleGenerator(totalSeries, labelCount, mint, maxt, 1, func(ts int64) tsdbutil.Sample {
 | 
						|
		return sample{t: ts, v: rand.Float64()}
 | 
						|
	})
 | 
						|
}
 | 
						|
 | 
						|
// genHistogramSeries generates series of histogram samples with a given number of labels and values.
 | 
						|
func genHistogramSeries(totalSeries, labelCount int, mint, maxt, step int64) []storage.Series {
 | 
						|
	return genSeriesFromSampleGenerator(totalSeries, labelCount, mint, maxt, step, func(ts int64) tsdbutil.Sample {
 | 
						|
		h := &histogram.Histogram{
 | 
						|
			Count:         5 + uint64(ts*4),
 | 
						|
			ZeroCount:     2 + uint64(ts),
 | 
						|
			ZeroThreshold: 0.001,
 | 
						|
			Sum:           18.4 * rand.Float64(),
 | 
						|
			Schema:        1,
 | 
						|
			PositiveSpans: []histogram.Span{
 | 
						|
				{Offset: 0, Length: 2},
 | 
						|
				{Offset: 1, Length: 2},
 | 
						|
			},
 | 
						|
			PositiveBuckets: []int64{int64(ts + 1), 1, -1, 0},
 | 
						|
		}
 | 
						|
		return sample{t: ts, h: h}
 | 
						|
	})
 | 
						|
}
 | 
						|
 | 
						|
// genHistogramAndFloatSeries generates series of mixed histogram and float64 samples with a given number of labels and values.
 | 
						|
func genHistogramAndFloatSeries(totalSeries, labelCount int, mint, maxt, step int64) []storage.Series {
 | 
						|
	floatSample := false
 | 
						|
	count := 0
 | 
						|
	return genSeriesFromSampleGenerator(totalSeries, labelCount, mint, maxt, step, func(ts int64) tsdbutil.Sample {
 | 
						|
		count++
 | 
						|
		var s sample
 | 
						|
		if floatSample {
 | 
						|
			s = sample{t: ts, v: rand.Float64()}
 | 
						|
		} else {
 | 
						|
			h := &histogram.Histogram{
 | 
						|
				Count:         5 + uint64(ts*4),
 | 
						|
				ZeroCount:     2 + uint64(ts),
 | 
						|
				ZeroThreshold: 0.001,
 | 
						|
				Sum:           18.4 * rand.Float64(),
 | 
						|
				Schema:        1,
 | 
						|
				PositiveSpans: []histogram.Span{
 | 
						|
					{Offset: 0, Length: 2},
 | 
						|
					{Offset: 1, Length: 2},
 | 
						|
				},
 | 
						|
				PositiveBuckets: []int64{int64(ts + 1), 1, -1, 0},
 | 
						|
			}
 | 
						|
			s = sample{t: ts, h: h}
 | 
						|
		}
 | 
						|
 | 
						|
		if count%5 == 0 {
 | 
						|
			// Flip the sample type for every 5 samples.
 | 
						|
			floatSample = !floatSample
 | 
						|
		}
 | 
						|
 | 
						|
		return s
 | 
						|
	})
 | 
						|
}
 | 
						|
 | 
						|
func genSeriesFromSampleGenerator(totalSeries, labelCount int, mint, maxt, step int64, generator func(ts int64) tsdbutil.Sample) []storage.Series {
 | 
						|
	if totalSeries == 0 || labelCount == 0 {
 | 
						|
		return nil
 | 
						|
	}
 | 
						|
 | 
						|
	series := make([]storage.Series, totalSeries)
 | 
						|
 | 
						|
	for i := 0; i < totalSeries; i++ {
 | 
						|
		lbls := make(map[string]string, labelCount)
 | 
						|
		lbls[defaultLabelName] = strconv.Itoa(i)
 | 
						|
		for j := 1; len(lbls) < labelCount; j++ {
 | 
						|
			lbls[defaultLabelName+strconv.Itoa(j)] = defaultLabelValue + strconv.Itoa(j)
 | 
						|
		}
 | 
						|
		samples := make([]tsdbutil.Sample, 0, (maxt-mint)/step+1)
 | 
						|
		for t := mint; t < maxt; t += step {
 | 
						|
			samples = append(samples, generator(t))
 | 
						|
		}
 | 
						|
		series[i] = storage.NewListSeries(labels.FromMap(lbls), samples)
 | 
						|
	}
 | 
						|
	return series
 | 
						|
}
 | 
						|
 | 
						|
// populateSeries generates series from given labels, mint and maxt.
 | 
						|
func populateSeries(lbls []map[string]string, mint, maxt int64) []storage.Series {
 | 
						|
	if len(lbls) == 0 {
 | 
						|
		return nil
 | 
						|
	}
 | 
						|
 | 
						|
	series := make([]storage.Series, 0, len(lbls))
 | 
						|
	for _, lbl := range lbls {
 | 
						|
		if len(lbl) == 0 {
 | 
						|
			continue
 | 
						|
		}
 | 
						|
		samples := make([]tsdbutil.Sample, 0, maxt-mint+1)
 | 
						|
		for t := mint; t <= maxt; t++ {
 | 
						|
			samples = append(samples, sample{t: t, v: rand.Float64()})
 | 
						|
		}
 | 
						|
		series = append(series, storage.NewListSeries(labels.FromMap(lbl), samples))
 | 
						|
	}
 | 
						|
	return series
 | 
						|
}
 |