mirror of
https://github.com/prometheus/prometheus.git
synced 2025-11-05 10:51:03 +01:00
scrape: "Modernize" test files
Applied the analyzer "modernize" to the test files. $ go run golang.org/x/tools/gopls/internal/analysis/modernize/cmd/modernize@latest -fix -test ./... Signed-off-by: beorn7 <beorn@grafana.com>
This commit is contained in:
parent
9a5bccbd4b
commit
72960c076d
@ -1277,8 +1277,7 @@ func requireTargets(
|
|||||||
// TestTargetDisappearsAfterProviderRemoved makes sure that when a provider is dropped, (only) its targets are dropped.
|
// TestTargetDisappearsAfterProviderRemoved makes sure that when a provider is dropped, (only) its targets are dropped.
|
||||||
func TestTargetDisappearsAfterProviderRemoved(t *testing.T) {
|
func TestTargetDisappearsAfterProviderRemoved(t *testing.T) {
|
||||||
t.Parallel()
|
t.Parallel()
|
||||||
ctx, cancel := context.WithCancel(context.Background())
|
ctx := t.Context()
|
||||||
defer cancel()
|
|
||||||
|
|
||||||
myJob := "my-job"
|
myJob := "my-job"
|
||||||
myJobSDTargetURL := "my:9876"
|
myJobSDTargetURL := "my:9876"
|
||||||
@ -1378,8 +1377,7 @@ scrape_configs:
|
|||||||
// and when the provider can no longer discover some of those targets, only those stale targets are dropped.
|
// and when the provider can no longer discover some of those targets, only those stale targets are dropped.
|
||||||
func TestOnlyProviderStaleTargetsAreDropped(t *testing.T) {
|
func TestOnlyProviderStaleTargetsAreDropped(t *testing.T) {
|
||||||
t.Parallel()
|
t.Parallel()
|
||||||
ctx, cancel := context.WithCancel(context.Background())
|
ctx := t.Context()
|
||||||
defer cancel()
|
|
||||||
|
|
||||||
jobName := "my-job"
|
jobName := "my-job"
|
||||||
jobTarget1URL := "foo:9876"
|
jobTarget1URL := "foo:9876"
|
||||||
@ -1441,8 +1439,7 @@ scrape_configs:
|
|||||||
// should no longer discover targets, the targets of that provider are dropped.
|
// should no longer discover targets, the targets of that provider are dropped.
|
||||||
// See: https://github.com/prometheus/prometheus/issues/12858
|
// See: https://github.com/prometheus/prometheus/issues/12858
|
||||||
func TestProviderStaleTargetsAreDropped(t *testing.T) {
|
func TestProviderStaleTargetsAreDropped(t *testing.T) {
|
||||||
ctx, cancel := context.WithCancel(context.Background())
|
ctx := t.Context()
|
||||||
defer cancel()
|
|
||||||
|
|
||||||
jobName := "my-job"
|
jobName := "my-job"
|
||||||
jobTargetURL := "foo:9876"
|
jobTargetURL := "foo:9876"
|
||||||
@ -1499,8 +1496,7 @@ scrape_configs:
|
|||||||
// TestOnlyStaleTargetsAreDropped makes sure that when a job has multiple providers, when one of them should no
|
// TestOnlyStaleTargetsAreDropped makes sure that when a job has multiple providers, when one of them should no
|
||||||
// longer discover targets, only the stale targets of that provider are dropped.
|
// longer discover targets, only the stale targets of that provider are dropped.
|
||||||
func TestOnlyStaleTargetsAreDropped(t *testing.T) {
|
func TestOnlyStaleTargetsAreDropped(t *testing.T) {
|
||||||
ctx, cancel := context.WithCancel(context.Background())
|
ctx := t.Context()
|
||||||
defer cancel()
|
|
||||||
|
|
||||||
myJob := "my-job"
|
myJob := "my-job"
|
||||||
myJobSDTargetURL := "my:9876"
|
myJobSDTargetURL := "my:9876"
|
||||||
|
|||||||
@ -1366,8 +1366,7 @@ func TestScrapeLoopSeriesAdded(t *testing.T) {
|
|||||||
func TestScrapeLoopFailWithInvalidLabelsAfterRelabel(t *testing.T) {
|
func TestScrapeLoopFailWithInvalidLabelsAfterRelabel(t *testing.T) {
|
||||||
s := teststorage.New(t)
|
s := teststorage.New(t)
|
||||||
defer s.Close()
|
defer s.Close()
|
||||||
ctx, cancel := context.WithCancel(context.Background())
|
ctx := t.Context()
|
||||||
defer cancel()
|
|
||||||
|
|
||||||
target := &Target{
|
target := &Target{
|
||||||
labels: labels.FromStrings("pod_label_invalid_012\xff", "test"),
|
labels: labels.FromStrings("pod_label_invalid_012\xff", "test"),
|
||||||
@ -1398,8 +1397,7 @@ func TestScrapeLoopFailLegacyUnderUTF8(t *testing.T) {
|
|||||||
// legacy.
|
// legacy.
|
||||||
s := teststorage.New(t)
|
s := teststorage.New(t)
|
||||||
defer s.Close()
|
defer s.Close()
|
||||||
ctx, cancel := context.WithCancel(context.Background())
|
ctx := t.Context()
|
||||||
defer cancel()
|
|
||||||
|
|
||||||
sl := newBasicScrapeLoop(t, ctx, &testScraper{}, s.Appender, 0)
|
sl := newBasicScrapeLoop(t, ctx, &testScraper{}, s.Appender, 0)
|
||||||
sl.validationScheme = model.LegacyValidation
|
sl.validationScheme = model.LegacyValidation
|
||||||
@ -1544,7 +1542,7 @@ func BenchmarkScrapeLoopAppend(b *testing.B) {
|
|||||||
|
|
||||||
b.ReportAllocs()
|
b.ReportAllocs()
|
||||||
b.ResetTimer()
|
b.ResetTimer()
|
||||||
for i := 0; i < b.N; i++ {
|
for b.Loop() {
|
||||||
ts = ts.Add(time.Second)
|
ts = ts.Add(time.Second)
|
||||||
_, _, _, err := sl.append(slApp, bcase.parsable, bcase.contentType, ts)
|
_, _, _, err := sl.append(slApp, bcase.parsable, bcase.contentType, ts)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
@ -3277,8 +3275,8 @@ func TestTargetScraperScrapeOK(t *testing.T) {
|
|||||||
"Expected Accept header to prefer application/vnd.google.protobuf.")
|
"Expected Accept header to prefer application/vnd.google.protobuf.")
|
||||||
}
|
}
|
||||||
|
|
||||||
contentTypes := strings.Split(accept, ",")
|
contentTypes := strings.SplitSeq(accept, ",")
|
||||||
for _, ct := range contentTypes {
|
for ct := range contentTypes {
|
||||||
match := qValuePattern.FindStringSubmatch(ct)
|
match := qValuePattern.FindStringSubmatch(ct)
|
||||||
require.Len(t, match, 3)
|
require.Len(t, match, 3)
|
||||||
qValue, err := strconv.ParseFloat(match[1], 64)
|
qValue, err := strconv.ParseFloat(match[1], 64)
|
||||||
@ -4100,8 +4098,7 @@ func TestScrapeReportLimit(t *testing.T) {
|
|||||||
// scrape have been inserted in the database.
|
// scrape have been inserted in the database.
|
||||||
}
|
}
|
||||||
|
|
||||||
ctx, cancel := context.WithCancel(context.Background())
|
ctx := t.Context()
|
||||||
defer cancel()
|
|
||||||
q, err := s.Querier(time.Time{}.UnixNano(), time.Now().UnixNano())
|
q, err := s.Querier(time.Time{}.UnixNano(), time.Now().UnixNano())
|
||||||
require.NoError(t, err)
|
require.NoError(t, err)
|
||||||
defer q.Close()
|
defer q.Close()
|
||||||
@ -4155,8 +4152,7 @@ func TestScrapeUTF8(t *testing.T) {
|
|||||||
// scrape have been inserted in the database.
|
// scrape have been inserted in the database.
|
||||||
}
|
}
|
||||||
|
|
||||||
ctx, cancel := context.WithCancel(context.Background())
|
ctx := t.Context()
|
||||||
defer cancel()
|
|
||||||
q, err := s.Querier(time.Time{}.UnixNano(), time.Now().UnixNano())
|
q, err := s.Querier(time.Time{}.UnixNano(), time.Now().UnixNano())
|
||||||
require.NoError(t, err)
|
require.NoError(t, err)
|
||||||
defer q.Close()
|
defer q.Close()
|
||||||
@ -4391,8 +4387,7 @@ test_summary_count 199
|
|||||||
case <-scrapedTwice:
|
case <-scrapedTwice:
|
||||||
}
|
}
|
||||||
|
|
||||||
ctx, cancel := context.WithCancel(context.Background())
|
ctx := t.Context()
|
||||||
defer cancel()
|
|
||||||
q, err := simpleStorage.Querier(time.Time{}.UnixNano(), time.Now().UnixNano())
|
q, err := simpleStorage.Querier(time.Time{}.UnixNano(), time.Now().UnixNano())
|
||||||
require.NoError(t, err)
|
require.NoError(t, err)
|
||||||
defer q.Close()
|
defer q.Close()
|
||||||
@ -4865,8 +4860,7 @@ metric: <
|
|||||||
sl.append(app, content, contentType, time.Now())
|
sl.append(app, content, contentType, time.Now())
|
||||||
require.NoError(t, app.Commit())
|
require.NoError(t, app.Commit())
|
||||||
|
|
||||||
ctx, cancel := context.WithCancel(context.Background())
|
ctx := t.Context()
|
||||||
defer cancel()
|
|
||||||
q, err := simpleStorage.Querier(time.Time{}.UnixNano(), time.Now().UnixNano())
|
q, err := simpleStorage.Querier(time.Time{}.UnixNano(), time.Now().UnixNano())
|
||||||
require.NoError(t, err)
|
require.NoError(t, err)
|
||||||
defer q.Close()
|
defer q.Close()
|
||||||
@ -4978,8 +4972,7 @@ disk_usage_bytes 456
|
|||||||
case <-scrapedTwice:
|
case <-scrapedTwice:
|
||||||
}
|
}
|
||||||
|
|
||||||
ctx, cancel := context.WithCancel(context.Background())
|
ctx := t.Context()
|
||||||
defer cancel()
|
|
||||||
q, err := simpleStorage.Querier(time.Time{}.UnixNano(), time.Now().UnixNano())
|
q, err := simpleStorage.Querier(time.Time{}.UnixNano(), time.Now().UnixNano())
|
||||||
require.NoError(t, err)
|
require.NoError(t, err)
|
||||||
defer q.Close()
|
defer q.Close()
|
||||||
@ -5244,7 +5237,7 @@ func BenchmarkTargetScraperGzip(b *testing.B) {
|
|||||||
timeout: time.Second,
|
timeout: time.Second,
|
||||||
}
|
}
|
||||||
b.ResetTimer()
|
b.ResetTimer()
|
||||||
for i := 0; i < b.N; i++ {
|
for b.Loop() {
|
||||||
_, err = ts.scrape(context.Background())
|
_, err = ts.scrape(context.Background())
|
||||||
require.NoError(b, err)
|
require.NoError(b, err)
|
||||||
}
|
}
|
||||||
|
|||||||
@ -510,7 +510,7 @@ scrape_configs:
|
|||||||
var tgets []*Target
|
var tgets []*Target
|
||||||
lb := labels.NewBuilder(labels.EmptyLabels())
|
lb := labels.NewBuilder(labels.EmptyLabels())
|
||||||
group := &targetgroup.Group{Targets: targets}
|
group := &targetgroup.Group{Targets: targets}
|
||||||
for i := 0; i < b.N; i++ {
|
for b.Loop() {
|
||||||
tgets, _ = TargetsFromGroup(group, config.ScrapeConfigs[0], tgets, lb)
|
tgets, _ = TargetsFromGroup(group, config.ScrapeConfigs[0], tgets, lb)
|
||||||
if len(targets) != nTargets {
|
if len(targets) != nTargets {
|
||||||
b.Fatalf("Expected %d targets, got %d", nTargets, len(targets))
|
b.Fatalf("Expected %d targets, got %d", nTargets, len(targets))
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user