prometheus/scrape/scrape_test.go
Bartlomiej Plotka 848b16d686
test: Add benchmark without storage + fix skipRecording mock feature (#17987)
* test: Add benchmark without storage

Signed-off-by: bwplotka <bwplotka@gmail.com>

make bench fair

Signed-off-by: bwplotka <bwplotka@gmail.com>

tmp

Signed-off-by: bwplotka <bwplotka@gmail.com>

* Apply suggestions from code review

Co-authored-by: Arve Knudsen <arve.knudsen@gmail.com>
Signed-off-by: Bartlomiej Plotka <bwplotka@gmail.com>

---------

Signed-off-by: bwplotka <bwplotka@gmail.com>
Signed-off-by: Bartlomiej Plotka <bwplotka@gmail.com>
Co-authored-by: Arve Knudsen <arve.knudsen@gmail.com>
2026-02-02 12:44:11 +00:00

6736 lines
212 KiB
Go

// Copyright The Prometheus Authors
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
package scrape
import (
"bytes"
"compress/gzip"
"context"
"encoding/binary"
"errors"
"fmt"
"io"
"maps"
"math"
"net/http"
"net/http/httptest"
"net/url"
"os"
"sort"
"strconv"
"strings"
"sync"
"testing"
"text/template"
"time"
"github.com/gogo/protobuf/proto"
"github.com/grafana/regexp"
"github.com/prometheus/client_golang/prometheus"
prom_testutil "github.com/prometheus/client_golang/prometheus/testutil"
dto "github.com/prometheus/client_model/go"
config_util "github.com/prometheus/common/config"
"github.com/prometheus/common/expfmt"
"github.com/prometheus/common/model"
"github.com/prometheus/common/promslog"
"github.com/stretchr/testify/require"
"go.opentelemetry.io/otel"
"go.opentelemetry.io/otel/propagation"
sdktrace "go.opentelemetry.io/otel/sdk/trace"
"go.uber.org/atomic"
"go.uber.org/goleak"
"github.com/prometheus/prometheus/config"
"github.com/prometheus/prometheus/discovery"
"github.com/prometheus/prometheus/discovery/targetgroup"
"github.com/prometheus/prometheus/model/exemplar"
"github.com/prometheus/prometheus/model/histogram"
"github.com/prometheus/prometheus/model/labels"
"github.com/prometheus/prometheus/model/metadata"
"github.com/prometheus/prometheus/model/relabel"
"github.com/prometheus/prometheus/model/textparse"
"github.com/prometheus/prometheus/model/timestamp"
"github.com/prometheus/prometheus/model/value"
"github.com/prometheus/prometheus/storage"
"github.com/prometheus/prometheus/tsdb"
"github.com/prometheus/prometheus/tsdb/chunkenc"
"github.com/prometheus/prometheus/util/pool"
"github.com/prometheus/prometheus/util/teststorage"
"github.com/prometheus/prometheus/util/testutil"
)
func TestMain(m *testing.M) {
testutil.TolerantVerifyLeak(m)
}
func newTestRegistryAndScrapeMetrics(t testing.TB) (*prometheus.Registry, *scrapeMetrics) {
reg := prometheus.NewRegistry()
metrics, err := newScrapeMetrics(reg)
require.NoError(t, err)
return reg, metrics
}
func newTestScrapeMetrics(t testing.TB) *scrapeMetrics {
_, metrics := newTestRegistryAndScrapeMetrics(t)
return metrics
}
func TestNewScrapePool(t *testing.T) {
foreachAppendable(t, func(t *testing.T, appV2 bool) {
testNewScrapePool(t, appV2)
})
}
func testNewScrapePool(t *testing.T, appV2 bool) {
var (
app = teststorage.NewAppendable()
sa = selectAppendable(app, appV2)
cfg = &config.ScrapeConfig{
MetricNameValidationScheme: model.UTF8Validation,
MetricNameEscapingScheme: model.AllowUTF8,
}
sp, err = newScrapePool(cfg, sa.V1(), sa.V2(), 0, nil, nil, &Options{}, newTestScrapeMetrics(t))
)
require.NoError(t, err)
if appV2 {
a, ok := sp.appendableV2.(*teststorage.Appendable)
require.True(t, ok, "Failure to append.")
require.Equal(t, app, a, "Wrong sample AppenderV2.")
require.Equal(t, cfg, sp.config, "Wrong scrape config.")
require.Nil(t, sp.appendable)
return
}
a, ok := sp.appendable.(*teststorage.Appendable)
require.True(t, ok, "Failure to append.")
require.Equal(t, app, a, "Wrong sample appender.")
require.Equal(t, cfg, sp.config, "Wrong scrape config.")
require.Nil(t, sp.appendableV2)
}
func TestStorageHandlesOutOfOrderTimestamps(t *testing.T) {
foreachAppendable(t, func(t *testing.T, appV2 bool) {
testStorageHandlesOutOfOrderTimestamps(t, appV2)
})
}
func testStorageHandlesOutOfOrderTimestamps(t *testing.T, appV2 bool) {
// Test with default OutOfOrderTimeWindow (0)
t.Run("Out-Of-Order Sample Disabled", func(t *testing.T) {
s := teststorage.New(t)
runScrapeLoopTest(t, appV2, s, false)
})
// Test with specific OutOfOrderTimeWindow (600000)
t.Run("Out-Of-Order Sample Enabled", func(t *testing.T) {
s := teststorage.New(t, func(opt *tsdb.Options) {
opt.OutOfOrderTimeWindow = 600000
})
runScrapeLoopTest(t, appV2, s, true)
})
}
func runScrapeLoopTest(t *testing.T, appV2 bool, s *teststorage.TestStorage, expectOutOfOrder bool) {
sl, _ := newTestScrapeLoop(t, withAppendable(s, appV2))
// Current time for generating timestamps.
now := time.Now()
// Calculate timestamps for the samples based on the current time.
now = now.Truncate(time.Minute) // round down the now timestamp to the nearest minute
timestampInorder1 := now
timestampOutOfOrder := now.Add(-5 * time.Minute)
timestampInorder2 := now.Add(5 * time.Minute)
app := sl.appender()
_, _, _, err := app.append([]byte(`metric_total{a="1",b="1"} 1`), "text/plain", timestampInorder1)
require.NoError(t, err)
_, _, _, err = app.append([]byte(`metric_total{a="1",b="1"} 2`), "text/plain", timestampOutOfOrder)
require.NoError(t, err)
_, _, _, err = app.append([]byte(`metric_total{a="1",b="1"} 3`), "text/plain", timestampInorder2)
require.NoError(t, err)
require.NoError(t, app.Commit())
// Query the samples back from the storage.
q, err := s.Querier(time.Time{}.UnixNano(), time.Now().UnixNano())
require.NoError(t, err)
t.Cleanup(func() { _ = q.Close() })
// Use a matcher to filter the metric name.
series := q.Select(t.Context(), false, nil, labels.MustNewMatcher(labels.MatchRegexp, "__name__", "metric_total"))
var results []sample
for series.Next() {
it := series.At().Iterator(nil)
for it.Next() == chunkenc.ValFloat {
t, v := it.At()
results = append(results, sample{
L: series.At().Labels(),
T: t,
V: v,
})
}
require.NoError(t, it.Err())
}
require.NoError(t, series.Err())
// Define the expected results
want := []sample{
{
L: labels.FromStrings("__name__", "metric_total", "a", "1", "b", "1"),
T: timestamp.FromTime(timestampInorder1),
V: 1,
},
{
L: labels.FromStrings("__name__", "metric_total", "a", "1", "b", "1"),
T: timestamp.FromTime(timestampInorder2),
V: 3,
},
}
if expectOutOfOrder {
teststorage.RequireNotEqual(t, want, results, "Expected results to include out-of-order sample:\n%s", results)
} else {
teststorage.RequireEqual(t, want, results, "Appended samples not as expected:\n%s", results)
}
}
// Regression test against https://github.com/prometheus/prometheus/issues/15831.
func TestScrapeAppend_MetadataUpdate(t *testing.T) {
foreachAppendable(t, func(t *testing.T, appV2 bool) {
testScrapeAppendMetadataUpdate(t, appV2)
})
}
func testScrapeAppendMetadataUpdate(t *testing.T, appV2 bool) {
const (
scrape1 = `# TYPE test_metric counter
# HELP test_metric some help text
# UNIT test_metric metric
test_metric_total 1
# TYPE test_metric2 gauge
# HELP test_metric2 other help text
test_metric2{foo="bar"} 2
# TYPE test_metric3 gauge
# HELP test_metric3 this represents tricky case of "broken" text that is not trivial to detect
test_metric3_metric4{foo="bar"} 2
# EOF`
scrape2 = `# TYPE test_metric counter
# HELP test_metric different help text
test_metric_total 11
# TYPE test_metric2 gauge
# HELP test_metric2 other help text
# UNIT test_metric2 metric2
test_metric2{foo="bar"} 22
# EOF`
)
appTest := teststorage.NewAppendable()
sl, _ := newTestScrapeLoop(t, withAppendable(appTest, appV2))
now := time.Now()
app := sl.appender()
_, _, _, err := app.append([]byte(scrape1), "application/openmetrics-text", now)
require.NoError(t, err)
require.NoError(t, app.Commit())
teststorage.RequireEqual(t, []sample{
{L: labels.FromStrings("__name__", "test_metric_total"), M: metadata.Metadata{Type: "counter", Unit: "metric", Help: "some help text"}},
{L: labels.FromStrings("__name__", "test_metric2", "foo", "bar"), M: metadata.Metadata{Type: "gauge", Unit: "", Help: "other help text"}},
}, appTest.ResultMetadata())
appTest.ResultReset()
app = sl.appender()
_, _, _, err = app.append([]byte(scrape1), "application/openmetrics-text", now.Add(15*time.Second))
require.NoError(t, err)
require.NoError(t, app.Commit())
if appV2 {
// Next (the same) scrape should pass new metadata entries as per always-on metadata Appendable V2 contract.
teststorage.RequireEqual(t, []sample{
{L: labels.FromStrings("__name__", "test_metric_total"), M: metadata.Metadata{Type: "counter", Unit: "metric", Help: "some help text"}},
{L: labels.FromStrings("__name__", "test_metric2", "foo", "bar"), M: metadata.Metadata{Type: "gauge", Unit: "", Help: "other help text"}},
}, appTest.ResultMetadata())
} else {
// Next (the same) scrape should not add new metadata entries.
require.Empty(t, appTest.ResultMetadata())
}
appTest.ResultReset()
app = sl.appender()
_, _, _, err = app.append([]byte(scrape2), "application/openmetrics-text", now.Add(15*time.Second))
require.NoError(t, err)
require.NoError(t, app.Commit())
teststorage.RequireEqual(t, []sample{
{L: labels.FromStrings("__name__", "test_metric_total"), M: metadata.Metadata{Type: "counter", Unit: "metric", Help: "different help text"}}, // Here, technically we should have no unit, but it's a known limitation of the current implementation.
{L: labels.FromStrings("__name__", "test_metric2", "foo", "bar"), M: metadata.Metadata{Type: "gauge", Unit: "metric2", Help: "other help text"}},
}, appTest.ResultMetadata())
appTest.ResultReset()
}
func TestScrapeReportMetadata(t *testing.T) {
foreachAppendable(t, func(t *testing.T, appV2 bool) {
testScrapeReportMetadata(t, appV2)
})
}
func testScrapeReportMetadata(t *testing.T, appV2 bool) {
appTest := teststorage.NewAppendable()
sl, _ := newTestScrapeLoop(t, withAppendable(appTest, appV2))
app := sl.appender()
now := time.Now()
require.NoError(t, sl.report(app, now, 2*time.Second, 1, 1, 1, 512, nil))
require.NoError(t, app.Commit())
teststorage.RequireEqual(t, []sample{
{L: labels.FromStrings("__name__", "up"), M: scrapeHealthMetric.Metadata},
{L: labels.FromStrings("__name__", "scrape_duration_seconds"), M: scrapeDurationMetric.Metadata},
{L: labels.FromStrings("__name__", "scrape_samples_scraped"), M: scrapeSamplesMetric.Metadata},
{L: labels.FromStrings("__name__", "scrape_samples_post_metric_relabeling"), M: samplesPostRelabelMetric.Metadata},
{L: labels.FromStrings("__name__", "scrape_series_added"), M: scrapeSeriesAddedMetric.Metadata},
}, appTest.ResultMetadata())
}
func TestIsSeriesPartOfFamily(t *testing.T) {
t.Run("counter", func(t *testing.T) {
require.True(t, isSeriesPartOfFamily("http_requests_total", []byte("http_requests_total"), model.MetricTypeCounter)) // Prometheus text style.
require.True(t, isSeriesPartOfFamily("http_requests_total", []byte("http_requests"), model.MetricTypeCounter)) // OM text style.
require.True(t, isSeriesPartOfFamily("http_requests_total", []byte("http_requests_total"), model.MetricTypeUnknown))
require.False(t, isSeriesPartOfFamily("http_requests_total", []byte("http_requests"), model.MetricTypeUnknown)) // We don't know.
require.False(t, isSeriesPartOfFamily("http_requests2_total", []byte("http_requests_total"), model.MetricTypeCounter))
require.False(t, isSeriesPartOfFamily("http_requests_requests_total", []byte("http_requests"), model.MetricTypeCounter))
})
t.Run("gauge", func(t *testing.T) {
require.True(t, isSeriesPartOfFamily("http_requests_count", []byte("http_requests_count"), model.MetricTypeGauge))
require.True(t, isSeriesPartOfFamily("http_requests_count", []byte("http_requests_count"), model.MetricTypeUnknown))
require.False(t, isSeriesPartOfFamily("http_requests_count2", []byte("http_requests_count"), model.MetricTypeCounter))
})
t.Run("histogram", func(t *testing.T) {
require.True(t, isSeriesPartOfFamily("http_requests_seconds_sum", []byte("http_requests_seconds"), model.MetricTypeHistogram))
require.True(t, isSeriesPartOfFamily("http_requests_seconds_count", []byte("http_requests_seconds"), model.MetricTypeHistogram))
require.True(t, isSeriesPartOfFamily("http_requests_seconds_bucket", []byte("http_requests_seconds"), model.MetricTypeHistogram))
require.True(t, isSeriesPartOfFamily("http_requests_seconds", []byte("http_requests_seconds"), model.MetricTypeHistogram))
require.False(t, isSeriesPartOfFamily("http_requests_seconds_sum", []byte("http_requests_seconds"), model.MetricTypeUnknown)) // We don't know.
require.False(t, isSeriesPartOfFamily("http_requests_seconds2_sum", []byte("http_requests_seconds"), model.MetricTypeHistogram))
})
t.Run("summary", func(t *testing.T) {
require.True(t, isSeriesPartOfFamily("http_requests_seconds_sum", []byte("http_requests_seconds"), model.MetricTypeSummary))
require.True(t, isSeriesPartOfFamily("http_requests_seconds_count", []byte("http_requests_seconds"), model.MetricTypeSummary))
require.True(t, isSeriesPartOfFamily("http_requests_seconds", []byte("http_requests_seconds"), model.MetricTypeSummary))
require.False(t, isSeriesPartOfFamily("http_requests_seconds_sum", []byte("http_requests_seconds"), model.MetricTypeUnknown)) // We don't know.
require.False(t, isSeriesPartOfFamily("http_requests_seconds2_sum", []byte("http_requests_seconds"), model.MetricTypeSummary))
})
t.Run("info", func(t *testing.T) {
require.True(t, isSeriesPartOfFamily("go_build_info", []byte("go_build_info"), model.MetricTypeInfo)) // Prometheus text style.
require.True(t, isSeriesPartOfFamily("go_build_info", []byte("go_build"), model.MetricTypeInfo)) // OM text style.
require.True(t, isSeriesPartOfFamily("go_build_info", []byte("go_build_info"), model.MetricTypeUnknown))
require.False(t, isSeriesPartOfFamily("go_build_info", []byte("go_build"), model.MetricTypeUnknown)) // We don't know.
require.False(t, isSeriesPartOfFamily("go_build2_info", []byte("go_build_info"), model.MetricTypeInfo))
require.False(t, isSeriesPartOfFamily("go_build_build_info", []byte("go_build_info"), model.MetricTypeInfo))
})
}
func TestDroppedTargetsList(t *testing.T) {
foreachAppendable(t, func(t *testing.T, appV2 bool) {
testDroppedTargetsList(t, appV2)
})
}
func testDroppedTargetsList(t *testing.T, appV2 bool) {
var (
app = teststorage.NewAppendable()
cfg = &config.ScrapeConfig{
JobName: "dropMe",
ScrapeInterval: model.Duration(1),
MetricNameValidationScheme: model.UTF8Validation,
MetricNameEscapingScheme: model.AllowUTF8,
RelabelConfigs: []*relabel.Config{
{
Action: relabel.Drop,
Regex: relabel.MustNewRegexp("dropMe"),
SourceLabels: model.LabelNames{"job"},
NameValidationScheme: model.UTF8Validation,
},
},
}
tgs = []*targetgroup.Group{
{
Targets: []model.LabelSet{
{model.AddressLabel: "127.0.0.1:9090"},
{model.AddressLabel: "127.0.0.1:9091"},
},
},
}
sa = selectAppendable(app, appV2)
sp, _ = newScrapePool(cfg, sa.V1(), sa.V2(), 0, nil, nil, &Options{}, newTestScrapeMetrics(t))
expectedLabelSetString = "{__address__=\"127.0.0.1:9090\", __scrape_interval__=\"0s\", __scrape_timeout__=\"0s\", job=\"dropMe\"}"
expectedLength = 2
)
sp.Sync(tgs)
sp.Sync(tgs)
require.Len(t, sp.droppedTargets, expectedLength)
require.Equal(t, expectedLength, sp.droppedTargetsCount)
lb := labels.NewBuilder(labels.EmptyLabels())
require.Equal(t, expectedLabelSetString, sp.droppedTargets[0].DiscoveredLabels(lb).String())
// Check that count is still correct when we don't retain all dropped targets.
sp.config.KeepDroppedTargets = 1
sp.Sync(tgs)
require.Len(t, sp.droppedTargets, 1)
require.Equal(t, expectedLength, sp.droppedTargetsCount)
}
// TestDiscoveredLabelsUpdate checks that DiscoveredLabels are updated
// even when new labels don't affect the target `hash`.
func TestDiscoveredLabelsUpdate(t *testing.T) {
sp := newTestScrapePool(t, nil, false, nil)
// These are used when syncing so need this to avoid a panic.
sp.config = &config.ScrapeConfig{
ScrapeInterval: model.Duration(1),
ScrapeTimeout: model.Duration(1),
MetricNameValidationScheme: model.UTF8Validation,
MetricNameEscapingScheme: model.AllowUTF8,
}
sp.activeTargets = make(map[uint64]*Target)
t1 := &Target{
tLabels: model.LabelSet{"label": "name"},
scrapeConfig: sp.config,
}
sp.activeTargets[t1.hash()] = t1
t2 := &Target{
tLabels: model.LabelSet{"labelNew": "nameNew"},
scrapeConfig: sp.config,
}
sp.sync([]*Target{t2})
lb := labels.NewBuilder(labels.EmptyLabels())
require.Equal(t, t2.DiscoveredLabels(lb), sp.activeTargets[t1.hash()].DiscoveredLabels(lb))
}
type testLoop struct {
startFunc func(interval, timeout time.Duration, errc chan<- error)
stopFunc func()
forcedErr error
forcedErrMtx sync.Mutex
runOnce bool
interval time.Duration
timeout time.Duration
}
func (*testLoop) setScrapeFailureLogger(FailureLogger) {
}
func (l *testLoop) run(errc chan<- error) {
if l.runOnce {
panic("loop must be started only once")
}
l.runOnce = true
l.startFunc(l.interval, l.timeout, errc)
}
func (*testLoop) disableEndOfRunStalenessMarkers() {
}
func (l *testLoop) setForcedError(err error) {
l.forcedErrMtx.Lock()
defer l.forcedErrMtx.Unlock()
l.forcedErr = err
}
func (l *testLoop) getForcedError() error {
l.forcedErrMtx.Lock()
defer l.forcedErrMtx.Unlock()
return l.forcedErr
}
func (l *testLoop) stop() {
l.stopFunc()
}
func (*testLoop) getCache() *scrapeCache {
return nil
}
func TestScrapePoolStop(t *testing.T) {
t.Parallel()
sp := newTestScrapePool(t, nil, false, nil)
var mtx sync.Mutex
stopped := map[uint64]bool{}
numTargets := 20
// Stopping the scrape pool must call stop() on all scrape loops,
// clean them and the respective targets up. It must wait until each loop's
// stop function returned before returning itself.
for i := range numTargets {
t := &Target{
labels: labels.FromStrings(model.AddressLabel, fmt.Sprintf("example.com:%d", i)),
scrapeConfig: &config.ScrapeConfig{},
}
l := &testLoop{}
d := time.Duration((i+1)*20) * time.Millisecond
l.stopFunc = func() {
time.Sleep(d)
mtx.Lock()
stopped[t.hash()] = true
mtx.Unlock()
}
sp.activeTargets[t.hash()] = t
sp.loops[t.hash()] = l
}
done := make(chan struct{})
stopTime := time.Now()
go func() {
sp.stop()
close(done)
}()
select {
case <-time.After(5 * time.Second):
require.Fail(t, "scrapeLoop.stop() did not return as expected")
case <-done:
// This should have taken at least as long as the last target slept.
require.GreaterOrEqual(t, time.Since(stopTime), time.Duration(numTargets*20)*time.Millisecond, "scrapeLoop.stop() exited before all targets stopped")
}
mtx.Lock()
require.Len(t, stopped, numTargets, "Unexpected number of stopped loops")
mtx.Unlock()
require.Empty(t, sp.activeTargets, "Targets were not cleared on stopping: %d left", len(sp.activeTargets))
require.Empty(t, sp.loops, "Loops were not cleared on stopping: %d left", len(sp.loops))
}
// TestScrapePoolReload tests reloading logic, so:
// * all loops are reloaded, reusing cache if scrape config changed.
// * reloaded loops are stopped before new ones are started.
// * new scrapeLoops are configured with the updated scrape config.
func TestScrapePoolReload(t *testing.T) {
t.Parallel()
var (
mtx sync.Mutex
numTargets = 20
stopped = map[uint64]bool{}
)
cfg0 := &config.ScrapeConfig{}
cfg1 := &config.ScrapeConfig{
ScrapeInterval: model.Duration(3 * time.Second),
ScrapeTimeout: model.Duration(2 * time.Second),
MetricNameValidationScheme: model.UTF8Validation,
MetricNameEscapingScheme: model.AllowUTF8,
// Test a few example options.
SampleLimit: 123,
ScrapeFallbackProtocol: "application/vnd.google.protobuf;proto=io.prometheus.client.MetricFamily;encoding=delimited",
}
newLoopCfg1 := func(opts scrapeLoopOptions) loop {
// Test cfg1 is being used.
require.Equal(t, cfg1, opts.sp.config)
// Inject out testLoop that allows mocking start and stop.
l := &testLoop{interval: opts.interval, timeout: opts.timeout}
// On start, expect previous loop instances for the same target to be stopped.
l.startFunc = func(interval, timeout time.Duration, _ chan<- error) {
// Ensure cfg1 interval and timeout are correctly configured.
require.Equal(t, time.Duration(cfg1.ScrapeInterval), interval, "Unexpected scrape interval")
require.Equal(t, time.Duration(cfg1.ScrapeTimeout), timeout, "Unexpected scrape timeout")
mtx.Lock()
targetScraper := opts.scraper.(*targetScraper)
require.True(t, stopped[targetScraper.hash()], "Scrape loop for %v not stopped yet", targetScraper)
mtx.Unlock()
}
return l
}
// Create test pool.
reg, metrics := newTestRegistryAndScrapeMetrics(t)
sp := newTestScrapePool(t, nil, false, newLoopCfg1)
sp.metrics = metrics
// Prefill pool with 20 loops, simulating 20 scrape targets.
for i := range numTargets {
t := &Target{
labels: labels.FromStrings(model.AddressLabel, fmt.Sprintf("example.com:%d", i)),
scrapeConfig: cfg0,
}
l := &testLoop{}
d := time.Duration((i+1)*20) * time.Millisecond
l.stopFunc = func() {
time.Sleep(d) // Sleep uneven time on stop.
mtx.Lock()
stopped[t.hash()] = true
mtx.Unlock()
}
sp.activeTargets[t.hash()] = t
sp.loops[t.hash()] = l
}
beforeTargets := map[uint64]*Target{}
maps.Copy(beforeTargets, sp.activeTargets)
// Reloading a scrape pool with a new scrape configuration must stop all scrape
// loops and start new ones. A new loop must not be started before the preceding
// one terminated.
require.NoError(t, sp.reload(cfg1))
var stoppedCount int
mtx.Lock()
stoppedCount = len(stopped)
mtx.Unlock()
require.Equal(t, numTargets, stoppedCount, "Unexpected number of stopped loops")
require.Equal(t, sp.activeTargets, beforeTargets, "Reloading affected target states unexpectedly")
require.Len(t, sp.loops, numTargets, "Unexpected number of loops after reload")
// Check if prometheus_target_reload_length_seconds points to cfg1.ScrapeInterval.
got, err := gatherLabels(reg, "prometheus_target_reload_length_seconds")
require.NoError(t, err)
expectedName, expectedValue := "interval", cfg1.ScrapeInterval.String()
require.Equal(t, [][]*dto.LabelPair{{{Name: &expectedName, Value: &expectedValue}}}, got)
require.Equal(t, 1.0, prom_testutil.ToFloat64(sp.metrics.targetScrapePoolReloads))
}
func TestScrapePoolReloadPreserveRelabeledIntervalTimeout(t *testing.T) {
reloadCfg := &config.ScrapeConfig{
ScrapeInterval: model.Duration(3 * time.Second),
ScrapeTimeout: model.Duration(2 * time.Second),
MetricNameValidationScheme: model.UTF8Validation,
MetricNameEscapingScheme: model.AllowUTF8,
}
newLoop := func(opts scrapeLoopOptions) loop {
l := &testLoop{interval: opts.interval, timeout: opts.timeout}
l.startFunc = func(interval, timeout time.Duration, _ chan<- error) {
require.Equal(t, 5*time.Second, interval, "Unexpected scrape interval")
require.Equal(t, 3*time.Second, timeout, "Unexpected scrape timeout")
}
return l
}
reg, metrics := newTestRegistryAndScrapeMetrics(t)
sp := newTestScrapePool(t, nil, false, newLoop)
sp.activeTargets[1] = &Target{
labels: labels.FromStrings(model.ScrapeIntervalLabel, "5s", model.ScrapeTimeoutLabel, "3s"),
}
sp.metrics = metrics
sp.loops[1] = noopLoop()
err := sp.reload(reloadCfg)
if err != nil {
t.Fatalf("unable to reload configuration: %s", err)
}
// Check that the reload metric is labeled with the pool interval, not the overridden interval.
got, err := gatherLabels(reg, "prometheus_target_reload_length_seconds")
require.NoError(t, err)
expectedName, expectedValue := "interval", "3s"
require.Equal(t, [][]*dto.LabelPair{{{Name: &expectedName, Value: &expectedValue}}}, got)
}
// Gather metrics from the provided Gatherer with specified familyName,
// and return all sets of name/value pairs.
func gatherLabels(g prometheus.Gatherer, familyName string) ([][]*dto.LabelPair, error) {
families, err := g.Gather()
if err != nil {
return nil, err
}
ret := make([][]*dto.LabelPair, 0)
for _, f := range families {
if f.GetName() == familyName {
for _, m := range f.GetMetric() {
ret = append(ret, m.GetLabel())
}
break
}
}
return ret, nil
}
func TestScrapePoolTargetLimit(t *testing.T) {
var wg sync.WaitGroup
// On starting to run, new loops created on reload check whether their preceding
// equivalents have been stopped.
newLoop := func(scrapeLoopOptions) loop {
wg.Add(1)
l := &testLoop{
startFunc: func(_, _ time.Duration, _ chan<- error) {
wg.Done()
},
stopFunc: func() {},
}
return l
}
sp := newTestScrapePool(t, nil, false, newLoop)
var tgs []*targetgroup.Group
for i := range 50 {
tgs = append(tgs,
&targetgroup.Group{
Targets: []model.LabelSet{
{model.AddressLabel: model.LabelValue(fmt.Sprintf("127.0.0.1:%d", 9090+i))},
},
},
)
}
var limit uint
reloadWithLimit := func(l uint) {
limit = l
require.NoError(t, sp.reload(&config.ScrapeConfig{
ScrapeInterval: model.Duration(3 * time.Second),
ScrapeTimeout: model.Duration(2 * time.Second),
MetricNameValidationScheme: model.UTF8Validation,
MetricNameEscapingScheme: model.AllowUTF8,
TargetLimit: l,
}))
}
var targets int
loadTargets := func(n int) {
targets = n
sp.Sync(tgs[:n])
}
validateIsRunning := func() {
wg.Wait()
for _, l := range sp.loops {
require.True(t, l.(*testLoop).runOnce, "loop should be running")
}
}
validateErrorMessage := func(shouldErr bool) {
for _, l := range sp.loops {
lerr := l.(*testLoop).getForcedError()
if shouldErr {
require.Error(t, lerr, "error was expected for %d targets with a limit of %d", targets, limit)
require.EqualError(t, lerr, fmt.Sprintf("target_limit exceeded (number of targets: %d, limit: %d)", targets, limit))
} else {
require.NoError(t, lerr)
}
}
}
reloadWithLimit(0)
loadTargets(50)
validateIsRunning()
// Simulate an initial config with a limit.
sp.config.TargetLimit = 30
limit = 30
loadTargets(50)
validateIsRunning()
validateErrorMessage(true)
reloadWithLimit(50)
validateIsRunning()
validateErrorMessage(false)
reloadWithLimit(40)
validateIsRunning()
validateErrorMessage(true)
loadTargets(30)
validateIsRunning()
validateErrorMessage(false)
loadTargets(40)
validateIsRunning()
validateErrorMessage(false)
loadTargets(41)
validateIsRunning()
validateErrorMessage(true)
reloadWithLimit(0)
validateIsRunning()
validateErrorMessage(false)
reloadWithLimit(51)
validateIsRunning()
validateErrorMessage(false)
tgs = append(tgs,
&targetgroup.Group{
Targets: []model.LabelSet{
{model.AddressLabel: "127.0.0.1:1090"},
},
},
&targetgroup.Group{
Targets: []model.LabelSet{
{model.AddressLabel: "127.0.0.1:1090"},
},
},
)
sp.Sync(tgs)
validateIsRunning()
validateErrorMessage(false)
}
func TestScrapePoolAppenderWithLimits(t *testing.T) {
// Create a unique value, to validate the correct chain of appenders.
baseAppender := struct{ storage.Appender }{}
appendable := appendableFunc(func(context.Context) storage.Appender { return baseAppender })
sl, _ := newTestScrapeLoop(t, func(sl *scrapeLoop) {
sl.appendable = appendable
})
wrapped := appenderWithLimits(sl.appendable.Appender(context.Background()), 0, 0, histogram.ExponentialSchemaMax)
tl, ok := wrapped.(*timeLimitAppender)
require.True(t, ok, "Expected timeLimitAppender but got %T", wrapped)
require.Equal(t, baseAppender, tl.Appender, "Expected base appender but got %T", tl.Appender)
sampleLimit := 100
sl, _ = newTestScrapeLoop(t, func(sl *scrapeLoop) {
sl.appendable = appendable
sl.sampleLimit = sampleLimit
})
wrapped = appenderWithLimits(sl.appendable.Appender(context.Background()), sampleLimit, 0, histogram.ExponentialSchemaMax)
la, ok := wrapped.(*limitAppender)
require.True(t, ok, "Expected limitAppender but got %T", wrapped)
tl, ok = la.Appender.(*timeLimitAppender)
require.True(t, ok, "Expected timeLimitAppender but got %T", la.Appender)
require.Equal(t, baseAppender, tl.Appender, "Expected base appender but got %T", tl.Appender)
wrapped = appenderWithLimits(sl.appendable.Appender(context.Background()), sampleLimit, 100, histogram.ExponentialSchemaMax)
bl, ok := wrapped.(*bucketLimitAppender)
require.True(t, ok, "Expected bucketLimitAppender but got %T", wrapped)
la, ok = bl.Appender.(*limitAppender)
require.True(t, ok, "Expected limitAppender but got %T", bl)
tl, ok = la.Appender.(*timeLimitAppender)
require.True(t, ok, "Expected timeLimitAppender but got %T", la.Appender)
require.Equal(t, baseAppender, tl.Appender, "Expected base appender but got %T", tl.Appender)
wrapped = appenderWithLimits(sl.appendable.Appender(context.Background()), sampleLimit, 100, 0)
ml, ok := wrapped.(*maxSchemaAppender)
require.True(t, ok, "Expected maxSchemaAppender but got %T", wrapped)
bl, ok = ml.Appender.(*bucketLimitAppender)
require.True(t, ok, "Expected bucketLimitAppender but got %T", wrapped)
la, ok = bl.Appender.(*limitAppender)
require.True(t, ok, "Expected limitAppender but got %T", bl)
tl, ok = la.Appender.(*timeLimitAppender)
require.True(t, ok, "Expected timeLimitAppender but got %T", la.Appender)
require.Equal(t, baseAppender, tl.Appender, "Expected base appender but got %T", tl.Appender)
}
type appendableV2Func func(ctx context.Context) storage.AppenderV2
func (a appendableV2Func) AppenderV2(ctx context.Context) storage.AppenderV2 { return a(ctx) }
func TestScrapePoolAppenderWithLimits_AppendV2(t *testing.T) {
// Create a unique value, to validate the correct chain of appenders.
baseAppender := struct{ storage.AppenderV2 }{}
appendable := appendableV2Func(func(context.Context) storage.AppenderV2 { return baseAppender })
sl, _ := newTestScrapeLoop(t, func(sl *scrapeLoop) {
sl.appendableV2 = appendable
})
wrapped := appenderV2WithLimits(sl.appendableV2.AppenderV2(context.Background()), 0, 0, histogram.ExponentialSchemaMax)
tl, ok := wrapped.(*timeLimitAppenderV2)
require.True(t, ok, "Expected timeLimitAppenderV2 but got %T", wrapped)
require.Equal(t, baseAppender, tl.AppenderV2, "Expected base AppenderV2 but got %T", tl.AppenderV2)
sampleLimit := 100
sl, _ = newTestScrapeLoop(t, func(sl *scrapeLoop) {
sl.appendableV2 = appendable
sl.sampleLimit = sampleLimit
})
wrapped = appenderV2WithLimits(sl.appendableV2.AppenderV2(context.Background()), sampleLimit, 0, histogram.ExponentialSchemaMax)
la, ok := wrapped.(*limitAppenderV2)
require.True(t, ok, "Expected limitAppenderV2 but got %T", wrapped)
tl, ok = la.AppenderV2.(*timeLimitAppenderV2)
require.True(t, ok, "Expected timeLimitAppenderV2 but got %T", la.AppenderV2)
require.Equal(t, baseAppender, tl.AppenderV2, "Expected base AppenderV2 but got %T", tl.AppenderV2)
wrapped = appenderV2WithLimits(sl.appendableV2.AppenderV2(context.Background()), sampleLimit, 100, histogram.ExponentialSchemaMax)
bl, ok := wrapped.(*bucketLimitAppenderV2)
require.True(t, ok, "Expected bucketLimitAppenderV2 but got %T", wrapped)
la, ok = bl.AppenderV2.(*limitAppenderV2)
require.True(t, ok, "Expected limitAppenderV2 but got %T", bl)
tl, ok = la.AppenderV2.(*timeLimitAppenderV2)
require.True(t, ok, "Expected timeLimitAppenderV2 but got %T", la.AppenderV2)
require.Equal(t, baseAppender, tl.AppenderV2, "Expected base AppenderV2 but got %T", tl.AppenderV2)
wrapped = appenderV2WithLimits(sl.appendableV2.AppenderV2(context.Background()), sampleLimit, 100, 0)
ml, ok := wrapped.(*maxSchemaAppenderV2)
require.True(t, ok, "Expected maxSchemaAppenderV2 but got %T", wrapped)
bl, ok = ml.AppenderV2.(*bucketLimitAppenderV2)
require.True(t, ok, "Expected bucketLimitAppenderV2 but got %T", wrapped)
la, ok = bl.AppenderV2.(*limitAppenderV2)
require.True(t, ok, "Expected limitAppenderV2 but got %T", bl)
tl, ok = la.AppenderV2.(*timeLimitAppenderV2)
require.True(t, ok, "Expected timeLimitAppenderV2 but got %T", la.AppenderV2)
require.Equal(t, baseAppender, tl.AppenderV2, "Expected base AppenderV2 but got %T", tl.AppenderV2)
}
func TestScrapePoolRaces(t *testing.T) {
foreachAppendable(t, func(t *testing.T, appV2 bool) {
testScrapePoolRaces(t, appV2)
})
}
func testScrapePoolRaces(t *testing.T, appV2 bool) {
t.Parallel()
interval, _ := model.ParseDuration("1s")
timeout, _ := model.ParseDuration("500ms")
newConfig := func() *config.ScrapeConfig {
return &config.ScrapeConfig{
ScrapeInterval: interval,
ScrapeTimeout: timeout,
MetricNameValidationScheme: model.UTF8Validation,
MetricNameEscapingScheme: model.AllowUTF8,
}
}
sa := selectAppendable(teststorage.NewAppendable(), appV2)
sp, _ := newScrapePool(newConfig(), sa.V1(), sa.V2(), 0, nil, nil, &Options{}, newTestScrapeMetrics(t))
tgts := []*targetgroup.Group{
{
Targets: []model.LabelSet{
{model.AddressLabel: "127.0.0.1:9090"},
{model.AddressLabel: "127.0.0.2:9090"},
{model.AddressLabel: "127.0.0.3:9090"},
{model.AddressLabel: "127.0.0.4:9090"},
{model.AddressLabel: "127.0.0.5:9090"},
{model.AddressLabel: "127.0.0.6:9090"},
{model.AddressLabel: "127.0.0.7:9090"},
{model.AddressLabel: "127.0.0.8:9090"},
},
},
}
sp.Sync(tgts)
active := sp.ActiveTargets()
dropped := sp.DroppedTargets()
expectedActive, expectedDropped := len(tgts[0].Targets), 0
require.Len(t, active, expectedActive, "Invalid number of active targets")
require.Len(t, dropped, expectedDropped, "Invalid number of dropped targets")
for range 20 {
time.Sleep(10 * time.Millisecond)
_ = sp.reload(newConfig())
}
sp.stop()
}
func TestScrapePoolScrapeLoopsStarted(t *testing.T) {
foreachAppendable(t, func(t *testing.T, appV2 bool) {
testScrapePoolScrapeLoopsStarted(t, appV2)
})
}
func testScrapePoolScrapeLoopsStarted(t *testing.T, appV2 bool) {
var wg sync.WaitGroup
newLoop := func(scrapeLoopOptions) loop {
wg.Add(1)
l := &testLoop{
startFunc: func(_, _ time.Duration, _ chan<- error) {
wg.Done()
},
stopFunc: func() {},
}
return l
}
sp := newTestScrapePool(t, teststorage.NewAppendable(), appV2, newLoop)
tgs := []*targetgroup.Group{
{
Targets: []model.LabelSet{
{model.AddressLabel: model.LabelValue("127.0.0.1:9090")},
},
},
{
Targets: []model.LabelSet{
{model.AddressLabel: model.LabelValue("127.0.0.1:9090")},
},
},
}
require.NoError(t, sp.reload(&config.ScrapeConfig{
ScrapeInterval: model.Duration(3 * time.Second),
ScrapeTimeout: model.Duration(2 * time.Second),
MetricNameValidationScheme: model.UTF8Validation,
MetricNameEscapingScheme: model.AllowUTF8,
}))
sp.Sync(tgs)
require.Len(t, sp.loops, 1)
wg.Wait()
for _, l := range sp.loops {
require.True(t, l.(*testLoop).runOnce, "loop should be running")
}
}
func TestScrapeLoopStopBeforeRun(t *testing.T) {
t.Parallel()
sl, scraper := newTestScrapeLoop(t)
// The scrape pool synchronizes on stopping scrape loops. However, new scrape
// loops are started asynchronously. Thus, it's possible, that a loop is stopped
// again before having started properly.
// Stopping not-yet-started loops must block until the run method was called and exited.
// The run method must exit immediately.
stopDone := make(chan struct{})
go func() {
sl.stop()
close(stopDone)
}()
select {
case <-stopDone:
require.FailNow(t, "Stopping terminated before run exited successfully.")
case <-time.After(500 * time.Millisecond):
}
// Running the scrape loop must exit before calling the scraper even once.
scraper.scrapeFunc = func(context.Context, io.Writer) error {
require.FailNow(t, "Scraper was called for terminated scrape loop.")
return nil
}
runDone := make(chan struct{})
go func() {
sl.run(nil)
close(runDone)
}()
select {
case <-runDone:
case <-time.After(1 * time.Second):
require.FailNow(t, "Running terminated scrape loop did not exit.")
}
select {
case <-stopDone:
case <-time.After(1 * time.Second):
require.FailNow(t, "Stopping did not terminate after running exited.")
}
}
func nopMutator(l labels.Labels) labels.Labels { return l }
func TestScrapeLoopStop(t *testing.T) {
foreachAppendable(t, func(t *testing.T, appV2 bool) {
testScrapeLoopStop(t, appV2)
})
}
func testScrapeLoopStop(t *testing.T, appV2 bool) {
signal := make(chan struct{}, 1)
appTest := teststorage.NewAppendable()
sl, scraper := newTestScrapeLoop(t, withAppendable(appTest, appV2), func(sl *scrapeLoop) {
// Since we're writing samples directly below we need to provide a protocol fallback.
sl.fallbackScrapeProtocol = "text/plain"
})
// Terminate loop after 2 scrapes.
numScrapes := 0
scraper.scrapeFunc = func(ctx context.Context, w io.Writer) error {
numScrapes++
if numScrapes == 2 {
go sl.stop()
<-sl.ctx.Done()
}
_, _ = w.Write([]byte("metric_a 42\n"))
return ctx.Err()
}
go func() {
sl.run(nil)
signal <- struct{}{}
}()
select {
case <-signal:
case <-time.After(5 * time.Second):
require.FailNow(t, "Scrape wasn't stopped.")
}
got := appTest.ResultSamples()
// We expected 1 actual sample for each scrape plus 5 for report samples.
// At least 2 scrapes were made, plus the final stale markers.
require.GreaterOrEqual(t, len(got), 6*3, "Expected at least 3 scrapes with 6 samples each.")
require.Zero(t, len(got)%6, "There is a scrape with missing samples.")
// All samples in a scrape must have the same timestamp.
var ts int64
for i, s := range got {
switch {
case i%6 == 0:
ts = s.T
case s.T != ts:
t.Fatalf("Unexpected multiple timestamps within single scrape")
}
}
// All samples from the last scrape must be stale markers.
for _, s := range got[len(got)-5:] {
require.True(t, value.IsStaleNaN(s.V), "Appended last sample not as expected. Wanted: stale NaN Got: %x", math.Float64bits(s.V))
}
}
func TestScrapeLoopRun(t *testing.T) {
foreachAppendable(t, func(t *testing.T, appV2 bool) {
testScrapeLoopRun(t, appV2)
})
}
func testScrapeLoopRun(t *testing.T, appV2 bool) {
t.Parallel()
var (
signal = make(chan struct{}, 1)
errc = make(chan error)
)
ctx, cancel := context.WithCancel(t.Context())
sl, scraper := newTestScrapeLoop(t, withCtx(ctx), withAppendable(teststorage.NewAppendable(), appV2))
// The loop must terminate during the initial offset if the context
// is canceled.
scraper.offsetDur = time.Hour
go func() {
sl.run(errc)
signal <- struct{}{}
}()
// Wait to make sure we are actually waiting on the offset.
time.Sleep(1 * time.Second)
cancel()
select {
case <-signal:
case <-time.After(5 * time.Second):
require.FailNow(t, "Cancellation during initial offset failed.")
case err := <-errc:
require.FailNow(t, "Unexpected error", "err: %s", err)
}
ctx, cancel = context.WithCancel(t.Context())
sl, scraper = newTestScrapeLoop(t, withAppendable(teststorage.NewAppendable(), appV2), func(sl *scrapeLoop) {
sl.ctx = ctx
sl.timeout = 100 * time.Millisecond
})
// The provided timeout must cause cancellation of the context passed down to the
// scraper. The scraper has to respect the context.
scraper.offsetDur = 0
blockCtx, blockCancel := context.WithCancel(t.Context())
scraper.scrapeFunc = func(ctx context.Context, _ io.Writer) error {
select {
case <-blockCtx.Done():
cancel()
case <-ctx.Done():
return ctx.Err()
}
return nil
}
go func() {
sl.run(errc)
signal <- struct{}{}
}()
select {
case err := <-errc:
require.ErrorIs(t, err, context.DeadlineExceeded)
case <-time.After(3 * time.Second):
require.FailNow(t, "Expected timeout error but got none.")
}
// We already caught the timeout error and are certainly in the loop.
// Let the scrapes returns immediately to cause no further timeout errors
// and check whether canceling the parent context terminates the loop.
blockCancel()
select {
case <-signal:
// Loop terminated as expected.
case err := <-errc:
require.FailNow(t, "Unexpected error", "err: %s", err)
case <-time.After(3 * time.Second):
require.FailNow(t, "Loop did not terminate on context cancellation")
}
}
func TestScrapeLoopForcedErr(t *testing.T) {
foreachAppendable(t, func(t *testing.T, appV2 bool) {
testScrapeLoopForcedErr(t, appV2)
})
}
func testScrapeLoopForcedErr(t *testing.T, appV2 bool) {
var (
signal = make(chan struct{}, 1)
errc = make(chan error)
)
ctx, cancel := context.WithCancel(t.Context())
sl, scraper := newTestScrapeLoop(t, withCtx(ctx), withAppendable(teststorage.NewAppendable(), appV2))
forcedErr := errors.New("forced err")
sl.setForcedError(forcedErr)
scraper.scrapeFunc = func(context.Context, io.Writer) error {
require.FailNow(t, "Should not be scraped.")
return nil
}
go func() {
sl.run(errc)
signal <- struct{}{}
}()
select {
case err := <-errc:
require.ErrorIs(t, err, forcedErr)
case <-time.After(3 * time.Second):
require.FailNow(t, "Expected forced error but got none.")
}
cancel()
select {
case <-signal:
case <-time.After(5 * time.Second):
require.FailNow(t, "Scrape not stopped.")
}
}
func TestScrapeLoopRun_ContextCancelTerminatesBlockedSend(t *testing.T) {
foreachAppendable(t, func(t *testing.T, appV2 bool) {
testScrapeLoopRunContextCancelTerminatesBlockedSend(t, appV2)
})
}
func testScrapeLoopRunContextCancelTerminatesBlockedSend(t *testing.T, appV2 bool) {
// Regression test for issue #17553
defer goleak.VerifyNone(t)
var (
signal = make(chan struct{})
errc = make(chan error)
)
ctx, cancel := context.WithCancel(t.Context())
sl, scraper := newTestScrapeLoop(t, withCtx(ctx), withAppendable(teststorage.NewAppendable(), appV2))
forcedErr := errors.New("forced err")
sl.setForcedError(forcedErr)
scraper.scrapeFunc = func(context.Context, io.Writer) error {
return nil
}
go func() {
sl.run(errc)
close(signal)
}()
time.Sleep(50 * time.Millisecond)
cancel()
select {
case <-signal:
// success case
case <-time.After(3 * time.Second):
require.FailNow(t, "Scrape loop failed to exit on context cancellation (goroutine leak detected)")
}
}
func TestScrapeLoopMetadata(t *testing.T) {
foreachAppendable(t, func(t *testing.T, appV2 bool) {
testScrapeLoopMetadata(t, appV2)
})
}
func testScrapeLoopMetadata(t *testing.T, appV2 bool) {
sl, _ := newTestScrapeLoop(t, withAppendable(teststorage.NewAppendable(), appV2))
app := sl.appender()
total, _, _, err := app.append([]byte(`# TYPE test_metric counter
# HELP test_metric some help text
# UNIT test_metric metric
test_metric_total 1
# TYPE test_metric_no_help gauge
# HELP test_metric_no_type other help text
# EOF`), "application/openmetrics-text", time.Now())
require.NoError(t, err)
require.NoError(t, app.Commit())
require.Equal(t, 1, total)
md, ok := sl.cache.GetMetadata("test_metric")
require.True(t, ok, "expected metadata to be present")
require.Equal(t, model.MetricTypeCounter, md.Type, "unexpected metric type")
require.Equal(t, "some help text", md.Help)
require.Equal(t, "metric", md.Unit)
md, ok = sl.cache.GetMetadata("test_metric_no_help")
require.True(t, ok, "expected metadata to be present")
require.Equal(t, model.MetricTypeGauge, md.Type, "unexpected metric type")
require.Empty(t, md.Help)
require.Empty(t, md.Unit)
md, ok = sl.cache.GetMetadata("test_metric_no_type")
require.True(t, ok, "expected metadata to be present")
require.Equal(t, model.MetricTypeUnknown, md.Type, "unexpected metric type")
require.Equal(t, "other help text", md.Help)
require.Empty(t, md.Unit)
}
func TestScrapeLoopSeriesAdded(t *testing.T) {
foreachAppendable(t, func(t *testing.T, appV2 bool) {
testScrapeLoopSeriesAdded(t, appV2)
})
}
func testScrapeLoopSeriesAdded(t *testing.T, appV2 bool) {
sl, _ := newTestScrapeLoop(t, withAppendable(teststorage.NewAppendable(), appV2))
app := sl.appender()
total, added, seriesAdded, err := app.append([]byte("test_metric 1\n"), "text/plain", time.Time{})
require.NoError(t, err)
require.NoError(t, app.Commit())
require.Equal(t, 1, total)
require.Equal(t, 1, added)
require.Equal(t, 1, seriesAdded)
app = sl.appender()
total, added, seriesAdded, err = app.append([]byte("test_metric 1\n"), "text/plain", time.Time{})
require.NoError(t, app.Commit())
require.NoError(t, err)
require.Equal(t, 1, total)
require.Equal(t, 1, added)
require.Equal(t, 0, seriesAdded)
}
func TestScrapeLoopFailWithInvalidLabelsAfterRelabel(t *testing.T) {
foreachAppendable(t, func(t *testing.T, appV2 bool) {
testScrapeLoopFailWithInvalidLabelsAfterRelabel(t, appV2)
})
}
func testScrapeLoopFailWithInvalidLabelsAfterRelabel(t *testing.T, appV2 bool) {
target := &Target{
labels: labels.FromStrings("pod_label_invalid_012\xff", "test"),
}
relabelConfig := []*relabel.Config{{
Action: relabel.LabelMap,
Regex: relabel.MustNewRegexp("pod_label_invalid_(.+)"),
Separator: ";",
Replacement: "$1",
NameValidationScheme: model.UTF8Validation,
}}
sl, _ := newTestScrapeLoop(t, withAppendable(teststorage.NewAppendable(), appV2), func(sl *scrapeLoop) {
sl.sampleMutator = func(l labels.Labels) labels.Labels {
return mutateSampleLabels(l, target, true, relabelConfig)
}
})
app := sl.appender()
total, added, seriesAdded, err := app.append([]byte("test_metric 1\n"), "text/plain", time.Time{})
require.ErrorContains(t, err, "invalid metric name or label names")
require.NoError(t, app.Rollback())
require.Equal(t, 1, total)
require.Equal(t, 0, added)
require.Equal(t, 0, seriesAdded)
}
func TestScrapeLoopFailLegacyUnderUTF8(t *testing.T) {
foreachAppendable(t, func(t *testing.T, appV2 bool) {
testScrapeLoopFailLegacyUnderUTF8(t, appV2)
})
}
func testScrapeLoopFailLegacyUnderUTF8(t *testing.T, appV2 bool) {
sl, _ := newTestScrapeLoop(t, withAppendable(teststorage.NewAppendable(), appV2), func(sl *scrapeLoop) {
sl.validationScheme = model.LegacyValidation
})
app := sl.appender()
total, added, seriesAdded, err := app.append([]byte("{\"test.metric\"} 1\n"), "text/plain", time.Time{})
require.ErrorContains(t, err, "invalid metric name or label names")
require.NoError(t, app.Rollback())
require.Equal(t, 1, total)
require.Equal(t, 0, added)
require.Equal(t, 0, seriesAdded)
// When scrapeloop has validation set to UTF-8, the metric is allowed.
sl, _ = newTestScrapeLoop(t, withAppendable(teststorage.NewAppendable(), appV2), func(sl *scrapeLoop) {
sl.validationScheme = model.UTF8Validation
})
app = sl.appender()
total, added, seriesAdded, err = app.append([]byte("{\"test.metric\"} 1\n"), "text/plain", time.Time{})
require.NoError(t, err)
require.Equal(t, 1, total)
require.Equal(t, 1, added)
require.Equal(t, 1, seriesAdded)
}
func readTextParseTestMetrics(t testing.TB) []byte {
t.Helper()
b, err := os.ReadFile("../model/textparse/testdata/alltypes.237mfs.prom.txt")
if err != nil {
t.Fatal(err)
}
// Replace all Carriage Return chars that appear when testing on windows.
return bytes.ReplaceAll(b, []byte{'\r'}, nil)
}
func makeTestGauges(n int) []byte {
sb := bytes.Buffer{}
sb.WriteString("# TYPE metric_a gauge\n")
sb.WriteString("# HELP metric_a help text\n")
for i := range n {
_, _ = fmt.Fprintf(&sb, "metric_a{foo=\"%d\",bar=\"%d\"} 1\n", i, i*100)
}
sb.WriteString("# EOF\n")
return sb.Bytes()
}
func makeTestHistogramsWithExemplars(n int) []byte {
sb := bytes.Buffer{}
for i := range n {
sb.WriteString(strings.ReplaceAll(`# HELP rpc_durations_histogram%d_seconds RPC latency distributions.
# TYPE rpc_durations_histogram%d_seconds histogram
rpc_durations_histogram%d_seconds_bucket{le="-0.00099"} 0
rpc_durations_histogram%d_seconds_bucket{le="-0.00089"} 1 # {dummyID="1242"} -0.00091 1.7268398142239082e+09
rpc_durations_histogram%d_seconds_bucket{le="-0.0007899999999999999"} 1 # {dummyID="17783"} -0.0003825067330956884 1.7268398142239082e+09
rpc_durations_histogram%d_seconds_bucket{le="-0.0006899999999999999"} 2 # {dummyID="17783"} -0.0003825067330956884 1.7268398142239082e+09
rpc_durations_histogram%d_seconds_bucket{le="-0.0005899999999999998"} 3 # {dummyID="17783"} -0.0003825067330956884 1.7268398142239082e+09
rpc_durations_histogram%d_seconds_bucket{le="-0.0004899999999999998"} 4 # {dummyID="17783"} -0.0003825067330956884 1.7268398142239082e+09
rpc_durations_histogram%d_seconds_bucket{le="-0.0003899999999999998"} 5 # {dummyID="17783"} -0.0003825067330956884 1.7268398142239082e+09
rpc_durations_histogram%d_seconds_bucket{le="-0.0002899999999999998"} 6 # {dummyID="17783"} -0.0003825067330956884 1.7268398142239082e+09
rpc_durations_histogram%d_seconds_bucket{le="-0.0001899999999999998"} 7 # {dummyID="84741"} -0.00020178290006788965 1.726839814829977e+09
rpc_durations_histogram%d_seconds_bucket{le="-8.999999999999979e-05"} 7
rpc_durations_histogram%d_seconds_bucket{le="1.0000000000000216e-05"} 8 # {dummyID="19206"} -4.6156147425468016e-05 1.7268398151337721e+09
rpc_durations_histogram%d_seconds_bucket{le="0.00011000000000000022"} 9 # {dummyID="3974"} 9.528436760156754e-05 1.726839814526797e+09
rpc_durations_histogram%d_seconds_bucket{le="0.00021000000000000023"} 11 # {dummyID="29640"} 0.00017459624183458996 1.7268398139220061e+09
rpc_durations_histogram%d_seconds_bucket{le="0.0003100000000000002"} 15 # {dummyID="9818"} 0.0002791130914009552 1.7268398149821382e+09
rpc_durations_histogram%d_seconds_bucket{le="0.0004100000000000002"} 15
rpc_durations_histogram%d_seconds_bucket{le="0.0005100000000000003"} 15
rpc_durations_histogram%d_seconds_bucket{le="0.0006100000000000003"} 15
rpc_durations_histogram%d_seconds_bucket{le="0.0007100000000000003"} 15
rpc_durations_histogram%d_seconds_bucket{le="0.0008100000000000004"} 15
rpc_durations_histogram%d_seconds_bucket{le="0.0009100000000000004"} 15
rpc_durations_histogram%d_seconds_bucket{le="+Inf"} 15
rpc_durations_histogram%d_seconds_sum -8.452185437166741e-05
rpc_durations_histogram%d_seconds_count 15
rpc_durations_histogram%d_seconds_created 1.726839813016302e+09
`, "%d", strconv.Itoa(i)))
}
sb.WriteString("# EOF\n")
return sb.Bytes()
}
// promTextToProto converts Prometheus text to proto.
// Given expfmt decoding limitations, it does not support OpenMetrics fully (e.g. exemplars).
func promTextToProto(tb testing.TB, text []byte) []byte {
tb.Helper()
p := expfmt.NewTextParser(model.UTF8Validation)
fams, err := p.TextToMetricFamilies(bytes.NewReader(text))
if err != nil {
tb.Fatal("TextToMetricFamilies:", err)
}
// Order by name for the deterministic tests.
var names []string
for n := range fams {
names = append(names, n)
}
sort.Strings(names)
buf := bytes.Buffer{}
for _, n := range names {
o, err := proto.Marshal(fams[n])
if err != nil {
tb.Fatal(err)
}
// Write first length, then binary protobuf.
varintBuf := binary.AppendUvarint(nil, uint64(len(o)))
buf.Write(varintBuf)
buf.Write(o)
}
return buf.Bytes()
}
func TestPromTextToProto(t *testing.T) {
metricsText := readTextParseTestMetrics(t)
// On windows \r is added when reading, but parsers do not support this. Kill it.
metricsText = bytes.ReplaceAll(metricsText, []byte("\r"), nil)
metricsProto := promTextToProto(t, metricsText)
d := expfmt.NewDecoder(bytes.NewReader(metricsProto), expfmt.NewFormat(expfmt.TypeProtoDelim))
var got []string
for {
mf := &dto.MetricFamily{}
if err := d.Decode(mf); err != nil {
if errors.Is(err, io.EOF) {
break
}
t.Fatal(err)
}
got = append(got, mf.GetName())
}
require.Len(t, got, 237)
// Check a few to see if those are not dups.
require.Equal(t, "go_gc_cycles_automatic_gc_cycles_total", got[0])
require.Equal(t, "prometheus_sd_kuma_fetch_duration_seconds", got[128])
require.Equal(t, "promhttp_metric_handler_requests_total", got[236])
}
// TestScrapeLoopAppend_WithStorage tests appends and storage integration for the
// large input files that are also used in benchmarks.
func TestScrapeLoopAppend_WithStorage(t *testing.T) {
ts := time.Now()
for _, appV2 := range []bool{false, true} {
for _, tc := range []struct {
name string
parsableText []byte
expectedSamplesLen int
testAppendedSamples func(t *testing.T, committed []sample)
testExemplars func(t *testing.T, er []exemplar.QueryResult)
}{
{
name: "1Fam2000Gauges",
parsableText: makeTestGauges(2000),
expectedSamplesLen: 2000,
testAppendedSamples: func(t *testing.T, committed []sample) {
var expectedMF string
if appV2 {
expectedMF = "metric_a" // Only AppenderV2 supports metric family passing.
}
// Verify a few samples.
testutil.RequireEqual(t, sample{
MF: expectedMF,
M: metadata.Metadata{Type: model.MetricTypeGauge, Help: "help text"},
L: labels.FromStrings(model.MetricNameLabel, "metric_a", "foo", "0", "bar", "0"), V: 1, T: timestamp.FromTime(ts),
}, committed[0])
testutil.RequireEqual(t, sample{
MF: expectedMF,
M: metadata.Metadata{Type: model.MetricTypeGauge, Help: "help text"},
L: labels.FromStrings(model.MetricNameLabel, "metric_a", "foo", "1245", "bar", "124500"), V: 1, T: timestamp.FromTime(ts),
}, committed[1245])
testutil.RequireEqual(t, sample{
MF: expectedMF,
M: metadata.Metadata{Type: model.MetricTypeGauge, Help: "help text"},
L: labels.FromStrings(model.MetricNameLabel, "metric_a", "foo", "1999", "bar", "199900"), V: 1, T: timestamp.FromTime(ts),
}, committed[len(committed)-1])
},
},
{
name: "237FamsAllTypes",
parsableText: readTextParseTestMetrics(t),
expectedSamplesLen: 1857,
testAppendedSamples: func(t *testing.T, committed []sample) {
// Verify a few samples.
testutil.RequireEqual(t, sample{
MF: func() string {
if !appV2 {
return ""
}
return "go_gc_gomemlimit_bytes"
}(),
M: metadata.Metadata{Type: model.MetricTypeGauge, Help: "Go runtime memory limit configured by the user, otherwise math.MaxInt64. This value is set by the GOMEMLIMIT environment variable, and the runtime/debug.SetMemoryLimit function. Sourced from /gc/gomemlimit:bytes"},
L: labels.FromStrings(model.MetricNameLabel, "go_gc_gomemlimit_bytes"), V: 9.03676723e+08, T: timestamp.FromTime(ts),
}, committed[11])
testutil.RequireEqual(t, sample{
MF: func() string {
if !appV2 {
return "" // Only AppenderV2 supports metric family passing.
}
return "prometheus_http_request_duration_seconds"
}(),
M: metadata.Metadata{Type: model.MetricTypeHistogram, Help: "Histogram of latencies for HTTP requests."},
L: labels.FromStrings(model.MetricNameLabel, "prometheus_http_request_duration_seconds_bucket", "handler", "/api/v1/query_range", "le", "120.0"), V: 118157, T: timestamp.FromTime(ts),
}, committed[448])
testutil.RequireEqual(t, sample{
MF: func() string {
if !appV2 {
return "" // Only AppenderV2 supports metric family passing.
}
return "promhttp_metric_handler_requests_total"
}(),
M: metadata.Metadata{Type: model.MetricTypeCounter, Help: "Total number of scrapes by HTTP status code."},
L: labels.FromStrings(model.MetricNameLabel, "promhttp_metric_handler_requests_total", "code", "503"), V: 0, T: timestamp.FromTime(ts),
}, committed[len(committed)-1])
},
},
{
name: "100HistsWithExemplars",
parsableText: makeTestHistogramsWithExemplars(100),
expectedSamplesLen: 24 * 100,
testAppendedSamples: func(t *testing.T, committed []sample) {
// Verify a few samples.
m := metadata.Metadata{Type: model.MetricTypeHistogram, Help: "RPC latency distributions."}
testutil.RequireEqual(t, sample{
MF: func() string {
if !appV2 {
return "" // Only AppenderV2 supports metric family passing.
}
return "rpc_durations_histogram0_seconds"
}(),
M: m, L: labels.FromStrings(model.MetricNameLabel, "rpc_durations_histogram0_seconds_bucket", "le", "0.0003100000000000002"), V: 15, T: timestamp.FromTime(ts),
ES: []exemplar.Exemplar{
{Labels: labels.FromStrings("dummyID", "9818"), Value: 0.0002791130914009552, Ts: 1726839814982, HasTs: true},
},
}, committed[13])
testutil.RequireEqual(t, sample{
MF: func() string {
if !appV2 {
return "" // Only AppenderV2 supports metric family passing.
}
return "rpc_durations_histogram49_seconds"
}(),
M: m, L: labels.FromStrings(model.MetricNameLabel, "rpc_durations_histogram49_seconds_sum"), V: -8.452185437166741e-05, T: timestamp.FromTime(ts),
}, committed[24*50-3])
// This series does not have metadata, nor metric family, because of isSeriesPartOfFamily bug and OpenMetric 1.0 limitations around _created series.
// TODO(bwplotka): Fix with https://github.com/prometheus/prometheus/issues/17900
testutil.RequireEqual(t, sample{
L: labels.FromStrings(model.MetricNameLabel, "rpc_durations_histogram99_seconds_created"), V: 1.726839813016302e+09, T: timestamp.FromTime(ts),
}, committed[len(committed)-1])
},
testExemplars: func(t *testing.T, er []exemplar.QueryResult) {
// 12 out of 24 histogram series have exemplars.
require.Len(t, er, 12*100)
testutil.RequireEqual(t, exemplar.QueryResult{
SeriesLabels: labels.FromStrings(model.MetricNameLabel, "rpc_durations_histogram0_seconds_bucket", "le", "0.0003100000000000002"),
Exemplars: []exemplar.Exemplar{
{Labels: labels.FromStrings("dummyID", "9818"), Value: 0.0002791130914009552, Ts: 1726839814982, HasTs: true},
},
}, er[10])
testutil.RequireEqual(t, exemplar.QueryResult{
SeriesLabels: labels.FromStrings(model.MetricNameLabel, "rpc_durations_histogram9_seconds_bucket", "le", "1.0000000000000216e-05"),
Exemplars: []exemplar.Exemplar{
{Labels: labels.FromStrings("dummyID", "19206"), Value: -4.6156147425468016e-05, Ts: 1726839815133, HasTs: true},
},
}, er[len(er)-1])
},
},
} {
t.Run(fmt.Sprintf("appV2=%v/data=%v", appV2, tc.name), func(t *testing.T) {
s := teststorage.New(t, func(opt *tsdb.Options) {
opt.EnableMetadataWALRecords = true
})
appTest := teststorage.NewAppendable().Then(s)
sl, _ := newTestScrapeLoop(t, withAppendable(appTest, appV2))
app := sl.appender()
_, _, _, err := app.append(tc.parsableText, "application/openmetrics-text", ts)
require.NoError(t, err)
require.NoError(t, app.Commit())
// Check the recorded samples on the Appender layer.
require.Nil(t, appTest.PendingSamples())
require.Nil(t, appTest.RolledbackSamples())
got := appTest.ResultSamples()
require.Len(t, got, tc.expectedSamplesLen)
tc.testAppendedSamples(t, got)
// Check basic storage stats.
stats := s.Head().Stats(model.MetricNameLabel, 2000)
require.Equal(t, tc.expectedSamplesLen, int(stats.NumSeries))
// Check exemplars.
eq, err := s.ExemplarQuerier(t.Context())
require.NoError(t, err)
er, err := eq.Select(math.MinInt64, math.MaxInt64, nil)
require.NoError(t, err)
if tc.testExemplars != nil {
tc.testExemplars(t, er)
} else {
// Expect no exemplars.
require.Empty(t, er, "%v is not empty", er)
}
})
}
}
}
// BenchmarkScrapeLoopAppend benchmarks scrape appends for typical cases.
//
// Benchmark compares append function run across 5 dimensions:
// * `withStorage`: without storage isolates the benchmark to the scrape loop append code. With storage is an
// integration benchmark with the TSDB head appender code. For acceptance criteria run with storage, without for debugging.
// * `appV2`: appender V1 or V2.
// * `appendMetadataToWAL`: metadata-wal-records feature enabled or not (problematic feature we might need to change
// soon, see https://github.com/prometheus/prometheus/issues/15911.
// * `data`: different sizes of metrics scraped e.g. one big gauge metric family
// with a thousand series and more realistic scenario with common types.
// * `fmt`: different scrape formats which will benchmark different parsers e.g.
// promtext, omtext and promproto.
//
// NOTE: withStorage=true uses sync.Pool buffers which is heavily non-deterministic and shared across go routines.
// As a result, it's recommended to run dimensions you want to compare with in e.g. separate go tool invocations.
// Recommended CLI invocation(s):
/*
# Acceptance: With storage with V1 and V2 in separate process:
export bench=appendV1 && go test ./scrape/... \
-run '^$' -bench '^BenchmarkScrapeLoopAppend/withStorage=true/appV2=false/$' \
-benchtime 2s -count 6 -cpu 2 -timeout 999m \
| tee ${bench}.txt
export bench=appendV2 && go test ./scrape/... \
-run '^$' -bench '^BenchmarkScrapeLoopAppend/withStorage=true/appV2=true/$' \
-benchtime 2s -count 6 -cpu 2 -timeout 999m \
| tee ${bench}.txt
# For debugging scrape overheads:
export bench=appendNoStorage && go test ./scrape/... \
-run '^$' -bench '^BenchmarkScrapeLoopAppend/withStorage=false/$' \
-benchtime 2s -count 6 -cpu 2 -timeout 999m \
| tee ${bench}.txt
*/
func BenchmarkScrapeLoopAppend(b *testing.B) {
for _, withStorage := range []bool{false, true} {
for _, appV2 := range []bool{false, true} {
for _, appendMetadataToWAL := range []bool{false, true} {
for _, data := range []struct {
name string
parsableText []byte
}{
{name: "1Fam2000Gauges", parsableText: makeTestGauges(2000)}, // ~68.1 KB, ~77.9 KB in proto.
{name: "237FamsAllTypes", parsableText: readTextParseTestMetrics(b)}, // ~185.7 KB, ~70.6 KB in proto.
} {
b.Run(fmt.Sprintf("withStorage=%v/appV2=%v/appendMetadataToWAL=%v/data=%v", withStorage, appV2, appendMetadataToWAL, data.name), func(b *testing.B) {
metricsProto := promTextToProto(b, data.parsableText)
for _, bcase := range []struct {
name string
contentType string
parsable []byte
}{
{name: "PromText", contentType: "text/plain", parsable: data.parsableText},
{name: "OMText", contentType: "application/openmetrics-text", parsable: data.parsableText},
{name: "PromProto", contentType: "application/vnd.google.protobuf", parsable: metricsProto},
} {
b.Run(fmt.Sprintf("fmt=%v", bcase.name), func(b *testing.B) {
benchScrapeLoopAppend(b, withStorage, appV2, bcase.parsable, bcase.contentType, appendMetadataToWAL, false)
})
}
})
}
}
}
}
}
func benchScrapeLoopAppend(
b *testing.B,
withStorage bool,
appV2 bool,
parsable []byte,
contentType string,
appendMetadataToWAL bool,
enableExemplarStorage bool,
) {
var a compatAppendable = teststorage.NewAppendable().SkipRecording(true) // Make it noop for benchmark purposes.
if withStorage {
a = teststorage.New(b, func(opt *tsdb.Options) {
opt.EnableMetadataWALRecords = appendMetadataToWAL
if enableExemplarStorage {
opt.EnableExemplarStorage = true
opt.MaxExemplars = 1e5
}
})
}
sl, _ := newTestScrapeLoop(b, withAppendable(a, appV2), func(sl *scrapeLoop) {
sl.appendMetadataToWAL = appendMetadataToWAL
})
ts := time.Time{}
b.ReportAllocs()
b.ResetTimer()
for b.Loop() {
app := sl.appender()
ts = ts.Add(time.Second)
_, _, _, err := app.append(parsable, contentType, ts)
if err != nil {
b.Fatal(err)
}
// Reset the appender so it doesn't grow indefinitely, and it mimics what prod scrape will do.
// We do rollback, because it's cheaper than Commit.
if err := app.Rollback(); err != nil {
b.Fatal(err)
}
}
}
// BenchmarkScrapeLoopAppend_HistogramsWithExemplars benchmarks OM scrapes with histograms full of exemplars.
//
// For e2e TSDB impact, we enable the TSDB exemplar storage
//
// Recommended CLI invocation:
/*
export bench=appendHistWithExemplars && go test ./scrape/... \
-run '^$' -bench '^BenchmarkScrapeLoopAppend_HistogramsWithExemplars' \
-benchtime 5s -count 6 -cpu 2 -timeout 999m \
| tee ${bench}.txt
*/
func BenchmarkScrapeLoopAppend_HistogramsWithExemplars(b *testing.B) {
for _, appV2 := range []bool{false, true} {
b.Run(fmt.Sprintf("appV2=%v", appV2), func(b *testing.B) {
parsable := makeTestHistogramsWithExemplars(100) // ~255.8 KB in OM text.
benchScrapeLoopAppend(b, true, appV2, parsable, "application/openmetrics-text", false, true)
})
}
}
func TestScrapeLoopScrapeAndReport(t *testing.T) {
foreachAppendable(t, func(t *testing.T, appV2 bool) {
testScrapeLoopScrapeAndReport(t, appV2)
})
}
func testScrapeLoopScrapeAndReport(t *testing.T, appV2 bool) {
parsableText := readTextParseTestMetrics(t)
// On windows \r is added when reading, but parsers do not support this. Kill it.
parsableText = bytes.ReplaceAll(parsableText, []byte("\r"), nil)
appTest := teststorage.NewAppendable()
sl, scraper := newTestScrapeLoop(t, withAppendable(appTest, appV2), func(sl *scrapeLoop) {
sl.fallbackScrapeProtocol = "application/openmetrics-text"
})
scraper.scrapeFunc = func(_ context.Context, writer io.Writer) error {
_, err := writer.Write(parsableText)
return err
}
ts := time.Time{}
sl.scrapeAndReport(time.Time{}, ts, nil)
require.NoError(t, scraper.lastError)
require.Len(t, appTest.ResultSamples(), 1862)
require.Len(t, appTest.ResultMetadata(), 1862)
}
// Recommended CLI invocation:
/*
export bench=scrapeAndReport && go test ./scrape/... \
-run '^$' -bench '^BenchmarkScrapeLoopScrapeAndReport$' \
-benchtime 5s -count 6 -cpu 2 -timeout 999m \
| tee ${bench}.txt
*/
func BenchmarkScrapeLoopScrapeAndReport(b *testing.B) {
for _, appV2 := range []bool{false, true} {
b.Run(fmt.Sprintf("appV2=%v", appV2), func(b *testing.B) {
parsableText := readTextParseTestMetrics(b)
s := teststorage.New(b)
sl, scraper := newTestScrapeLoop(b, withAppendable(s, appV2), func(sl *scrapeLoop) {
sl.fallbackScrapeProtocol = "application/openmetrics-text"
})
scraper.scrapeFunc = func(_ context.Context, writer io.Writer) error {
_, err := writer.Write(parsableText)
return err
}
ts := time.Time{}
b.ReportAllocs()
b.ResetTimer()
for b.Loop() {
ts = ts.Add(time.Second)
sl.scrapeAndReport(time.Time{}, ts, nil)
require.NoError(b, scraper.lastError)
}
})
}
}
func TestSetOptionsHandlingStaleness(t *testing.T) {
foreachAppendable(t, func(t *testing.T, appV2 bool) {
testSetOptionsHandlingStaleness(t, appV2)
})
}
func testSetOptionsHandlingStaleness(t *testing.T, appV2 bool) {
s := teststorage.New(t, func(opt *tsdb.Options) {
opt.OutOfOrderTimeWindow = 600000
})
signal := make(chan struct{}, 1)
ctx, cancel := context.WithCancel(t.Context())
defer cancel()
// Function to run the scrape loop
runScrapeLoop := func(ctx context.Context, t *testing.T, cue int, action func(*scrapeLoop)) {
sl, scraper := newTestScrapeLoop(t, withAppendable(s, appV2), func(sl *scrapeLoop) {
sl.ctx = ctx
})
numScrapes := 0
scraper.scrapeFunc = func(_ context.Context, w io.Writer) error {
numScrapes++
if numScrapes == cue {
action(sl)
}
_, _ = fmt.Fprintf(w, "metric_a{a=\"1\",b=\"1\"} %d\n", 42+numScrapes)
return nil
}
sl.run(nil)
}
go func() {
runScrapeLoop(ctx, t, 2, func(sl *scrapeLoop) {
go sl.stop()
// Wait a bit then start a new target.
time.Sleep(100 * time.Millisecond)
go func() {
runScrapeLoop(ctx, t, 4, func(*scrapeLoop) {
cancel()
})
signal <- struct{}{}
}()
})
}()
select {
case <-signal:
case <-time.After(10 * time.Second):
t.Fatalf("Scrape wasn't stopped.")
}
ctx1, cancel := context.WithCancel(t.Context())
defer cancel()
q, err := s.Querier(0, time.Now().UnixNano())
require.NoError(t, err)
t.Cleanup(func() { _ = q.Close() })
series := q.Select(ctx1, false, nil, labels.MustNewMatcher(labels.MatchRegexp, "__name__", "metric_a"))
var results []sample
for series.Next() {
it := series.At().Iterator(nil)
for it.Next() == chunkenc.ValFloat {
t, v := it.At()
results = append(results, sample{
L: series.At().Labels(),
T: t,
V: v,
})
}
require.NoError(t, it.Err())
}
require.NoError(t, series.Err())
var c int
for _, s := range results {
if value.IsStaleNaN(s.V) {
c++
}
}
require.Zero(t, c, "invalid count of staleness markers after stopping the engine")
}
func TestScrapeLoopRunCreatesStaleMarkersOnFailedScrape(t *testing.T) {
foreachAppendable(t, func(t *testing.T, appV2 bool) {
testScrapeLoopRunCreatesStaleMarkersOnFailedScrape(t, appV2)
})
}
func testScrapeLoopRunCreatesStaleMarkersOnFailedScrape(t *testing.T, appV2 bool) {
signal := make(chan struct{}, 1)
ctx, cancel := context.WithCancel(t.Context())
appTest := teststorage.NewAppendable()
sl, scraper := newTestScrapeLoop(t, withAppendable(appTest, appV2), func(sl *scrapeLoop) {
sl.ctx = ctx
// Since we're writing samples directly below we need to provide a protocol fallback.
sl.fallbackScrapeProtocol = "text/plain"
})
// Succeed once, several failures, then stop.
numScrapes := 0
scraper.scrapeFunc = func(_ context.Context, w io.Writer) error {
numScrapes++
switch numScrapes {
case 1:
_, _ = w.Write([]byte("metric_a 42\n"))
return nil
case 5:
cancel()
}
return errors.New("scrape failed")
}
go func() {
sl.run(nil)
signal <- struct{}{}
}()
select {
case <-signal:
case <-time.After(5 * time.Second):
require.FailNow(t, "Scrape wasn't stopped.")
}
got := appTest.ResultSamples()
// 1 successfully scraped sample
// 1 stale marker after first fail
// 5x 5 report samples for each scrape successful or not.
require.Len(t, got, 27, "Appended samples not as expected:\n%s", appTest)
require.Equal(t, 42.0, got[0].V, "Appended first sample not as expected")
require.True(t, value.IsStaleNaN(got[6].V),
"Appended second sample not as expected. Wanted: stale NaN Got: %x", math.Float64bits(got[6].V))
}
func TestScrapeLoopRunCreatesStaleMarkersOnParseFailure(t *testing.T) {
foreachAppendable(t, func(t *testing.T, appV2 bool) {
testScrapeLoopRunCreatesStaleMarkersOnParseFailure(t, appV2)
})
}
func testScrapeLoopRunCreatesStaleMarkersOnParseFailure(t *testing.T, appV2 bool) {
signal := make(chan struct{}, 1)
ctx, cancel := context.WithCancel(t.Context())
appTest := teststorage.NewAppendable()
sl, scraper := newTestScrapeLoop(t, withAppendable(appTest, appV2), func(sl *scrapeLoop) {
sl.ctx = ctx
// Since we're writing samples directly below we need to provide a protocol fallback.
sl.fallbackScrapeProtocol = "text/plain"
})
// Succeed once, several failures, then stop.
numScrapes := 0
scraper.scrapeFunc = func(_ context.Context, w io.Writer) error {
numScrapes++
switch numScrapes {
case 1:
_, _ = w.Write([]byte("metric_a 42\n"))
return nil
case 2:
_, _ = w.Write([]byte("7&-\n"))
return nil
case 3:
cancel()
}
return errors.New("scrape failed")
}
go func() {
sl.run(nil)
signal <- struct{}{}
}()
select {
case <-signal:
// TODO(bwplotka): Prone to flakiness, depend on atomic numScrapes.
case <-time.After(5 * time.Second):
require.FailNow(t, "Scrape wasn't stopped.")
}
got := appTest.ResultSamples()
// 1 successfully scraped sample
// 1 stale marker after first fail
// 3x 5 report samples for each scrape successful or not.
require.Len(t, got, 17, "Appended samples not as expected:\n%s", appTest)
require.Equal(t, 42.0, got[0].V, "Appended first sample not as expected")
require.True(t, value.IsStaleNaN(got[6].V),
"Appended second sample not as expected. Wanted: stale NaN Got: %x", math.Float64bits(got[6].V))
}
// If we have a target with sample_limit set and scrape initially works, but then we hit the sample_limit error,
// then we don't expect to see any StaleNaNs appended for the series that disappeared due to sample_limit error.
func TestScrapeLoopRunCreatesStaleMarkersOnSampleLimit(t *testing.T) {
foreachAppendable(t, func(t *testing.T, appV2 bool) {
testScrapeLoopRunCreatesStaleMarkersOnSampleLimit(t, appV2)
})
}
func testScrapeLoopRunCreatesStaleMarkersOnSampleLimit(t *testing.T, appV2 bool) {
signal := make(chan struct{}, 1)
ctx, cancel := context.WithCancel(t.Context())
appTest := teststorage.NewAppendable()
sl, scraper := newTestScrapeLoop(t, withAppendable(appTest, appV2), func(sl *scrapeLoop) {
sl.ctx = ctx
// Since we're writing samples directly below we need to provide a protocol fallback.
sl.fallbackScrapeProtocol = "text/plain"
sl.sampleLimit = 4
})
// Succeed once, several failures, then stop.
numScrapes := 0
scraper.scrapeFunc = func(_ context.Context, w io.Writer) error {
numScrapes++
switch numScrapes {
case 1:
_, _ = w.Write([]byte("metric_a 10\nmetric_b 10\nmetric_c 10\nmetric_d 10\n"))
return nil
case 2:
_, _ = w.Write([]byte("metric_a 20\nmetric_b 20\nmetric_c 20\nmetric_d 20\nmetric_e 999\n"))
return nil
case 3:
_, _ = w.Write([]byte("metric_a 30\nmetric_b 30\nmetric_c 30\nmetric_d 30\n"))
return nil
case 4:
cancel()
}
return errors.New("scrape failed")
}
go func() {
sl.run(nil)
signal <- struct{}{}
}()
select {
case <-signal:
case <-time.After(5 * time.Second):
require.FailNow(t, "Scrape wasn't stopped.")
}
got := appTest.ResultSamples()
// 4 scrapes in total:
// #1 - success - 4 samples appended + 5 report series
// #2 - sample_limit exceeded - no samples appended, only 5 report series
// #3 - success - 4 samples appended + 5 report series
// #4 - scrape canceled - 4 StaleNaNs appended because of scrape error + 5 report series
require.Len(t, got, (4+5)+5+(4+5)+(4+5), "Appended samples not as expected:\n%s", appTest)
// Expect first 4 samples to be metric_X [0-3].
for i := range 4 {
require.Equal(t, 10.0, got[i].V, "Appended %d sample not as expected", i)
}
// Next 5 samples are report series [4-8].
// Next 5 samples are report series for the second scrape [9-13].
// Expect first 4 samples to be metric_X from the third scrape [14-17].
for i := 14; i <= 17; i++ {
require.Equal(t, 30.0, got[i].V, "Appended %d sample not as expected", i)
}
// Next 5 samples are report series [18-22].
// Next 5 samples are report series [23-26].
for i := 23; i <= 26; i++ {
require.True(t, value.IsStaleNaN(got[i].V),
"Appended second sample not as expected. Wanted: stale NaN Got: %x", math.Float64bits(got[i].V))
}
}
func TestScrapeLoopCache(t *testing.T) {
foreachAppendable(t, func(t *testing.T, appV2 bool) {
testScrapeLoopCache(t, appV2)
})
}
func testScrapeLoopCache(t *testing.T, appV2 bool) {
s := teststorage.New(t)
signal := make(chan struct{}, 1)
ctx, cancel := context.WithCancel(t.Context())
appTest := teststorage.NewAppendable().Then(s)
sl, scraper := newTestScrapeLoop(t, withAppendable(appTest, appV2), func(sl *scrapeLoop) {
sl.ctx = ctx
sl.l = promslog.New(&promslog.Config{})
// Since we're writing samples directly below we need to provide a protocol fallback.
sl.fallbackScrapeProtocol = "text/plain"
// Decreasing the scrape interval could make the test fail, as multiple scrapes might be initiated at identical millisecond timestamps.
// See https://github.com/prometheus/prometheus/issues/12727.
sl.interval = 100 * time.Millisecond
})
numScrapes := 0
scraper.scrapeFunc = func(_ context.Context, w io.Writer) error {
switch numScrapes {
case 1, 2:
_, ok := sl.cache.series["metric_a"]
require.True(t, ok, "metric_a missing from cache after scrape %d", numScrapes)
_, ok = sl.cache.series["metric_b"]
require.True(t, ok, "metric_b missing from cache after scrape %d", numScrapes)
case 3:
_, ok := sl.cache.series["metric_a"]
require.True(t, ok, "metric_a missing from cache after scrape %d", numScrapes)
_, ok = sl.cache.series["metric_b"]
require.False(t, ok, "metric_b present in cache after scrape %d", numScrapes)
}
numScrapes++
switch numScrapes {
case 1:
_, _ = w.Write([]byte("metric_a 42\nmetric_b 43\n"))
return nil
case 3:
_, _ = w.Write([]byte("metric_a 44\n"))
return nil
case 4:
cancel()
}
return errors.New("scrape failed")
}
go func() {
sl.run(nil)
signal <- struct{}{}
}()
select {
case <-signal:
case <-time.After(5 * time.Second):
require.FailNow(t, "Scrape wasn't stopped.")
}
// 3 successfully scraped samples
// 3 stale marker after samples were missing.
// 4x 5 report samples for each scrape successful or not.
require.Len(t, appTest.ResultSamples(), 26, "Appended samples not as expected:\n%s", appTest)
}
func TestScrapeLoopCacheMemoryExhaustionProtection(t *testing.T) {
foreachAppendable(t, func(t *testing.T, appV2 bool) {
testScrapeLoopCacheMemoryExhaustionProtection(t, appV2)
})
}
func testScrapeLoopCacheMemoryExhaustionProtection(t *testing.T, appV2 bool) {
s := teststorage.New(t)
signal := make(chan struct{}, 1)
ctx, cancel := context.WithCancel(t.Context())
sl, scraper := newTestScrapeLoop(t, withAppendable(teststorage.NewAppendable().Then(s), appV2), func(sl *scrapeLoop) {
sl.ctx = ctx
})
numScrapes := 0
scraper.scrapeFunc = func(_ context.Context, w io.Writer) error {
numScrapes++
if numScrapes < 5 {
s := ""
for i := range 500 {
s = fmt.Sprintf("%smetric_%d_%d 42\n", s, i, numScrapes)
}
_, _ = w.Write([]byte(s + "&"))
} else {
cancel()
}
return nil
}
go func() {
sl.run(nil)
signal <- struct{}{}
}()
select {
case <-signal:
case <-time.After(5 * time.Second):
require.FailNow(t, "Scrape wasn't stopped.")
}
require.LessOrEqual(t, len(sl.cache.series), 2000, "More than 2000 series cached.")
}
func TestScrapeLoopAppend_HonorLabels(t *testing.T) {
foreachAppendable(t, func(t *testing.T, appV2 bool) {
testScrapeLoopAppendHonorLabels(t, appV2)
})
}
func testScrapeLoopAppendHonorLabels(t *testing.T, appV2 bool) {
for _, test := range []struct {
title string
honorLabels bool
scrapeText string
discoveryLabels []string
expLset labels.Labels
}{
{
// On label collision, when "honor_labels" is not set, prefix is added.
title: "HonorLabels=false",
scrapeText: `metric{n="1"} 1`,
discoveryLabels: []string{"n", "2"},
expLset: labels.FromStrings("__name__", "metric", "exported_n", "1", "n", "2"),
},
{
// Case where SD already has the prefixed label - it shouldn't be overridden.
title: "HonorLabels=false;exported prefix already exists in SD",
scrapeText: `metric{n="1"} 1`,
discoveryLabels: []string{"n", "2", "exported_n", "2"},
expLset: labels.FromStrings("__name__", "metric", "n", "2", "exported_n", "2", "exported_exported_n", "1"),
},
{
// Labels with no value need to be removed as these should not be ingested.
title: "HonorLabels=false;empty label",
scrapeText: `metric{n=""} 1`,
discoveryLabels: nil,
expLset: labels.FromStrings("__name__", "metric"),
},
{
// On label collision, when "honor_labels" is true, label is overridden.
title: "HonorLabels=true",
honorLabels: true,
scrapeText: `metric{n="1"} 1`,
discoveryLabels: []string{"n", "2"},
expLset: labels.FromStrings("__name__", "metric", "n", "1"),
},
} {
t.Run(test.title, func(t *testing.T) {
discoveryLabels := &Target{
labels: labels.FromStrings(test.discoveryLabels...),
}
appTest := teststorage.NewAppendable()
sl, _ := newTestScrapeLoop(t, withAppendable(appTest, appV2), func(sl *scrapeLoop) {
sl.sampleMutator = func(l labels.Labels) labels.Labels {
return mutateSampleLabels(l, discoveryLabels, test.honorLabels, nil)
}
sl.reportSampleMutator = func(l labels.Labels) labels.Labels {
return mutateReportSampleLabels(l, discoveryLabels)
}
})
now := time.Now()
app := sl.appender()
_, _, _, err := app.append([]byte(test.scrapeText), "text/plain", now)
require.NoError(t, err)
require.NoError(t, app.Commit())
expected := []sample{
{
L: test.expLset,
T: timestamp.FromTime(now),
V: 1,
},
}
teststorage.RequireEqual(t, expected, appTest.ResultSamples())
})
}
}
func TestScrapeLoopAppendForConflictingPrefixedLabels(t *testing.T) {
foreachAppendable(t, func(t *testing.T, appV2 bool) {
testScrapeLoopAppendForConflictingPrefixedLabels(t, appV2)
})
}
func testScrapeLoopAppendForConflictingPrefixedLabels(t *testing.T, appV2 bool) {
for _, tc := range []struct {
name string
targetLabels []string
exposedLabels string
expected []string
}{
{
name: "One target label collides with existing label",
targetLabels: []string{"foo", "2"},
exposedLabels: `metric{foo="1"} 0`,
expected: []string{"__name__", "metric", "exported_foo", "1", "foo", "2"},
},
{
name: "One target label collides with existing label, plus target label already with prefix 'exported'",
targetLabels: []string{"foo", "2", "exported_foo", "3"},
exposedLabels: `metric{foo="1"} 0`,
expected: []string{"__name__", "metric", "exported_exported_foo", "1", "exported_foo", "3", "foo", "2"},
},
{
name: "One target label collides with existing label, plus existing label already with prefix 'exported",
targetLabels: []string{"foo", "3"},
exposedLabels: `metric{foo="1", exported_foo="2"} 0`,
expected: []string{"__name__", "metric", "exported_exported_foo", "1", "exported_foo", "2", "foo", "3"},
},
{
name: "One target label collides with existing label, both already with prefix 'exported'",
targetLabels: []string{"exported_foo", "2"},
exposedLabels: `metric{exported_foo="1"} 0`,
expected: []string{"__name__", "metric", "exported_exported_foo", "1", "exported_foo", "2"},
},
{
name: "Two target labels collide with existing labels, both with and without prefix 'exported'",
targetLabels: []string{"foo", "3", "exported_foo", "4"},
exposedLabels: `metric{foo="1", exported_foo="2"} 0`,
expected: []string{
"__name__", "metric", "exported_exported_foo", "1", "exported_exported_exported_foo",
"2", "exported_foo", "4", "foo", "3",
},
},
{
name: "Extreme example",
targetLabels: []string{"foo", "0", "exported_exported_foo", "1", "exported_exported_exported_foo", "2"},
exposedLabels: `metric{foo="3", exported_foo="4", exported_exported_exported_foo="5"} 0`,
expected: []string{
"__name__", "metric",
"exported_exported_exported_exported_exported_foo", "5",
"exported_exported_exported_exported_foo", "3",
"exported_exported_exported_foo", "2",
"exported_exported_foo", "1",
"exported_foo", "4",
"foo", "0",
},
},
} {
t.Run(tc.name, func(t *testing.T) {
appTest := teststorage.NewAppendable()
sl, _ := newTestScrapeLoop(t, withAppendable(appTest, appV2), func(sl *scrapeLoop) {
sl.sampleMutator = func(l labels.Labels) labels.Labels {
return mutateSampleLabels(l, &Target{labels: labels.FromStrings(tc.targetLabels...)}, false, nil)
}
})
app := sl.appender()
_, _, _, err := app.append([]byte(tc.exposedLabels), "text/plain", time.Date(2000, 1, 1, 1, 0, 0, 0, time.UTC))
require.NoError(t, err)
require.NoError(t, app.Commit())
teststorage.RequireEqual(t, []sample{
{
L: labels.FromStrings(tc.expected...),
T: timestamp.FromTime(time.Date(2000, 1, 1, 1, 0, 0, 0, time.UTC)),
V: 0,
},
}, appTest.ResultSamples())
})
}
}
func TestScrapeLoopAppendCacheEntryButErrNotFound(t *testing.T) {
foreachAppendable(t, func(t *testing.T, appV2 bool) {
testScrapeLoopAppendCacheEntryButErrNotFound(t, appV2)
})
}
func testScrapeLoopAppendCacheEntryButErrNotFound(t *testing.T, appV2 bool) {
appTest := teststorage.NewAppendable()
sl, _ := newTestScrapeLoop(t, withAppendable(appTest, appV2))
fakeRef := storage.SeriesRef(1)
expValue := float64(1)
metric := []byte(`metric{n="1"} 1`)
p, warning := textparse.New(metric, "text/plain", labels.NewSymbolTable(), textparse.ParserOptions{})
require.NotNil(t, p)
require.NoError(t, warning)
var lset labels.Labels
_, err := p.Next()
require.NoError(t, err)
p.Labels(&lset)
hash := lset.Hash()
// Create a fake entry in the cache
sl.cache.addRef(metric, fakeRef, lset, hash)
now := time.Now()
app := sl.appender()
_, _, _, err = app.append(metric, "text/plain", now)
require.NoError(t, err)
require.NoError(t, app.Commit())
expected := []sample{
{
L: lset,
T: timestamp.FromTime(now),
V: expValue,
},
}
teststorage.RequireEqual(t, expected, appTest.ResultSamples())
}
type appendableFunc func(ctx context.Context) storage.Appender
func (a appendableFunc) Appender(ctx context.Context) storage.Appender { return a(ctx) }
func TestScrapeLoopAppendSampleLimit(t *testing.T) {
foreachAppendable(t, func(t *testing.T, appV2 bool) {
testScrapeLoopAppendSampleLimit(t, appV2)
})
}
func testScrapeLoopAppendSampleLimit(t *testing.T, appV2 bool) {
appTest := teststorage.NewAppendable()
sl, _ := newTestScrapeLoop(t, func(sl *scrapeLoop) {
if appV2 {
sl.appendableV2 = appendableV2Func(func(ctx context.Context) storage.AppenderV2 {
// Chain appTest to verify what samples passed through.
return &limitAppenderV2{AppenderV2: appTest.AppenderV2(ctx), limit: 1}
})
} else {
sl.appendable = appendableFunc(func(ctx context.Context) storage.Appender {
// Chain appTest to verify what samples passed through.
return &limitAppender{Appender: appTest.Appender(ctx), limit: 1}
})
}
sl.sampleMutator = func(l labels.Labels) labels.Labels {
if l.Has("deleteme") {
return labels.EmptyLabels()
}
return l
}
sl.sampleLimit = 1 // Same as limitAppender.limit
})
// Get the value of the Counter before performing append.
beforeMetric := dto.Metric{}
err := sl.metrics.targetScrapeSampleLimit.Write(&beforeMetric)
require.NoError(t, err)
beforeMetricValue := beforeMetric.GetCounter().GetValue()
now := time.Now()
app := sl.appender()
total, added, seriesAdded, err := app.append([]byte("metric_a 1\nmetric_b 1\nmetric_c 1\n"), "text/plain", now)
require.ErrorIs(t, err, errSampleLimit)
require.NoError(t, app.Rollback())
require.Equal(t, 3, total)
require.Equal(t, 3, added)
require.Equal(t, 1, seriesAdded)
// Check that the Counter has been incremented a single time for the scrape,
// not multiple times for each sample.
metric := dto.Metric{}
err = sl.metrics.targetScrapeSampleLimit.Write(&metric)
require.NoError(t, err)
v := metric.GetCounter().GetValue()
change := v - beforeMetricValue
require.Equal(t, 1.0, change, "Unexpected change of sample limit metric: %f", change)
// And verify that we got the samples that fit under the limit.
want := []sample{
{
L: labels.FromStrings(model.MetricNameLabel, "metric_a"),
T: timestamp.FromTime(now),
V: 1,
},
}
teststorage.RequireEqual(t, want, appTest.RolledbackSamples(), "Appended samples not as expected:\n%s", appTest)
now = time.Now()
app = sl.appender()
total, added, seriesAdded, err = app.append([]byte("metric_a 1\nmetric_b 1\nmetric_c{deleteme=\"yes\"} 1\nmetric_d 1\nmetric_e 1\nmetric_f 1\nmetric_g 1\nmetric_h{deleteme=\"yes\"} 1\nmetric_i{deleteme=\"yes\"} 1\n"), "text/plain", now)
require.ErrorIs(t, err, errSampleLimit)
require.NoError(t, app.Rollback())
require.Equal(t, 9, total)
require.Equal(t, 6, added)
require.Equal(t, 1, seriesAdded)
}
func TestScrapeLoop_HistogramBucketLimit(t *testing.T) {
foreachAppendable(t, func(t *testing.T, appV2 bool) {
testScrapeLoopHistogramBucketLimit(t, appV2)
})
}
func testScrapeLoopHistogramBucketLimit(t *testing.T, appV2 bool) {
sl, _ := newTestScrapeLoop(t, func(sl *scrapeLoop) {
if appV2 {
sl.appendableV2 = appendableV2Func(func(ctx context.Context) storage.AppenderV2 {
return &bucketLimitAppenderV2{AppenderV2: teststorage.NewAppendable().AppenderV2(ctx), limit: 2}
})
} else {
sl.appendable = appendableFunc(func(ctx context.Context) storage.Appender {
return &bucketLimitAppender{Appender: teststorage.NewAppendable().Appender(ctx), limit: 2}
})
}
sl.enableNativeHistogramScraping = true
sl.sampleMutator = func(l labels.Labels) labels.Labels {
if l.Has("deleteme") {
return labels.EmptyLabels()
}
return l
}
})
app := sl.appender()
metric := dto.Metric{}
err := sl.metrics.targetScrapeNativeHistogramBucketLimit.Write(&metric)
require.NoError(t, err)
beforeMetricValue := metric.GetCounter().GetValue()
nativeHistogram := prometheus.NewHistogramVec(
prometheus.HistogramOpts{
Namespace: "testing",
Name: "example_native_histogram",
Help: "This is used for testing",
ConstLabels: map[string]string{"some": "value"},
NativeHistogramBucketFactor: 1.1, // 10% increase from bucket to bucket
NativeHistogramMaxBucketNumber: 100, // intentionally higher than the limit we'll use in the scraper
},
[]string{"size"},
)
registry := prometheus.NewRegistry()
require.NoError(t, registry.Register(nativeHistogram))
nativeHistogram.WithLabelValues("S").Observe(1.0)
nativeHistogram.WithLabelValues("M").Observe(1.0)
nativeHistogram.WithLabelValues("L").Observe(1.0)
nativeHistogram.WithLabelValues("M").Observe(10.0)
nativeHistogram.WithLabelValues("L").Observe(10.0) // in different bucket since > 1*1.1
gathered, err := registry.Gather()
require.NoError(t, err)
require.NotEmpty(t, gathered)
histogramMetricFamily := gathered[0]
msg, err := MetricFamilyToProtobuf(histogramMetricFamily)
require.NoError(t, err)
now := time.Now()
total, added, seriesAdded, err := app.append(msg, "application/vnd.google.protobuf", now)
require.NoError(t, err)
require.Equal(t, 3, total)
require.Equal(t, 3, added)
require.Equal(t, 3, seriesAdded)
err = sl.metrics.targetScrapeNativeHistogramBucketLimit.Write(&metric)
require.NoError(t, err)
metricValue := metric.GetCounter().GetValue()
require.Equal(t, beforeMetricValue, metricValue)
beforeMetricValue = metricValue
nativeHistogram.WithLabelValues("L").Observe(100.0) // in different bucket since > 10*1.1
gathered, err = registry.Gather()
require.NoError(t, err)
require.NotEmpty(t, gathered)
histogramMetricFamily = gathered[0]
msg, err = MetricFamilyToProtobuf(histogramMetricFamily)
require.NoError(t, err)
now = time.Now()
total, added, seriesAdded, err = app.append(msg, "application/vnd.google.protobuf", now)
require.NoError(t, err)
require.Equal(t, 3, total)
require.Equal(t, 3, added)
require.Equal(t, 0, seriesAdded) // Series are cached.
err = sl.metrics.targetScrapeNativeHistogramBucketLimit.Write(&metric)
require.NoError(t, err)
metricValue = metric.GetCounter().GetValue()
require.Equal(t, beforeMetricValue, metricValue)
beforeMetricValue = metricValue
nativeHistogram.WithLabelValues("L").Observe(100000.0) // in different bucket since > 10*1.1
gathered, err = registry.Gather()
require.NoError(t, err)
require.NotEmpty(t, gathered)
histogramMetricFamily = gathered[0]
msg, err = MetricFamilyToProtobuf(histogramMetricFamily)
require.NoError(t, err)
now = time.Now()
total, added, seriesAdded, err = app.append(msg, "application/vnd.google.protobuf", now)
if !errors.Is(err, errBucketLimit) {
t.Fatalf("Did not see expected histogram bucket limit error: %s", err)
}
require.NoError(t, app.Rollback())
require.Equal(t, 3, total)
require.Equal(t, 3, added)
require.Equal(t, 0, seriesAdded) // Series are cached.
err = sl.metrics.targetScrapeNativeHistogramBucketLimit.Write(&metric)
require.NoError(t, err)
metricValue = metric.GetCounter().GetValue()
require.Equal(t, beforeMetricValue+1, metricValue)
}
func TestScrapeLoop_ChangingMetricString(t *testing.T) {
foreachAppendable(t, func(t *testing.T, appV2 bool) {
testScrapeLoopChangingMetricString(t, appV2)
})
}
func testScrapeLoopChangingMetricString(t *testing.T, appV2 bool) {
// This is a regression test for the scrape loop cache not properly maintaining
// IDs when the string representation of a metric changes across a scrape. Thus,
// we use a real storage appender here.
appTest := teststorage.NewAppendable()
sl, _ := newTestScrapeLoop(t, withAppendable(appTest, appV2))
now := time.Now()
app := sl.appender()
_, _, _, err := app.append([]byte(`metric_a{a="1",b="1"} 1`), "text/plain", now)
require.NoError(t, err)
require.NoError(t, app.Commit())
app = sl.appender()
_, _, _, err = app.append([]byte(`metric_a{b="1",a="1"} 2`), "text/plain", now.Add(time.Minute))
require.NoError(t, err)
require.NoError(t, app.Commit())
want := []sample{
{
L: labels.FromStrings("__name__", "metric_a", "a", "1", "b", "1"),
T: timestamp.FromTime(now),
V: 1,
},
{
L: labels.FromStrings("__name__", "metric_a", "a", "1", "b", "1"),
T: timestamp.FromTime(now.Add(time.Minute)),
V: 2,
},
}
teststorage.RequireEqual(t, want, appTest.ResultSamples(), "Appended samples not as expected:\n%s", appTest)
}
func TestScrapeLoopAppendFailsWithNoContentType(t *testing.T) {
foreachAppendable(t, func(t *testing.T, appV2 bool) {
testScrapeLoopAppendFailsWithNoContentType(t, appV2)
})
}
func testScrapeLoopAppendFailsWithNoContentType(t *testing.T, appV2 bool) {
sl, _ := newTestScrapeLoop(t, withAppendable(teststorage.NewAppendable(), appV2), func(sl *scrapeLoop) {
// Explicitly setting the lack of fallback protocol here to make it obvious.
sl.fallbackScrapeProtocol = ""
})
now := time.Now()
app := sl.appender()
_, _, _, err := app.append([]byte("metric_a 1\n"), "", now)
// We expected the appropriate error.
require.ErrorContains(t, err, "non-compliant scrape target sending blank Content-Type and no fallback_scrape_protocol specified for target", "Expected \"non-compliant scrape\" error but got: %s", err)
}
// TestScrapeLoopAppendEmptyWithNoContentType ensures we there are no errors when we get a blank scrape or just want to append a stale marker.
func TestScrapeLoopAppendEmptyWithNoContentType(t *testing.T) {
foreachAppendable(t, func(t *testing.T, appV2 bool) {
testScrapeLoopAppendEmptyWithNoContentType(t, appV2)
})
}
func testScrapeLoopAppendEmptyWithNoContentType(t *testing.T, appV2 bool) {
sl, _ := newTestScrapeLoop(t, withAppendable(teststorage.NewAppendable(), appV2), func(sl *scrapeLoop) {
// Explicitly setting the lack of fallback protocol here to make it obvious.
sl.fallbackScrapeProtocol = ""
})
now := time.Now()
app := sl.appender()
_, _, _, err := app.append([]byte(""), "", now)
require.NoError(t, err)
require.NoError(t, app.Commit())
}
func TestScrapeLoopAppendStaleness(t *testing.T) {
foreachAppendable(t, func(t *testing.T, appV2 bool) {
testScrapeLoopAppendStaleness(t, appV2)
})
}
func testScrapeLoopAppendStaleness(t *testing.T, appV2 bool) {
appTest := teststorage.NewAppendable()
sl, _ := newTestScrapeLoop(t, withAppendable(appTest, appV2))
now := time.Now()
app := sl.appender()
_, _, _, err := app.append([]byte("metric_a 1\n"), "text/plain", now)
require.NoError(t, err)
require.NoError(t, app.Commit())
app = sl.appender()
_, _, _, err = app.append([]byte(""), "", now.Add(time.Second))
require.NoError(t, err)
require.NoError(t, app.Commit())
want := []sample{
{
L: labels.FromStrings(model.MetricNameLabel, "metric_a"),
T: timestamp.FromTime(now),
V: 1,
},
{
L: labels.FromStrings(model.MetricNameLabel, "metric_a"),
T: timestamp.FromTime(now.Add(time.Second)),
V: math.Float64frombits(value.StaleNaN),
},
}
teststorage.RequireEqual(t, want, appTest.ResultSamples(), "Appended samples not as expected:\n%s", appTest)
}
func TestScrapeLoopAppendNoStalenessIfTimestamp(t *testing.T) {
foreachAppendable(t, func(t *testing.T, appV2 bool) {
testScrapeLoopAppendNoStalenessIfTimestamp(t, appV2)
})
}
func testScrapeLoopAppendNoStalenessIfTimestamp(t *testing.T, appV2 bool) {
appTest := teststorage.NewAppendable()
sl, _ := newTestScrapeLoop(t, withAppendable(appTest, appV2))
now := time.Now()
app := sl.appender()
_, _, _, err := app.append([]byte("metric_a 1 1000\n"), "text/plain", now)
require.NoError(t, err)
require.NoError(t, app.Commit())
app = sl.appender()
_, _, _, err = app.append([]byte(""), "", now.Add(time.Second))
require.NoError(t, err)
require.NoError(t, app.Commit())
want := []sample{
{
L: labels.FromStrings(model.MetricNameLabel, "metric_a"),
T: 1000,
V: 1,
},
}
teststorage.RequireEqual(t, want, appTest.ResultSamples(), "Appended samples not as expected:\n%s", appTest)
}
func TestScrapeLoopAppendStalenessIfTrackTimestampStaleness(t *testing.T) {
foreachAppendable(t, func(t *testing.T, appV2 bool) {
testScrapeLoopAppendStalenessIfTrackTimestampStaleness(t, appV2)
})
}
func testScrapeLoopAppendStalenessIfTrackTimestampStaleness(t *testing.T, appV2 bool) {
appTest := teststorage.NewAppendable()
sl, _ := newTestScrapeLoop(t, withAppendable(appTest, appV2), func(sl *scrapeLoop) {
sl.trackTimestampsStaleness = true
})
now := time.Now()
app := sl.appender()
_, _, _, err := app.append([]byte("metric_a 1 1000\n"), "text/plain", now)
require.NoError(t, err)
require.NoError(t, app.Commit())
app = sl.appender()
_, _, _, err = app.append([]byte(""), "", now.Add(time.Second))
require.NoError(t, err)
require.NoError(t, app.Commit())
want := []sample{
{
L: labels.FromStrings(model.MetricNameLabel, "metric_a"),
T: 1000,
V: 1,
},
{
L: labels.FromStrings(model.MetricNameLabel, "metric_a"),
T: timestamp.FromTime(now.Add(time.Second)),
V: math.Float64frombits(value.StaleNaN),
},
}
teststorage.RequireEqual(t, want, appTest.ResultSamples(), "Appended samples not as expected:\n%s", appTest)
}
// TestScrapeLoopAppend is the main table test testing the scrape appends, including histograms, exemplar and metadata.
func TestScrapeLoopAppend(t *testing.T) {
foreachAppendable(t, func(t *testing.T, appV2 bool) {
testScrapeLoopAppend(t, appV2)
})
}
func testScrapeLoopAppend(t *testing.T, appV2 bool) {
for _, test := range []struct {
title string
alwaysScrapeClassicHist bool
enableNativeHistogramsIngestion bool
scrapeText string
contentType string
discoveryLabels []string
samples []sample
}{
{
title: "Normal NaN scraped",
scrapeText: "metric_total{n=\"1\"} NaN\n# EOF",
contentType: "application/openmetrics-text",
samples: []sample{{
L: labels.FromStrings("__name__", "metric_total", "n", "1"),
V: math.Float64frombits(value.NormalNaN),
}},
},
{
title: "Metric without exemplars",
scrapeText: "metric_total{n=\"1\"} 0\n# EOF",
contentType: "application/openmetrics-text",
discoveryLabels: []string{"n", "2"},
samples: []sample{{
L: labels.FromStrings("__name__", "metric_total", "exported_n", "1", "n", "2"),
V: 0,
}},
},
{
title: "Metric with exemplars",
scrapeText: "metric_total{n=\"1\"} 0 # {a=\"abc\"} 1.0\n# EOF",
contentType: "application/openmetrics-text",
discoveryLabels: []string{"n", "2"},
samples: []sample{{
L: labels.FromStrings("__name__", "metric_total", "exported_n", "1", "n", "2"),
V: 0,
ES: []exemplar.Exemplar{
{Labels: labels.FromStrings("a", "abc"), Value: 1},
},
}},
},
{
title: "Metric with exemplars and TS",
scrapeText: "metric_total{n=\"1\"} 0 # {a=\"abc\"} 1.0 10000\n# EOF",
contentType: "application/openmetrics-text",
discoveryLabels: []string{"n", "2"},
samples: []sample{{
L: labels.FromStrings("__name__", "metric_total", "exported_n", "1", "n", "2"),
V: 0,
ES: []exemplar.Exemplar{{Labels: labels.FromStrings("a", "abc"), Value: 1, Ts: 10000000, HasTs: true}},
}},
},
{
title: "Two metrics and exemplars",
scrapeText: `metric_total{n="1"} 1 # {t="1"} 1.0 10000
metric_total{n="2"} 2 # {t="2"} 2.0 20000
# EOF`,
contentType: "application/openmetrics-text",
samples: []sample{{
L: labels.FromStrings("__name__", "metric_total", "n", "1"),
V: 1,
ES: []exemplar.Exemplar{{Labels: labels.FromStrings("t", "1"), Value: 1, Ts: 10000000, HasTs: true}},
}, {
L: labels.FromStrings("__name__", "metric_total", "n", "2"),
V: 2,
ES: []exemplar.Exemplar{{Labels: labels.FromStrings("t", "2"), Value: 2, Ts: 20000000, HasTs: true}},
}},
},
{
title: "Native histogram with three exemplars from classic buckets",
enableNativeHistogramsIngestion: true,
scrapeText: `name: "test_histogram"
help: "Test histogram with many buckets removed to keep it manageable in size."
type: HISTOGRAM
metric: <
histogram: <
sample_count: 175
sample_sum: 0.0008280461746287094
bucket: <
cumulative_count: 2
upper_bound: -0.0004899999999999998
>
bucket: <
cumulative_count: 4
upper_bound: -0.0003899999999999998
exemplar: <
label: <
name: "dummyID"
value: "59727"
>
value: -0.00039
timestamp: <
seconds: 1625851155
nanos: 146848499
>
>
>
bucket: <
cumulative_count: 16
upper_bound: -0.0002899999999999998
exemplar: <
label: <
name: "dummyID"
value: "5617"
>
value: -0.00029
>
>
bucket: <
cumulative_count: 32
upper_bound: -0.0001899999999999998
exemplar: <
label: <
name: "dummyID"
value: "58215"
>
value: -0.00019
timestamp: <
seconds: 1625851055
nanos: 146848599
>
>
>
schema: 3
zero_threshold: 2.938735877055719e-39
zero_count: 2
negative_span: <
offset: -162
length: 1
>
negative_span: <
offset: 23
length: 4
>
negative_delta: 1
negative_delta: 3
negative_delta: -2
negative_delta: -1
negative_delta: 1
positive_span: <
offset: -161
length: 1
>
positive_span: <
offset: 8
length: 3
>
positive_delta: 1
positive_delta: 2
positive_delta: -1
positive_delta: -1
>
timestamp_ms: 1234568
>
`,
contentType: "application/vnd.google.protobuf",
samples: []sample{{
T: 1234568,
L: labels.FromStrings("__name__", "test_histogram"),
H: &histogram.Histogram{
Count: 175,
ZeroCount: 2,
Sum: 0.0008280461746287094,
ZeroThreshold: 2.938735877055719e-39,
Schema: 3,
PositiveSpans: []histogram.Span{
{Offset: -161, Length: 1},
{Offset: 8, Length: 3},
},
NegativeSpans: []histogram.Span{
{Offset: -162, Length: 1},
{Offset: 23, Length: 4},
},
PositiveBuckets: []int64{1, 2, -1, -1},
NegativeBuckets: []int64{1, 3, -2, -1, 1},
},
ES: []exemplar.Exemplar{
// Native histogram exemplars are arranged by timestamp, and those with missing timestamps are dropped.
{Labels: labels.FromStrings("dummyID", "58215"), Value: -0.00019, Ts: 1625851055146, HasTs: true},
{Labels: labels.FromStrings("dummyID", "59727"), Value: -0.00039, Ts: 1625851155146, HasTs: true},
},
}},
},
{
title: "Native histogram with three exemplars scraped as classic histogram",
enableNativeHistogramsIngestion: true,
scrapeText: `name: "test_histogram"
help: "Test histogram with many buckets removed to keep it manageable in size."
type: HISTOGRAM
metric: <
histogram: <
sample_count: 175
sample_sum: 0.0008280461746287094
bucket: <
cumulative_count: 2
upper_bound: -0.0004899999999999998
>
bucket: <
cumulative_count: 4
upper_bound: -0.0003899999999999998
exemplar: <
label: <
name: "dummyID"
value: "59727"
>
value: -0.00039
timestamp: <
seconds: 1625851155
nanos: 146848499
>
>
>
bucket: <
cumulative_count: 16
upper_bound: -0.0002899999999999998
exemplar: <
label: <
name: "dummyID"
value: "5617"
>
value: -0.00029
>
>
bucket: <
cumulative_count: 32
upper_bound: -0.0001899999999999998
exemplar: <
label: <
name: "dummyID"
value: "58215"
>
value: -0.00019
timestamp: <
seconds: 1625851055
nanos: 146848599
>
>
>
schema: 3
zero_threshold: 2.938735877055719e-39
zero_count: 2
negative_span: <
offset: -162
length: 1
>
negative_span: <
offset: 23
length: 4
>
negative_delta: 1
negative_delta: 3
negative_delta: -2
negative_delta: -1
negative_delta: 1
positive_span: <
offset: -161
length: 1
>
positive_span: <
offset: 8
length: 3
>
positive_delta: 1
positive_delta: 2
positive_delta: -1
positive_delta: -1
>
timestamp_ms: 1234568
>
`,
alwaysScrapeClassicHist: true,
contentType: "application/vnd.google.protobuf",
samples: []sample{
{
T: 1234568,
L: labels.FromStrings("__name__", "test_histogram"),
H: &histogram.Histogram{
Count: 175,
ZeroCount: 2,
Sum: 0.0008280461746287094,
ZeroThreshold: 2.938735877055719e-39,
Schema: 3,
PositiveSpans: []histogram.Span{
{Offset: -161, Length: 1},
{Offset: 8, Length: 3},
},
NegativeSpans: []histogram.Span{
{Offset: -162, Length: 1},
{Offset: 23, Length: 4},
},
PositiveBuckets: []int64{1, 2, -1, -1},
NegativeBuckets: []int64{1, 3, -2, -1, 1},
},
ES: []exemplar.Exemplar{
// Native histogram one is arranged by timestamp.
// Exemplars with missing timestamps are dropped for native histograms.
{Labels: labels.FromStrings("dummyID", "58215"), Value: -0.00019, Ts: 1625851055146, HasTs: true},
{Labels: labels.FromStrings("dummyID", "59727"), Value: -0.00039, Ts: 1625851155146, HasTs: true},
},
},
{L: labels.FromStrings("__name__", "test_histogram_count"), T: 1234568, V: 175},
{L: labels.FromStrings("__name__", "test_histogram_sum"), T: 1234568, V: 0.0008280461746287094},
{L: labels.FromStrings("__name__", "test_histogram_bucket", "le", "-0.0004899999999999998"), T: 1234568, V: 2},
{
L: labels.FromStrings("__name__", "test_histogram_bucket", "le", "-0.0003899999999999998"), T: 1234568, V: 4,
ES: []exemplar.Exemplar{{Labels: labels.FromStrings("dummyID", "59727"), Value: -0.00039, Ts: 1625851155146, HasTs: true}},
},
{
L: labels.FromStrings("__name__", "test_histogram_bucket", "le", "-0.0002899999999999998"), T: 1234568, V: 16,
ES: []exemplar.Exemplar{{Labels: labels.FromStrings("dummyID", "5617"), Value: -0.00029, Ts: 1234568, HasTs: false}},
},
{
L: labels.FromStrings("__name__", "test_histogram_bucket", "le", "-0.0001899999999999998"), T: 1234568, V: 32,
ES: []exemplar.Exemplar{{Labels: labels.FromStrings("dummyID", "58215"), Value: -0.00019, Ts: 1625851055146, HasTs: true}},
},
{L: labels.FromStrings("__name__", "test_histogram_bucket", "le", "+Inf"), T: 1234568, V: 175},
},
},
{
title: "Native histogram with exemplars and no classic buckets",
contentType: "application/vnd.google.protobuf",
enableNativeHistogramsIngestion: true,
scrapeText: `name: "test_histogram"
help: "Test histogram."
type: HISTOGRAM
metric: <
histogram: <
sample_count: 175
sample_sum: 0.0008280461746287094
schema: 3
zero_threshold: 2.938735877055719e-39
zero_count: 2
negative_span: <
offset: -162
length: 1
>
negative_span: <
offset: 23
length: 4
>
negative_delta: 1
negative_delta: 3
negative_delta: -2
negative_delta: -1
negative_delta: 1
positive_span: <
offset: -161
length: 1
>
positive_span: <
offset: 8
length: 3
>
positive_delta: 1
positive_delta: 2
positive_delta: -1
positive_delta: -1
exemplars: <
label: <
name: "dummyID"
value: "59732"
>
value: -0.00039
timestamp: <
seconds: 1625851155
nanos: 146848499
>
>
exemplars: <
label: <
name: "dummyID"
value: "58242"
>
value: -0.00019
timestamp: <
seconds: 1625851055
nanos: 146848599
>
>
exemplars: <
label: <
name: "dummyID"
value: "5617"
>
value: -0.00029
>
>
timestamp_ms: 1234568
>
`,
samples: []sample{{
T: 1234568,
L: labels.FromStrings("__name__", "test_histogram"),
H: &histogram.Histogram{
Count: 175,
ZeroCount: 2,
Sum: 0.0008280461746287094,
ZeroThreshold: 2.938735877055719e-39,
Schema: 3,
PositiveSpans: []histogram.Span{
{Offset: -161, Length: 1},
{Offset: 8, Length: 3},
},
NegativeSpans: []histogram.Span{
{Offset: -162, Length: 1},
{Offset: 23, Length: 4},
},
PositiveBuckets: []int64{1, 2, -1, -1},
NegativeBuckets: []int64{1, 3, -2, -1, 1},
},
ES: []exemplar.Exemplar{
// Exemplars with missing timestamps are dropped for native histograms.
{Labels: labels.FromStrings("dummyID", "58242"), Value: -0.00019, Ts: 1625851055146, HasTs: true},
{Labels: labels.FromStrings("dummyID", "59732"), Value: -0.00039, Ts: 1625851155146, HasTs: true},
},
}},
},
{
title: "Native histogram with exemplars but ingestion disabled",
contentType: "application/vnd.google.protobuf",
enableNativeHistogramsIngestion: false,
scrapeText: `name: "test_histogram"
help: "Test histogram."
type: HISTOGRAM
metric: <
histogram: <
sample_count: 175
sample_sum: 0.0008280461746287094
schema: 3
zero_threshold: 2.938735877055719e-39
zero_count: 2
negative_span: <
offset: -162
length: 1
>
negative_span: <
offset: 23
length: 4
>
negative_delta: 1
negative_delta: 3
negative_delta: -2
negative_delta: -1
negative_delta: 1
positive_span: <
offset: -161
length: 1
>
positive_span: <
offset: 8
length: 3
>
positive_delta: 1
positive_delta: 2
positive_delta: -1
positive_delta: -1
exemplars: <
label: <
name: "dummyID"
value: "59732"
>
value: -0.00039
timestamp: <
seconds: 1625851155
nanos: 146848499
>
>
exemplars: <
label: <
name: "dummyID"
value: "58242"
>
value: -0.00019
timestamp: <
seconds: 1625851055
nanos: 146848599
>
>
exemplars: <
label: <
name: "dummyID"
value: "5617"
>
value: -0.00029
>
>
timestamp_ms: 1234568
>
`,
samples: []sample{
{L: labels.FromStrings("__name__", "test_histogram_count"), T: 1234568, V: 175},
{L: labels.FromStrings("__name__", "test_histogram_sum"), T: 1234568, V: 0.0008280461746287094},
{L: labels.FromStrings("__name__", "test_histogram_bucket", "le", "+Inf"), T: 1234568, V: 175},
},
},
} {
t.Run(test.title, func(t *testing.T) {
discoveryLabels := &Target{
labels: labels.FromStrings(test.discoveryLabels...),
}
appTest := teststorage.NewAppendable()
sl, _ := newTestScrapeLoop(t, withAppendable(appTest, appV2), func(sl *scrapeLoop) {
sl.enableNativeHistogramScraping = test.enableNativeHistogramsIngestion
sl.sampleMutator = func(l labels.Labels) labels.Labels {
return mutateSampleLabels(l, discoveryLabels, false, nil)
}
sl.reportSampleMutator = func(l labels.Labels) labels.Labels {
return mutateReportSampleLabels(l, discoveryLabels)
}
sl.alwaysScrapeClassicHist = test.alwaysScrapeClassicHist
// This test does not care about metadata.
// Having this true would mean we need to add metadata to sample
// expectations.
// TODO(bwplotka): Add cases for append metadata to WAL and pass metadata
sl.appendMetadataToWAL = false
})
app := sl.appender()
now := time.Now()
// Process expected samples.
for i := range test.samples {
if !appV2 && test.samples[i].MF != "" {
// AppenderV1 does not support metric family passing.
test.samples[i].MF = ""
}
if test.samples[i].T != 0 {
continue
}
test.samples[i].T = timestamp.FromTime(now)
// We need to set the timestamp for expected exemplars that does not have a timestamp.
for j := range test.samples[i].ES {
if test.samples[i].ES[j].Ts == 0 {
test.samples[i].ES[j].Ts = timestamp.FromTime(now)
}
}
}
buf := &bytes.Buffer{}
if test.contentType == "application/vnd.google.protobuf" {
require.NoError(t, textToProto(test.scrapeText, buf))
} else {
buf.WriteString(test.scrapeText)
}
_, _, _, err := app.append(buf.Bytes(), test.contentType, now)
require.NoError(t, err)
require.NoError(t, app.Commit())
teststorage.RequireEqual(t, test.samples, appTest.ResultSamples())
})
}
}
func textToProto(text string, buf *bytes.Buffer) error {
// In case of protobuf, we have to create the binary representation.
pb := &dto.MetricFamily{}
// From text to proto message.
err := proto.UnmarshalText(text, pb)
if err != nil {
return err
}
// From proto message to binary protobuf.
protoBuf, err := proto.Marshal(pb)
if err != nil {
return err
}
// Write first length, then binary protobuf.
varintBuf := binary.AppendUvarint(nil, uint64(len(protoBuf)))
buf.Write(varintBuf)
buf.Write(protoBuf)
return nil
}
func TestScrapeLoopAppendExemplarSeries(t *testing.T) {
foreachAppendable(t, func(t *testing.T, appV2 bool) {
testScrapeLoopAppendExemplarSeries(t, appV2)
})
}
func testScrapeLoopAppendExemplarSeries(t *testing.T, appV2 bool) {
scrapeText := []string{`metric_total{n="1"} 1 # {t="1"} 1.0 10000
# EOF`, `metric_total{n="1"} 2 # {t="2"} 2.0 20000
# EOF`}
samples := []sample{{
L: labels.FromStrings("__name__", "metric_total", "n", "1"),
V: 1,
ES: []exemplar.Exemplar{
{Labels: labels.FromStrings("t", "1"), Value: 1, Ts: 10000000, HasTs: true},
},
}, {
L: labels.FromStrings("__name__", "metric_total", "n", "1"),
V: 2,
ES: []exemplar.Exemplar{
{Labels: labels.FromStrings("t", "2"), Value: 2, Ts: 20000000, HasTs: true},
},
}}
discoveryLabels := &Target{
labels: labels.FromStrings(),
}
appTest := teststorage.NewAppendable()
sl, _ := newTestScrapeLoop(t, withAppendable(appTest, appV2), func(sl *scrapeLoop) {
sl.sampleMutator = func(l labels.Labels) labels.Labels {
return mutateSampleLabels(l, discoveryLabels, false, nil)
}
sl.reportSampleMutator = func(l labels.Labels) labels.Labels {
return mutateReportSampleLabels(l, discoveryLabels)
}
// This test does not care about metadata. Having this true would mean we need to add metadata to sample
// expectations.
sl.appendMetadataToWAL = false
})
now := time.Now()
for i := range samples {
ts := now.Add(time.Second * time.Duration(i))
samples[i].T = timestamp.FromTime(ts)
}
for i, st := range scrapeText {
app := sl.appender()
_, _, _, err := app.append([]byte(st), "application/openmetrics-text", timestamp.Time(samples[i].T))
require.NoError(t, err)
require.NoError(t, app.Commit())
}
teststorage.RequireEqual(t, samples, appTest.ResultSamples())
}
func TestScrapeLoopRunReportsTargetDownOnScrapeError(t *testing.T) {
foreachAppendable(t, func(t *testing.T, appV2 bool) {
testScrapeLoopRunReportsTargetDownOnScrapeError(t, appV2)
})
}
func testScrapeLoopRunReportsTargetDownOnScrapeError(t *testing.T, appV2 bool) {
ctx, cancel := context.WithCancel(t.Context())
appTest := teststorage.NewAppendable()
sl, scraper := newTestScrapeLoop(t, withAppendable(appTest, appV2), func(sl *scrapeLoop) {
sl.ctx = ctx
})
scraper.scrapeFunc = func(context.Context, io.Writer) error {
cancel()
return errors.New("scrape failed")
}
sl.run(nil)
require.Equal(t, 0.0, appTest.ResultSamples()[0].V, "bad 'up' value")
}
func TestScrapeLoopRunReportsTargetDownOnInvalidUTF8(t *testing.T) {
foreachAppendable(t, func(t *testing.T, appV2 bool) {
testScrapeLoopRunReportsTargetDownOnInvalidUTF8(t, appV2)
})
}
func testScrapeLoopRunReportsTargetDownOnInvalidUTF8(t *testing.T, appV2 bool) {
ctx, cancel := context.WithCancel(t.Context())
appTest := teststorage.NewAppendable()
sl, scraper := newTestScrapeLoop(t, withAppendable(appTest, appV2), func(sl *scrapeLoop) {
sl.ctx = ctx
})
scraper.scrapeFunc = func(_ context.Context, w io.Writer) error {
cancel()
_, _ = w.Write([]byte("a{l=\"\xff\"} 1\n"))
return nil
}
sl.run(nil)
require.Equal(t, 0.0, appTest.ResultSamples()[0].V, "bad 'up' value")
}
func TestScrapeLoopAppendGracefullyIfAmendOrOutOfOrderOrOutOfBounds(t *testing.T) {
foreachAppendable(t, func(t *testing.T, appV2 bool) {
testScrapeLoopAppendGracefullyIfAmendOrOutOfOrderOrOutOfBounds(t, appV2)
})
}
func testScrapeLoopAppendGracefullyIfAmendOrOutOfOrderOrOutOfBounds(t *testing.T, appV2 bool) {
appTest := teststorage.NewAppendable().WithErrs(
func(ls labels.Labels) error {
switch ls.Get(model.MetricNameLabel) {
case "out_of_order":
return storage.ErrOutOfOrderSample
case "amend":
return storage.ErrDuplicateSampleForTimestamp
case "out_of_bounds":
return storage.ErrOutOfBounds
default:
return nil
}
}, nil, nil)
sl, _ := newTestScrapeLoop(t, withAppendable(appTest, appV2))
now := time.Unix(1, 0)
app := sl.appender()
total, added, seriesAdded, err := app.append([]byte("out_of_order 1\namend 1\nnormal 1\nout_of_bounds 1\n"), "text/plain", now)
require.NoError(t, err)
require.NoError(t, app.Commit())
want := []sample{
{
L: labels.FromStrings(model.MetricNameLabel, "normal"),
T: timestamp.FromTime(now),
V: 1,
},
}
teststorage.RequireEqual(t, want, appTest.ResultSamples(), "Appended samples not as expected:\n%s", appTest)
require.Equal(t, 4, total)
require.Equal(t, 4, added)
require.Equal(t, 1, seriesAdded)
}
func TestScrapeLoopOutOfBoundsTimeError(t *testing.T) {
foreachAppendable(t, func(t *testing.T, appV2 bool) {
testScrapeLoopOutOfBoundsTimeError(t, appV2)
})
}
func testScrapeLoopOutOfBoundsTimeError(t *testing.T, appV2 bool) {
sl, _ := newTestScrapeLoop(t, func(sl *scrapeLoop) {
if appV2 {
sl.appendableV2 = appendableV2Func(func(ctx context.Context) storage.AppenderV2 {
return &timeLimitAppenderV2{
AppenderV2: teststorage.NewAppendable().AppenderV2(ctx),
maxTime: timestamp.FromTime(time.Now().Add(10 * time.Minute)),
}
})
} else {
sl.appendable = appendableFunc(func(ctx context.Context) storage.Appender {
return &timeLimitAppender{
Appender: teststorage.NewAppendable().Appender(ctx),
maxTime: timestamp.FromTime(time.Now().Add(10 * time.Minute)),
}
})
}
})
now := time.Now().Add(20 * time.Minute)
app := sl.appender()
total, added, seriesAdded, err := app.append([]byte("normal 1\n"), "text/plain", now)
require.NoError(t, err)
require.NoError(t, app.Commit())
require.Equal(t, 1, total)
require.Equal(t, 1, added)
require.Equal(t, 0, seriesAdded)
}
func TestAcceptHeader(t *testing.T) {
tests := []struct {
name string
scrapeProtocols []config.ScrapeProtocol
scheme model.EscapingScheme
expectedHeader string
}{
{
name: "default scrape protocols with underscore escaping",
scrapeProtocols: config.DefaultScrapeProtocols,
scheme: model.UnderscoreEscaping,
expectedHeader: "application/openmetrics-text;version=1.0.0;escaping=underscores;q=0.6,application/openmetrics-text;version=0.0.1;q=0.5,text/plain;version=1.0.0;escaping=underscores;q=0.4,text/plain;version=0.0.4;q=0.3,*/*;q=0.2",
},
{
name: "default proto first scrape protocols with underscore escaping",
scrapeProtocols: config.DefaultProtoFirstScrapeProtocols,
scheme: model.DotsEscaping,
expectedHeader: "application/vnd.google.protobuf;proto=io.prometheus.client.MetricFamily;encoding=delimited;q=0.6,application/openmetrics-text;version=1.0.0;escaping=dots;q=0.5,application/openmetrics-text;version=0.0.1;q=0.4,text/plain;version=1.0.0;escaping=dots;q=0.3,text/plain;version=0.0.4;q=0.2,*/*;q=0.1",
},
{
name: "default scrape protocols with no escaping",
scrapeProtocols: config.DefaultScrapeProtocols,
scheme: model.NoEscaping,
expectedHeader: "application/openmetrics-text;version=1.0.0;escaping=allow-utf-8;q=0.6,application/openmetrics-text;version=0.0.1;q=0.5,text/plain;version=1.0.0;escaping=allow-utf-8;q=0.4,text/plain;version=0.0.4;q=0.3,*/*;q=0.2",
},
{
name: "default proto first scrape protocols with no escaping",
scrapeProtocols: config.DefaultProtoFirstScrapeProtocols,
scheme: model.NoEscaping,
expectedHeader: "application/vnd.google.protobuf;proto=io.prometheus.client.MetricFamily;encoding=delimited;q=0.6,application/openmetrics-text;version=1.0.0;escaping=allow-utf-8;q=0.5,application/openmetrics-text;version=0.0.1;q=0.4,text/plain;version=1.0.0;escaping=allow-utf-8;q=0.3,text/plain;version=0.0.4;q=0.2,*/*;q=0.1",
},
}
for _, tc := range tests {
t.Run(tc.name, func(t *testing.T) {
header := acceptHeader(tc.scrapeProtocols, tc.scheme)
require.Equal(t, tc.expectedHeader, header)
})
}
}
// setupTracing temporarily sets the global TracerProvider and Propagator
// and restores the original state after the test completes.
func setupTracing(t *testing.T) {
t.Helper()
origTracerProvider := otel.GetTracerProvider()
origPropagator := otel.GetTextMapPropagator()
tp := sdktrace.NewTracerProvider(sdktrace.WithSampler(sdktrace.AlwaysSample()))
otel.SetTracerProvider(tp)
otel.SetTextMapPropagator(propagation.TraceContext{})
t.Cleanup(func() {
otel.SetTracerProvider(origTracerProvider)
otel.SetTextMapPropagator(origPropagator)
})
}
// TestRequestTraceparentHeader verifies that the HTTP client used by the target scraper
// propagates the OpenTelemetry "traceparent" header correctly.
func TestRequestTraceparentHeader(t *testing.T) {
setupTracing(t)
server := httptest.NewServer(http.HandlerFunc(func(_ http.ResponseWriter, r *http.Request) {
// the traceparent header is sent.
require.NotEmpty(t, r.Header.Get("traceparent"))
}))
defer server.Close()
serverURL, err := url.Parse(server.URL)
require.NoError(t, err)
client, err := newScrapeClient(config_util.DefaultHTTPClientConfig, "test")
require.NoError(t, err)
ts := &targetScraper{
Target: &Target{
labels: labels.FromStrings(
model.SchemeLabel, serverURL.Scheme,
model.AddressLabel, serverURL.Host,
),
scrapeConfig: &config.ScrapeConfig{},
},
client: client,
}
resp, err := ts.scrape(context.Background())
require.NoError(t, err)
require.NotNil(t, resp)
t.Cleanup(func() { _ = resp.Body.Close() })
}
func TestTargetScraperScrapeOK(t *testing.T) {
const (
configTimeout = 1500 * time.Millisecond
expectedTimeout = "1.5"
)
var (
protobufParsing bool
allowUTF8 bool
qValuePattern = regexp.MustCompile(`q=([0-9]+(\.\d+)?)`)
)
server := httptest.NewServer(
http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {
accept := r.Header.Get("Accept")
if allowUTF8 {
require.Containsf(t, accept, "escaping=allow-utf-8", "Expected Accept header to allow utf8, got %q", accept)
} else {
require.NotContainsf(t, accept, "escaping=allow-utf-8", "Expected Accept header to not allow utf8, got %q", accept)
}
if protobufParsing {
require.True(t, strings.HasPrefix(accept, "application/vnd.google.protobuf;"),
"Expected Accept header to prefer application/vnd.google.protobuf.")
}
contentTypes := strings.SplitSeq(accept, ",")
for st := range contentTypes {
match := qValuePattern.FindStringSubmatch(st)
require.Len(t, match, 3)
qValue, err := strconv.ParseFloat(match[1], 64)
require.NoError(t, err, "Error parsing q value")
require.GreaterOrEqual(t, qValue, float64(0))
require.LessOrEqual(t, qValue, float64(1))
require.LessOrEqual(t, len(strings.Split(match[1], ".")[1]), 3, "q value should have at most 3 decimal places")
}
timeout := r.Header.Get("X-Prometheus-Scrape-Timeout-Seconds")
require.Equal(t, expectedTimeout, timeout, "Expected scrape timeout header.")
if allowUTF8 {
w.Header().Set("Content-Type", `text/plain; version=1.0.0; escaping=allow-utf-8`)
} else {
w.Header().Set("Content-Type", `text/plain; version=0.0.4`)
}
_, _ = w.Write([]byte("metric_a 1\nmetric_b 2\n"))
}),
)
defer server.Close()
serverURL, err := url.Parse(server.URL)
if err != nil {
panic(err)
}
runTest := func(t *testing.T, acceptHeader string) {
ts := &targetScraper{
Target: &Target{
labels: labels.FromStrings(
model.SchemeLabel, serverURL.Scheme,
model.AddressLabel, serverURL.Host,
),
scrapeConfig: &config.ScrapeConfig{},
},
client: http.DefaultClient,
timeout: configTimeout,
acceptHeader: acceptHeader,
}
var buf bytes.Buffer
resp, err := ts.scrape(context.Background())
require.NoError(t, err)
contentType, err := ts.readResponse(context.Background(), resp, &buf)
require.NoError(t, err)
if allowUTF8 {
require.Equal(t, "text/plain; version=1.0.0; escaping=allow-utf-8", contentType)
} else {
require.Equal(t, "text/plain; version=0.0.4", contentType)
}
require.Equal(t, "metric_a 1\nmetric_b 2\n", buf.String())
}
for _, tc := range []struct {
scrapeProtocols []config.ScrapeProtocol
scheme model.EscapingScheme
protobufParsing bool
allowUTF8 bool
}{
{
scrapeProtocols: config.DefaultScrapeProtocols,
scheme: model.UnderscoreEscaping,
protobufParsing: false,
allowUTF8: false,
},
{
scrapeProtocols: config.DefaultProtoFirstScrapeProtocols,
scheme: model.UnderscoreEscaping,
protobufParsing: true,
allowUTF8: false,
},
{
scrapeProtocols: config.DefaultScrapeProtocols,
scheme: model.NoEscaping,
protobufParsing: false,
allowUTF8: true,
},
{
scrapeProtocols: config.DefaultProtoFirstScrapeProtocols,
scheme: model.NoEscaping,
protobufParsing: true,
allowUTF8: true,
},
} {
t.Run(fmt.Sprintf("%+v", tc), func(t *testing.T) {
protobufParsing = tc.protobufParsing
allowUTF8 = tc.allowUTF8
runTest(t, acceptHeader(tc.scrapeProtocols, tc.scheme))
})
}
}
func TestTargetScrapeScrapeCancel(t *testing.T) {
block := make(chan struct{})
server := httptest.NewServer(
http.HandlerFunc(func(http.ResponseWriter, *http.Request) {
<-block
}),
)
defer server.Close()
serverURL, err := url.Parse(server.URL)
if err != nil {
panic(err)
}
ts := &targetScraper{
Target: &Target{
labels: labels.FromStrings(
model.SchemeLabel, serverURL.Scheme,
model.AddressLabel, serverURL.Host,
),
scrapeConfig: &config.ScrapeConfig{},
},
client: http.DefaultClient,
acceptHeader: acceptHeader(config.DefaultGlobalConfig.ScrapeProtocols, model.UnderscoreEscaping),
}
ctx, cancel := context.WithCancel(context.Background())
errc := make(chan error, 1)
go func() {
time.Sleep(1 * time.Second)
cancel()
}()
go func() {
_, err := ts.scrape(ctx)
switch {
case err == nil:
errc <- errors.New("expected error but got nil")
case !errors.Is(ctx.Err(), context.Canceled):
errc <- fmt.Errorf("expected context cancellation error but got: %w", ctx.Err())
default:
close(errc)
}
}()
select {
case <-time.After(5 * time.Second):
require.FailNow(t, "Scrape function did not return unexpectedly.")
case err := <-errc:
require.NoError(t, err)
}
// If this is closed in a defer above the function the test server
// doesn't terminate and the test doesn't complete.
close(block)
}
func TestTargetScrapeScrapeNotFound(t *testing.T) {
server := httptest.NewServer(
http.HandlerFunc(func(w http.ResponseWriter, _ *http.Request) {
w.WriteHeader(http.StatusNotFound)
}),
)
defer server.Close()
serverURL, err := url.Parse(server.URL)
if err != nil {
panic(err)
}
ts := &targetScraper{
Target: &Target{
labels: labels.FromStrings(
model.SchemeLabel, serverURL.Scheme,
model.AddressLabel, serverURL.Host,
),
scrapeConfig: &config.ScrapeConfig{},
},
client: http.DefaultClient,
acceptHeader: acceptHeader(config.DefaultGlobalConfig.ScrapeProtocols, model.UnderscoreEscaping),
}
resp, err := ts.scrape(context.Background())
require.NoError(t, err)
_, err = ts.readResponse(context.Background(), resp, io.Discard)
require.ErrorContains(t, err, "404", "Expected \"404 NotFound\" error but got: %s", err)
}
func TestTargetScraperBodySizeLimit(t *testing.T) {
const (
bodySizeLimit = 15
responseBody = "metric_a 1\nmetric_b 2\n"
)
var gzipResponse bool
server := httptest.NewServer(
http.HandlerFunc(func(w http.ResponseWriter, _ *http.Request) {
w.Header().Set("Content-Type", `text/plain; version=0.0.4`)
if gzipResponse {
w.Header().Set("Content-Encoding", "gzip")
gw := gzip.NewWriter(w)
defer func() { _ = gw.Close() }()
_, _ = gw.Write([]byte(responseBody))
return
}
_, _ = w.Write([]byte(responseBody))
}),
)
defer server.Close()
serverURL, err := url.Parse(server.URL)
if err != nil {
panic(err)
}
ts := &targetScraper{
Target: &Target{
labels: labels.FromStrings(
model.SchemeLabel, serverURL.Scheme,
model.AddressLabel, serverURL.Host,
),
scrapeConfig: &config.ScrapeConfig{},
},
client: http.DefaultClient,
bodySizeLimit: bodySizeLimit,
acceptHeader: acceptHeader(config.DefaultGlobalConfig.ScrapeProtocols, model.UnderscoreEscaping),
metrics: newTestScrapeMetrics(t),
}
var buf bytes.Buffer
// Target response uncompressed body, scrape with body size limit.
resp, err := ts.scrape(context.Background())
require.NoError(t, err)
_, err = ts.readResponse(context.Background(), resp, &buf)
require.ErrorIs(t, err, errBodySizeLimit)
require.Equal(t, bodySizeLimit, buf.Len())
// Target response gzip compressed body, scrape with body size limit.
gzipResponse = true
buf.Reset()
resp, err = ts.scrape(context.Background())
require.NoError(t, err)
_, err = ts.readResponse(context.Background(), resp, &buf)
require.ErrorIs(t, err, errBodySizeLimit)
require.Equal(t, bodySizeLimit, buf.Len())
// Target response uncompressed body, scrape without body size limit.
gzipResponse = false
buf.Reset()
ts.bodySizeLimit = 0
resp, err = ts.scrape(context.Background())
require.NoError(t, err)
_, err = ts.readResponse(context.Background(), resp, &buf)
require.NoError(t, err)
require.Len(t, responseBody, buf.Len())
// Target response gzip compressed body, scrape without body size limit.
gzipResponse = true
buf.Reset()
resp, err = ts.scrape(context.Background())
require.NoError(t, err)
_, err = ts.readResponse(context.Background(), resp, &buf)
require.NoError(t, err)
require.Len(t, responseBody, buf.Len())
}
// testScraper implements the scraper interface and allows setting values
// returned by its methods. It also allows setting a custom scrape function.
type testScraper struct {
offsetDur time.Duration
lastStart time.Time
lastDuration time.Duration
lastError error
scrapeErr error
scrapeFunc func(context.Context, io.Writer) error
}
func (ts *testScraper) offset(time.Duration, uint64) time.Duration {
return ts.offsetDur
}
func (ts *testScraper) Report(start time.Time, duration time.Duration, err error) {
ts.lastStart = start
ts.lastDuration = duration
ts.lastError = err
}
func (ts *testScraper) scrape(context.Context) (*http.Response, error) {
return nil, ts.scrapeErr
}
func (ts *testScraper) readResponse(ctx context.Context, _ *http.Response, w io.Writer) (string, error) {
if ts.scrapeFunc != nil {
return "", ts.scrapeFunc(ctx, w)
}
return "", ts.scrapeErr
}
func TestScrapeLoop_RespectTimestamps(t *testing.T) {
foreachAppendable(t, func(t *testing.T, appV2 bool) {
testScrapeLoopRespectTimestamps(t, appV2)
})
}
func testScrapeLoopRespectTimestamps(t *testing.T, appV2 bool) {
s := teststorage.New(t)
appTest := teststorage.NewAppendable().Then(s)
sl, _ := newTestScrapeLoop(t, withAppendable(appTest, appV2))
now := time.Now()
app := sl.appender()
_, _, _, err := app.append([]byte(`metric_a{a="1",b="1"} 1 0`), "text/plain", now)
require.NoError(t, err)
require.NoError(t, app.Commit())
want := []sample{
{
L: labels.FromStrings("__name__", "metric_a", "a", "1", "b", "1"),
T: 0,
V: 1,
},
}
teststorage.RequireEqual(t, want, appTest.ResultSamples(), "Appended samples not as expected:\n%s", appTest)
}
func TestScrapeLoop_DiscardTimestamps(t *testing.T) {
foreachAppendable(t, func(t *testing.T, appV2 bool) {
testScrapeLoopDiscardTimestamps(t, appV2)
})
}
func testScrapeLoopDiscardTimestamps(t *testing.T, appV2 bool) {
s := teststorage.New(t)
appTest := teststorage.NewAppendable().Then(s)
sl, _ := newTestScrapeLoop(t, withAppendable(appTest, appV2), func(sl *scrapeLoop) {
sl.honorTimestamps = false
})
now := time.Now()
app := sl.appender()
_, _, _, err := app.append([]byte(`metric_a{a="1",b="1"} 1 0`), "text/plain", now)
require.NoError(t, err)
require.NoError(t, app.Commit())
want := []sample{
{
L: labels.FromStrings("__name__", "metric_a", "a", "1", "b", "1"),
T: timestamp.FromTime(now),
V: 1,
},
}
teststorage.RequireEqual(t, want, appTest.ResultSamples(), "Appended samples not as expected:\n%s", appTest)
}
func TestScrapeLoopDiscardDuplicateLabels(t *testing.T) {
foreachAppendable(t, func(t *testing.T, appV2 bool) {
testScrapeLoopDiscardDuplicateLabels(t, appV2)
})
}
func testScrapeLoopDiscardDuplicateLabels(t *testing.T, appV2 bool) {
s := teststorage.New(t)
appTest := teststorage.NewAppendable().Then(s)
sl, _ := newTestScrapeLoop(t, withAppendable(appTest, appV2))
// We add a good and a bad metric to check that both are discarded.
app := sl.appender()
_, _, _, err := app.append([]byte("test_metric{le=\"500\"} 1\ntest_metric{le=\"600\",le=\"700\"} 1\n"), "text/plain", time.Time{})
require.Error(t, err)
require.NoError(t, app.Rollback())
// We need to cycle staleness cache maps after a manual rollback. Otherwise, they will have old entries in them,
// which would cause ErrDuplicateSampleForTimestamp errors on the next append.
sl.cache.iterDone(true)
q, err := s.Querier(time.Time{}.UnixNano(), 0)
require.NoError(t, err)
series := q.Select(sl.ctx, false, nil, labels.MustNewMatcher(labels.MatchRegexp, "__name__", ".*"))
require.False(t, series.Next(), "series found in tsdb")
require.NoError(t, series.Err())
// We add a good metric to check that it is recorded.
app = sl.appender()
_, _, _, err = app.append([]byte("test_metric{le=\"500\"} 1\n"), "text/plain", time.Time{})
require.NoError(t, err)
require.NoError(t, app.Commit())
q, err = s.Querier(time.Time{}.UnixNano(), 0)
require.NoError(t, err)
series = q.Select(sl.ctx, false, nil, labels.MustNewMatcher(labels.MatchEqual, "le", "500"))
require.True(t, series.Next(), "series not found in tsdb")
require.NoError(t, series.Err())
require.False(t, series.Next(), "more than one series found in tsdb")
}
func TestScrapeLoopDiscardUnnamedMetrics(t *testing.T) {
foreachAppendable(t, func(t *testing.T, appV2 bool) {
testScrapeLoopDiscardUnnamedMetrics(t, appV2)
})
}
func testScrapeLoopDiscardUnnamedMetrics(t *testing.T, appV2 bool) {
s := teststorage.New(t)
appTest := teststorage.NewAppendable().Then(s)
sl, _ := newTestScrapeLoop(t, withAppendable(appTest, appV2), func(sl *scrapeLoop) {
sl.sampleMutator = func(l labels.Labels) labels.Labels {
if l.Has("drop") {
return labels.FromStrings("no", "name") // This label set will trigger an error.
}
return l
}
})
app := sl.appender()
_, _, _, err := app.append([]byte("nok 1\nnok2{drop=\"drop\"} 1\n"), "text/plain", time.Time{})
require.Error(t, err)
require.NoError(t, app.Rollback())
require.Equal(t, errNameLabelMandatory, err)
q, err := s.Querier(time.Time{}.UnixNano(), 0)
require.NoError(t, err)
series := q.Select(sl.ctx, false, nil, labels.MustNewMatcher(labels.MatchRegexp, "__name__", ".*"))
require.False(t, series.Next(), "series found in tsdb")
require.NoError(t, series.Err())
}
func TestReusableConfig(t *testing.T) {
variants := []*config.ScrapeConfig{
{
JobName: "prometheus",
ScrapeTimeout: model.Duration(15 * time.Second),
},
{
JobName: "httpd",
ScrapeTimeout: model.Duration(15 * time.Second),
},
{
JobName: "prometheus",
ScrapeTimeout: model.Duration(5 * time.Second),
},
{
JobName: "prometheus",
MetricsPath: "/metrics",
},
{
JobName: "prometheus",
MetricsPath: "/metrics2",
},
{
JobName: "prometheus",
ScrapeTimeout: model.Duration(5 * time.Second),
MetricsPath: "/metrics2",
},
{
JobName: "prometheus",
ScrapeInterval: model.Duration(5 * time.Second),
MetricsPath: "/metrics2",
},
{
JobName: "prometheus",
ScrapeInterval: model.Duration(5 * time.Second),
SampleLimit: 1000,
MetricsPath: "/metrics2",
},
}
match := [][]int{
{0, 2},
{4, 5},
{4, 6},
{4, 7},
{5, 6},
{5, 7},
{6, 7},
}
noMatch := [][]int{
{1, 2},
{0, 4},
{3, 4},
}
for i, m := range match {
require.True(t, reusableCache(variants[m[0]], variants[m[1]]), "match test %d", i)
require.True(t, reusableCache(variants[m[1]], variants[m[0]]), "match test %d", i)
require.True(t, reusableCache(variants[m[1]], variants[m[1]]), "match test %d", i)
require.True(t, reusableCache(variants[m[0]], variants[m[0]]), "match test %d", i)
}
for i, m := range noMatch {
require.False(t, reusableCache(variants[m[0]], variants[m[1]]), "not match test %d", i)
require.False(t, reusableCache(variants[m[1]], variants[m[0]]), "not match test %d", i)
}
}
func TestReuseScrapeCache(t *testing.T) {
foreachAppendable(t, func(t *testing.T, appV2 bool) {
testReuseScrapeCache(t, appV2)
})
}
func testReuseScrapeCache(t *testing.T, appV2 bool) {
var (
app = teststorage.NewAppendable()
cfg = &config.ScrapeConfig{
JobName: "Prometheus",
ScrapeTimeout: model.Duration(5 * time.Second),
ScrapeInterval: model.Duration(5 * time.Second),
MetricsPath: "/metrics",
MetricNameValidationScheme: model.UTF8Validation,
MetricNameEscapingScheme: model.AllowUTF8,
}
sa = selectAppendable(app, appV2)
sp, _ = newScrapePool(cfg, sa.V1(), sa.V2(), 0, nil, nil, &Options{}, newTestScrapeMetrics(t))
t1 = &Target{
labels: labels.FromStrings("labelNew", "nameNew", "labelNew1", "nameNew1", "labelNew2", "nameNew2"),
scrapeConfig: &config.ScrapeConfig{
MetricNameValidationScheme: model.UTF8Validation,
MetricNameEscapingScheme: model.AllowUTF8,
},
}
proxyURL, _ = url.Parse("http://localhost:2128")
)
defer sp.stop()
sp.sync([]*Target{t1})
steps := []struct {
keep bool
newConfig *config.ScrapeConfig
}{
{
keep: true,
newConfig: &config.ScrapeConfig{
JobName: "Prometheus",
ScrapeInterval: model.Duration(5 * time.Second),
ScrapeTimeout: model.Duration(5 * time.Second),
MetricsPath: "/metrics",
MetricNameValidationScheme: model.UTF8Validation,
MetricNameEscapingScheme: model.AllowUTF8,
},
},
{
keep: false,
newConfig: &config.ScrapeConfig{
JobName: "Prometheus",
ScrapeInterval: model.Duration(5 * time.Second),
ScrapeTimeout: model.Duration(15 * time.Second),
MetricsPath: "/metrics2",
MetricNameValidationScheme: model.UTF8Validation,
MetricNameEscapingScheme: model.AllowUTF8,
},
},
{
keep: true,
newConfig: &config.ScrapeConfig{
JobName: "Prometheus",
SampleLimit: 400,
ScrapeInterval: model.Duration(5 * time.Second),
ScrapeTimeout: model.Duration(15 * time.Second),
MetricsPath: "/metrics2",
MetricNameValidationScheme: model.UTF8Validation,
MetricNameEscapingScheme: model.AllowUTF8,
},
},
{
keep: false,
newConfig: &config.ScrapeConfig{
JobName: "Prometheus",
HonorTimestamps: true,
SampleLimit: 400,
ScrapeInterval: model.Duration(5 * time.Second),
ScrapeTimeout: model.Duration(15 * time.Second),
MetricsPath: "/metrics2",
MetricNameValidationScheme: model.UTF8Validation,
MetricNameEscapingScheme: model.AllowUTF8,
},
},
{
keep: true,
newConfig: &config.ScrapeConfig{
JobName: "Prometheus",
HonorTimestamps: true,
SampleLimit: 400,
HTTPClientConfig: config_util.HTTPClientConfig{
ProxyConfig: config_util.ProxyConfig{ProxyURL: config_util.URL{URL: proxyURL}},
},
ScrapeInterval: model.Duration(5 * time.Second),
ScrapeTimeout: model.Duration(15 * time.Second),
MetricsPath: "/metrics2",
MetricNameValidationScheme: model.UTF8Validation,
MetricNameEscapingScheme: model.AllowUTF8,
},
},
{
keep: false,
newConfig: &config.ScrapeConfig{
JobName: "Prometheus",
HonorTimestamps: true,
HonorLabels: true,
SampleLimit: 400,
ScrapeInterval: model.Duration(5 * time.Second),
ScrapeTimeout: model.Duration(15 * time.Second),
MetricsPath: "/metrics2",
MetricNameValidationScheme: model.UTF8Validation,
MetricNameEscapingScheme: model.AllowUTF8,
},
},
{
keep: false,
newConfig: &config.ScrapeConfig{
JobName: "Prometheus",
ScrapeInterval: model.Duration(5 * time.Second),
ScrapeTimeout: model.Duration(15 * time.Second),
MetricsPath: "/metrics",
LabelLimit: 1,
MetricNameValidationScheme: model.UTF8Validation,
MetricNameEscapingScheme: model.AllowUTF8,
},
},
{
keep: false,
newConfig: &config.ScrapeConfig{
JobName: "Prometheus",
ScrapeInterval: model.Duration(5 * time.Second),
ScrapeTimeout: model.Duration(15 * time.Second),
MetricsPath: "/metrics",
LabelLimit: 15,
MetricNameValidationScheme: model.UTF8Validation,
MetricNameEscapingScheme: model.AllowUTF8,
},
},
{
keep: false,
newConfig: &config.ScrapeConfig{
JobName: "Prometheus",
ScrapeInterval: model.Duration(5 * time.Second),
ScrapeTimeout: model.Duration(15 * time.Second),
MetricsPath: "/metrics",
LabelLimit: 15,
LabelNameLengthLimit: 5,
MetricNameValidationScheme: model.UTF8Validation,
MetricNameEscapingScheme: model.AllowUTF8,
},
},
{
keep: false,
newConfig: &config.ScrapeConfig{
JobName: "Prometheus",
ScrapeInterval: model.Duration(5 * time.Second),
ScrapeTimeout: model.Duration(15 * time.Second),
MetricsPath: "/metrics",
LabelLimit: 15,
LabelNameLengthLimit: 5,
LabelValueLengthLimit: 7,
MetricNameValidationScheme: model.UTF8Validation,
MetricNameEscapingScheme: model.AllowUTF8,
},
},
}
cacheAddr := func(sp *scrapePool) map[uint64]string {
r := make(map[uint64]string)
for fp, l := range sp.loops {
r[fp] = fmt.Sprintf("%p", l.getCache())
}
return r
}
for i, s := range steps {
initCacheAddr := cacheAddr(sp)
require.NoError(t, sp.reload(s.newConfig))
for fp, newCacheAddr := range cacheAddr(sp) {
if s.keep {
require.Equal(t, initCacheAddr[fp], newCacheAddr, "step %d: old cache and new cache are not the same", i)
} else {
require.NotEqual(t, initCacheAddr[fp], newCacheAddr, "step %d: old cache and new cache are the same", i)
}
}
initCacheAddr = cacheAddr(sp)
require.NoError(t, sp.reload(s.newConfig))
for fp, newCacheAddr := range cacheAddr(sp) {
require.Equal(t, initCacheAddr[fp], newCacheAddr, "step %d: reloading the exact config invalidates the cache", i)
}
}
}
func TestScrapeAddFast(t *testing.T) {
foreachAppendable(t, func(t *testing.T, appV2 bool) {
testScrapeAddFast(t, appV2)
})
}
func testScrapeAddFast(t *testing.T, appV2 bool) {
s := teststorage.New(t)
sl, _ := newTestScrapeLoop(t, withAppendable(s, appV2))
app := sl.appender()
_, _, _, err := app.append([]byte("up 1\n"), "text/plain", time.Time{})
require.NoError(t, err)
require.NoError(t, app.Commit())
// Poison the cache. There is just one entry, and one series in the
// storage. Changing the ref will create a 'not found' error.
for _, v := range sl.getCache().series {
v.ref++
}
app = sl.appender()
_, _, _, err = app.append([]byte("up 1\n"), "text/plain", time.Time{}.Add(time.Second))
require.NoError(t, err)
require.NoError(t, app.Commit())
}
func TestReuseCacheRace(t *testing.T) {
foreachAppendable(t, func(t *testing.T, appV2 bool) {
testReuseCacheRace(t, appV2)
})
}
func testReuseCacheRace(t *testing.T, appV2 bool) {
var (
cfg = &config.ScrapeConfig{
JobName: "Prometheus",
ScrapeTimeout: model.Duration(5 * time.Second),
ScrapeInterval: model.Duration(5 * time.Second),
MetricsPath: "/metrics",
MetricNameValidationScheme: model.UTF8Validation,
MetricNameEscapingScheme: model.AllowUTF8,
}
buffers = pool.New(1e3, 100e6, 3, func(sz int) any { return make([]byte, 0, sz) })
sa = selectAppendable(teststorage.NewAppendable(), appV2)
sp, _ = newScrapePool(cfg, sa.V1(), sa.V2(), 0, nil, buffers, &Options{}, newTestScrapeMetrics(t))
t1 = &Target{
labels: labels.FromStrings("labelNew", "nameNew"),
scrapeConfig: &config.ScrapeConfig{},
}
)
defer sp.stop()
sp.sync([]*Target{t1})
start := time.Now()
for i := uint(1); i > 0; i++ {
if time.Since(start) > 5*time.Second {
break
}
require.NoError(t, sp.reload(&config.ScrapeConfig{
JobName: "Prometheus",
ScrapeTimeout: model.Duration(1 * time.Millisecond),
ScrapeInterval: model.Duration(1 * time.Millisecond),
MetricsPath: "/metrics",
SampleLimit: i,
MetricNameValidationScheme: model.UTF8Validation,
MetricNameEscapingScheme: model.AllowUTF8,
}))
}
}
func TestCheckAddError(t *testing.T) {
var appErrs appendErrors
sl, _ := newTestScrapeLoop(t)
// TODO: Check err etc
_, _ = sl.checkAddError(nil, nil, storage.ErrOutOfOrderSample, nil, nil, &appErrs)
require.Equal(t, 1, appErrs.numOutOfOrder)
// TODO(bwplotka): Test partial error check and other cases
}
func TestScrapeReportSingleAppender(t *testing.T) {
foreachAppendable(t, func(t *testing.T, appV2 bool) {
testScrapeReportSingleAppender(t, appV2)
})
}
func testScrapeReportSingleAppender(t *testing.T, appV2 bool) {
t.Parallel()
s := teststorage.New(t)
signal := make(chan struct{}, 1)
ctx, cancel := context.WithCancel(t.Context())
sl, scraper := newTestScrapeLoop(t, withAppendable(s, appV2), func(sl *scrapeLoop) {
sl.ctx = ctx
// Since we're writing samples directly below we need to provide a protocol fallback.
sl.fallbackScrapeProtocol = "text/plain"
})
numScrapes := 0
scraper.scrapeFunc = func(_ context.Context, w io.Writer) error {
numScrapes++
if numScrapes%4 == 0 {
return errors.New("scrape failed")
}
_, _ = w.Write([]byte("metric_a 44\nmetric_b 44\nmetric_c 44\nmetric_d 44\n"))
return nil
}
go func() {
sl.run(nil)
signal <- struct{}{}
}()
start := time.Now()
for time.Since(start) < 3*time.Second {
q, err := s.Querier(time.Time{}.UnixNano(), time.Now().UnixNano())
require.NoError(t, err)
series := q.Select(ctx, false, nil, labels.MustNewMatcher(labels.MatchRegexp, "__name__", ".+"))
c := 0
for series.Next() {
i := series.At().Iterator(nil)
for i.Next() != chunkenc.ValNone {
c++
}
}
require.Equal(t, 0, c%9, "Appended samples not as expected: %d", c)
require.NoError(t, q.Close())
}
cancel()
select {
case <-signal:
case <-time.After(5 * time.Second):
require.FailNow(t, "Scrape wasn't stopped.")
}
}
func TestScrapeReportLimit(t *testing.T) {
foreachAppendable(t, func(t *testing.T, appV2 bool) {
testScrapeReportLimit(t, appV2)
})
}
func testScrapeReportLimit(t *testing.T, appV2 bool) {
s := teststorage.New(t)
cfg := &config.ScrapeConfig{
JobName: "test",
SampleLimit: 5,
Scheme: "http",
ScrapeInterval: model.Duration(100 * time.Millisecond),
ScrapeTimeout: model.Duration(100 * time.Millisecond),
MetricNameValidationScheme: model.UTF8Validation,
MetricNameEscapingScheme: model.AllowUTF8,
}
ts, scrapedTwice := newScrapableServer("metric_a 44\nmetric_b 44\nmetric_c 44\nmetric_d 44\n")
defer ts.Close()
sa := selectAppendable(s, appV2)
sp, err := newScrapePool(cfg, sa.V1(), sa.V2(), 0, nil, nil, &Options{}, newTestScrapeMetrics(t))
require.NoError(t, err)
defer sp.stop()
testURL, err := url.Parse(ts.URL)
require.NoError(t, err)
sp.Sync([]*targetgroup.Group{
{
Targets: []model.LabelSet{{model.AddressLabel: model.LabelValue(testURL.Host)}},
},
})
select {
case <-time.After(5 * time.Second):
t.Fatalf("target was not scraped twice")
case <-scrapedTwice:
// If the target has been scraped twice, report samples from the first
// scrape have been inserted in the database.
}
ctx := t.Context()
q, err := s.Querier(time.Time{}.UnixNano(), time.Now().UnixNano())
require.NoError(t, err)
t.Cleanup(func() { _ = q.Close() })
series := q.Select(ctx, false, nil, labels.MustNewMatcher(labels.MatchRegexp, "__name__", "up"))
var found bool
for series.Next() {
i := series.At().Iterator(nil)
for i.Next() == chunkenc.ValFloat {
_, v := i.At()
require.Equal(t, 1.0, v)
found = true
}
}
require.True(t, found)
}
func TestScrapeUTF8(t *testing.T) {
foreachAppendable(t, func(t *testing.T, appV2 bool) {
testScrapeUTF8(t, appV2)
})
}
func testScrapeUTF8(t *testing.T, appV2 bool) {
s := teststorage.New(t)
cfg := &config.ScrapeConfig{
JobName: "test",
Scheme: "http",
ScrapeInterval: model.Duration(100 * time.Millisecond),
ScrapeTimeout: model.Duration(100 * time.Millisecond),
MetricNameValidationScheme: model.UTF8Validation,
MetricNameEscapingScheme: model.AllowUTF8,
}
ts, scrapedTwice := newScrapableServer("{\"with.dots\"} 42\n")
defer ts.Close()
sa := selectAppendable(s, appV2)
sp, err := newScrapePool(cfg, sa.V1(), sa.V2(), 0, nil, nil, &Options{}, newTestScrapeMetrics(t))
require.NoError(t, err)
defer sp.stop()
testURL, err := url.Parse(ts.URL)
require.NoError(t, err)
sp.Sync([]*targetgroup.Group{
{
Targets: []model.LabelSet{{model.AddressLabel: model.LabelValue(testURL.Host)}},
},
})
select {
case <-time.After(5 * time.Second):
t.Fatalf("target was not scraped twice")
case <-scrapedTwice:
// If the target has been scraped twice, report samples from the first
// scrape have been inserted in the database.
}
ctx := t.Context()
q, err := s.Querier(time.Time{}.UnixNano(), time.Now().UnixNano())
require.NoError(t, err)
t.Cleanup(func() { _ = q.Close() })
series := q.Select(ctx, false, nil, labels.MustNewMatcher(labels.MatchRegexp, "__name__", "with.dots"))
require.True(t, series.Next(), "series not found in tsdb")
}
func TestScrapeLoopLabelLimit(t *testing.T) {
foreachAppendable(t, func(t *testing.T, appV2 bool) {
testScrapeLoopLabelLimit(t, appV2)
})
}
func testScrapeLoopLabelLimit(t *testing.T, appV2 bool) {
for _, test := range []struct {
title string
scrapeLabels string
discoveryLabels []string
labelLimits labelLimits
expectErr bool
}{
{
title: "Valid number of labels",
scrapeLabels: `metric{l1="1", l2="2"} 0`,
discoveryLabels: nil,
labelLimits: labelLimits{labelLimit: 5},
expectErr: false,
}, {
title: "Too many labels",
scrapeLabels: `metric{l1="1", l2="2", l3="3", l4="4", l5="5", l6="6"} 0`,
discoveryLabels: nil,
labelLimits: labelLimits{labelLimit: 5},
expectErr: true,
}, {
title: "Too many labels including discovery labels",
scrapeLabels: `metric{l1="1", l2="2", l3="3", l4="4"} 0`,
discoveryLabels: []string{"l5", "5", "l6", "6"},
labelLimits: labelLimits{labelLimit: 5},
expectErr: true,
}, {
title: "Valid labels name length",
scrapeLabels: `metric{l1="1", l2="2"} 0`,
discoveryLabels: nil,
labelLimits: labelLimits{labelNameLengthLimit: 10},
expectErr: false,
}, {
title: "Label name too long",
scrapeLabels: `metric{label_name_too_long="0"} 0`,
discoveryLabels: nil,
labelLimits: labelLimits{labelNameLengthLimit: 10},
expectErr: true,
}, {
title: "Discovery label name too long",
scrapeLabels: `metric{l1="1", l2="2"} 0`,
discoveryLabels: []string{"label_name_too_long", "0"},
labelLimits: labelLimits{labelNameLengthLimit: 10},
expectErr: true,
}, {
title: "Valid labels value length",
scrapeLabels: `metric{l1="1", l2="2"} 0`,
discoveryLabels: nil,
labelLimits: labelLimits{labelValueLengthLimit: 10},
expectErr: false,
}, {
title: "Label value too long",
scrapeLabels: `metric{l1="label_value_too_long"} 0`,
discoveryLabels: nil,
labelLimits: labelLimits{labelValueLengthLimit: 10},
expectErr: true,
}, {
title: "Discovery label value too long",
scrapeLabels: `metric{l1="1", l2="2"} 0`,
discoveryLabels: []string{"l1", "label_value_too_long"},
labelLimits: labelLimits{labelValueLengthLimit: 10},
expectErr: true,
},
} {
discoveryLabels := &Target{
labels: labels.FromStrings(test.discoveryLabels...),
}
sl, _ := newTestScrapeLoop(t, withAppendable(teststorage.NewAppendable(), appV2), func(sl *scrapeLoop) {
sl.sampleMutator = func(l labels.Labels) labels.Labels {
return mutateSampleLabels(l, discoveryLabels, false, nil)
}
sl.reportSampleMutator = func(l labels.Labels) labels.Labels {
return mutateReportSampleLabels(l, discoveryLabels)
}
sl.labelLimits = &test.labelLimits
})
app := sl.appender()
_, _, _, err := app.append([]byte(test.scrapeLabels), "text/plain", time.Now())
t.Logf("Test:%s", test.title)
if test.expectErr {
require.Error(t, err)
} else {
require.NoError(t, err)
require.NoError(t, app.Commit())
}
}
}
func TestTargetScrapeIntervalAndTimeoutRelabel(t *testing.T) {
foreachAppendable(t, func(t *testing.T, appV2 bool) {
testTargetScrapeIntervalAndTimeoutRelabel(t, appV2)
})
}
func testTargetScrapeIntervalAndTimeoutRelabel(t *testing.T, appV2 bool) {
interval, _ := model.ParseDuration("2s")
timeout, _ := model.ParseDuration("500ms")
cfg := &config.ScrapeConfig{
ScrapeInterval: interval,
ScrapeTimeout: timeout,
MetricNameValidationScheme: model.UTF8Validation,
MetricNameEscapingScheme: model.AllowUTF8,
RelabelConfigs: []*relabel.Config{
{
SourceLabels: model.LabelNames{model.ScrapeIntervalLabel},
Regex: relabel.MustNewRegexp("2s"),
Replacement: "3s",
TargetLabel: model.ScrapeIntervalLabel,
Action: relabel.Replace,
NameValidationScheme: model.UTF8Validation,
},
{
SourceLabels: model.LabelNames{model.ScrapeTimeoutLabel},
Regex: relabel.MustNewRegexp("500ms"),
Replacement: "750ms",
TargetLabel: model.ScrapeTimeoutLabel,
Action: relabel.Replace,
NameValidationScheme: model.UTF8Validation,
},
},
}
sa := selectAppendable(teststorage.NewAppendable(), appV2)
sp, _ := newScrapePool(cfg, sa.V1(), sa.V2(), 0, nil, nil, &Options{}, newTestScrapeMetrics(t))
tgts := []*targetgroup.Group{
{
Targets: []model.LabelSet{{model.AddressLabel: "127.0.0.1:9090"}},
},
}
sp.Sync(tgts)
defer sp.stop()
require.Equal(t, "3s", sp.ActiveTargets()[0].labels.Get(model.ScrapeIntervalLabel))
require.Equal(t, "750ms", sp.ActiveTargets()[0].labels.Get(model.ScrapeTimeoutLabel))
}
// Testing whether we can remove trailing .0 from histogram 'le' and summary 'quantile' labels.
func TestLeQuantileReLabel(t *testing.T) {
foreachAppendable(t, func(t *testing.T, appV2 bool) {
testLeQuantileReLabel(t, appV2)
})
}
func testLeQuantileReLabel(t *testing.T, appV2 bool) {
s := teststorage.New(t)
cfg := &config.ScrapeConfig{
JobName: "test",
MetricRelabelConfigs: []*relabel.Config{
{
SourceLabels: model.LabelNames{"le", "__name__"},
Regex: relabel.MustNewRegexp("(\\d+)\\.0+;.*_bucket"),
Replacement: relabel.DefaultRelabelConfig.Replacement,
Separator: relabel.DefaultRelabelConfig.Separator,
TargetLabel: "le",
Action: relabel.Replace,
NameValidationScheme: model.UTF8Validation,
},
{
SourceLabels: model.LabelNames{"quantile"},
Regex: relabel.MustNewRegexp("(\\d+)\\.0+"),
Replacement: relabel.DefaultRelabelConfig.Replacement,
Separator: relabel.DefaultRelabelConfig.Separator,
TargetLabel: "quantile",
Action: relabel.Replace,
NameValidationScheme: model.UTF8Validation,
},
},
SampleLimit: 100,
Scheme: "http",
ScrapeInterval: model.Duration(100 * time.Millisecond),
ScrapeTimeout: model.Duration(100 * time.Millisecond),
MetricNameValidationScheme: model.UTF8Validation,
MetricNameEscapingScheme: model.AllowUTF8,
}
metricsText := `
# HELP test_histogram This is a histogram with default buckets
# TYPE test_histogram histogram
test_histogram_bucket{address="0.0.0.0",port="5001",le="0.005"} 0
test_histogram_bucket{address="0.0.0.0",port="5001",le="0.01"} 0
test_histogram_bucket{address="0.0.0.0",port="5001",le="0.025"} 0
test_histogram_bucket{address="0.0.0.0",port="5001",le="0.05"} 0
test_histogram_bucket{address="0.0.0.0",port="5001",le="0.1"} 0
test_histogram_bucket{address="0.0.0.0",port="5001",le="0.25"} 0
test_histogram_bucket{address="0.0.0.0",port="5001",le="0.5"} 0
test_histogram_bucket{address="0.0.0.0",port="5001",le="1.0"} 0
test_histogram_bucket{address="0.0.0.0",port="5001",le="2.5"} 0
test_histogram_bucket{address="0.0.0.0",port="5001",le="5.0"} 0
test_histogram_bucket{address="0.0.0.0",port="5001",le="10.0"} 0
test_histogram_bucket{address="0.0.0.0",port="5001",le="+Inf"} 0
test_histogram_sum{address="0.0.0.0",port="5001"} 0
test_histogram_count{address="0.0.0.0",port="5001"} 0
# HELP test_summary Number of inflight requests sampled at a regular interval. Quantile buckets keep track of inflight requests over the last 60s.
# TYPE test_summary summary
test_summary{quantile="0.5"} 0
test_summary{quantile="0.9"} 0
test_summary{quantile="0.95"} 0
test_summary{quantile="0.99"} 0
test_summary{quantile="1.0"} 1
test_summary_sum 1
test_summary_count 199
`
// The expected "le" values do not have the trailing ".0".
expectedLeValues := []string{"0.005", "0.01", "0.025", "0.05", "0.1", "0.25", "0.5", "1", "2.5", "5", "10", "+Inf"}
// The expected "quantile" values do not have the trailing ".0".
expectedQuantileValues := []string{"0.5", "0.9", "0.95", "0.99", "1"}
ts, scrapedTwice := newScrapableServer(metricsText)
defer ts.Close()
sa := selectAppendable(s, appV2)
sp, err := newScrapePool(cfg, sa.V1(), sa.V2(), 0, nil, nil, &Options{}, newTestScrapeMetrics(t))
require.NoError(t, err)
defer sp.stop()
testURL, err := url.Parse(ts.URL)
require.NoError(t, err)
sp.Sync([]*targetgroup.Group{
{
Targets: []model.LabelSet{{model.AddressLabel: model.LabelValue(testURL.Host)}},
},
})
require.Len(t, sp.ActiveTargets(), 1)
select {
case <-time.After(5 * time.Second):
t.Fatalf("target was not scraped")
case <-scrapedTwice:
}
ctx := t.Context()
q, err := s.Querier(time.Time{}.UnixNano(), time.Now().UnixNano())
require.NoError(t, err)
t.Cleanup(func() { _ = q.Close() })
checkValues := func(labelName string, expectedValues []string, series storage.SeriesSet) {
foundLeValues := map[string]bool{}
for series.Next() {
s := series.At()
v := s.Labels().Get(labelName)
require.NotContains(t, foundLeValues, v, "duplicate label value found")
foundLeValues[v] = true
}
require.Len(t, foundLeValues, len(expectedValues), "number of label values not as expected")
for _, v := range expectedValues {
require.Contains(t, foundLeValues, v, "label value not found")
}
}
series := q.Select(ctx, false, nil, labels.MustNewMatcher(labels.MatchRegexp, "__name__", "test_histogram_bucket"))
checkValues("le", expectedLeValues, series)
series = q.Select(ctx, false, nil, labels.MustNewMatcher(labels.MatchRegexp, "__name__", "test_summary"))
checkValues("quantile", expectedQuantileValues, series)
}
// Testing whether we can automatically convert scraped classic histograms into native histograms with custom buckets.
func TestConvertClassicHistogramsToNHCB(t *testing.T) {
foreachAppendable(t, func(t *testing.T, appV2 bool) {
testConvertClassicHistogramsToNHCB(t, appV2)
})
}
func testConvertClassicHistogramsToNHCB(t *testing.T, appV2 bool) {
t.Parallel()
genTestCounterText := func(name string) string {
return fmt.Sprintf(`
# HELP %s some help text
# TYPE %s counter
%s{address="0.0.0.0",port="5001"} 1
`, name, name, name)
}
genTestHistText := func(name string) string {
data := map[string]any{
"name": name,
}
b := &bytes.Buffer{}
require.NoError(t, template.Must(template.New("").Parse(`
# HELP {{.name}} This is a histogram with default buckets
# TYPE {{.name}} histogram
{{.name}}_bucket{address="0.0.0.0",port="5001",le="0.005"} 0
{{.name}}_bucket{address="0.0.0.0",port="5001",le="0.01"} 0
{{.name}}_bucket{address="0.0.0.0",port="5001",le="0.025"} 0
{{.name}}_bucket{address="0.0.0.0",port="5001",le="0.05"} 0
{{.name}}_bucket{address="0.0.0.0",port="5001",le="0.1"} 0
{{.name}}_bucket{address="0.0.0.0",port="5001",le="0.25"} 0
{{.name}}_bucket{address="0.0.0.0",port="5001",le="0.5"} 0
{{.name}}_bucket{address="0.0.0.0",port="5001",le="1"} 0
{{.name}}_bucket{address="0.0.0.0",port="5001",le="2.5"} 0
{{.name}}_bucket{address="0.0.0.0",port="5001",le="5"} 0
{{.name}}_bucket{address="0.0.0.0",port="5001",le="10"} 1
{{.name}}_bucket{address="0.0.0.0",port="5001",le="+Inf"} 1
{{.name}}_sum{address="0.0.0.0",port="5001"} 10
{{.name}}_count{address="0.0.0.0",port="5001"} 1
`)).Execute(b, data))
return b.String()
}
genTestCounterProto := func(name string) string {
return fmt.Sprintf(`
name: "%s"
help: "some help text"
type: COUNTER
metric: <
label: <
name: "address"
value: "0.0.0.0"
>
label: <
name: "port"
value: "5001"
>
counter: <
value: %d
>
>
`, name, 1)
}
genTestHistProto := func(name string, hasClassic, hasExponential bool) string {
var classic string
if hasClassic {
classic = `
bucket: <
cumulative_count: 0
upper_bound: 0.005
>
bucket: <
cumulative_count: 0
upper_bound: 0.01
>
bucket: <
cumulative_count: 0
upper_bound: 0.025
>
bucket: <
cumulative_count: 0
upper_bound: 0.05
>
bucket: <
cumulative_count: 0
upper_bound: 0.1
>
bucket: <
cumulative_count: 0
upper_bound: 0.25
>
bucket: <
cumulative_count: 0
upper_bound: 0.5
>
bucket: <
cumulative_count: 0
upper_bound: 1
>
bucket: <
cumulative_count: 0
upper_bound: 2.5
>
bucket: <
cumulative_count: 0
upper_bound: 5
>
bucket: <
cumulative_count: 1
upper_bound: 10
>`
}
var expo string
if hasExponential {
expo = `
schema: 3
zero_threshold: 2.938735877055719e-39
zero_count: 0
positive_span: <
offset: 2
length: 1
>
positive_delta: 1`
}
return fmt.Sprintf(`
name: "%s"
help: "This is a histogram with default buckets"
type: HISTOGRAM
metric: <
label: <
name: "address"
value: "0.0.0.0"
>
label: <
name: "port"
value: "5001"
>
histogram: <
sample_count: 1
sample_sum: 10
%s
%s
>
>
`, name, classic, expo)
}
metricsTexts := map[string]struct {
text []string
contentType string
hasClassic bool
hasExponential bool
}{
"text": {
text: []string{
genTestCounterText("test_metric_1"),
genTestCounterText("test_metric_1_count"),
genTestCounterText("test_metric_1_sum"),
genTestCounterText("test_metric_1_bucket"),
genTestHistText("test_histogram_1"),
genTestCounterText("test_metric_2"),
genTestCounterText("test_metric_2_count"),
genTestCounterText("test_metric_2_sum"),
genTestCounterText("test_metric_2_bucket"),
genTestHistText("test_histogram_2"),
genTestCounterText("test_metric_3"),
genTestCounterText("test_metric_3_count"),
genTestCounterText("test_metric_3_sum"),
genTestCounterText("test_metric_3_bucket"),
genTestHistText("test_histogram_3"),
},
hasClassic: true,
},
"text, in different order": {
text: []string{
genTestCounterText("test_metric_1"),
genTestCounterText("test_metric_1_count"),
genTestCounterText("test_metric_1_sum"),
genTestCounterText("test_metric_1_bucket"),
genTestHistText("test_histogram_1"),
genTestCounterText("test_metric_2"),
genTestCounterText("test_metric_2_count"),
genTestCounterText("test_metric_2_sum"),
genTestCounterText("test_metric_2_bucket"),
genTestHistText("test_histogram_2"),
genTestHistText("test_histogram_3"),
genTestCounterText("test_metric_3"),
genTestCounterText("test_metric_3_count"),
genTestCounterText("test_metric_3_sum"),
genTestCounterText("test_metric_3_bucket"),
},
hasClassic: true,
},
"protobuf": {
text: []string{
genTestCounterProto("test_metric_1"),
genTestCounterProto("test_metric_1_count"),
genTestCounterProto("test_metric_1_sum"),
genTestCounterProto("test_metric_1_bucket"),
genTestHistProto("test_histogram_1", true, false),
genTestCounterProto("test_metric_2"),
genTestCounterProto("test_metric_2_count"),
genTestCounterProto("test_metric_2_sum"),
genTestCounterProto("test_metric_2_bucket"),
genTestHistProto("test_histogram_2", true, false),
genTestCounterProto("test_metric_3"),
genTestCounterProto("test_metric_3_count"),
genTestCounterProto("test_metric_3_sum"),
genTestCounterProto("test_metric_3_bucket"),
genTestHistProto("test_histogram_3", true, false),
},
contentType: "application/vnd.google.protobuf",
hasClassic: true,
},
"protobuf, in different order": {
text: []string{
genTestHistProto("test_histogram_1", true, false),
genTestCounterProto("test_metric_1"),
genTestCounterProto("test_metric_1_count"),
genTestCounterProto("test_metric_1_sum"),
genTestCounterProto("test_metric_1_bucket"),
genTestHistProto("test_histogram_2", true, false),
genTestCounterProto("test_metric_2"),
genTestCounterProto("test_metric_2_count"),
genTestCounterProto("test_metric_2_sum"),
genTestCounterProto("test_metric_2_bucket"),
genTestHistProto("test_histogram_3", true, false),
genTestCounterProto("test_metric_3"),
genTestCounterProto("test_metric_3_count"),
genTestCounterProto("test_metric_3_sum"),
genTestCounterProto("test_metric_3_bucket"),
},
contentType: "application/vnd.google.protobuf",
hasClassic: true,
},
"protobuf, with additional native exponential histogram": {
text: []string{
genTestCounterProto("test_metric_1"),
genTestCounterProto("test_metric_1_count"),
genTestCounterProto("test_metric_1_sum"),
genTestCounterProto("test_metric_1_bucket"),
genTestHistProto("test_histogram_1", true, true),
genTestCounterProto("test_metric_2"),
genTestCounterProto("test_metric_2_count"),
genTestCounterProto("test_metric_2_sum"),
genTestCounterProto("test_metric_2_bucket"),
genTestHistProto("test_histogram_2", true, true),
genTestCounterProto("test_metric_3"),
genTestCounterProto("test_metric_3_count"),
genTestCounterProto("test_metric_3_sum"),
genTestCounterProto("test_metric_3_bucket"),
genTestHistProto("test_histogram_3", true, true),
},
contentType: "application/vnd.google.protobuf",
hasClassic: true,
hasExponential: true,
},
"protobuf, with only native exponential histogram": {
text: []string{
genTestCounterProto("test_metric_1"),
genTestCounterProto("test_metric_1_count"),
genTestCounterProto("test_metric_1_sum"),
genTestCounterProto("test_metric_1_bucket"),
genTestHistProto("test_histogram_1", false, true),
genTestCounterProto("test_metric_2"),
genTestCounterProto("test_metric_2_count"),
genTestCounterProto("test_metric_2_sum"),
genTestCounterProto("test_metric_2_bucket"),
genTestHistProto("test_histogram_2", false, true),
genTestCounterProto("test_metric_3"),
genTestCounterProto("test_metric_3_count"),
genTestCounterProto("test_metric_3_sum"),
genTestCounterProto("test_metric_3_bucket"),
genTestHistProto("test_histogram_3", false, true),
},
contentType: "application/vnd.google.protobuf",
hasExponential: true,
},
}
checkBucketValues := func(t testing.TB, expectedCount int, series storage.SeriesSet) {
labelName := "le"
var expectedValues []string
if expectedCount > 0 {
expectedValues = []string{"0.005", "0.01", "0.025", "0.05", "0.1", "0.25", "0.5", "1.0", "2.5", "5.0", "10.0", "+Inf"}
}
foundLeValues := map[string]bool{}
for series.Next() {
s := series.At()
v := s.Labels().Get(labelName)
require.NotContains(t, foundLeValues, v, "duplicate label value found")
foundLeValues[v] = true
}
require.Len(t, foundLeValues, len(expectedValues), "unexpected number of label values, expected %v but found %v", expectedValues, foundLeValues)
for _, v := range expectedValues {
require.Contains(t, foundLeValues, v, "label value not found")
}
}
// Checks that the expected series is present and runs a basic sanity check of the float values.
checkFloatSeries := func(t testing.TB, series storage.SeriesSet, expectedCount int, expectedFloat float64) {
count := 0
for series.Next() {
i := series.At().Iterator(nil)
loop:
for {
switch i.Next() {
case chunkenc.ValNone:
break loop
case chunkenc.ValFloat:
_, f := i.At()
require.Equal(t, expectedFloat, f)
case chunkenc.ValHistogram:
panic("unexpected value type: histogram")
case chunkenc.ValFloatHistogram:
panic("unexpected value type: float histogram")
default:
panic("unexpected value type")
}
}
count++
}
require.Equal(t, expectedCount, count, "number of float series not as expected")
}
// Checks that the expected series is present and runs a basic sanity check of the histogram values.
checkHistSeries := func(t testing.TB, series storage.SeriesSet, expectedCount int, expectedSchema int32) {
count := 0
for series.Next() {
i := series.At().Iterator(nil)
loop:
for {
switch i.Next() {
case chunkenc.ValNone:
break loop
case chunkenc.ValFloat:
panic("unexpected value type: float")
case chunkenc.ValHistogram:
_, h := i.AtHistogram(nil)
require.Equal(t, expectedSchema, h.Schema)
require.Equal(t, uint64(1), h.Count)
require.Equal(t, 10.0, h.Sum)
case chunkenc.ValFloatHistogram:
_, h := i.AtFloatHistogram(nil)
require.Equal(t, expectedSchema, h.Schema)
require.Equal(t, uint64(1), h.Count)
require.Equal(t, 10.0, h.Sum)
default:
panic("unexpected value type")
}
}
count++
}
require.Equal(t, expectedCount, count, "number of histogram series not as expected")
}
for metricsTextName, metricsText := range metricsTexts {
for name, tc := range map[string]struct {
alwaysScrapeClassicHistograms bool
convertClassicHistToNHCB bool
}{
"convert with scrape": {
alwaysScrapeClassicHistograms: true,
convertClassicHistToNHCB: true,
},
"convert without scrape": {
alwaysScrapeClassicHistograms: false,
convertClassicHistToNHCB: true,
},
"scrape without convert": {
alwaysScrapeClassicHistograms: true,
convertClassicHistToNHCB: false,
},
"scrape with nil convert": {
alwaysScrapeClassicHistograms: true,
},
"neither scrape nor convert": {
alwaysScrapeClassicHistograms: false,
convertClassicHistToNHCB: false,
},
} {
var expectedClassicHistCount, expectedNativeHistCount int
var expectCustomBuckets bool
if metricsText.hasExponential {
expectedNativeHistCount = 1
expectCustomBuckets = false
expectedClassicHistCount = 0
if metricsText.hasClassic && tc.alwaysScrapeClassicHistograms {
expectedClassicHistCount = 1
}
} else if metricsText.hasClassic {
switch {
case !tc.convertClassicHistToNHCB:
expectedClassicHistCount = 1
expectedNativeHistCount = 0
case tc.alwaysScrapeClassicHistograms && tc.convertClassicHistToNHCB:
expectedClassicHistCount = 1
expectedNativeHistCount = 1
expectCustomBuckets = true
case !tc.alwaysScrapeClassicHistograms && tc.convertClassicHistToNHCB:
expectedClassicHistCount = 0
expectedNativeHistCount = 1
expectCustomBuckets = true
}
}
t.Run(fmt.Sprintf("%s with %s", name, metricsTextName), func(t *testing.T) {
t.Parallel()
s := teststorage.New(t)
sl, _ := newTestScrapeLoop(t, withAppendable(s, appV2), func(sl *scrapeLoop) {
sl.alwaysScrapeClassicHist = tc.alwaysScrapeClassicHistograms
sl.convertClassicHistToNHCB = tc.convertClassicHistToNHCB
sl.enableNativeHistogramScraping = true
})
var content []byte
contentType := metricsText.contentType
switch contentType {
case "application/vnd.google.protobuf":
buf := &bytes.Buffer{}
for _, text := range metricsText.text {
// In case of protobuf, we have to create the binary representation.
pb := &dto.MetricFamily{}
// From text to proto message.
require.NoError(t, proto.UnmarshalText(text, pb))
// From proto message to binary protobuf.
buf.Write(protoMarshalDelimited(t, pb))
}
content = buf.Bytes()
case "text/plain", "":
// The input text fragments already have a newline at the
// end, so we just concatenate them without separator.
content = []byte(strings.Join(metricsText.text, ""))
contentType = "text/plain"
default:
t.Error("unexpected content type")
}
now := time.Now()
app := sl.appender()
_, _, _, err := app.append(content, contentType, now)
require.NoError(t, err)
require.NoError(t, app.Commit())
var expectedSchema int32
if expectCustomBuckets {
expectedSchema = histogram.CustomBucketsSchema
} else {
expectedSchema = 3
}
// Validated what was appended can be queried.
ctx := t.Context()
q, err := s.Querier(time.Time{}.UnixNano(), time.Now().UnixNano())
require.NoError(t, err)
t.Cleanup(func() { _ = q.Close() })
var series storage.SeriesSet
for i := 1; i <= 3; i++ {
series = q.Select(ctx, false, nil, labels.MustNewMatcher(labels.MatchRegexp, "__name__", fmt.Sprintf("test_metric_%d", i)))
checkFloatSeries(t, series, 1, 1.)
series = q.Select(ctx, false, nil, labels.MustNewMatcher(labels.MatchRegexp, "__name__", fmt.Sprintf("test_metric_%d_count", i)))
checkFloatSeries(t, series, 1, 1.)
series = q.Select(ctx, false, nil, labels.MustNewMatcher(labels.MatchRegexp, "__name__", fmt.Sprintf("test_metric_%d_sum", i)))
checkFloatSeries(t, series, 1, 1.)
series = q.Select(ctx, false, nil, labels.MustNewMatcher(labels.MatchRegexp, "__name__", fmt.Sprintf("test_metric_%d_bucket", i)))
checkFloatSeries(t, series, 1, 1.)
series = q.Select(ctx, false, nil, labels.MustNewMatcher(labels.MatchRegexp, "__name__", fmt.Sprintf("test_histogram_%d_count", i)))
checkFloatSeries(t, series, expectedClassicHistCount, 1.)
series = q.Select(ctx, false, nil, labels.MustNewMatcher(labels.MatchRegexp, "__name__", fmt.Sprintf("test_histogram_%d_sum", i)))
checkFloatSeries(t, series, expectedClassicHistCount, 10.)
series = q.Select(ctx, false, nil, labels.MustNewMatcher(labels.MatchRegexp, "__name__", fmt.Sprintf("test_histogram_%d_bucket", i)))
checkBucketValues(t, expectedClassicHistCount, series)
series = q.Select(ctx, false, nil, labels.MustNewMatcher(labels.MatchRegexp, "__name__", fmt.Sprintf("test_histogram_%d", i)))
checkHistSeries(t, series, expectedNativeHistCount, expectedSchema)
}
})
}
}
}
func TestTypeUnitReLabel(t *testing.T) {
foreachAppendable(t, func(t *testing.T, appV2 bool) {
testTypeUnitReLabel(t, appV2)
})
}
func testTypeUnitReLabel(t *testing.T, appV2 bool) {
s := teststorage.New(t)
cfg := &config.ScrapeConfig{
JobName: "test",
MetricRelabelConfigs: []*relabel.Config{
{
SourceLabels: model.LabelNames{"__name__"},
Regex: relabel.MustNewRegexp(".*_total$"),
Replacement: "counter",
TargetLabel: "__type__",
Action: relabel.Replace,
NameValidationScheme: model.UTF8Validation,
},
{
SourceLabels: model.LabelNames{"__name__"},
Regex: relabel.MustNewRegexp(".*_bytes$"),
Replacement: "bytes",
TargetLabel: "__unit__",
Action: relabel.Replace,
NameValidationScheme: model.UTF8Validation,
},
},
SampleLimit: 100,
Scheme: "http",
ScrapeInterval: model.Duration(100 * time.Millisecond),
ScrapeTimeout: model.Duration(100 * time.Millisecond),
MetricNameValidationScheme: model.UTF8Validation,
MetricNameEscapingScheme: model.AllowUTF8,
}
metricsText := `
# HELP test_metric_1_total This is a counter
# TYPE test_metric_1_total counter
test_metric_1_total 123
# HELP test_metric_2_total This is a counter
# TYPE test_metric_2_total counter
test_metric_2_total 234
# HELP disk_usage_bytes This is a gauge
# TYPE disk_usage_bytes gauge
disk_usage_bytes 456
`
ts, scrapedTwice := newScrapableServer(metricsText)
defer ts.Close()
sa := selectAppendable(s, appV2)
sp, err := newScrapePool(cfg, sa.V1(), sa.V2(), 0, nil, nil, &Options{}, newTestScrapeMetrics(t))
require.NoError(t, err)
defer sp.stop()
testURL, err := url.Parse(ts.URL)
require.NoError(t, err)
sp.Sync([]*targetgroup.Group{
{
Targets: []model.LabelSet{{model.AddressLabel: model.LabelValue(testURL.Host)}},
},
})
require.Len(t, sp.ActiveTargets(), 1)
select {
case <-time.After(5 * time.Second):
t.Fatalf("target was not scraped")
case <-scrapedTwice:
}
ctx := t.Context()
q, err := s.Querier(time.Time{}.UnixNano(), time.Now().UnixNano())
require.NoError(t, err)
t.Cleanup(func() { _ = q.Close() })
series := q.Select(ctx, false, nil, labels.MustNewMatcher(labels.MatchRegexp, "__name__", ".*_total$"))
for series.Next() {
s := series.At()
require.Equal(t, "counter", s.Labels().Get("__type__"))
}
series = q.Select(ctx, false, nil, labels.MustNewMatcher(labels.MatchRegexp, "__name__", "disk_usage_bytes"))
for series.Next() {
s := series.At()
require.Equal(t, "bytes", s.Labels().Get("__unit__"))
}
}
func TestScrapeLoopRunCreatesStaleMarkersOnFailedScrapeForTimestampedMetrics(t *testing.T) {
foreachAppendable(t, func(t *testing.T, appV2 bool) {
testScrapeLoopRunCreatesStaleMarkersOnFailedScrapeForTimestampedMetrics(t, appV2)
})
}
func testScrapeLoopRunCreatesStaleMarkersOnFailedScrapeForTimestampedMetrics(t *testing.T, appV2 bool) {
signal := make(chan struct{}, 1)
ctx, cancel := context.WithCancel(t.Context())
appTest := teststorage.NewAppendable()
sl, scraper := newTestScrapeLoop(t, withAppendable(appTest, appV2), func(sl *scrapeLoop) {
sl.ctx = ctx // Since we're writing samples directly below we need to provide a protocol fallback.
sl.fallbackScrapeProtocol = "text/plain"
sl.trackTimestampsStaleness = true
})
// Succeed once, several failures, then stop.
numScrapes := 0
scraper.scrapeFunc = func(_ context.Context, w io.Writer) error {
numScrapes++
switch numScrapes {
case 1:
_, _ = fmt.Fprintf(w, "metric_a 42 %d\n", time.Now().UnixNano()/int64(time.Millisecond))
return nil
case 5:
cancel()
}
return errors.New("scrape failed")
}
go func() {
sl.run(nil)
signal <- struct{}{}
}()
select {
case <-signal:
case <-time.After(5 * time.Second):
t.Fatalf("Scrape wasn't stopped.")
}
got := appTest.ResultSamples()
// 1 successfully scraped sample, 1 stale marker after first fail, 5 report samples for
// each scrape successful or not.
require.Len(t, got, 27, "Appended samples not as expected:\n%s", appTest)
require.Equal(t, 42.0, got[0].V, "Appended first sample not as expected")
require.True(t, value.IsStaleNaN(got[6].V),
"Appended second sample not as expected. Wanted: stale NaN Got: %x", math.Float64bits(got[6].V))
}
func TestScrapeLoopCompression(t *testing.T) {
foreachAppendable(t, func(t *testing.T, appV2 bool) {
testScrapeLoopCompression(t, appV2)
})
}
func testScrapeLoopCompression(t *testing.T, appV2 bool) {
s := teststorage.New(t)
metricsText := makeTestGauges(10)
for _, tc := range []struct {
enableCompression bool
acceptEncoding string
}{
{
enableCompression: true,
acceptEncoding: "gzip",
},
{
enableCompression: false,
acceptEncoding: "identity",
},
} {
t.Run(fmt.Sprintf("compression=%v,acceptEncoding=%s", tc.enableCompression, tc.acceptEncoding), func(t *testing.T) {
scraped := make(chan bool)
ts := httptest.NewServer(http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {
require.Equal(t, tc.acceptEncoding, r.Header.Get("Accept-Encoding"), "invalid value of the Accept-Encoding header")
_, _ = fmt.Fprint(w, string(metricsText))
close(scraped)
}))
defer ts.Close()
cfg := &config.ScrapeConfig{
JobName: "test",
SampleLimit: 100,
Scheme: "http",
ScrapeInterval: model.Duration(100 * time.Millisecond),
ScrapeTimeout: model.Duration(100 * time.Millisecond),
EnableCompression: tc.enableCompression,
MetricNameValidationScheme: model.UTF8Validation,
MetricNameEscapingScheme: model.AllowUTF8,
}
sa := selectAppendable(s, appV2)
sp, err := newScrapePool(cfg, sa.V1(), sa.V2(), 0, nil, nil, &Options{}, newTestScrapeMetrics(t))
require.NoError(t, err)
defer sp.stop()
testURL, err := url.Parse(ts.URL)
require.NoError(t, err)
sp.Sync([]*targetgroup.Group{
{
Targets: []model.LabelSet{{model.AddressLabel: model.LabelValue(testURL.Host)}},
},
})
require.Len(t, sp.ActiveTargets(), 1)
select {
case <-time.After(5 * time.Second):
t.Fatalf("target was not scraped")
case <-scraped:
}
})
}
}
func TestPickSchema(t *testing.T) {
tcs := []struct {
factor float64
schema int32
}{
{
factor: 65536,
schema: -4,
},
{
factor: 256,
schema: -3,
},
{
factor: 16,
schema: -2,
},
{
factor: 4,
schema: -1,
},
{
factor: 2,
schema: 0,
},
{
factor: 1.4,
schema: 1,
},
{
factor: 1.1,
schema: 2,
},
{
factor: 1.09,
schema: 3,
},
{
factor: 1.04,
schema: 4,
},
{
factor: 1.02,
schema: 5,
},
{
factor: 1.01,
schema: 6,
},
{
factor: 1.005,
schema: 7,
},
{
factor: 1.002,
schema: 8,
},
// The default value of native_histogram_min_bucket_factor
{
factor: 0,
schema: 8,
},
}
for _, tc := range tcs {
schema := pickSchema(tc.factor)
require.Equal(t, tc.schema, schema)
}
}
func BenchmarkTargetScraperGzip(b *testing.B) {
scenarios := []struct {
metricsCount int
body []byte
}{
{metricsCount: 1},
{metricsCount: 100},
{metricsCount: 1000},
{metricsCount: 10000},
{metricsCount: 100000},
}
for i := range scenarios {
var buf bytes.Buffer
var name string
gw := gzip.NewWriter(&buf)
for j := 0; j < scenarios[i].metricsCount; j++ {
name = fmt.Sprintf("go_memstats_alloc_bytes_total_%d", j)
_, _ = fmt.Fprintf(gw, "# HELP %s Total number of bytes allocated, even if freed.\n", name)
_, _ = fmt.Fprintf(gw, "# TYPE %s counter\n", name)
_, _ = fmt.Fprintf(gw, "%s %d\n", name, i*j)
}
require.NoError(b, gw.Close())
scenarios[i].body = buf.Bytes()
}
handler := http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {
w.Header().Set("Content-Type", `text/plain; version=0.0.4`)
w.Header().Set("Content-Encoding", "gzip")
for _, scenario := range scenarios {
if strconv.Itoa(scenario.metricsCount) == r.URL.Query()["count"][0] {
_, _ = w.Write(scenario.body)
return
}
}
w.WriteHeader(http.StatusBadRequest)
})
server := httptest.NewServer(handler)
defer server.Close()
serverURL, err := url.Parse(server.URL)
if err != nil {
panic(err)
}
client, err := config_util.NewClientFromConfig(config_util.DefaultHTTPClientConfig, "test_job")
if err != nil {
panic(err)
}
for _, scenario := range scenarios {
b.Run(fmt.Sprintf("metrics=%d", scenario.metricsCount), func(b *testing.B) {
ts := &targetScraper{
Target: &Target{
labels: labels.FromStrings(
model.SchemeLabel, serverURL.Scheme,
model.AddressLabel, serverURL.Host,
),
scrapeConfig: &config.ScrapeConfig{
MetricNameValidationScheme: model.UTF8Validation,
MetricNameEscapingScheme: model.AllowUTF8,
Params: url.Values{"count": []string{strconv.Itoa(scenario.metricsCount)}},
},
},
client: client,
timeout: time.Second,
}
b.ResetTimer()
for b.Loop() {
_, err = ts.scrape(context.Background())
require.NoError(b, err)
}
})
}
}
// When a scrape contains multiple instances for the same time series we should increment
// prometheus_target_scrapes_sample_duplicate_timestamp_total metric.
func TestScrapeLoopSeriesAddedDuplicates(t *testing.T) {
foreachAppendable(t, func(t *testing.T, appV2 bool) {
testScrapeLoopSeriesAddedDuplicates(t, appV2)
})
}
func testScrapeLoopSeriesAddedDuplicates(t *testing.T, appV2 bool) {
sl, _ := newTestScrapeLoop(t, withAppendable(teststorage.NewAppendable(), appV2))
app := sl.appender()
total, added, seriesAdded, err := app.append([]byte("test_metric 1\ntest_metric 2\ntest_metric 3\n"), "text/plain", time.Time{})
require.NoError(t, err)
require.NoError(t, app.Commit())
require.Equal(t, 3, total)
require.Equal(t, 3, added)
require.Equal(t, 1, seriesAdded)
require.Equal(t, 2.0, prom_testutil.ToFloat64(sl.metrics.targetScrapeSampleDuplicate))
app = sl.appender()
total, added, seriesAdded, err = app.append([]byte("test_metric 1\ntest_metric 1\ntest_metric 1\n"), "text/plain", time.Time{})
require.NoError(t, err)
require.NoError(t, app.Commit())
require.Equal(t, 3, total)
require.Equal(t, 3, added)
require.Equal(t, 0, seriesAdded)
require.Equal(t, 4.0, prom_testutil.ToFloat64(sl.metrics.targetScrapeSampleDuplicate))
// When different timestamps are supplied, multiple samples are accepted.
app = sl.appender()
total, added, seriesAdded, err = app.append([]byte("test_metric 1 1001\ntest_metric 1 1002\ntest_metric 1 1003\n"), "text/plain", time.Time{})
require.NoError(t, err)
require.NoError(t, app.Commit())
require.Equal(t, 3, total)
require.Equal(t, 3, added)
require.Equal(t, 0, seriesAdded)
// Metric is not higher than last time.
require.Equal(t, 4.0, prom_testutil.ToFloat64(sl.metrics.targetScrapeSampleDuplicate))
}
// This tests running a full scrape loop and checking that the scrape option
// `native_histogram_min_bucket_factor` is used correctly.
func TestNativeHistogramMaxSchemaSet(t *testing.T) {
foreachAppendable(t, func(t *testing.T, appV2 bool) {
for _, tc := range []struct {
name string
minBucketFactor string
expectedSchema int32
}{
{
name: "min factor not specified",
minBucketFactor: "",
expectedSchema: 3, // Factor 1.09.
},
{
name: "min factor 1",
minBucketFactor: "native_histogram_min_bucket_factor: 1",
expectedSchema: 3, // Factor 1.09.
},
{
name: "min factor 2",
minBucketFactor: "native_histogram_min_bucket_factor: 2",
expectedSchema: 0, // Factor 2.00.
},
} {
t.Run(tc.name, func(t *testing.T) {
t.Parallel()
testNativeHistogramMaxSchemaSet(t, tc.minBucketFactor, tc.expectedSchema, appV2)
})
}
})
}
func testNativeHistogramMaxSchemaSet(t *testing.T, minBucketFactor string, expectedSchema int32, appV2 bool) {
// Create a ProtoBuf message to serve as a Prometheus metric.
nativeHistogram := prometheus.NewHistogram(
prometheus.HistogramOpts{
Namespace: "testing",
Name: "example_native_histogram",
Help: "This is used for testing",
NativeHistogramBucketFactor: 1.1,
NativeHistogramMaxBucketNumber: 100,
},
)
registry := prometheus.NewRegistry()
require.NoError(t, registry.Register(nativeHistogram))
nativeHistogram.Observe(1.0)
nativeHistogram.Observe(1.0)
nativeHistogram.Observe(1.0)
nativeHistogram.Observe(10.0) // in different bucket since > 1*1.1.
nativeHistogram.Observe(10.0)
gathered, err := registry.Gather()
require.NoError(t, err)
require.NotEmpty(t, gathered)
histogramMetricFamily := gathered[0]
buffer := protoMarshalDelimited(t, histogramMetricFamily)
// Create an HTTP server to serve /metrics via ProtoBuf
metricsServer := httptest.NewServer(http.HandlerFunc(func(w http.ResponseWriter, _ *http.Request) {
w.Header().Set("Content-Type", `application/vnd.google.protobuf; proto=io.prometheus.client.MetricFamily; encoding=delimited`)
_, _ = w.Write(buffer)
}))
defer metricsServer.Close()
// Create a scrape loop with the HTTP server as the target.
configStr := fmt.Sprintf(`
global:
metric_name_validation_scheme: legacy
scrape_interval: 50ms
scrape_timeout: 25ms
scrape_configs:
- job_name: test
scrape_native_histograms: true
%s
static_configs:
- targets: [%s]
`, minBucketFactor, strings.ReplaceAll(metricsServer.URL, "http://", ""))
s := teststorage.New(t)
reg := prometheus.NewRegistry()
sa := selectAppendable(s, appV2)
mng, err := NewManager(&Options{DiscoveryReloadInterval: model.Duration(10 * time.Millisecond)}, nil, nil, sa.V1(), sa.V2(), reg)
require.NoError(t, err)
cfg, err := config.Load(configStr, promslog.NewNopLogger())
require.NoError(t, err)
require.NoError(t, mng.ApplyConfig(cfg))
tsets := make(chan map[string][]*targetgroup.Group)
go func() {
require.NoError(t, mng.Run(tsets))
}()
defer mng.Stop()
// Get the static targets and apply them to the scrape manager.
require.Len(t, cfg.ScrapeConfigs, 1)
scrapeCfg := cfg.ScrapeConfigs[0]
require.Len(t, scrapeCfg.ServiceDiscoveryConfigs, 1)
staticDiscovery, ok := scrapeCfg.ServiceDiscoveryConfigs[0].(discovery.StaticConfig)
require.True(t, ok)
require.Len(t, staticDiscovery, 1)
tsets <- map[string][]*targetgroup.Group{"test": staticDiscovery}
// Wait for the scrape loop to scrape the target.
require.Eventually(t, func() bool {
q, err := s.Querier(0, math.MaxInt64)
require.NoError(t, err)
seriesS := q.Select(context.Background(), false, nil, labels.MustNewMatcher(labels.MatchEqual, "__name__", "testing_example_native_histogram"))
countSeries := 0
for seriesS.Next() {
countSeries++
}
return countSeries > 0
}, 5*time.Second, 100*time.Millisecond)
// Check that native histogram schema is as expected.
q, err := s.Querier(0, math.MaxInt64)
require.NoError(t, err)
seriesS := q.Select(context.Background(), false, nil, labels.MustNewMatcher(labels.MatchEqual, "__name__", "testing_example_native_histogram"))
var histogramSamples []*histogram.Histogram
for seriesS.Next() {
series := seriesS.At()
it := series.Iterator(nil)
for vt := it.Next(); vt != chunkenc.ValNone; vt = it.Next() {
if vt != chunkenc.ValHistogram {
// don't care about other samples
continue
}
_, h := it.AtHistogram(nil)
histogramSamples = append(histogramSamples, h)
}
}
require.NoError(t, seriesS.Err())
require.NotEmpty(t, histogramSamples)
for _, h := range histogramSamples {
require.Equal(t, expectedSchema, h.Schema)
}
}
func TestTargetScrapeConfigWithLabels(t *testing.T) {
foreachAppendable(t, func(t *testing.T, appV2 bool) {
testTargetScrapeConfigWithLabels(t, appV2)
})
}
func testTargetScrapeConfigWithLabels(t *testing.T, appV2 bool) {
t.Parallel()
const (
configTimeout = 1500 * time.Millisecond
expectedTimeout = "1.5"
expectedTimeoutLabel = "1s500ms"
secondTimeout = 500 * time.Millisecond
secondTimeoutLabel = "500ms"
expectedParam = "value1"
secondParam = "value2"
expectedPath = "/metric-ok"
secondPath = "/metric-nok"
httpScheme = "http"
paramLabel = "__param_param"
jobName = "test"
)
createTestServer := func(t *testing.T, done chan struct{}) *url.URL {
server := httptest.NewServer(
http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {
defer close(done)
require.Equal(t, expectedTimeout, r.Header.Get("X-Prometheus-Scrape-Timeout-Seconds"))
require.Equal(t, expectedParam, r.URL.Query().Get("param"))
require.Equal(t, expectedPath, r.URL.Path)
w.Header().Set("Content-Type", `text/plain; version=0.0.4`)
_, _ = w.Write([]byte("metric_a 1\nmetric_b 2\n"))
}),
)
t.Cleanup(server.Close)
serverURL, err := url.Parse(server.URL)
require.NoError(t, err)
return serverURL
}
run := func(t *testing.T, cfg *config.ScrapeConfig, targets []*targetgroup.Group) chan struct{} {
done := make(chan struct{})
srvURL := createTestServer(t, done)
// Update target addresses to use the dynamically created server URL.
for _, target := range targets {
for i := range target.Targets {
target.Targets[i][model.AddressLabel] = model.LabelValue(srvURL.Host)
}
}
sa := selectAppendable(teststorage.NewAppendable(), appV2)
sp, err := newScrapePool(cfg, sa.V1(), sa.V2(), 0, nil, nil, &Options{}, newTestScrapeMetrics(t))
require.NoError(t, err)
t.Cleanup(sp.stop)
sp.Sync(targets)
return done
}
cases := []struct {
name string
cfg *config.ScrapeConfig
targets []*targetgroup.Group
}{
{
name: "Everything in scrape config",
cfg: &config.ScrapeConfig{
ScrapeInterval: model.Duration(2 * time.Second),
ScrapeTimeout: model.Duration(configTimeout),
Params: url.Values{"param": []string{expectedParam}},
JobName: jobName,
Scheme: httpScheme,
MetricsPath: expectedPath,
MetricNameValidationScheme: model.UTF8Validation,
MetricNameEscapingScheme: model.AllowUTF8,
},
targets: []*targetgroup.Group{
{
Targets: []model.LabelSet{
{model.AddressLabel: model.LabelValue("")},
},
},
},
},
{
name: "Overridden in target",
cfg: &config.ScrapeConfig{
ScrapeInterval: model.Duration(2 * time.Second),
ScrapeTimeout: model.Duration(secondTimeout),
JobName: jobName,
Scheme: httpScheme,
MetricsPath: secondPath,
MetricNameValidationScheme: model.UTF8Validation,
MetricNameEscapingScheme: model.AllowUTF8,
Params: url.Values{"param": []string{secondParam}},
},
targets: []*targetgroup.Group{
{
Targets: []model.LabelSet{
{
model.AddressLabel: model.LabelValue(""),
model.ScrapeTimeoutLabel: expectedTimeoutLabel,
model.MetricsPathLabel: expectedPath,
paramLabel: expectedParam,
},
},
},
},
},
{
name: "Overridden in relabel_config",
cfg: &config.ScrapeConfig{
ScrapeInterval: model.Duration(2 * time.Second),
ScrapeTimeout: model.Duration(secondTimeout),
JobName: jobName,
Scheme: httpScheme,
MetricsPath: secondPath,
MetricNameValidationScheme: model.UTF8Validation,
MetricNameEscapingScheme: model.AllowUTF8,
Params: url.Values{"param": []string{secondParam}},
RelabelConfigs: []*relabel.Config{
{
Action: relabel.DefaultRelabelConfig.Action,
Regex: relabel.DefaultRelabelConfig.Regex,
SourceLabels: relabel.DefaultRelabelConfig.SourceLabels,
TargetLabel: model.ScrapeTimeoutLabel,
Replacement: expectedTimeoutLabel,
NameValidationScheme: model.UTF8Validation,
},
{
Action: relabel.DefaultRelabelConfig.Action,
Regex: relabel.DefaultRelabelConfig.Regex,
SourceLabels: relabel.DefaultRelabelConfig.SourceLabels,
TargetLabel: paramLabel,
Replacement: expectedParam,
NameValidationScheme: model.UTF8Validation,
},
{
Action: relabel.DefaultRelabelConfig.Action,
Regex: relabel.DefaultRelabelConfig.Regex,
SourceLabels: relabel.DefaultRelabelConfig.SourceLabels,
TargetLabel: model.MetricsPathLabel,
Replacement: expectedPath,
NameValidationScheme: model.UTF8Validation,
},
},
},
targets: []*targetgroup.Group{
{
Targets: []model.LabelSet{
{
model.AddressLabel: model.LabelValue(""),
model.ScrapeTimeoutLabel: secondTimeoutLabel,
model.MetricsPathLabel: secondPath,
paramLabel: secondParam,
},
},
},
},
},
}
for _, c := range cases {
t.Run(c.name, func(t *testing.T) {
t.Parallel()
select {
case <-run(t, c.cfg, c.targets):
case <-time.After(10 * time.Second):
t.Fatal("timeout after 10 seconds")
}
})
}
}
func newScrapableServer(scrapeText string) (s *httptest.Server, scrapedTwice chan bool) {
var scrapes int
scrapedTwice = make(chan bool)
return httptest.NewServer(http.HandlerFunc(func(w http.ResponseWriter, _ *http.Request) {
_, _ = fmt.Fprint(w, scrapeText)
scrapes++
if scrapes == 2 {
close(scrapedTwice)
}
})), scrapedTwice
}
// Regression test for the panic fixed in https://github.com/prometheus/prometheus/pull/15523.
func TestScrapePoolScrapeAfterReload(t *testing.T) {
foreachAppendable(t, func(t *testing.T, appV2 bool) {
testScrapePoolScrapeAfterReload(t, appV2)
})
}
func testScrapePoolScrapeAfterReload(t *testing.T, appV2 bool) {
h := httptest.NewServer(http.HandlerFunc(
func(w http.ResponseWriter, _ *http.Request) {
_, _ = w.Write([]byte{0x42, 0x42})
},
))
t.Cleanup(h.Close)
cfg := &config.ScrapeConfig{
BodySizeLimit: 1,
JobName: "test",
Scheme: "http",
ScrapeInterval: model.Duration(100 * time.Millisecond),
ScrapeTimeout: model.Duration(100 * time.Millisecond),
MetricNameValidationScheme: model.UTF8Validation,
MetricNameEscapingScheme: model.AllowUTF8,
EnableCompression: false,
ServiceDiscoveryConfigs: discovery.Configs{
&discovery.StaticConfig{
{
Targets: []model.LabelSet{{model.AddressLabel: model.LabelValue(h.URL)}},
},
},
},
}
sa := selectAppendable(teststorage.NewAppendable(), appV2)
p, err := newScrapePool(cfg, sa.V1(), sa.V2(), 0, nil, nil, &Options{}, newTestScrapeMetrics(t))
require.NoError(t, err)
t.Cleanup(p.stop)
p.Sync([]*targetgroup.Group{
{
Targets: []model.LabelSet{{model.AddressLabel: model.LabelValue(strings.TrimPrefix(h.URL, "http://"))}},
Source: "test",
},
})
require.NoError(t, p.reload(cfg))
<-time.After(1 * time.Second)
}
// Regression test against https://github.com/prometheus/prometheus/issues/16160.
// The first scrape fails with a parsing error, but the second should
// succeed and cause `metric_1=11` to appear in the appender.
func TestScrapeAppendWithParseError(t *testing.T) {
foreachAppendable(t, func(t *testing.T, appV2 bool) {
testScrapeAppendWithParseError(t, appV2)
})
}
func testScrapeAppendWithParseError(t *testing.T, appV2 bool) {
const (
scrape1 = `metric_a 1
`
scrape2 = `metric_a 11
# EOF`
)
appTest := teststorage.NewAppendable()
sl, _ := newTestScrapeLoop(t, withAppendable(appTest, appV2))
now := time.Now()
app := sl.appender()
_, _, _, err := app.append([]byte(scrape1), "application/openmetrics-text", now)
require.Error(t, err)
require.NoError(t, app.Rollback())
app = sl.appender()
_, _, _, err = app.append(nil, "application/openmetrics-text", now)
require.NoError(t, err)
require.NoError(t, app.Commit())
require.Empty(t, appTest.ResultSamples())
app = sl.appender()
_, _, _, err = app.append([]byte(scrape2), "application/openmetrics-text", now.Add(15*time.Second))
require.NoError(t, err)
require.NoError(t, app.Commit())
want := []sample{
{
L: labels.FromStrings(model.MetricNameLabel, "metric_a"),
T: timestamp.FromTime(now.Add(15 * time.Second)),
V: 11,
},
}
teststorage.RequireEqual(t, want, appTest.ResultSamples(), "Appended samples not as expected:\n%s", appTest)
}
// This test covers a case where there's a target with sample_limit set and some samples
// changes between scrapes.
func TestScrapeLoopAppendSampleLimitWithDisappearingSeries(t *testing.T) {
foreachAppendable(t, func(t *testing.T, appV2 bool) {
testScrapeLoopAppendSampleLimitWithDisappearingSeries(t, appV2)
})
}
func testScrapeLoopAppendSampleLimitWithDisappearingSeries(t *testing.T, appV2 bool) {
const sampleLimit = 4
appTest := teststorage.NewAppendable()
sl, _ := newTestScrapeLoop(t, withAppendable(appTest, appV2), func(sl *scrapeLoop) {
sl.sampleLimit = sampleLimit
})
now := time.Now()
app := sl.appender()
samplesScraped, samplesAfterRelabel, createdSeries, err := app.append(
// Start with 3 samples, all accepted.
[]byte("metric_a 1\nmetric_b 1\nmetric_c 1\n"),
"text/plain",
now,
)
require.NoError(t, err)
require.NoError(t, app.Commit())
require.Equal(t, 3, samplesScraped) // All on scrape.
require.Equal(t, 3, samplesAfterRelabel) // This is series after relabeling.
require.Equal(t, 3, createdSeries) // Newly added to TSDB.
want := []sample{
{
L: labels.FromStrings(model.MetricNameLabel, "metric_a"),
T: timestamp.FromTime(now),
V: 1,
},
{
L: labels.FromStrings(model.MetricNameLabel, "metric_b"),
T: timestamp.FromTime(now),
V: 1,
},
{
L: labels.FromStrings(model.MetricNameLabel, "metric_c"),
T: timestamp.FromTime(now),
V: 1,
},
}
teststorage.RequireEqual(t, want, appTest.ResultSamples(), "Appended samples not as expected:\n%s", app)
now = now.Add(time.Minute)
app = sl.appender()
samplesScraped, samplesAfterRelabel, createdSeries, err = app.append(
// Start exporting 3 more samples, so we're over the limit now.
[]byte("metric_a 1\nmetric_b 1\nmetric_c 1\nmetric_d 1\nmetric_e 1\nmetric_f 1\n"),
"text/plain",
now,
)
require.ErrorIs(t, err, errSampleLimit)
require.NoError(t, app.Rollback())
require.Equal(t, 6, samplesScraped)
require.Equal(t, 6, samplesAfterRelabel)
require.Equal(t, 1, createdSeries) // We've added one series before hitting the limit.
testutil.RequireEqual(t, want, appTest.ResultSamples(), "Appended samples not as expected:\n%s", app)
sl.cache.iterDone(false)
now = now.Add(time.Minute)
app = sl.appender()
samplesScraped, samplesAfterRelabel, createdSeries, err = app.append(
// Remove all samples except first 2.
[]byte("metric_a 1\nmetric_b 1\n"),
"text/plain",
now,
)
require.NoError(t, err)
require.NoError(t, app.Commit())
require.Equal(t, 2, samplesScraped)
require.Equal(t, 2, samplesAfterRelabel)
require.Equal(t, 0, createdSeries)
// This is where important things happen. We should now see:
// - Appends for samples from metric_a & metric_b.
// - Append with stale markers for metric_c - this series was added during first scrape but disappeared during last scrape.
// - Append with stale marker for metric_d - this series was added during second scrape before we hit the sample_limit.
// We should NOT see:
// - Appends with stale markers for metric_e & metric_f - both over the limit during second scrape, and so they never made it into TSDB.
want = append(want, []sample{
{
L: labels.FromStrings(model.MetricNameLabel, "metric_a"),
T: timestamp.FromTime(now),
V: 1,
},
{
L: labels.FromStrings(model.MetricNameLabel, "metric_b"),
T: timestamp.FromTime(now),
V: 1,
},
{
L: labels.FromStrings(model.MetricNameLabel, "metric_c"),
T: timestamp.FromTime(now),
V: math.Float64frombits(value.StaleNaN),
},
{
L: labels.FromStrings(model.MetricNameLabel, "metric_d"),
T: timestamp.FromTime(now),
V: math.Float64frombits(value.StaleNaN),
},
}...)
teststorage.RequireEqual(t, want, appTest.ResultSamples(), "Appended samples not as expected:\n%s", appTest)
}
// This test covers a case where there's a target with sample_limit set and each scrape sees a completely
// different set of samples.
func TestScrapeLoopAppendSampleLimitReplaceAllSamples(t *testing.T) {
foreachAppendable(t, func(t *testing.T, appV2 bool) {
testScrapeLoopAppendSampleLimitReplaceAllSamples(t, appV2)
})
}
func testScrapeLoopAppendSampleLimitReplaceAllSamples(t *testing.T, appV2 bool) {
const sampleLimit = 4
appTest := teststorage.NewAppendable()
sl, _ := newTestScrapeLoop(t, withAppendable(appTest, appV2), func(sl *scrapeLoop) {
sl.sampleLimit = sampleLimit
})
now := time.Now()
app := sl.appender()
samplesScraped, samplesAfterRelabel, createdSeries, err := app.append(
// Start with 4 samples, all accepted.
[]byte("metric_a 1\nmetric_b 1\nmetric_c 1\nmetric_d 1\n"),
"text/plain",
now,
)
require.NoError(t, err)
require.NoError(t, app.Commit())
require.Equal(t, 4, samplesScraped) // All on scrape.
require.Equal(t, 4, samplesAfterRelabel) // This is series after relabeling.
require.Equal(t, 4, createdSeries) // Newly added to TSDB.
want := []sample{
{
L: labels.FromStrings(model.MetricNameLabel, "metric_a"),
T: timestamp.FromTime(now),
V: 1,
},
{
L: labels.FromStrings(model.MetricNameLabel, "metric_b"),
T: timestamp.FromTime(now),
V: 1,
},
{
L: labels.FromStrings(model.MetricNameLabel, "metric_c"),
T: timestamp.FromTime(now),
V: 1,
},
{
L: labels.FromStrings(model.MetricNameLabel, "metric_d"),
T: timestamp.FromTime(now),
V: 1,
},
}
teststorage.RequireEqual(t, want, appTest.ResultSamples(), "Appended samples not as expected:\n%s", app)
now = now.Add(time.Minute)
app = sl.appender()
samplesScraped, samplesAfterRelabel, createdSeries, err = app.append(
// Replace all samples with new time series.
[]byte("metric_e 1\nmetric_f 1\nmetric_g 1\nmetric_h 1\n"),
"text/plain",
now,
)
require.NoError(t, err)
require.NoError(t, app.Commit())
require.Equal(t, 4, samplesScraped)
require.Equal(t, 4, samplesAfterRelabel)
require.Equal(t, 4, createdSeries)
// We replaced all samples from first scrape with new set of samples.
// We expected to see:
// - 4 appends for new samples.
// - 4 appends with staleness markers for old samples.
want = append(want, []sample{
{
L: labels.FromStrings(model.MetricNameLabel, "metric_e"),
T: timestamp.FromTime(now),
V: 1,
},
{
L: labels.FromStrings(model.MetricNameLabel, "metric_f"),
T: timestamp.FromTime(now),
V: 1,
},
{
L: labels.FromStrings(model.MetricNameLabel, "metric_g"),
T: timestamp.FromTime(now),
V: 1,
},
{
L: labels.FromStrings(model.MetricNameLabel, "metric_h"),
T: timestamp.FromTime(now),
V: 1,
},
{
L: labels.FromStrings(model.MetricNameLabel, "metric_a"),
T: timestamp.FromTime(now),
V: math.Float64frombits(value.StaleNaN),
},
{
L: labels.FromStrings(model.MetricNameLabel, "metric_b"),
T: timestamp.FromTime(now),
V: math.Float64frombits(value.StaleNaN),
},
{
L: labels.FromStrings(model.MetricNameLabel, "metric_c"),
T: timestamp.FromTime(now),
V: math.Float64frombits(value.StaleNaN),
},
{
L: labels.FromStrings(model.MetricNameLabel, "metric_d"),
T: timestamp.FromTime(now),
V: math.Float64frombits(value.StaleNaN),
},
}...)
teststorage.RequireEqual(t, want, appTest.ResultSamples(), "Appended samples not as expected:\n%s", app)
}
func TestScrapeLoopDisableStalenessMarkerInjection(t *testing.T) {
foreachAppendable(t, func(t *testing.T, appV2 bool) {
testScrapeLoopDisableStalenessMarkerInjection(t, appV2)
})
}
func testScrapeLoopDisableStalenessMarkerInjection(t *testing.T, appV2 bool) {
loopDone := atomic.NewBool(false)
appTest := teststorage.NewAppendable()
sl, scraper := newTestScrapeLoop(t, withAppendable(appTest, appV2))
scraper.scrapeFunc = func(ctx context.Context, w io.Writer) error {
if _, err := w.Write([]byte("metric_a 42\n")); err != nil {
return err
}
return ctx.Err()
}
// Start the scrape loop.
go func() {
sl.run(nil)
loopDone.Store(true)
}()
// Wait for some samples to be appended.
require.Eventually(t, func() bool {
return len(appTest.ResultSamples()) > 2
}, 5*time.Second, 100*time.Millisecond, "Scrape loop didn't append any samples.")
// Disable end of run staleness markers and stop the loop.
sl.disableEndOfRunStalenessMarkers()
sl.stop()
require.Eventually(t, func() bool {
return loopDone.Load()
}, 5*time.Second, 100*time.Millisecond, "Scrape loop didn't stop.")
// No stale markers should be appended, since they were disabled.
for _, s := range appTest.ResultSamples() {
if value.IsStaleNaN(s.V) {
t.Fatalf("Got stale NaN samples while end of run staleness is disabled: %x", math.Float64bits(s.V))
}
}
}
// Recommended CLI invocation:
/*
export bench=restartLoops && go test ./scrape/... \
-run '^$' -bench '^BenchmarkScrapePoolRestartLoops' \
-benchtime 5s -count 6 -cpu 2 -timeout 999m \
| tee ${bench}.txt
*/
func BenchmarkScrapePoolRestartLoops(b *testing.B) {
sp, err := newScrapePool(
&config.ScrapeConfig{
MetricNameValidationScheme: model.UTF8Validation,
ScrapeInterval: model.Duration(1 * time.Hour),
ScrapeTimeout: model.Duration(1 * time.Hour),
},
nil,
nil,
0,
nil,
nil,
&Options{},
newTestScrapeMetrics(b),
)
require.NoError(b, err)
b.Cleanup(sp.stop)
for i := range 1000 {
sp.activeTargets[uint64(i)] = &Target{scrapeConfig: &config.ScrapeConfig{}}
sp.loops[uint64(i)] = noopLoop() // First restart will supplement those with proper scrapeLoops.
}
sp.restartLoops(true)
b.ReportAllocs()
b.ResetTimer()
for b.Loop() {
sp.restartLoops(true)
}
}
// TestNewScrapeLoopHonorLabelsWiring verifies that newScrapeLoop correctly wires
// HonorLabels (not HonorTimestamps) to the sampleMutator.
func TestNewScrapeLoopHonorLabelsWiring(t *testing.T) {
foreachAppendable(t, func(t *testing.T, appV2 bool) {
testNewScrapeLoopHonorLabelsWiring(t, appV2)
})
}
func testNewScrapeLoopHonorLabelsWiring(t *testing.T, appV2 bool) {
// Scraped metric has label "lbl" with value "scraped".
// Discovery target has label "lbl" with value "discovery".
// With honor_labels=true, the scraped value should win.
// With honor_labels=false, the discovery value should win and scraped moves to exported_lbl.
testCases := []struct {
name string
honorLabels bool
expectedLbl string
expectedExpLbl string // exported_lbl value, empty if not expected
}{
{
name: "honor_labels=true",
honorLabels: true,
expectedLbl: "scraped",
},
{
name: "honor_labels=false",
honorLabels: false,
expectedLbl: "discovery",
expectedExpLbl: "scraped",
},
}
for _, tc := range testCases {
t.Run(tc.name, func(t *testing.T) {
ts, scrapedTwice := newScrapableServer(`metric{lbl="scraped"} 1`)
defer ts.Close()
testURL, err := url.Parse(ts.URL)
require.NoError(t, err)
s := teststorage.New(t)
cfg := &config.ScrapeConfig{
JobName: "test",
Scheme: "http",
HonorLabels: tc.honorLabels,
HonorTimestamps: !tc.honorLabels, // Opposite of HonorLabels to catch wiring bugs
ScrapeInterval: model.Duration(1 * time.Second),
ScrapeTimeout: model.Duration(100 * time.Millisecond),
MetricNameValidationScheme: model.UTF8Validation,
}
sa := selectAppendable(s, appV2)
sp, err := newScrapePool(cfg, sa.V1(), sa.V2(), 0, nil, nil, &Options{skipOffsetting: true}, newTestScrapeMetrics(t))
require.NoError(t, err)
defer sp.stop()
// Sync with a target that has a conflicting label.
sp.Sync([]*targetgroup.Group{{
Targets: []model.LabelSet{{
model.AddressLabel: model.LabelValue(testURL.Host),
"lbl": "discovery",
}},
}})
require.Len(t, sp.ActiveTargets(), 1)
// Wait for scrape to complete.
select {
case <-time.After(5 * time.Second):
t.Fatal("scrape did not complete in time")
case <-scrapedTwice:
}
// Query the storage to verify label values.
q, err := s.Querier(time.Time{}.UnixNano(), time.Now().UnixNano())
require.NoError(t, err)
defer q.Close()
series := q.Select(t.Context(), false, nil, labels.MustNewMatcher(labels.MatchEqual, "__name__", "metric"))
require.True(t, series.Next(), "metric series not found")
require.Equal(t, tc.expectedLbl, series.At().Labels().Get("lbl"))
require.Equal(t, tc.expectedExpLbl, series.At().Labels().Get("exported_lbl"))
})
}
}
func TestDropsSeriesFromMetricRelabeling(t *testing.T) {
foreachAppendable(t, func(t *testing.T, appV2 bool) {
testDropsSeriesFromMetricRelabeling(t, appV2)
})
}
func testDropsSeriesFromMetricRelabeling(t *testing.T, appV2 bool) {
target := &Target{}
relabelConfig := []*relabel.Config{
{
SourceLabels: model.LabelNames{"__name__"},
Regex: relabel.MustNewRegexp("test_metric.*$"),
Action: relabel.Keep,
NameValidationScheme: model.UTF8Validation,
},
{
SourceLabels: model.LabelNames{"__name__"},
Regex: relabel.MustNewRegexp("test_metric_2$"),
Action: relabel.Drop,
NameValidationScheme: model.UTF8Validation,
},
}
sl, _ := newTestScrapeLoop(t, withAppendable(teststorage.NewAppendable(), appV2), func(sl *scrapeLoop) {
sl.sampleMutator = func(l labels.Labels) labels.Labels {
return mutateSampleLabels(l, target, true, relabelConfig)
}
})
app := sl.appender()
total, added, seriesAdded, err := app.append([]byte("test_metric_1 1\n"), "text/plain", time.Time{})
require.NoError(t, err)
require.Equal(t, 1, total)
require.Equal(t, 1, added)
require.Equal(t, 1, seriesAdded)
total, added, seriesAdded, err = app.append([]byte("test_metric_2 1\n"), "text/plain", time.Time{})
require.NoError(t, err)
require.Equal(t, 1, total)
require.Equal(t, 0, added)
require.Equal(t, 0, seriesAdded)
total, added, seriesAdded, err = app.append([]byte("unwanted_metric 1\n"), "text/plain", time.Time{})
require.NoError(t, err)
require.Equal(t, 1, total)
require.Equal(t, 0, added)
require.Equal(t, 0, seriesAdded)
require.NoError(t, app.Commit())
}