Skip to content

Commit

Permalink
stats/cloud: stop sending cloud metrics when limit reached
Browse files Browse the repository at this point in the history
Once the cloud test gets aborted by limit, we should not send metrics
anymore.

Close #1074
  • Loading branch information
cuonglm committed Sep 4, 2019
1 parent d385166 commit ff003be
Show file tree
Hide file tree
Showing 2 changed files with 149 additions and 17 deletions.
61 changes: 44 additions & 17 deletions stats/cloud/collector.go
Original file line number Diff line number Diff line change
Expand Up @@ -23,6 +23,7 @@ package cloud
import (
"context"
"encoding/json"
"net/http"
"path/filepath"
"sync"
"time"
Expand Down Expand Up @@ -69,6 +70,8 @@ type Collector struct {
// checks basically O(1). And even if for some reason there are occasional metrics with past times that
// don't fit in the chosen ring buffer size, we could just send them along to the buffer unaggregated
aggrBuckets map[int64]aggregationBucket

stopSendingMetricsCh chan struct{}
}

// Verify that Collector implements lib.Collector
Expand Down Expand Up @@ -136,13 +139,14 @@ func New(conf Config, src *loader.SourceData, opts lib.Options, version string)
}

return &Collector{
config: conf,
thresholds: thresholds,
client: NewClient(conf.Token.String, conf.Host.String, version),
anonymous: !conf.Token.Valid,
duration: duration,
opts: opts,
aggrBuckets: map[int64]aggregationBucket{},
config: conf,
thresholds: thresholds,
client: NewClient(conf.Token.String, conf.Host.String, version),
anonymous: !conf.Token.Valid,
duration: duration,
opts: opts,
aggrBuckets: map[int64]aggregationBucket{},
stopSendingMetricsCh: make(chan struct{}),
}, nil
}

Expand Down Expand Up @@ -196,22 +200,24 @@ func (c *Collector) Link() string {
// at regular intervals and when the context is terminated.
func (c *Collector) Run(ctx context.Context) {
wg := sync.WaitGroup{}

quit := make(chan struct{})
// If enabled, start periodically aggregating the collected HTTP trails
if c.config.AggregationPeriod.Duration > 0 {
wg.Add(1)
aggregationTicker := time.NewTicker(time.Duration(c.config.AggregationCalcInterval.Duration))

go func() {
defer wg.Done()
for {
select {
case <-c.stopSendingMetricsCh:
return
case <-aggregationTicker.C:
c.aggregateHTTPTrails(time.Duration(c.config.AggregationWaitPeriod.Duration))
case <-ctx.Done():
c.aggregateHTTPTrails(0)
c.flushHTTPTrails()
c.pushMetrics()
wg.Done()
close(quit)
return
}
}
Expand All @@ -226,11 +232,16 @@ func (c *Collector) Run(ctx context.Context) {
pushTicker := time.NewTicker(time.Duration(c.config.MetricPushInterval.Duration))
for {
select {
case <-pushTicker.C:
c.pushMetrics()
case <-ctx.Done():
case <-c.stopSendingMetricsCh:
return
default:
}
select {
case <-quit:
c.pushMetrics()
return
case <-pushTicker.C:
c.pushMetrics()
}
}
}
Expand Down Expand Up @@ -441,6 +452,19 @@ func (c *Collector) flushHTTPTrails() {
c.aggrBuckets = map[int64]aggregationBucket{}
c.bufferSamples = append(c.bufferSamples, newSamples...)
}

func (c *Collector) shouldStopSendingMetrics(err error) bool {
if err == nil {
return false
}

if errResp, ok := err.(ErrorResponse); ok && errResp.Response != nil {
return errResp.Response.StatusCode == http.StatusForbidden && errResp.Code == 4
}

return false
}

func (c *Collector) pushMetrics() {
c.bufferMutex.Lock()
if len(c.bufferSamples) == 0 {
Expand All @@ -462,9 +486,12 @@ func (c *Collector) pushMetrics() {
}
err := c.client.PushMetric(c.referenceID, c.config.NoCompress.Bool, buffer[:size])
if err != nil {
logrus.WithFields(logrus.Fields{
"error": err,
}).Warn("Failed to send metrics to cloud")
if c.shouldStopSendingMetrics(err) {
logrus.WithError(err).Warn("Stopped sending metrics to cloud due to an error")
close(c.stopSendingMetricsCh)
break
}
logrus.WithError(err).Warn("Failed to send metrics to cloud")
}
buffer = buffer[size:]
}
Expand Down Expand Up @@ -493,7 +520,7 @@ func (c *Collector) testFinished() {
}).Debug("Sending test finished")

runStatus := lib.RunStatusFinished
if c.runStatus != 0 {
if c.runStatus != lib.RunStatusQueued {
runStatus = c.runStatus
}

Expand Down
105 changes: 105 additions & 0 deletions stats/cloud/collector_test.go
Original file line number Diff line number Diff line change
Expand Up @@ -355,3 +355,108 @@ func TestCloudCollectorMaxPerPacket(t *testing.T) {
wg.Wait()
require.True(t, gotTheLimit)
}

func TestCloudCollectorStopSendingMetric(t *testing.T) {
t.Parallel()
tb := testutils.NewHTTPMultiBin(t)
tb.Mux.HandleFunc("/v1/tests", http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {
_, err := fmt.Fprint(w, `{
"reference_id": "12",
"config": {
"metricPushInterval": "200ms",
"aggregationPeriod": "100ms",
"maxMetricSamplesPerPackage": 20,
"aggregationCalcInterval": "100ms",
"aggregationWaitPeriod": "100ms"
}
}`)
require.NoError(t, err)
}))
defer tb.Cleanup()

script := &loader.SourceData{
Data: []byte(""),
URL: &url.URL{Path: "/script.js"},
}

options := lib.Options{
Duration: types.NullDurationFrom(1 * time.Second),
}

config := NewConfig().Apply(Config{
Host: null.StringFrom(tb.ServerHTTP.URL),
NoCompress: null.BoolFrom(true),
})
collector, err := New(config, script, options, "1.0")
require.NoError(t, err)
now := time.Now()
tags := stats.IntoSampleTags(&map[string]string{"test": "mest", "a": "b"})

count := 1
max := 5
tb.Mux.HandleFunc(fmt.Sprintf("/v1/metrics/%s", collector.referenceID),
func(w http.ResponseWriter, r *http.Request) {
count++
if count == max {
type payload struct {
Error ErrorResponse `json:"error"`
}
res := &payload{}
res.Error = ErrorResponse{Code: 4}
w.Header().Set("Content-Type", "application/json")
data, err := json.Marshal(res)
if err != nil {
t.Fatal(err)
}
w.WriteHeader(http.StatusForbidden)
_, _ = w.Write(data)
return
}
body, err := ioutil.ReadAll(r.Body)
assert.NoError(t, err)
receivedSamples := []Sample{}
assert.NoError(t, json.Unmarshal(body, &receivedSamples))
})

require.NoError(t, collector.Init())
ctx, cancel := context.WithCancel(context.Background())
wg := sync.WaitGroup{}
wg.Add(1)
go func() {
collector.Run(ctx)
wg.Done()
}()

collector.Collect([]stats.SampleContainer{stats.Sample{
Time: now,
Metric: metrics.VUs,
Tags: stats.NewSampleTags(tags.CloneTags()),
Value: 1.0,
}})
for j := time.Duration(1); j <= 200; j++ {
var container = make([]stats.SampleContainer, 0, 500)
for i := time.Duration(1); i <= 50; i++ {
container = append(container, &httpext.Trail{
Blocked: i % 200 * 100 * time.Millisecond,
Connecting: i % 200 * 200 * time.Millisecond,
TLSHandshaking: i % 200 * 300 * time.Millisecond,
Sending: i * i * 400 * time.Millisecond,
Waiting: 500 * time.Millisecond,
Receiving: 600 * time.Millisecond,

EndTime: now.Add(i * 100),
ConnDuration: 500 * time.Millisecond,
Duration: j * i * 1500 * time.Millisecond,
Tags: stats.NewSampleTags(tags.CloneTags()),
})
}
collector.Collect(container)
}

cancel()
wg.Wait()
require.Equal(t, lib.RunStatusQueued, collector.runStatus)
_, ok := <-collector.stopSendingMetricsCh
require.False(t, ok)
require.Equal(t, max, count)
}

0 comments on commit ff003be

Please sign in to comment.