Skip to content

Commit

Permalink
Fix linting issues (#7789)
Browse files Browse the repository at this point in the history
Signed-off-by: Arve Knudsen <[email protected]>
  • Loading branch information
aknuds1 authored Apr 4, 2024
1 parent 383d904 commit 9fccbac
Show file tree
Hide file tree
Showing 47 changed files with 167 additions and 168 deletions.
2 changes: 1 addition & 1 deletion cmd/mimirtool/main.go
Original file line number Diff line number Diff line change
Expand Up @@ -49,7 +49,7 @@ func main() {
remoteReadCommand.Register(app, envVars)
ruleCommand.Register(app, envVars, prometheus.DefaultRegisterer)

app.Command("version", "Get the version of the mimirtool CLI").Action(func(k *kingpin.ParseContext) error {
app.Command("version", "Get the version of the mimirtool CLI").Action(func(*kingpin.ParseContext) error {
fmt.Fprintln(os.Stdout, mimirversion.Print("Mimirtool"))
version.CheckLatest(mimirversion.Version)
return nil
Expand Down
2 changes: 1 addition & 1 deletion integration/e2emimir/services.go
Original file line number Diff line number Diff line change
Expand Up @@ -407,7 +407,7 @@ func WithConfigFile(configFile string) Option {
}

// WithNoopOption returns an option that doesn't change anything.
func WithNoopOption() Option { return func(options *Options) {} }
func WithNoopOption() Option { return func(*Options) {} }

// FlagMapper is the type of function that maps flags, just to reduce some verbosity.
type FlagMapper func(flags map[string]string) map[string]string
Expand Down
10 changes: 5 additions & 5 deletions integration/kv_test.go
Original file line number Diff line number Diff line change
Expand Up @@ -31,7 +31,7 @@ func TestKVList(t *testing.T) {
// Create keys to list back
keysToCreate := []string{"key-a", "key-b", "key-c"}
for _, key := range keysToCreate {
err := client.CAS(context.Background(), key, func(in interface{}) (out interface{}, retry bool, err error) {
err := client.CAS(context.Background(), key, func(interface{}) (out interface{}, retry bool, err error) {
return key, false, nil
})
require.NoError(t, err, "could not create key")
Expand All @@ -53,7 +53,7 @@ func TestKVList(t *testing.T) {
func TestKVDelete(t *testing.T) {
testKVs(t, func(t *testing.T, client kv.Client, reg *prometheus.Registry) {
// Create a key
err := client.CAS(context.Background(), "key-to-delete", func(in interface{}) (out interface{}, retry bool, err error) {
err := client.CAS(context.Background(), "key-to-delete", func(interface{}) (out interface{}, retry bool, err error) {
return "key-to-delete", false, nil
})
require.NoError(t, err, "object could not be created")
Expand All @@ -76,11 +76,11 @@ func TestKVDelete(t *testing.T) {
}

func TestKVWatchAndDelete(t *testing.T) {
testKVs(t, func(t *testing.T, client kv.Client, reg *prometheus.Registry) {
testKVs(t, func(t *testing.T, client kv.Client, _ *prometheus.Registry) {
ctx, cancel := context.WithCancel(context.Background())
defer cancel()

err := client.CAS(context.Background(), "key-before-watch", func(in interface{}) (out interface{}, retry bool, err error) {
err := client.CAS(context.Background(), "key-before-watch", func(interface{}) (out interface{}, retry bool, err error) {
return "value-before-watch", false, nil
})
require.NoError(t, err)
Expand All @@ -93,7 +93,7 @@ func TestKVWatchAndDelete(t *testing.T) {
w.watch(ctx, client)
}()

err = client.CAS(context.Background(), "key-to-delete", func(in interface{}) (out interface{}, retry bool, err error) {
err = client.CAS(context.Background(), "key-to-delete", func(interface{}) (out interface{}, retry bool, err error) {
return "value-to-delete", false, nil
})
require.NoError(t, err, "object could not be created")
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -114,7 +114,7 @@ func (c *ConfigExtractor) ResolveConfigs() ([]*yaml.RNode, error) {
return nil, err
}

err = concurrency.ForEachJob(context.Background(), len(c.allItems), runtime.NumCPU(), func(ctx context.Context, idx int) error {
err = concurrency.ForEachJob(context.Background(), len(c.allItems), runtime.NumCPU(), func(_ context.Context, idx int) error {
pod, ok, err := extractPodSpec(c.allItems[idx])
if err != nil {
return err
Expand Down
2 changes: 1 addition & 1 deletion pkg/querier/block_test.go
Original file line number Diff line number Diff line change
Expand Up @@ -304,7 +304,7 @@ func TestBlockQuerierSeriesSet(t *testing.T) {
t.Run(fmt.Sprintf("consume with .Next() method, perform .At() after every %dth call to .Next()", callAtEvery), func(t *testing.T) {
t.Parallel()

advance := func(it chunkenc.Iterator, wantTs int64) chunkenc.ValueType { return it.Next() }
advance := func(it chunkenc.Iterator, _ int64) chunkenc.ValueType { return it.Next() }
ss := getSeriesSet()

verifyNextSeries(t, ss, labels.FromStrings("__name__", "first", "a", "a"), 3*time.Millisecond, []timeRange{
Expand Down
6 changes: 3 additions & 3 deletions pkg/querier/blocks_store_queryable_test.go
Original file line number Diff line number Diff line change
Expand Up @@ -1046,7 +1046,7 @@ func TestBlocksStoreQuerier_ShouldReturnContextCanceledIfContextWasCanceledWhile

srv, q := prepareTestCase(t)

srv.onSeries = func(req *storepb.SeriesRequest, srv storegatewaypb.StoreGateway_SeriesServer) error {
srv.onSeries = func(*storepb.SeriesRequest, storegatewaypb.StoreGateway_SeriesServer) error {
if numExecutions.Inc() == 1 {
close(waitExecution)
<-continueExecution
Expand Down Expand Up @@ -1082,7 +1082,7 @@ func TestBlocksStoreQuerier_ShouldReturnContextCanceledIfContextWasCanceledWhile

srv, q := prepareTestCase(t)

srv.onLabelNames = func(ctx context.Context, req *storepb.LabelNamesRequest) (*storepb.LabelNamesResponse, error) {
srv.onLabelNames = func(context.Context, *storepb.LabelNamesRequest) (*storepb.LabelNamesResponse, error) {
if numExecutions.Inc() == 1 {
close(waitExecution)
<-continueExecution
Expand Down Expand Up @@ -1117,7 +1117,7 @@ func TestBlocksStoreQuerier_ShouldReturnContextCanceledIfContextWasCanceledWhile

srv, q := prepareTestCase(t)

srv.onLabelValues = func(ctx context.Context, req *storepb.LabelValuesRequest) (*storepb.LabelValuesResponse, error) {
srv.onLabelValues = func(context.Context, *storepb.LabelValuesRequest) (*storepb.LabelValuesResponse, error) {
if numExecutions.Inc() == 1 {
close(waitExecution)
<-continueExecution
Expand Down
4 changes: 2 additions & 2 deletions pkg/querier/blocks_store_replicated_set_test.go
Original file line number Diff line number Diff line change
Expand Up @@ -322,7 +322,7 @@ func TestBlocksStoreReplicationSet_GetClientsFor(t *testing.T) {
ringStore, closer := consul.NewInMemoryClient(ring.GetCodec(), log.NewNopLogger(), nil)
t.Cleanup(func() { assert.NoError(t, closer.Close()) })

require.NoError(t, ringStore.CAS(ctx, "test", func(in interface{}) (interface{}, bool, error) {
require.NoError(t, ringStore.CAS(ctx, "test", func(interface{}) (interface{}, bool, error) {
d := ring.NewDesc()
testData.setup(d)
return d, true, nil
Expand Down Expand Up @@ -389,7 +389,7 @@ func TestBlocksStoreReplicationSet_GetClientsFor_ShouldSupportRandomLoadBalancin
ringStore, closer := consul.NewInMemoryClient(ring.GetCodec(), log.NewNopLogger(), nil)
t.Cleanup(func() { assert.NoError(t, closer.Close()) })

require.NoError(t, ringStore.CAS(ctx, "test", func(in interface{}) (interface{}, bool, error) {
require.NoError(t, ringStore.CAS(ctx, "test", func(interface{}) (interface{}, bool, error) {
d := ring.NewDesc()
for n := 1; n <= numInstances; n++ {
d.AddIngester(fmt.Sprintf("instance-%d", n), fmt.Sprintf("127.0.0.%d", n), "", []uint32{uint32(n)}, ring.ACTIVE, registeredAt)
Expand Down
4 changes: 2 additions & 2 deletions pkg/querier/querier_test.go
Original file line number Diff line number Diff line change
Expand Up @@ -1284,10 +1284,10 @@ func TestConfig_ValidateLimits(t *testing.T) {
expected error
}{
"should pass with default config": {
setup: func(cfg *Config, limits *validation.Limits) {},
setup: func(*Config, *validation.Limits) {},
},
"should pass if 'query store after' is enabled and shuffle-sharding is disabled": {
setup: func(cfg *Config, limits *validation.Limits) {
setup: func(cfg *Config, _ *validation.Limits) {
cfg.QueryStoreAfter = time.Hour
},
},
Expand Down
8 changes: 4 additions & 4 deletions pkg/querier/remote_read_test.go
Original file line number Diff line number Diff line change
Expand Up @@ -102,7 +102,7 @@ func (p *partiallyFailingSeriesSet) Warnings() annotations.Annotations {

func TestSampledRemoteRead(t *testing.T) {
q := &mockSampleAndChunkQueryable{
queryableFn: func(mint, maxt int64) (storage.Querier, error) {
queryableFn: func(int64, int64) (storage.Querier, error) {
return mockQuerier{
seriesSet: series.NewConcreteSeriesSetFromUnsortedSeries([]storage.Series{
series.NewConcreteSeries(
Expand Down Expand Up @@ -326,7 +326,7 @@ func TestStreamedRemoteRead(t *testing.T) {
for tn, tc := range tcs {
t.Run(tn, func(t *testing.T) {
q := &mockSampleAndChunkQueryable{
chunkQueryableFn: func(mint, maxt int64) (storage.ChunkQuerier, error) {
chunkQueryableFn: func(int64, int64) (storage.ChunkQuerier, error) {
return mockChunkQuerier{
seriesSet: series.NewConcreteSeriesSetFromUnsortedSeries([]storage.Series{
series.NewConcreteSeries(
Expand Down Expand Up @@ -519,7 +519,7 @@ func TestRemoteReadErrorParsing(t *testing.T) {
for tn, tc := range testCases {
t.Run(tn, func(t *testing.T) {
q := &mockSampleAndChunkQueryable{
queryableFn: func(mint, maxt int64) (storage.Querier, error) {
queryableFn: func(int64, int64) (storage.Querier, error) {
return mockQuerier{
seriesSet: tc.seriesSet,
}, tc.getQuerierErr
Expand Down Expand Up @@ -555,7 +555,7 @@ func TestRemoteReadErrorParsing(t *testing.T) {
for tn, tc := range testCases {
t.Run(tn, func(t *testing.T) {
q := &mockSampleAndChunkQueryable{
chunkQueryableFn: func(mint, maxt int64) (storage.ChunkQuerier, error) {
chunkQueryableFn: func(int64, int64) (storage.ChunkQuerier, error) {
return mockChunkQuerier{
seriesSet: tc.seriesSet,
}, tc.getQuerierErr
Expand Down
12 changes: 6 additions & 6 deletions pkg/querier/tenantfederation/merge_exemplar_queryable.go
Original file line number Diff line number Diff line change
Expand Up @@ -176,7 +176,7 @@ func (m *mergeExemplarQuerier) Select(start, end int64, matchers ...[]*labels.Ma
// Each task grabs a job object from the slice and stores its results in the corresponding
// index in the results slice. The job handles performing a tenant-specific exemplar query
// and adding a tenant ID label to each of the results.
run := func(ctx context.Context, idx int) error {
run := func(_ context.Context, idx int) error {
job := jobs[idx]

res, err := job.querier.Select(start, end, job.matchers...)
Expand Down Expand Up @@ -217,21 +217,21 @@ func filterTenantsAndRewriteMatchers(idLabelName string, ids []string, allMatche
return sliceToSet(ids), allMatchers
}

outIds := make(map[string]struct{})
outIDs := make(map[string]struct{})
outMatchers := make([][]*labels.Matcher, len(allMatchers))

// The ExemplarQuerier.Select method accepts a slice of slices of matchers. The matchers within
// a single slice are AND'd together (intersection) and each outer slice is OR'd together (union).
// In order to support that, we start with a set of 0 tenant IDs and add any tenant IDs that remain
// after filtering (based on the inner slice of matchers), for each outer slice.
for i, matchers := range allMatchers {
filteredIds, unrelatedMatchers := filterValuesByMatchers(idLabelName, ids, matchers...)
for k := range filteredIds {
outIds[k] = struct{}{}
filteredIDs, unrelatedMatchers := filterValuesByMatchers(idLabelName, ids, matchers...)
for k := range filteredIDs {
outIDs[k] = struct{}{}
}

outMatchers[i] = unrelatedMatchers
}

return outIds, outMatchers
return outIDs, outMatchers
}
4 changes: 2 additions & 2 deletions pkg/querier/worker/frontend_processor_test.go
Original file line number Diff line number Diff line change
Expand Up @@ -72,12 +72,12 @@ func TestFrontendProcessor_processQueriesOnSingleStream(t *testing.T) {

workerCtx, workerCancel := context.WithCancel(context.Background())

requestHandler.On("Handle", mock.Anything, mock.Anything).Run(func(args mock.Arguments) {
requestHandler.On("Handle", mock.Anything, mock.Anything).Run(func(mock.Arguments) {
// Cancel the worker context while the query execution is in progress.
workerCancel()

// Ensure the execution context hasn't been canceled yet.
require.Nil(t, processClient.Context().Err())
require.NoError(t, processClient.Context().Err())

// Intentionally slow down the query execution, to double check the worker waits until done.
time.Sleep(time.Second)
Expand Down
4 changes: 2 additions & 2 deletions pkg/querier/worker/scheduler_processor_test.go
Original file line number Diff line number Diff line change
Expand Up @@ -83,7 +83,7 @@ func TestSchedulerProcessor_processQueriesOnSingleStream(t *testing.T) {

workerCtx, workerCancel := context.WithCancel(context.Background())

requestHandler.On("Handle", mock.Anything, mock.Anything).Run(func(args mock.Arguments) {
requestHandler.On("Handle", mock.Anything, mock.Anything).Run(func(mock.Arguments) {
// Cancel the worker context while the query execution is in progress.
workerCancel()

Expand Down Expand Up @@ -405,7 +405,7 @@ func TestSchedulerProcessor_ResponseStream(t *testing.T) {
ctx, cancel := context.WithCancel(context.Background())

requestHandler.On("Handle", mock.Anything, mock.Anything).Run(
func(arguments mock.Arguments) { cancel() },
func(mock.Arguments) { cancel() },
).Return(returnResponses(responses)())

reqProcessor.processQueriesOnSingleStream(ctx, nil, "127.0.0.1")
Expand Down
2 changes: 1 addition & 1 deletion pkg/querier/worker/worker_test.go
Original file line number Diff line number Diff line change
Expand Up @@ -29,7 +29,7 @@ func TestConfig_Validate(t *testing.T) {
expectedErr string
}{
"should pass with default config": {
setup: func(cfg *Config) {},
setup: func(*Config) {},
},
"should pass if frontend address is configured, but not scheduler address": {
setup: func(cfg *Config) {
Expand Down
6 changes: 3 additions & 3 deletions pkg/ruler/api_test.go
Original file line number Diff line number Diff line change
Expand Up @@ -298,7 +298,7 @@ func TestRuler_PrometheusRules(t *testing.T) {
},
},
expectedConfigured: 1,
limits: validation.MockOverrides(func(defaults *validation.Limits, tenantLimits map[string]*validation.Limits) {
limits: validation.MockOverrides(func(_ *validation.Limits, tenantLimits map[string]*validation.Limits) {
tenantLimits[userID] = validation.MockDefaultLimits()
tenantLimits[userID].RulerRecordingRulesEvaluationEnabled = true
tenantLimits[userID].RulerAlertingRulesEvaluationEnabled = false
Expand Down Expand Up @@ -330,7 +330,7 @@ func TestRuler_PrometheusRules(t *testing.T) {
},
},
expectedConfigured: 1,
limits: validation.MockOverrides(func(defaults *validation.Limits, tenantLimits map[string]*validation.Limits) {
limits: validation.MockOverrides(func(_ *validation.Limits, tenantLimits map[string]*validation.Limits) {
tenantLimits[userID] = validation.MockDefaultLimits()
tenantLimits[userID].RulerRecordingRulesEvaluationEnabled = false
tenantLimits[userID].RulerAlertingRulesEvaluationEnabled = true
Expand Down Expand Up @@ -364,7 +364,7 @@ func TestRuler_PrometheusRules(t *testing.T) {
},
},
expectedConfigured: 0,
limits: validation.MockOverrides(func(defaults *validation.Limits, tenantLimits map[string]*validation.Limits) {
limits: validation.MockOverrides(func(_ *validation.Limits, tenantLimits map[string]*validation.Limits) {
tenantLimits[userID] = validation.MockDefaultLimits()
tenantLimits[userID].RulerRecordingRulesEvaluationEnabled = false
tenantLimits[userID].RulerAlertingRulesEvaluationEnabled = false
Expand Down
4 changes: 2 additions & 2 deletions pkg/ruler/compat_test.go
Original file line number Diff line number Diff line change
Expand Up @@ -379,7 +379,7 @@ func TestMetricsQueryFuncErrors(t *testing.T) {
queries := promauto.With(nil).NewCounter(prometheus.CounterOpts{})
failures := promauto.With(nil).NewCounter(prometheus.CounterOpts{})

mockFunc := func(ctx context.Context, q string, t time.Time) (promql.Vector, error) {
mockFunc := func(context.Context, string, time.Time) (promql.Vector, error) {
return promql.Vector{}, tc.returnedError
}
qf := MetricsQueryFunc(mockFunc, queries, failures, tc.remoteQuerier)
Expand All @@ -397,7 +397,7 @@ func TestRecordAndReportRuleQueryMetrics(t *testing.T) {
queryTime := promauto.With(nil).NewCounterVec(prometheus.CounterOpts{}, []string{"user"})
zeroFetchedSeriesCount := promauto.With(nil).NewCounterVec(prometheus.CounterOpts{}, []string{"user"})

mockFunc := func(ctx context.Context, q string, t time.Time) (promql.Vector, error) {
mockFunc := func(context.Context, string, time.Time) (promql.Vector, error) {
time.Sleep(1 * time.Second)
return promql.Vector{}, nil
}
Expand Down
18 changes: 9 additions & 9 deletions pkg/ruler/remotequerier_test.go
Original file line number Diff line number Diff line change
Expand Up @@ -42,7 +42,7 @@ func TestRemoteQuerier_Read(t *testing.T) {
setup := func() (mockHTTPGRPCClient, *httpgrpc.HTTPRequest) {
var inReq httpgrpc.HTTPRequest

mockClientFn := func(ctx context.Context, req *httpgrpc.HTTPRequest, _ ...grpc.CallOption) (*httpgrpc.HTTPResponse, error) {
mockClientFn := func(_ context.Context, req *httpgrpc.HTTPRequest, _ ...grpc.CallOption) (*httpgrpc.HTTPResponse, error) {
inReq = *req

b, err := proto.Marshal(&prompb.ReadResponse{
Expand Down Expand Up @@ -97,7 +97,7 @@ func TestRemoteQuerier_Read(t *testing.T) {
}

func TestRemoteQuerier_ReadReqTimeout(t *testing.T) {
mockClientFn := func(ctx context.Context, req *httpgrpc.HTTPRequest, _ ...grpc.CallOption) (*httpgrpc.HTTPResponse, error) {
mockClientFn := func(ctx context.Context, _ *httpgrpc.HTTPRequest, _ ...grpc.CallOption) (*httpgrpc.HTTPResponse, error) {
<-ctx.Done()
return nil, ctx.Err()
}
Expand All @@ -117,7 +117,7 @@ func TestRemoteQuerier_Query(t *testing.T) {
setup := func() (mockHTTPGRPCClient, *httpgrpc.HTTPRequest) {
var inReq httpgrpc.HTTPRequest

mockClientFn := func(ctx context.Context, req *httpgrpc.HTTPRequest, _ ...grpc.CallOption) (*httpgrpc.HTTPResponse, error) {
mockClientFn := func(_ context.Context, req *httpgrpc.HTTPRequest, _ ...grpc.CallOption) (*httpgrpc.HTTPResponse, error) {
inReq = *req

return &httpgrpc.HTTPResponse{
Expand Down Expand Up @@ -266,7 +266,7 @@ func TestRemoteQuerier_QueryRetryOnFailure(t *testing.T) {
var count atomic.Int64

ctx, cancel := context.WithCancel(context.Background())
mockClientFn := func(ctx context.Context, req *httpgrpc.HTTPRequest, _ ...grpc.CallOption) (*httpgrpc.HTTPResponse, error) {
mockClientFn := func(context.Context, *httpgrpc.HTTPRequest, ...grpc.CallOption) (*httpgrpc.HTTPResponse, error) {
count.Add(1)
if testCase.err != nil {
if grpcutil.IsCanceled(testCase.err) {
Expand Down Expand Up @@ -396,7 +396,7 @@ func TestRemoteQuerier_QueryJSONDecoding(t *testing.T) {

for name, scenario := range scenarios {
t.Run(name, func(t *testing.T) {
mockClientFn := func(ctx context.Context, req *httpgrpc.HTTPRequest, _ ...grpc.CallOption) (*httpgrpc.HTTPResponse, error) {
mockClientFn := func(context.Context, *httpgrpc.HTTPRequest, ...grpc.CallOption) (*httpgrpc.HTTPResponse, error) {
return &httpgrpc.HTTPResponse{
Code: http.StatusOK,
Headers: []*httpgrpc.Header{
Expand Down Expand Up @@ -664,7 +664,7 @@ func TestRemoteQuerier_QueryProtobufDecoding(t *testing.T) {

for name, scenario := range scenarios {
t.Run(name, func(t *testing.T) {
mockClientFn := func(ctx context.Context, req *httpgrpc.HTTPRequest, _ ...grpc.CallOption) (*httpgrpc.HTTPResponse, error) {
mockClientFn := func(context.Context, *httpgrpc.HTTPRequest, ...grpc.CallOption) (*httpgrpc.HTTPResponse, error) {
b, err := scenario.body.Marshal()
if err != nil {
return nil, err
Expand Down Expand Up @@ -692,7 +692,7 @@ func TestRemoteQuerier_QueryProtobufDecoding(t *testing.T) {
}

func TestRemoteQuerier_QueryUnknownResponseContentType(t *testing.T) {
mockClientFn := func(ctx context.Context, req *httpgrpc.HTTPRequest, _ ...grpc.CallOption) (*httpgrpc.HTTPResponse, error) {
mockClientFn := func(context.Context, *httpgrpc.HTTPRequest, ...grpc.CallOption) (*httpgrpc.HTTPResponse, error) {
return &httpgrpc.HTTPResponse{
Code: http.StatusOK,
Headers: []*httpgrpc.Header{
Expand All @@ -709,7 +709,7 @@ func TestRemoteQuerier_QueryUnknownResponseContentType(t *testing.T) {
}

func TestRemoteQuerier_QueryReqTimeout(t *testing.T) {
mockClientFn := func(ctx context.Context, req *httpgrpc.HTTPRequest, _ ...grpc.CallOption) (*httpgrpc.HTTPResponse, error) {
mockClientFn := func(ctx context.Context, _ *httpgrpc.HTTPRequest, _ ...grpc.CallOption) (*httpgrpc.HTTPResponse, error) {
<-ctx.Done()
return nil, ctx.Err()
}
Expand Down Expand Up @@ -767,7 +767,7 @@ func TestRemoteQuerier_StatusErrorResponses(t *testing.T) {
}
for testName, testCase := range testCases {
t.Run(testName, func(t *testing.T) {
mockClientFn := func(ctx context.Context, req *httpgrpc.HTTPRequest, _ ...grpc.CallOption) (*httpgrpc.HTTPResponse, error) {
mockClientFn := func(context.Context, *httpgrpc.HTTPRequest, ...grpc.CallOption) (*httpgrpc.HTTPResponse, error) {
return testCase.resp, testCase.err
}
logger := newLoggerWithCounter()
Expand Down
2 changes: 1 addition & 1 deletion pkg/ruler/ruler.go
Original file line number Diff line number Diff line change
Expand Up @@ -1290,7 +1290,7 @@ func (r *Ruler) notifySyncRules(ctx context.Context, userIDs []string) {
// the client-side gRPC instrumentation fails.
ctx = user.InjectOrgID(ctx, "")

errs.Add(r.forEachRulerInTheRing(ctx, r.ring, RuleSyncRingOp, func(ctx context.Context, inst *ring.InstanceDesc, rulerClient RulerClient, rulerClientErr error) error {
errs.Add(r.forEachRulerInTheRing(ctx, r.ring, RuleSyncRingOp, func(ctx context.Context, _ *ring.InstanceDesc, rulerClient RulerClient, rulerClientErr error) error {
var err error

if rulerClientErr != nil {
Expand Down
Loading

0 comments on commit 9fccbac

Please sign in to comment.