Skip to content

Commit

Permalink
storage: rebalance replicas based on load at the store level
Browse files Browse the repository at this point in the history
Follow-up to cockroachdb#28340, which did this for just leases.

Fixes cockroachdb#17979

Release note (performance improvement): Range replicas will be
automatically rebalanced throughout the cluster to even out the amount
of QPS being handled by each node.
  • Loading branch information
a-robinson committed Aug 31, 2018
1 parent 61325cd commit 2414b16
Show file tree
Hide file tree
Showing 7 changed files with 340 additions and 52 deletions.
4 changes: 2 additions & 2 deletions docs/generated/settings/settings.html
Original file line number Diff line number Diff line change
Expand Up @@ -18,8 +18,8 @@
<tr><td><code>kv.allocator.lease_rebalancing_aggressiveness</code></td><td>float</td><td><code>1</code></td><td>set greater than 1.0 to rebalance leases toward load more aggressively, or between 0 and 1.0 to be more conservative about rebalancing leases</td></tr>
<tr><td><code>kv.allocator.load_based_lease_rebalancing.enabled</code></td><td>boolean</td><td><code>true</code></td><td>set to enable rebalancing of range leases based on load and latency</td></tr>
<tr><td><code>kv.allocator.range_rebalance_threshold</code></td><td>float</td><td><code>0.05</code></td><td>minimum fraction away from the mean a store's range count can be before it is considered overfull or underfull</td></tr>
<tr><td><code>kv.allocator.stat_based_rebalancing.enabled</code></td><td>boolean</td><td><code>false</code></td><td>set to enable rebalancing range replicas and leases to more evenly distribute read and write load across the stores in a cluster</td></tr>
<tr><td><code>kv.allocator.stat_rebalance_threshold</code></td><td>float</td><td><code>0.2</code></td><td>minimum fraction away from the mean a store's stats (like disk usage or writes per second) can be before it is considered overfull or underfull</td></tr>
<tr><td><code>kv.allocator.stat_based_rebalancing.enabled</code></td><td>boolean</td><td><code>true</code></td><td>set to enable rebalancing range replicas and leases to more evenly distribute read and write load across the stores in a cluster</td></tr>
<tr><td><code>kv.allocator.stat_rebalance_threshold</code></td><td>float</td><td><code>0.2</code></td><td>minimum fraction away from the mean a store's stats (such as queries per second) can be before it is considered overfull or underfull</td></tr>
<tr><td><code>kv.bulk_io_write.concurrent_export_requests</code></td><td>integer</td><td><code>5</code></td><td>number of export requests a store will handle concurrently before queuing</td></tr>
<tr><td><code>kv.bulk_io_write.concurrent_import_requests</code></td><td>integer</td><td><code>1</code></td><td>number of import requests a store will handle concurrently before queuing</td></tr>
<tr><td><code>kv.bulk_io_write.max_rate</code></td><td>byte size</td><td><code>8.0 EiB</code></td><td>the rate limit (bytes/sec) to use for writes to disk on behalf of bulk io ops</td></tr>
Expand Down
43 changes: 30 additions & 13 deletions pkg/storage/allocator.go
Original file line number Diff line number Diff line change
Expand Up @@ -361,9 +361,36 @@ func (a *Allocator) AllocateTarget(
) (*roachpb.StoreDescriptor, string, error) {
sl, aliveStoreCount, throttledStoreCount := a.storePool.getStoreList(rangeInfo.Desc.RangeID, storeFilterThrottled)

target, details := a.allocateTargetFromList(
ctx, sl, zone, existing, rangeInfo, a.scorerOptions(disableStatsBasedRebalancing))

if target != nil {
return target, details, nil
}

// When there are throttled stores that do match, we shouldn't send
// the replica to purgatory.
if throttledStoreCount > 0 {
return nil, "", errors.Errorf("%d matching stores are currently throttled", throttledStoreCount)
}
return nil, "", &allocatorError{
constraints: zone.Constraints,
existingReplicas: len(existing),
aliveStores: aliveStoreCount,
throttledStores: throttledStoreCount,
}
}

func (a *Allocator) allocateTargetFromList(
ctx context.Context,
sl StoreList,
zone config.ZoneConfig,
existing []roachpb.ReplicaDescriptor,
rangeInfo RangeInfo,
options scorerOptions,
) (*roachpb.StoreDescriptor, string) {
analyzedConstraints := analyzeConstraints(
ctx, a.storePool.getStoreDescriptor, existing, zone)
options := a.scorerOptions(disableStatsBasedRebalancing)
candidates := allocateCandidates(
sl, analyzedConstraints, existing, rangeInfo, a.storePool.getLocalities(existing), options,
)
Expand All @@ -379,20 +406,10 @@ func (a *Allocator) AllocateTarget(
if err != nil {
log.Warningf(ctx, "failed to marshal details for choosing allocate target: %s", err)
}
return &target.store, string(detailsBytes), nil
return &target.store, string(detailsBytes)
}

// When there are throttled stores that do match, we shouldn't send
// the replica to purgatory.
if throttledStoreCount > 0 {
return nil, "", errors.Errorf("%d matching stores are currently throttled", throttledStoreCount)
}
return nil, "", &allocatorError{
constraints: zone.Constraints,
existingReplicas: len(existing),
aliveStores: aliveStoreCount,
throttledStores: throttledStoreCount,
}
return nil, ""
}

func (a Allocator) simulateRemoveTarget(
Expand Down
16 changes: 14 additions & 2 deletions pkg/storage/allocator_scorer.go
Original file line number Diff line number Diff line change
Expand Up @@ -60,7 +60,7 @@ const (
var EnableStatsBasedRebalancing = settings.RegisterBoolSetting(
"kv.allocator.stat_based_rebalancing.enabled",
"set to enable rebalancing range replicas and leases to more evenly distribute read and write load across the stores in a cluster",
false, // TODO(a-robinson): switch to true for v2.1 once the store-rebalancer is done
true,
)

// rangeRebalanceThreshold is the minimum ratio of a store's range count to
Expand All @@ -85,12 +85,13 @@ var rangeRebalanceThreshold = settings.RegisterNonNegativeFloatSetting(
// TODO(a-robinson): Should disk usage be held to a higher standard than this?
var statRebalanceThreshold = settings.RegisterNonNegativeFloatSetting(
"kv.allocator.stat_rebalance_threshold",
"minimum fraction away from the mean a store's stats (like disk usage or writes per second) can be before it is considered overfull or underfull",
"minimum fraction away from the mean a store's stats (such as queries per second) can be before it is considered overfull or underfull",
0.20,
)

type scorerOptions struct {
deterministic bool
balanceQPSInsteadOfCount bool
statsBasedRebalancingEnabled bool
rangeRebalanceThreshold float64
statRebalanceThreshold float64
Expand Down Expand Up @@ -442,11 +443,22 @@ func allocateCandidates(
}
diversityScore := diversityAllocateScore(s, existingNodeLocalities)
balanceScore := balanceScore(sl, s.Capacity, rangeInfo, options)
var convergesScore int
if options.balanceQPSInsteadOfCount {
if s.Capacity.QueriesPerSecond < underfullStatThreshold(options, sl.candidateQueriesPerSecond.mean) {
convergesScore = 1
} else if s.Capacity.QueriesPerSecond < sl.candidateQueriesPerSecond.mean {
convergesScore = 0
} else {
convergesScore = -1
}
}
candidates = append(candidates, candidate{
store: s,
valid: constraintsOK,
necessary: necessary,
diversityScore: diversityScore,
convergesScore: convergesScore,
balanceScore: balanceScore,
rangeCount: int(s.Capacity.RangeCount),
})
Expand Down
1 change: 1 addition & 0 deletions pkg/storage/replica_rankings.go
Original file line number Diff line number Diff line change
Expand Up @@ -28,6 +28,7 @@ const (
type replicaWithStats struct {
repl *Replica
qps float64
// TODO(a-robinson): Include writes-per-second and logicalBytes of storage?
}

// replicaRankings maintains top-k orderings of the replicas in a store along
Expand Down
2 changes: 1 addition & 1 deletion pkg/storage/store.go
Original file line number Diff line number Diff line change
Expand Up @@ -1546,7 +1546,7 @@ func (s *Store) Start(ctx context.Context, stopper *stop.Stopper) error {
s.startLeaseRenewer(ctx)
}

s.storeRebalancer.Start(ctx, s.stopper, s.StoreID())
s.storeRebalancer.Start(ctx, s.stopper)

// Start the storage engine compactor.
if envutil.EnvOrDefaultBool("COCKROACH_ENABLE_COMPACTOR", true) {
Expand Down
Loading

0 comments on commit 2414b16

Please sign in to comment.