Skip to content

Commit

Permalink
*: s/the the /the /
Browse files Browse the repository at this point in the history
This is a surprisingly common typo!

Release note: None
  • Loading branch information
jordanlewis committed Nov 30, 2020
1 parent b7de6c9 commit 293c931
Show file tree
Hide file tree
Showing 54 changed files with 90 additions and 90 deletions.
2 changes: 1 addition & 1 deletion pkg/base/test_server_args.go
Original file line number Diff line number Diff line change
Expand Up @@ -68,7 +68,7 @@ type TestServerArgs struct {
// JoinAddr is the address of a node we are joining.
//
// If left empty and the TestServer is being added to a nonempty cluster, this
// will be set to the the address of the cluster's first node.
// will be set to the address of the cluster's first node.
JoinAddr string

// StoreSpecs define the stores for this server. If you want more than
Expand Down
2 changes: 1 addition & 1 deletion pkg/ccl/backupccl/restore_job.go
Original file line number Diff line number Diff line change
Expand Up @@ -275,7 +275,7 @@ rangeLoop:
return requestEntries, maxEndTime, nil
}

// WriteDescriptors writes all the the new descriptors: First the ID ->
// WriteDescriptors writes all the new descriptors: First the ID ->
// TableDescriptor for the new table, then flip (or initialize) the name -> ID
// entry so any new queries will use the new one. The tables are assigned the
// permissions of their parent database and the user must have CREATE permission
Expand Down
2 changes: 1 addition & 1 deletion pkg/ccl/backupccl/targets.go
Original file line number Diff line number Diff line change
Expand Up @@ -485,7 +485,7 @@ func descriptorsMatchingTargets(
// descriptor revision matches if it is an earlier revision of a descriptor in
// descs (same ID) or has parentID in `expanded`. Deleted descriptors are
// represented as nil. Fills in the `priorIDs` map in the process, which maps
// a descriptor the the ID by which it was previously known (e.g pre-TRUNCATE).
// a descriptor the ID by which it was previously known (e.g pre-TRUNCATE).
func getRelevantDescChanges(
ctx context.Context,
codec keys.SQLCodec,
Expand Down
2 changes: 1 addition & 1 deletion pkg/ccl/changefeedccl/sink_cloudstorage.go
Original file line number Diff line number Diff line change
Expand Up @@ -180,7 +180,7 @@ func (f *cloudStorageSinkFile) Write(p []byte) (int, error) {
//
// The resolved timestamp files are named `<timestamp>.RESOLVED`. This is
// carefully done so that we can offer the following external guarantee: At any
// given time, if the the files are iterated in lexicographic filename order,
// given time, if the files are iterated in lexicographic filename order,
// then encountering any filename containing `RESOLVED` means that everything
// before it is finalized (and thus can be ingested into some other system and
// deleted, included in hive queries, etc). A typical user of cloudStorageSink
Expand Down
2 changes: 1 addition & 1 deletion pkg/cli/demo.go
Original file line number Diff line number Diff line change
Expand Up @@ -193,7 +193,7 @@ func checkDemoConfiguration(
}

demoCtx.disableTelemetry = cluster.TelemetryOptOut()
// disableLicenseAcquisition can also be set by the the user as an
// disableLicenseAcquisition can also be set by the user as an
// input flag, so make sure it include it when considering the final
// value of disableLicenseAcquisition.
demoCtx.disableLicenseAcquisition =
Expand Down
4 changes: 2 additions & 2 deletions pkg/cmd/roachprod/install/cluster_synced.go
Original file line number Diff line number Diff line change
Expand Up @@ -625,7 +625,7 @@ tar cf - .ssh/id_rsa .ssh/id_rsa.pub .ssh/authorized_keys
return nil, nil
})

// Skip the the first node which is where we generated the key.
// Skip the first node which is where we generated the key.
nodes := c.Nodes[1:]
c.Parallel("distributing ssh key", len(nodes), 0, func(i int) ([]byte, error) {
sess, err := c.newSession(nodes[i])
Expand Down Expand Up @@ -915,7 +915,7 @@ tar cvf certs.tar certs
os.Exit(1)
}

// Skip the the first node which is where we generated the certs.
// Skip the first node which is where we generated the certs.
display = c.Name + ": distributing certs"
nodes = nodes[1:]
c.Parallel(display, len(nodes), 0, func(i int) ([]byte, error) {
Expand Down
2 changes: 1 addition & 1 deletion pkg/geo/geodist/geodist.go
Original file line number Diff line number Diff line change
Expand Up @@ -370,7 +370,7 @@ func onPolygonToPolygon(c DistanceCalculator, a Polygon, b Polygon) bool {
bFirstPoint := b.LinearRing(0).Vertex(0)

// MinDistance:
// If there is at least one point on the the exterior ring of B that is outside the exterior ring
// If there is at least one point on the exterior ring of B that is outside the exterior ring
// of A, then we have one of these two cases:
// * The exterior rings of A and B intersect. The distance can always be found by comparing
// the exterior rings.
Expand Down
2 changes: 1 addition & 1 deletion pkg/internal/client/requestbatcher/batcher_test.go
Original file line number Diff line number Diff line change
Expand Up @@ -311,7 +311,7 @@ func TestPanicWithNilStopper(t *testing.T) {
New(Config{Sender: make(chanSender)})
}

// TestBatchTimeout verifies the the RequestBatcher uses the context with the
// TestBatchTimeout verifies the RequestBatcher uses the context with the
// deadline from the latest call to send.
func TestBatchTimeout(t *testing.T) {
defer leaktest.AfterTest(t)()
Expand Down
2 changes: 1 addition & 1 deletion pkg/kv/db.go
Original file line number Diff line number Diff line change
Expand Up @@ -133,7 +133,7 @@ type Result struct {
// Keys is set by some operations instead of returning the rows themselves.
Keys []roachpb.Key

// ResumeSpan is the the span to be used on the next operation in a
// ResumeSpan is the span to be used on the next operation in a
// sequence of operations. It is returned whenever an operation over a
// span of keys is bounded and the operation returns before completely
// running over the span. It allows the operation to be called again with
Expand Down
2 changes: 1 addition & 1 deletion pkg/kv/kvclient/kvcoord/dist_sender.go
Original file line number Diff line number Diff line change
Expand Up @@ -1049,7 +1049,7 @@ func (ds *DistSender) detectIntentMissingDueToIntentResolution(
// finalized as committed and already had its transaction record GCed. Both
// these cases return an ABORTED txn; in the GC case the record has been
// synthesized.
// If the the record has been GC'ed, then we can't distinguish between the
// If the record has been GC'ed, then we can't distinguish between the
// two cases, and so we're forced to return an ambiguous error. On the other
// hand, if the record exists, then we know that the transaction did not
// commit because a committed record cannot be GC'ed and the recreated as
Expand Down
2 changes: 1 addition & 1 deletion pkg/kv/kvserver/batcheval/cmd_lease_request.go
Original file line number Diff line number Diff line change
Expand Up @@ -112,7 +112,7 @@ func RequestLease(
// same store.
//
// Note also that leasePostApply makes sure to update the timestamp cache in
// this case: even though the lease holder does not change, the the sequence
// this case: even though the lease holder does not change, the sequence
// number does and this triggers a low water mark bump.
//
// The bug prevented with this is unlikely to occur in practice
Expand Down
10 changes: 5 additions & 5 deletions pkg/kv/kvserver/client_merge_test.go
Original file line number Diff line number Diff line change
Expand Up @@ -1100,7 +1100,7 @@ func TestStoreRangeMergeInFlightTxns(t *testing.T) {
}
rhsKey := roachpb.Key("cc")

// Set a timeout, and set the the transaction liveness threshold to
// Set a timeout, and set the transaction liveness threshold to
// something much larger than our timeout. We want transactions to get stuck
// in the transaction wait queue and trigger the timeout if we forget to
// clear it.
Expand Down Expand Up @@ -1484,7 +1484,7 @@ func TestStoreRangeMergeRHSLeaseExpiration(t *testing.T) {
// considered live. The automatic heartbeat might not come for a while.
require.NoError(t, mtc.heartbeatLiveness(ctx, 0))

// Send several get and put requests to the the RHS. The first of these to
// Send several get and put requests to the RHS. The first of these to
// arrive will acquire the lease; the remaining requests will wait for that
// lease acquisition to complete. Then all requests should block waiting for
// the Subsume request to complete. By sending several of these requests in
Expand Down Expand Up @@ -1569,7 +1569,7 @@ func TestStoreRangeMergeRHSLeaseExpiration(t *testing.T) {
}
}

// TestStoreRangeMergeCheckConsistencyAfterSubsumption verifies the the following:
// TestStoreRangeMergeCheckConsistencyAfterSubsumption verifies the following:
// 1. While a range is subsumed, ComputeChecksum requests wait until the merge
// is complete before proceeding.
// 2. Once a merge is aborted, pending (and future) requests will be allowed to
Expand Down Expand Up @@ -2235,7 +2235,7 @@ func TestStoreRangeMergeSlowAbandonedFollower(t *testing.T) {
// it may require the replica GC queue. In rare cases the LHS will never
// hear about the merge and may need to be GC'd on its own.
testutils.SucceedsSoon(t, func() error {
// Make the the LHS gets destroyed.
// Make the LHS gets destroyed.
if lhsRepl, err := store2.GetReplica(lhsDesc.RangeID); err == nil {
if err := store2.ManualReplicaGC(lhsRepl); err != nil {
t.Fatal(err)
Expand Down Expand Up @@ -3130,7 +3130,7 @@ func TestStoreRangeMergeRaftSnapshot(t *testing.T) {
return err
}

// Keep adding kv data to the SST until the the key exceeds the
// Keep adding kv data to the SST until the key exceeds the
// bounds of the range, then proceed to the next range.
for ; ; it.Next() {
valid, err := it.Valid()
Expand Down
2 changes: 1 addition & 1 deletion pkg/kv/kvserver/client_replica_test.go
Original file line number Diff line number Diff line change
Expand Up @@ -2189,7 +2189,7 @@ func TestConcurrentAdminChangeReplicasRequests(t *testing.T) {

// TestRandomConcurrentAdminChangeReplicasRequests ensures that when multiple
// AdminChangeReplicasRequests are issued concurrently, so long as requests
// provide the the value of the RangeDescriptor they will not accidentally
// provide the value of the RangeDescriptor they will not accidentally
// perform replication changes. In particular this test runs a number of
// concurrent actors which all use the same expectations of the RangeDescriptor
// and verifies that at most one actor succeeds in making all of its changes.
Expand Down
2 changes: 1 addition & 1 deletion pkg/kv/kvserver/closedts/closedts.go
Original file line number Diff line number Diff line change
Expand Up @@ -129,7 +129,7 @@ type Producer interface {
// resulting entries to all of its subscribers.
// 3. it accepts notifications from other nodes, passing these updates through
// to its local storage, so that
// 4. the CanServe method determines via the the underlying storage whether a
// 4. the CanServe method determines via the underlying storage whether a
// given read can be satisfied via follower reads.
// 5. the MaxClosed method determines via the underlying storage what the maximum
// closed timestamp is for the specified LAI.
Expand Down
2 changes: 1 addition & 1 deletion pkg/kv/kvserver/replica.go
Original file line number Diff line number Diff line change
Expand Up @@ -911,7 +911,7 @@ func (r *Replica) getFreezeStartRLocked() hlc.Timestamp {
return r.mu.freezeStart
}

// setLastReplicaDescriptors sets the the most recently seen replica
// setLastReplicaDescriptors sets the most recently seen replica
// descriptors to those contained in the *RaftMessageRequest, acquiring r.mu
// to do so.
func (r *Replica) setLastReplicaDescriptors(req *RaftMessageRequest) {
Expand Down
2 changes: 1 addition & 1 deletion pkg/kv/kvserver/replica_raftstorage.go
Original file line number Diff line number Diff line change
Expand Up @@ -467,7 +467,7 @@ type OutgoingSnapshot struct {
Iter *rditer.ReplicaEngineDataIterator
// The replica state within the snapshot.
State kvserverpb.ReplicaState
// Allows access the the original Replica's sideloaded storage. Note that
// Allows access the original Replica's sideloaded storage. Note that
// this isn't a snapshot of the sideloaded storage congruent with EngineSnap
// or RaftSnap -- a log truncation could have removed files from the
// sideloaded storage in the meantime.
Expand Down
2 changes: 1 addition & 1 deletion pkg/kv/kvserver/replica_stats.go
Original file line number Diff line number Diff line change
Expand Up @@ -179,7 +179,7 @@ func (rs *replicaStats) perLocalityDecayingQPS() (perLocalityCounts, time.Durati
// avgQPS returns the average requests-per-second and the amount of time
// over which the stat was accumulated. Note that these averages are exact,
// not exponentially decayed (there isn't a ton of justification for going
// one way or the the other, but not decaying makes the average more stable,
// one way or the other, but not decaying makes the average more stable,
// which is probably better for avoiding rebalance thrashing).
func (rs *replicaStats) avgQPS() (float64, time.Duration) {
now := timeutil.Unix(0, rs.clock.PhysicalNow())
Expand Down
2 changes: 1 addition & 1 deletion pkg/kv/kvserver/store.go
Original file line number Diff line number Diff line change
Expand Up @@ -2773,7 +2773,7 @@ func WriteClusterVersion(
return storage.MVCCPutProto(ctx, writer, nil, keys.StoreClusterVersionKey(), hlc.Timestamp{}, nil, &cv)
}

// ReadClusterVersion reads the the cluster version from the store-local version
// ReadClusterVersion reads the cluster version from the store-local version
// key. Returns an empty version if the key is not found.
func ReadClusterVersion(
ctx context.Context, reader storage.Reader,
Expand Down
2 changes: 1 addition & 1 deletion pkg/kv/kvserver/store_init.go
Original file line number Diff line number Diff line change
Expand Up @@ -106,7 +106,7 @@ func WriteInitialClusterData(
// The last range has id = len(splits) + 1
rangeIDVal.SetInt(int64(len(splits) + 1))

// We're the the first node in the cluster, let's seed our liveness record.
// We're the first node in the cluster, let's seed our liveness record.
// It's crucial that we do to maintain the invariant that there's always a
// liveness record for a given node. We'll do something similar through the
// join RPC when adding new nodes to an already bootstrapped cluster [1].
Expand Down
2 changes: 1 addition & 1 deletion pkg/kv/kvserver/tscache/interval_skl.go
Original file line number Diff line number Diff line change
Expand Up @@ -536,7 +536,7 @@ func (s *intervalSkl) LookupTimestampRange(from, to []byte, opt rangeOptions) ca
val, _ = ratchetValue(val, val2)
}

// Return the higher value from the the page lookups and the floor
// Return the higher value from the page lookups and the floor
// timestamp.
floorVal := cacheValue{ts: s.floorTS, txnID: noTxnID}
val, _ = ratchetValue(val, floorVal)
Expand Down
2 changes: 1 addition & 1 deletion pkg/server/admin_test.go
Original file line number Diff line number Diff line change
Expand Up @@ -1870,7 +1870,7 @@ func TestEnqueueRange(t *testing.T) {

// Up-replicate r1 to all 3 nodes. We use manual replication to avoid lease
// transfers causing temporary conditions in which no store is the
// leaseholder, which can break the the tests below.
// leaseholder, which can break the tests below.
_, err := testCluster.AddVoters(roachpb.KeyMin, testCluster.Target(1), testCluster.Target(2))
if err != nil {
t.Fatal(err)
Expand Down
2 changes: 1 addition & 1 deletion pkg/server/server_sql.go
Original file line number Diff line number Diff line change
Expand Up @@ -606,7 +606,7 @@ func newSQLServer(ctx context.Context, cfg sqlServerArgs) (*sqlServer, error) {
// Now that we have a pgwire.Server (which has a sql.Server), we can close a
// circular dependency between the rowexec.Server and sql.Server and set
// SessionBoundInternalExecutorFactory. The same applies for setting a
// SessionBoundInternalExecutor on the the job registry.
// SessionBoundInternalExecutor on the job registry.
ieFactory := func(
ctx context.Context, sessionData *sessiondata.SessionData,
) sqlutil.InternalExecutor {
Expand Down
2 changes: 1 addition & 1 deletion pkg/sql/backfill.go
Original file line number Diff line number Diff line change
Expand Up @@ -955,7 +955,7 @@ func (sc *SchemaChanger) distBackfill(
if err := sc.db.Txn(ctx, func(ctx context.Context, txn *kv.Txn) error {
updatedTodoSpans = todoSpans
// Report schema change progress. We define progress at this point
// as the the fraction of fully-backfilled ranges of the primary index of
// as the fraction of fully-backfilled ranges of the primary index of
// the table being scanned. Since we may have already modified the
// fraction completed of our job from the 10% allocated to completing the
// schema change state machine or from a previous backfill attempt,
Expand Down
2 changes: 1 addition & 1 deletion pkg/sql/colexec/execgen/inline.go
Original file line number Diff line number Diff line change
Expand Up @@ -316,7 +316,7 @@ func extractReturnValues(decl *dst.FuncDecl) (retValDeclStmt dst.Stmt, retValNam
// b string = y
// )
//
// In the case where the formal parameter name is the same as the the input
// In the case where the formal parameter name is the same as the input
// parameter name, no extra assignment is created.
func getFormalParamReassignments(decl *dst.FuncDecl, callExpr *dst.CallExpr) dst.Stmt {
formalParams := decl.Type.Params.List
Expand Down
2 changes: 1 addition & 1 deletion pkg/sql/colexec/hashtable_tmpl.go
Original file line number Diff line number Diff line change
Expand Up @@ -52,7 +52,7 @@ const _RIGHT_CANONICAL_TYPE_FAMILY = types.UnknownFamily
const _RIGHT_TYPE_WIDTH = 0

// _ASSIGN_NE is the template equality function for assigning the first input
// to the result of the the second input != the third input.
// to the result of the second input != the third input.
func _ASSIGN_NE(_, _, _, _, _, _ interface{}) int {
colexecerror.InternalError(errors.AssertionFailedf(""))
}
Expand Down
2 changes: 1 addition & 1 deletion pkg/sql/colexec/mergejoinbase_tmpl.go
Original file line number Diff line number Diff line change
Expand Up @@ -47,7 +47,7 @@ const _CANONICAL_TYPE_FAMILY = types.UnknownFamily
const _TYPE_WIDTH = 0

// _ASSIGN_EQ is the template equality function for assigning the first input
// to the result of the the second input == the third input.
// to the result of the second input == the third input.
func _ASSIGN_EQ(_, _, _, _, _, _ interface{}) int {
colexecerror.InternalError(errors.AssertionFailedf(""))
}
Expand Down
2 changes: 1 addition & 1 deletion pkg/sql/colexec/quicksort.eg.go

Some generated files are not rendered by default. Learn more about how customized files appear on GitHub.

2 changes: 1 addition & 1 deletion pkg/sql/colexec/quicksort_tmpl.go
Original file line number Diff line number Diff line change
Expand Up @@ -20,7 +20,7 @@ package colexec

import "context"

// This file is copied from the the Go standard library's sort
// This file is copied from the Go standard library's sort
// implementation, found in https://golang.org/src/sort/sort.go. The only
// modifications are to template each function into each sort_* struct, so
// that the sort methods can be called without incurring any interface overhead
Expand Down
2 changes: 1 addition & 1 deletion pkg/sql/colexec/routers.go
Original file line number Diff line number Diff line change
Expand Up @@ -368,7 +368,7 @@ func (o *routerOutputOp) forwardErr(err error) {
// an empty selection slice will add no elements. Note that the selection vector
// on the batch is ignored. This is so that callers of addBatch can push the
// same batch with different selection vectors to many different outputs.
// True is returned if the the output changes state to blocked (note: if the
// True is returned if the output changes state to blocked (note: if the
// output is already blocked, false is returned).
// TODO(asubiotto): We should explore pipelining addBatch if disk-spilling
// performance becomes a concern. The main router goroutine will be writing to
Expand Down
2 changes: 1 addition & 1 deletion pkg/sql/covering/overlap_merge.go
Original file line number Diff line number Diff line change
Expand Up @@ -83,7 +83,7 @@ func OverlapCoveringMerge(coverings []Covering) []Range {

// TODO(dan): Verify that the ranges in each covering are non-overlapping.

// We would like to flatten all coverings on the the single line. Assume that total
// We would like to flatten all coverings on the single line. Assume that total
// amount of coverings is N, then overall there are 2*N endpoints. Flatten coverings
// on single line and sort them will take O(NlogN) time. Later we pass over endpoints
// one by one and append payloads.
Expand Down
2 changes: 1 addition & 1 deletion pkg/sql/crdb_internal_test.go
Original file line number Diff line number Diff line change
Expand Up @@ -96,7 +96,7 @@ func TestRangeLocalityBasedOnNodeIDs(t *testing.T) {
defer tc.Stopper().Stop(ctx)
assert.EqualValues(t, 1, tc.Servers[len(tc.Servers)-1].GetFirstStoreID())

// Set to 2 so the the next store id will be 3.
// Set to 2 so the next store id will be 3.
assert.NoError(t, tc.Servers[0].DB().Put(ctx, keys.StoreIDGenerator, 2))

// NodeID=2, StoreID=3
Expand Down
2 changes: 1 addition & 1 deletion pkg/sql/delayed.go
Original file line number Diff line number Diff line change
Expand Up @@ -55,7 +55,7 @@ func (d *delayedNode) startExec(params runParams) error {

// Recursively invoke startExec on new plan. Normally, startExec doesn't
// recurse - calling children is handled by the planNode walker. The reason
// this won't suffice here is that the the child of this node doesn't exist
// this won't suffice here is that the child of this node doesn't exist
// until after startExec is invoked.
return startExec(params, plan)
}
2 changes: 1 addition & 1 deletion pkg/sql/delegate/show_range_for_row.go
Original file line number Diff line number Diff line change
Expand Up @@ -42,7 +42,7 @@ func (d *delegator) delegateShowRangeForRow(n *tree.ShowRangeForRow) (tree.State

// Format the expressions into a string to be passed into the
// crdb_internal.encode_key function. We have to be sneaky here and special
// case when exprs has length 1 and place a comma after the the single tuple
// case when exprs has length 1 and place a comma after the single tuple
// element so that we can deduce the expression actually has a tuple type for
// the crdb_internal.encode_key function.
// Example: exprs = (1)
Expand Down
Loading

0 comments on commit 293c931

Please sign in to comment.