There are two types of sampling in tracing; head and tail.
- Head sampling is performed at the source, the choice is made within the traced component as to whether the trace should be sampled.
- Tail sampling is performed in infrastructure downstream from the traced component. This could be in a collector or the tracing system itself.
As mentioned in the design document, there are three kinds of head sampler provided out of the box,
although you are welcome to create other implementations of the SpanSampler
or TailSpanSampler
interfaces:
- Always
- Never
- Probabilistic
- Rate
This sampler ensures all spans will be sent to downstream components. You can use the sampler with the code block below:
import io.janstenpickle.trace4cats.kernel.SpanSampler
SpanSampler.always
This sampler ensures no spans will be sent to downstream components. You can use the sampler with the code block below:
import io.janstenpickle.trace4cats.kernel.SpanSampler
SpanSampler.never
This sampler uses the trace ID to decide whether a trace should be sampled, based on a probability parameter, which must
be between 0.0
and 1.0
, where 0.0
cause no spans to be sent downstream and 1.0
will cause all spans to be
forwarded.
The parameter rootSpansOnly
, which defaults to true
, configures whether to apply the sampler to spans which have
parents. You may want to set this to false
if upstream systems are making many requests, having a negative
performance impact on the traced application.
You can use the sampler with the code block below:
import io.janstenpickle.trace4cats.kernel.SpanSampler
SpanSampler.probabilistic(probability = 0.05, rootSpansOnly = false)
The rate span sampler uses a token bucket to control rate and ensure bursts of spans don't overwhelm downstream components. There are two parameters required to configure this sampler;
- Bucket size: maximum number of tokens that spans that may be allocated in one go
- Token rate: how often tokens are added back to the bucket
In order to use the rate sampler, add the following dependency to your build.sbt
:
"io.janstenpickle" %% "trace4cats-rate-sampling" % "0.14.0"
You can use the sampler with the code block below. For a more complete example see here.
import io.janstenpickle.trace4cats.rate.sampling.RateSpanSampler
import scala.concurrent.duration._
RateSpanSampler.create[IO](bucketSize = 100, tokenRate = 10.millis)
This allows for sampling to be configured by some dynamic means, this could be some central control plane or, in the future, Jaeger adaptive sampling.
A couple of low level implementations are provided that may be integrated with dynamic configuration sources and used by higher level implementations.
HotSwapSpanSampler
exposes methods for updating the sampler implementation at runtime. When updating, if the ID of the sampler does not change, the sampler will not get updated. This is indicated by the return type ofF[Boolean]
on theupdateSampler
methodPollingSpanSampler
given some updating source at construction, polls it for updates and changes implementation accordingly
To access these implementations, include the following module in your project:
"io.janstenpickle" %% "trace4cats-dynamic-sampling" % "0.14.0"
This module provides an ADT called
SamplerConfig
that allows the sampler type and configuration to be changed at runtime. Companions to the low level implementations
described above are available in this module, which take SamplerConfig
as an argument:
To access these implementations, include the following module in your project:
"io.janstenpickle" %% "trace4cats-dynamic-sampling-config" % "0.14.0"
A couple of modules provide a solid implementation to allow samplers to be configured to runtime via HTTP endpoints. If
you just want Http4s endpoints in order to include them alongside other routes then just use the following module and
see
SamplerHttpRoutes
to get started
"io.janstenpickle" %% "trace4cats-dynamic-sampling-http4s" % "0.14.0"
If you want a complete implementation of a HTTP server then use the following module, and see the example.
"io.janstenpickle" %% "trace4cats-dynamic-sampling-http-server" % "0.14.0"
Assuming the Http4s endpoints are mounted on /trace4cats
or you are using the bundled HTTP server, the following
commands should work:
Get the current sampler config:
curl http://localhost:8080/trace4cats/config
Set the sampler to "always", to trace every request:
curl -XPOST -d '{ "samplerType": "Always" }' http://localhost:8080/trace4cats/config
Set the sampler to "never", to stop tracing completely:
curl -XPOST -d '{ "samplerType": "Never" }' http://localhost:8080/trace4cats/config
Set the sampler to probabilistic, with a probability parameter between 0 and 1:
curl -XPOST -d '{ "samplerType": "Probabilistic", "probability": 0.3 }' http://localhost:8080/trace4cats/config
Set the sampler to rate, with token bucket parameters:
curl -XPOST -d '{ "samplerType": "Rate", "bucketSize": 100, "tokenRate": 2.0 }' http://localhost:8080/trace4cats/config
Use a kill switch to stop collecting traces (same as "never" sampler):
curl -XPOST http://localhost:8080/trace4cats/killswitch
Tail sampling is slightly more complicated than head sampling, as sample decisions have to be applied post-hoc and therefore remembered for when subsequent spans for a trace arrive after the initial decision has been made.
Tail sampling allows you to centrally control the rate or span production and even which traces get exported to downstream systems. Depending on your topology, you may be exporting spans to a cloud tracing system such as Stackdriver Trace or Datadog where you are charged by the number of spans stored.
Your internal topology and configuration will influence the kind of decision store you use and whether it can be internal to the Collector, or external.
If you just use a probability based sampler, decisions are based on the trace ID and will always result in the same decision for the same trace ID. So if you are just using the probability sampler you will not need a sample decision store.
If you use Kafka, you can guarantee all spans for a trace will end up at the same collector and use the default cache based decision store. This is because the trace ID is used as the message key and the producer is set up to send all spans for the same trace ID to the same partition.
If you do not use Kafka, you should use an external sample store so that if you have more than one collector that can receive spans for the same trace, the sample decision may be shared between all instances. If you enable sampling in this kind of topology, then you could end up with incomplete traces in you tracing system. At present the only external sample store implementation available is Redis.
By using the tail sampling config on the Trace4Cats Collector you can centrally configure sampling for all of your traces.
The configuration fragment below sets up the following:
- A probability sampler, with a 50% chance the trace will be sampled
- A span name sampler, which filters traces whose name contain the configured strings
- This sampler only applies to root spans, subsequent spans will look up decisions for the trace against the decision store
- Note that the greater number of names, the more performance of the collector may be negatively impacted
- A rate sampler, which ensures the number of spans passing through the collector doesn't exceed a defined
rate
- Take care ensure that the configuration does not conflict with batching config
- A local cache decision store with
- TTL per trace of 10 minutes
- Maximum number of trace sample decision entries of 500000
sampling:
sample-probability: 0.05 # Optional - must be between 0 and 1.0. 1.0 being always sample, and 0.0 being never
drop-span-names: # Optional - name of spans to drop (may be partial match)
- healthcheck
- readiness
- metrics
rate: # Optional - rate sampling
max-batch-size: 1000
token-rate-millis: 10
cache-ttl-minutes: 10 # Cache duration for sample decision, defaults to 2 mins
max-cache-size: 500000 # Max number of entries in the sample decision cache, defaults to 1000000
redis: # Optional - use redis as a sample decision store
host: redis-host
port: 6378 # defaults to 6379
Alternatively you can configure a connection to a Redis cluster
sampling:
redis:
cluster:
- host: redis1
port: 6378 # defaults to 6379
- host: redis2