Skip to content

Commit

Permalink
Update 2022-09-08-thanos-at-medallia.md (thanos-io#5884)
Browse files Browse the repository at this point in the history
Signed-off-by: Vic Thomas <[email protected]>

Signed-off-by: Vic Thomas <[email protected]>
  • Loading branch information
VicThomas authored and Nathaniel Graham committed May 18, 2023
1 parent 194d3c8 commit 23ae1f1
Showing 1 changed file with 3 additions and 1 deletion.
4 changes: 3 additions & 1 deletion docs/blog/2022-09-08-thanos-at-medallia.md
Original file line number Diff line number Diff line change
Expand Up @@ -70,7 +70,7 @@ One of the key principles of the architectural design was to construct the Thano

This solution was elegant and relatively easy to support. The colocation data centers are sufficiently large to easily accommodate all the Thanos and Prometheus components.

However, by late 2020 a shift in operating strategy for Medallia was to make use of public cloud providers such as Amazon Web Services (AWS) and Oracle Cloud Infrastructure (OCI). So, the POE team faced a new mandate – incorporate metrics generated within these virtual data centers into the global view.
However, by late 2020 Medallia had expanded its strategy to make more use of public cloud providers such as Amazon Web Services (AWS) and Oracle Cloud Infrastructure (OCI). So, the POE team faced a new mandate – incorporate metrics generated within these virtual data centers into the global view.

In general, these virtual data centers are small Kubernetes clusters, with compute nodes that are much smaller in terms of CPU and memory capacity than what we enjoy in the colocation data centers. Deploying the full complement of Thanos and Prometheus components in such an environment would require an unacceptably high percentage of capacity within those environments.

Expand All @@ -86,6 +86,8 @@ For the past two years, this hybrid solution – both federated queries and remo

![Giraffe Architecture](img/architecture.png)

Note that the architecture diagram is representative, intended so show how it can, by design, easily be extended to all the major public clouds.

### Traffic and Data Size

As mentioned earlier, at any given moment in time, the current number of active series is in the ballpark of 1 Billion, with our ingestion throughput at approximately 15 million samples per second due to our once-per-minute scrape frequency. We retain samples for 366 days – long enough to accommodate a full leap year.
Expand Down

0 comments on commit 23ae1f1

Please sign in to comment.