-
Notifications
You must be signed in to change notification settings - Fork 142
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Skip the mass transit test to see if it solves flake issues (#5861 -> v2) #5911
Conversation
## Summary of changes Skip the mass transit smoke test as it seems to be a cause of a lot of flakiness ## Reason for change We've seen a lot of errors in the `CheckBuildlogsForErr` stage: ``` CheckBuildLogsForErr: 03:08:39 [Error] An error occurred while sending data to the agent at http://127.0.0.1:39573/v0.4/traces. If the error isn't transient, please check https://docs.datadoghq.com/tracing/troubleshooting/connection_errors/?code-lang=dotnet for guidance. System.Net.Http.HttpRequestException: Error while copying content to a stream. ``` These seemed to get a lot worse after we disabled keep-alive, but that's anecdotal. ## Implementation details It's not entirely clear if the problem is just coincidentally related to the MassTransit test (i.e. it's a test ordering process) or if it's actually something about the test. As a check I tried skipping the test in this branch and did 4 full (all TFM) integration tests runs, and didn't see the issue again. It's all still anecdotal, but rather trade off flakiness here. If the problem reappears subsequently, we can look into it again further. ## Test coverage Did 4 full runs, and didn't see the issue again
Datadog ReportBranch report: ✅ 0 Failed, 353745 Passed, 1797 Skipped, 14h 35m 6.14s Total Time |
Execution-Time Benchmarks Report ⏱️Execution-time results for samples comparing the following branches/commits: Execution-time benchmarks measure the whole time it takes to execute a program. And are intended to measure the one-off costs. Cases where the execution time results for the PR are worse than latest master results are shown in red. The following thresholds were used for comparing the execution times:
Note that these results are based on a single point-in-time result for each branch. For full results, see the dashboard. Graphs show the p99 interval based on the mean and StdDev of the test run, as well as the mean value of the run (shown as a diamond below the graph). gantt
title Execution time (ms) FakeDbCommand (.NET Framework 4.6.2)
dateFormat X
axisFormat %s
todayMarker off
section Baseline
This PR (5911) - mean (77ms) : 62, 92
. : milestone, 77,
section CallTarget+Inlining+NGEN
This PR (5911) - mean (1,072ms) : 1050, 1093
. : milestone, 1072,
gantt
title Execution time (ms) FakeDbCommand (.NET Core 3.1)
dateFormat X
axisFormat %s
todayMarker off
section Baseline
This PR (5911) - mean (111ms) : 107, 114
. : milestone, 111,
section CallTarget+Inlining+NGEN
This PR (5911) - mean (782ms) : 763, 801
. : milestone, 782,
gantt
title Execution time (ms) FakeDbCommand (.NET 6)
dateFormat X
axisFormat %s
todayMarker off
section Baseline
This PR (5911) - mean (94ms) : 92, 97
. : milestone, 94,
section CallTarget+Inlining+NGEN
This PR (5911) - mean (723ms) : 705, 742
. : milestone, 723,
gantt
title Execution time (ms) HttpMessageHandler (.NET Framework 4.6.2)
dateFormat X
axisFormat %s
todayMarker off
section Baseline
This PR (5911) - mean (191ms) : 188, 195
. : milestone, 191,
section CallTarget+Inlining+NGEN
This PR (5911) - mean (1,155ms) : 1128, 1182
. : milestone, 1155,
gantt
title Execution time (ms) HttpMessageHandler (.NET Core 3.1)
dateFormat X
axisFormat %s
todayMarker off
section Baseline
This PR (5911) - mean (277ms) : 272, 281
. : milestone, 277,
section CallTarget+Inlining+NGEN
This PR (5911) - mean (947ms) : 926, 969
. : milestone, 947,
gantt
title Execution time (ms) HttpMessageHandler (.NET 6)
dateFormat X
axisFormat %s
todayMarker off
section Baseline
This PR (5911) - mean (265ms) : 260, 270
. : milestone, 265,
section CallTarget+Inlining+NGEN
This PR (5911) - mean (930ms) : 908, 951
. : milestone, 930,
|
Benchmarks Report for tracer 🐌Benchmarks for #5911 compared to master:
The following thresholds were used for comparing the benchmark speeds:
Allocation changes below 0.5% are ignored. Benchmark detailsBenchmarks.Trace.ActivityBenchmark - Same speed ✔️ Same allocations ✔️Raw results
Benchmarks.Trace.AgentWriterBenchmark - Same speed ✔️ Same allocations ✔️Raw results
Benchmarks.Trace.AspNetCoreBenchmark - Same speed ✔️ Same allocations ✔️Raw results
Benchmarks.Trace.CIVisibilityProtocolWriterBenchmark - Same speed ✔️ Same allocations ✔️Raw results
Benchmarks.Trace.DbCommandBenchmark - Same speed ✔️ Same allocations ✔️Raw results
Benchmarks.Trace.ElasticsearchBenchmark - Same speed ✔️ Same allocations ✔️Raw results
Benchmarks.Trace.GraphQLBenchmark - Same speed ✔️ Same allocations ✔️Raw results
Benchmarks.Trace.HttpClientBenchmark - Same speed ✔️ Same allocations ✔️Raw results
Benchmarks.Trace.ILoggerBenchmark - Same speed ✔️ Same allocations ✔️Raw results
Benchmarks.Trace.Log4netBenchmark - Same speed ✔️ Same allocations ✔️Raw results
Benchmarks.Trace.NLogBenchmark - Same speed ✔️ Same allocations ✔️Raw results
Benchmarks.Trace.RedisBenchmark - Same speed ✔️ Same allocations ✔️Raw results
Benchmarks.Trace.SerilogBenchmark - Same speed ✔️ Same allocations ✔️Raw results
Benchmarks.Trace.SpanBenchmark - Faster 🎉 Same allocations ✔️
|
Benchmark | base/diff | Base Median (ns) | Diff Median (ns) | Modality |
---|---|---|---|---|
Benchmarks.Trace.SpanBenchmark.StartFinishScope‑net6.0 | 1.128 | 547.60 | 485.28 |
Raw results
Branch | Method | Toolchain | Mean | StdError | StdDev | Gen 0 | Gen 1 | Gen 2 | Allocated |
---|---|---|---|---|---|---|---|---|---|
master | StartFinishSpan |
net6.0 | 401ns | 0.224ns | 0.868ns | 0.00809 | 0 | 0 | 576 B |
master | StartFinishSpan |
netcoreapp3.1 | 567ns | 0.387ns | 1.5ns | 0.0079 | 0 | 0 | 576 B |
master | StartFinishSpan |
net472 | 598ns | 0.647ns | 2.5ns | 0.0916 | 0 | 0 | 578 B |
master | StartFinishScope |
net6.0 | 547ns | 0.227ns | 0.878ns | 0.00964 | 0 | 0 | 696 B |
master | StartFinishScope |
netcoreapp3.1 | 747ns | 0.654ns | 2.53ns | 0.0093 | 0 | 0 | 696 B |
master | StartFinishScope |
net472 | 889ns | 0.598ns | 2.32ns | 0.104 | 0 | 0 | 658 B |
#5911 | StartFinishSpan |
net6.0 | 399ns | 0.285ns | 1.07ns | 0.00802 | 0 | 0 | 576 B |
#5911 | StartFinishSpan |
netcoreapp3.1 | 554ns | 0.63ns | 2.27ns | 0.00768 | 0 | 0 | 576 B |
#5911 | StartFinishSpan |
net472 | 627ns | 0.515ns | 2ns | 0.0916 | 0 | 0 | 578 B |
#5911 | StartFinishScope |
net6.0 | 485ns | 0.385ns | 1.49ns | 0.00978 | 0 | 0 | 696 B |
#5911 | StartFinishScope |
netcoreapp3.1 | 702ns | 0.985ns | 3.81ns | 0.00919 | 0 | 0 | 696 B |
#5911 | StartFinishScope |
net472 | 835ns | 0.816ns | 3.16ns | 0.104 | 0 | 0 | 658 B |
Benchmarks.Trace.TraceAnnotationsBenchmark - Same speed ✔️ Same allocations ✔️
Raw results
Branch | Method | Toolchain | Mean | StdError | StdDev | Gen 0 | Gen 1 | Gen 2 | Allocated |
---|---|---|---|---|---|---|---|---|---|
master | RunOnMethodBegin |
net6.0 | 652ns | 0.672ns | 2.6ns | 0.00986 | 0 | 0 | 696 B |
master | RunOnMethodBegin |
netcoreapp3.1 | 929ns | 0.986ns | 3.82ns | 0.00924 | 0 | 0 | 696 B |
master | RunOnMethodBegin |
net472 | 1.06μs | 0.959ns | 3.71ns | 0.104 | 0 | 0 | 658 B |
#5911 | RunOnMethodBegin |
net6.0 | 589ns | 0.49ns | 1.9ns | 0.00971 | 0 | 0 | 696 B |
#5911 | RunOnMethodBegin |
netcoreapp3.1 | 900ns | 0.845ns | 3.27ns | 0.00947 | 0 | 0 | 696 B |
#5911 | RunOnMethodBegin |
net472 | 1.09μs | 1.38ns | 5.35ns | 0.104 | 0 | 0 | 658 B |
Summary of changes
Skip the mass transit smoke test as it seems to be a cause of a lot of flakiness
Reason for change
We've seen a lot of errors in the
CheckBuildlogsForErr
stage:These seemed to get a lot worse after we disabled keep-alive, but that's anecdotal.
Implementation details
It's not entirely clear if the problem is just coincidentally related to the MassTransit test (i.e. it's a test ordering process) or if it's actually something about the test.
As a check I tried skipping the test in this branch and did 4 full (all TFM) integration tests runs, and didn't see the issue again. It's all still anecdotal, but rather trade off flakiness here. If the problem reappears subsequently, we can look into it again further.
Test coverage
Did 4 full runs, and didn't see the issue again
Other details
Backport of #5861 (as still getting a lot of flake on the release/2.x branch)