Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Make google cloud output split requests if they are too large #263

Merged
merged 5 commits into from
Apr 2, 2021

Conversation

djaglowski
Copy link
Member

@djaglowski djaglowski commented Mar 30, 2021

resolves #257

@codecov
Copy link

codecov bot commented Mar 31, 2021

Codecov Report

Merging #263 (94da7b4) into master (6749dc9) will increase coverage by 0.03%.
The diff coverage is 95.00%.

Impacted file tree graph

@@            Coverage Diff             @@
##           master     #263      +/-   ##
==========================================
+ Coverage   71.67%   71.70%   +0.03%     
==========================================
  Files         103      103              
  Lines        5676     5689      +13     
==========================================
+ Hits         4068     4079      +11     
- Misses       1172     1174       +2     
  Partials      436      436              
Impacted Files Coverage Δ
operator/buffer/memory.go 76.52% <0.00%> (ø)
...perator/builtin/output/googlecloud/google_cloud.go 54.55% <100.00%> (+5.20%) ⬆️
operator/flusher/flusher.go 84.00% <0.00%> (-4.00%) ⬇️
operator/builtin/output/forward/forward.go 56.52% <0.00%> (-2.90%) ⬇️

Continue to review full report at Codecov.

Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update 6749dc9...94da7b4. Read the comment docs.

@djaglowski djaglowski marked this pull request as ready for review March 31, 2021 13:20
@djaglowski djaglowski requested a review from jsirianni March 31, 2021 13:20
@djaglowski
Copy link
Member Author

Log Files Logs / Second CPU Avg (%) CPU Avg Δ (%) Memory Avg (MB) Memory Avg Δ (MB)
1 1000 1.8793484 +0.15516508 124.44774 +0.32569885
1 5000 6.569074 +0.4310422 132.93803 -0.7412567
1 10000 12.482988 +0.15513515 139.85965 -0.36206055
1 50000 59.757153 -4.1068764 176.33473 +0.53556824
1 100000 118.36555 +2.1602554 230.69559 +6.6095123
10 100 2.3449147 +0.01728487 128.44571 +1.4744034
10 500 7.500286 +0.1380024 138.7065 +2.5995483
10 1000 13.396839 -0.36199284 138.35184 -3.577179
10 5000 64.7081 +0.97166824 185.57382 +6.9484253
10 10000 136.8979 +9.948463 237.59227 +9.695587

Copy link
Member

@jsirianni jsirianni left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Looks great, and resolves the issue I was seeing in my environment.

@djaglowski
Copy link
Member Author

Log Files Logs / Second CPU Avg (%) CPU Avg Δ (%) Memory Avg (MB) Memory Avg Δ (MB)
1 1000 1.9310892 +0.20690584 123.20407 -0.91796875
1 5000 6.3277926 +0.18976068 135.1906 +1.5113068
1 10000 13.15556 +0.82770634 140.23007 +0.008361816
1 50000 59.673717 -4.1903114 177.9437 +2.1445312
1 100000 123.60719 +7.4019012 232.4363 +8.35022
10 100 2.5345833 +0.20695353 128.32045 +1.349144
10 500 7.000205 -0.36207867 137.01671 +0.9097595
10 1000 14.293792 +0.5349598 140.04243 -1.8865814
10 5000 65.67377 +1.937336 182.38052 +3.755127
10 10000 127.6933 +0.74385834 215.9154 -11.981277

@jsirianni jsirianni merged commit 4d0a4bc into master Apr 2, 2021
@jsirianni jsirianni deleted the not-master branch April 2, 2021 14:08
@KonradSchieban
Copy link
Contributor

👍 thanks, looks good!

@KonradSchieban
Copy link
Contributor

What is the message that is being logged in the case a larger message has been split? Is there a way to find that out? @djaglowski @jsirianni

@jsirianni
Copy link
Member

Large single messages are not being split. This is what we are doing:

  • If bulk submission (multiple log entries) too big, split the bulk entries in half and retry. Continue doing this until all messages are submitted.
  • If a single message is too big, replace it's record with the error given by the Google API.

For example, I submitted a log entry that was well over the limit, the result in Cloud Logging looks like this:
Screen Shot 2021-04-05 at 3 09 55 PM

@KonradSchieban
Copy link
Contributor

Thanks @jsirianni , this is what I was looking for!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

Stanza retries forever when log entry is too big for Google Cloud Logging
3 participants