-
Notifications
You must be signed in to change notification settings - Fork 236
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[BUG] netty OOM with MULTITHREADED shuffle #9153
Comments
Note that I also see this with Parquet when the number of cores is high. As a test, I changed |
By setting Normally this should default to Xmx according to what I read in a similar issue in Spark https://issues.apache.org/jira/browse/SPARK-27991. In this case, the value is lower than Xmx, as I started with a 16GB heap, so it's not clear why that would be the case. The Spark strategy to deal with this issue, which we also follow, is to perform retries when something like this happens. We may need to decrease our default |
Great work @abellina we really need to figure out how to document and figure out how this fits in with our off heap memory limit story. |
I added the reliability tag because we would like to document the behavior of the maxBytesInFlight setting relative to performance in addition to coming up with better defaults and also likely providing a way to tie this in with the host allocator interface. |
Given the runs I have made I believe setting this config to somewhere between 64MB-128MB is going to be a good enough default. I am running more iterations, but that's the current status. |
Here are findings at 3TB running in the cloud (dataproc) with 4 x n1-standard-16 nodes and using 16 threads for the MT shuffle. I see a lot of noise in here, but I do have another datapoint in our performance cluster for query67 specifically (this query is particularly sensitive to MT shuffle). Given the dataproc and perf cluster data, I am going to set the config to 128MB, and we'll need to continue this work as part of the host memory framework #8900 |
I ran NDS @100TB and ran into a netty OOM in query50. This needs further investigation. It is likely that either our limits are not working or there is a leak somewhere:
The text was updated successfully, but these errors were encountered: