-
Notifications
You must be signed in to change notification settings - Fork 4.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
unexpected EOF on client connection with an open transaction #5641
Comments
An error that occurs because the db client (sqlalchemy?) terminated abnormally. The same occurs even if localhost and (committed internal) IPs are additionally set in "/var/lib/postgresql/data/pg_hba.conf". It seems to happen right after the connection is closed or rollback failed. |
Solved? |
Thanks for the ping. No solution so far. Because I've not been able to reproduce this in version 10.1 locally. Does anyone have steps to reproduce? Are specific requests occurring around which these errors are emitted? |
I'm not sure about steps to reproduce, but it does seem to be related to the volume of records. Currently experiencing the issue on a couple queries. If i set a smaller date range of records to return, it doesnt hit the error. Once I walked up the date range, it started hitting this issue somewhere after 22,000 records returned. Also confirmed it wasn't a specific bad record by taking multiple sample datasets from various date ranges. |
Is this a regression from V8? |
Also hitting this in v10 |
Also seeing this in Redash 10.0.0 (9c928bd) with PostgreSQL 9.6.24 Logs seems to come in 15s or 30s intervals in batches of 1-3. The clear interval making it feel like some keepalive ping or such. |
Facing the same issue while migrating from v8 to v10. Is there any solution to fix this? |
@OnkarVO7 Copy and paste your docker-compose.yml file script. |
Hello, i have the same issue when i am trying to upgrade from 8V to 10V
|
Hi. same issue for me on a fresh install via Helm on a K8s cluster |
@OnkarVO7 i have the same issue, did you solve it? |
Has this problem been resolved so far? I am also troubled by this problem. |
Has this been resolved? |
Hi folks, unless you give us instructions to reproduce the error, we can't resolve it. |
As stupid as this sounds, it's by following the walkthrough video for updating to v10 from v8. Have you or someone on the team tried to follow those instructions recently to see if they are still correct? I tried doing it 3 or 4 times, each time on a fresh AWS EC2 using the pre-baked AMI. Here's the link: https://redash.io/help/open-source/setup |
To troubleshoot I tried several things - force closing any open processes that might be running postgres (which wasn't installed), I tried installing postgres, I made sure there weren't any active connections, I tried setting it up once by only logging in after the upgrade, etc. |
Well; a couple of things.
Redash is community owned now, there is no team of dedicated developers.
Both of these indicate that you're using a pretty old version of redash; we don't really maintain the AMIs anymore, and v10 is from several years ago (let alone v8). My recommendation would be to back up your database and set redash up using either the v10 docker image directly, or the I don't personally have access to running any AMI things, so I'm unable to reproduce this bug if you're only using the AMIs. Does the same issue happen if you use the docker images? |
@guidopetri For me the reproduction is |
@ribtoks are you starting from scratch? or an existing db? |
@guidopetri An existing DB. But "from scratch" instruction also requires you to "init" the DB first so it's kind of always from existing DB in such sense. |
I can repro! 🎉 I can also assert that it's definitely related to refreshing schemas. The screenshot below has the From an initial look, it appears the postgres runner doesn't have the |
Some more investigation:
I only have access to a postgres data source; can anyone confirm if something similar happens on different data sources? |
@guidopetri nope, only using Postgres so far. Great progress btw! |
@ribtoks @alavi-sorrek - to be clear - do you see this message on the postgres db backing redash, or on a postgres data source? (again, in my case I only have one postgres server, so I can't tell) |
I don’t have a postgres server at all and I see it lol
…On Sat, Feb 3, 2024 at 7:29 AM Guido Petri ***@***.***> wrote:
@ribtoks <https://github.com/ribtoks> @alavi-sorrek
<https://github.com/alavi-sorrek> - to be clear - do you see this message
on the postgres db backing redash, or on a postgres data source?
(again, in my case I only have one postgres server, so I can't tell)
—
Reply to this email directly, view it on GitHub
<#5641 (comment)>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/A2QBB56AGSFX4JA75DT3MWLYRYUUBAVCNFSM5HR6DVU2U5DIOJSWCZC7NNSXTN2JONZXKZKDN5WW2ZLOOQ5TCOJSGUZTAOBVG42Q>
.
You are receiving this because you were mentioned.Message ID:
***@***.***>
|
Hmm. What data source are you using then? |
I’m using Redshift, which I suppose could cause a similar message, but I
received the error during upgrading to v10 before I even connected to a
data source.
…On Sat, Feb 3, 2024 at 1:18 PM Guido Petri ***@***.***> wrote:
Hmm. What data source are you using then?
—
Reply to this email directly, view it on GitHub
<#5641 (comment)>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/A2QBB5YWNI2GZQJJF6C3JRTYRZ5QNAVCNFSM5HR6DVU2U5DIOJSWCZC7NNSXTN2JONZXKZKDN5WW2ZLOOQ5TCOJSGU2DEMBUGU4Q>
.
You are receiving this because you were mentioned.Message ID:
***@***.***>
|
I found a potentially related issue. If you have hit the graceful shutdown request count, then the processing loop could be terminated early, closing the connection. I think this situation could also cause this log error if you've opened any database connections for running the query. I'm about to test an upgrade of NOTE: maybe only with async workers, not sync workers (for Keep Alive) |
Seems to be stable in production with |
…#4) * Fix 502 Bad Gateway error from gunicorn keep-alive default setting of 2 seconds See https://www.ikw.cz/aws-alb-gunicorn-error-502-bad-gateway-fix * Log gunicorn errors * Ungrade gunicorn to latest See getredash#5641 (comment) and benoitc/gunicorn#2297
I was having the same problem and resolved it by using the Bitnami AMI. https://bitnami.com/stack/redash |
wdyt if we migrate Gunicorn to Granian for better performance and stability? And from poetry to UV for faster dependency resolver? |
Let's see, I will try it |
When we upgrade v8.0.0 to v10.0.0, I have a error log
unexpected EOF on client connection with an open transaction
with postgres for exec cmd docker logs -f xxxx.Not sure if it will affect the service?
The text was updated successfully, but these errors were encountered: