-
Notifications
You must be signed in to change notification settings - Fork 2.4k
This issue was moved to a discussion.
You can continue the conversation there. Go to discussion →
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Using podman.socket from inside container results in very long 'podman ps' latency #20270
Comments
Thanks for reaching out, @tghansen. Can you please share an exact reproducer? I am bit lost in which socket is being used where and why Docker is being used. |
Podman is running on the OL8.8 host. podman.socket is also running which maps /var/run/docker.sock to /run/podman/podman.socket. This should get the podman daemon running. |
I am hoping thereis some logging from podman.socket that I can find to help diagnose what it is waiting for. The latency is seen both on the host VM and inside of the container. |
Producing the problem outside of our environment will take some work, since we need the docker-in-docker container. It may be possible to create a simple stripped down version of that container that just comes up and loops doing 'docker ps ' commands. |
I had considered installing PODMAN CLI inside of the OL7 container. If I do that, will it still use /var/run/podman/podman.socket? It seems like a bit of an anti pattern running docker cli with podman over /var/run/docker.sock. |
There are official podman container images, for instance Maybe there is a specific reason for why you're mounting the socket into the container but you're loosing quite some security when doing it as the containers inside will have root access to your host. |
Unfortinately I am required to use the Oracle official base image for containers. This is also why I am stuck with OL7. This was obviously all running on docker, not podman, but is now moving to podman with the transition to OL8. |
also note that I do NOT want the containers to run inside my container, I just want to manage them from inside of the container. |
@vrothberg this looks like it could have been some of the fixes you did to improve latency on ps and images commands. @tghansen Can you try with a new podman 4.6 or later. |
Let me try that. |
are there any instructions for installing newer podman distributions on centos? So far google hits a bunch of non-functional instructions. thanks |
FYI: watchtower is a good example of exactly what I am doing. https://github.com/containrrr/watchtower |
Regarding OL8 specific questions, it's probably best to reach out to Oracle.
Can you elaborate on that? |
I was just hoping folks had the steps to get 4.6-4.7 setup on CENTOS distribution. I found some steps for CENTOS 7 with older PODMAN versions but since I now want to test 4.6+ on OL8 (4.6.1 is going to be included in 8.9 according to folks here at Oracle) I need to do the install manually. |
I can download the podman-remote executable. But it doesn't support 'podman system service' anymore, so when systemd tries to start podman.socket, it gets errors. So guessing there is a new mechanism to launched that. |
No mechanism has changed, so something else might be going on |
Interesting. podman system service returns unknown command with 4.7 and works with 4.1.1. |
podman-remote is not podman. podman-remote is just a client to connect to the podman service. |
Ah, I think I need to clarify the difference. |
got it. so my original question: how do I build or get a prebuilt binary for centos distributions for podman 4.7? Do I need to build it? Hitting lots of issues with that, but slowly working through them. |
I only see the remote binary linked to the update |
Progress. podman version |
I am testing with this version and so far so good. I will do some more testing and then close the issue. Hopefully OL 8.9 which should include podman 4.6.1 behaves as well as I am seeing so far with podman 4.8.0. Thanks for all the help. |
This issue was moved to a discussion.
You can continue the conversation there. Go to discussion →
Issue Description
I am running podman on OL8 8.8 virtual machine with three containers (using either overlap of vfs storage, makes no difference). One of the containers maps /var/run/podman/podman/socket and uses docker CLI to list containers, etc. This is classic docker-in-docker model that works in docker. It also works with podman, however, the CLI degrades to VERY high latency.
time docker ps
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
985d2cd1bd34 docker.io/library/escs-proxy-manager-xpercps:231010.056.0 45 minutes ago Up 38 minutes escs-proxy-manager-xpercps-231010.056.0
0430b9346a15 docker.io/library/escs-telemetry-proxy-xpercps:231010.054.0 43 minutes ago Up 28 minutes escs-telemetry-proxy_xpercps_231010.054.0
f2c774084649 docker.io/library/escs-mgmt-proxy-xpercps:231010.056.0 41 minutes ago Up 28 minutes escs-mgmt-proxy_xpercps_231010.056.0
real 2m14.452s
user 0m0.108s
sys 0m0.041s
2-3 minutes is common. Note that this happens over a period of a few minutes, so something seems to build up insice of podman daemon resulting in degradation. If I run 'systemctl restart podman.socket', the problem clears up and then returns after some number of minutes.
'podman info' struggles too with same kind of latency.
Steps to reproduce the issue
I can make this problem occur consistently with:
Describe the results you received
When I run these steps, I observe very high latency from podman ps, podman info, etc. Note that the commands do complete, but after several minutes.
Describe the results you expected
Since we use the docker cli to monitor the health of all services, this high latency will mean we cannot manage the service keepalive in a reasonable time period.
podman info output
Podman in a container
Yes
Privileged Or Rootless
Privileged
Upstream Latest Release
No
Additional environment details
OL8 U8 VM.
cat /etc/system-release
Oracle Linux Server release
df -h
Filesystem Size Used Avail Use% Mounted on
devtmpfs 3.8G 0 3.8G 0% /dev
tmpfs 3.8G 84K 3.8G 1% /dev/shm
tmpfs 3.8G 201M 3.6G 6% /run
tmpfs 3.8G 0 3.8G 0% /sys/fs/cgroup
/dev/vda3 74G 8.3G 66G 12% /
/dev/vda5 20G 172M 20G 1% /home
/dev/vdb1 98G 1.8G 96G 2% /state
/dev/vda1 1014M 308M 707M 31% /boot
tmpfs 766M 0 766M 0% /run/user/0
shm 63M 0 63M 0% /var/lib/containers/storage/vfs-containers/985d2cd1bd3458b6fd298ae0b473d016182c5e8a4a92d7e7c603cf299fcf7480/userdata/shm
shm 63M 0 63M 0% /var/lib/containers/storage/vfs-containers/f2c77408464979fc2755b5c2bdea4855ddd2e43846fff03785ed5843b7b38dac/userdata/shm
shm 63M 0 63M 0% /var/lib/containers/storage/vfs-containers/0430b9346a1500a2b6812207cc9c1911688650d8856ca7a441e9d8ef476519e6/userdata/shm
Additional information
Any ideas of where I can look at log output to get some ideas of what is happening during this long period of latency?
The text was updated successfully, but these errors were encountered: