You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I encountered a situation that required an upgrade of Podman from version 4.1, which comes with Kirkstone by default, to version 4.4.4 in order to utilize the Quadlet feature. To test this, I created a Python-based "Hello World" image. Initially, when I tried to load and run the image, the container executed successfully without any delays.
However, after upgrading Podman to 4.4.4 to use Quadlet, I encountered an issue where, upon completion of the container's execution, it took an additional 20 seconds, and I received the following error:
Exceeded conmon timeout waiting for container e6ac3bb749659a82023fbe35db7b972be8af7d2cc755a32095df04b2beb58e5f to exit
To resolve this, I upgraded not only the Conmon package but also the runc packages and libseccomp. Despite these updates, the issue persisted.
Steps to reproduce the issue
Steps to reproduce the issue
Load the image with podman load -i "tar file name"
Run the image with podman run -ti "image name"
Describe the results you received
root:# podman run -ti --log-level debug hello_world_py
INFO[0000] podman filtering at log level debug
DEBU[0000] Called run.PersistentPreRunE(podman run -ti --log-level debug hello_world_py)
DEBU[0000] Using conmon: "/usr/bin/conmon"
DEBU[0000] Initializing boltdb state at /mnt/internal_storage/podman/lib/containers/storage/libpod/bolt_state.db
DEBU[0000] Using graph driver overlay
DEBU[0000] Using graph root /mnt/internal_storage/podman/lib/containers/storage
DEBU[0000] Using run root /mnt/internal_storage/podman/run/containers/storage
DEBU[0000] Using static dir /mnt/internal_storage/podman/lib/containers/storage/libpod
DEBU[0000] Using tmp dir /run/libpod
DEBU[0000] Using volume path /mnt/internal_storage/podman/lib/containers/storage/volumes
DEBU[0000] Using transient store: false
DEBU[0000] Set libpod namespace to ""
DEBU[0000] [graphdriver] trying provided driver "overlay"
DEBU[0000] Cached value indicated that overlay is supported
DEBU[0000] Cached value indicated that overlay is supported
DEBU[0000] Cached value indicated that metacopy is not being used
DEBU[0000] Cached value indicated that native-diff is usable
DEBU[0000] backingFs=extfs, projectQuotaSupported=false, useNativeDiff=true, usingMetacopy=false
DEBU[0000] Initializing event backend journald
DEBU[0000] Configured OCI runtime runj initialization failed: no valid executable found for OCI runtime runj: invalid argument
DEBU[0000] Configured OCI runtime kata initialization failed: no valid executable found for OCI runtime kata: invalid argument
DEBU[0000] Configured OCI runtime runsc initialization failed: no valid executable found for OCI runtime runsc: invalid argument
DEBU[0000] Configured OCI runtime krun initialization failed: no valid executable found for OCI runtime krun: invalid argument
DEBU[0000] Configured OCI runtime ocijail initialization failed: no valid executable found for OCI runtime ocijail: invalid argument
DEBU[0000] Configured OCI runtime crun-wasm initialization failed: no valid executable found for OCI runtime crun-wasm: invalid argument
DEBU[0000] Configured OCI runtime youki initialization failed: no valid executable found for OCI runtime youki: invalid argument
DEBU[0000] Configured OCI runtime crun initialization failed: no valid executable found for OCI runtime crun: invalid argument
DEBU[0000] Using OCI runtime "/usr/bin/runc"
INFO[0000] Setting parallel job count to 13
DEBU[0000] Successfully loaded 1 networks
DEBU[0000] Pulling image hello_world_py (policy: missing)
DEBU[0000] Looking up image "hello_world_py" in local containers storage
DEBU[0000] Normalized platform linux/arm64 to {arm64 linux [] }
DEBU[0000] Loading registries configuration "/etc/containers/registries.conf"
DEBU[0000] Trying "localhost/hello_world_py:latest" ...
DEBU[0000] Trying "docker.io/library/hello_world_py:latest" ...
DEBU[0000] Trying "registry.fedoraproject.org/hello_world_py:latest" ...
DEBU[0000] Trying "quay.io/hello_world_py:latest" ...
DEBU[0000] Trying "registry.access.redhat.com/hello_world_py:latest" ...
DEBU[0000] Trying "registry.centos.org/hello_world_py:latest" ...
DEBU[0000] Trying "docker.io/library/hello_world_py:latest" ...
DEBU[0000] Trying "hello_world_py" ...
DEBU[0000] Trying "docker.io/custom/hello_world_py:latest" ...
DEBU[0000] parsed reference into "[overlay@/mnt/internal_storage/podman/lib/containers/storage+/mnt/internal_storage/podman/run/containers/storage]@be4fc274aebce36f0c519b2f942e918793796d52de4b8f81071017ffbd661255"
DEBU[0000] Found image "hello_world_py" as "docker.io/custom/hello_world_py:latest" in local containers storage
DEBU[0000] Found image "hello_world_py" as "docker.io/custom/hello_world_py:latest" in local containers storage ([overlay@/mnt/internal_storage/podman/lib/containers/storage+/mnt/internal_storage/podman/run/containers/storage]@be4fc274aebce36f0c519b2f942e918793796d52de4b8f81071017ffbd661255)
DEBU[0000] exporting opaque data as blob "sha256:be4fc274aebce36f0c519b2f942e918793796d52de4b8f81071017ffbd661255"
DEBU[0000] Looking up image "docker.io/custom/hello_world_py:latest" in local containers storage
DEBU[0000] Normalized platform linux/arm64 to {arm64 linux [] }
DEBU[0000] Trying "docker.io/custom/hello_world_py:latest" ...
DEBU[0000] parsed reference into "[overlay@/mnt/internal_storage/podman/lib/containers/storage+/mnt/internal_storage/podman/run/containers/storage]@be4fc274aebce36f0c519b2f942e918793796d52de4b8f81071017ffbd661255"
DEBU[0000] Found image "docker.io/custom/hello_world_py:latest" as "docker.io/custom/hello_world_py:latest" in local containers storage
DEBU[0000] Found image "docker.io/custom/hello_world_py:latest" as "docker.io/custom/hello_world_py:latest" in local containers storage ([overlay@/mnt/internal_storage/podman/lib/containers/storage+/mnt/internal_storage/podman/run/containers/storage]@be4fc274aebce36f0c519b2f942e918793796d52de4b8f81071017ffbd661255)
DEBU[0000] exporting opaque data as blob "sha256:be4fc274aebce36f0c519b2f942e918793796d52de4b8f81071017ffbd661255"
DEBU[0000] Looking up image "hello_world_py" in local containers storage
DEBU[0000] Normalized platform linux/arm64 to {arm64 linux [] }
DEBU[0000] Trying "localhost/hello_world_py:latest" ...
DEBU[0000] Trying "docker.io/library/hello_world_py:latest" ...
DEBU[0000] Trying "registry.fedoraproject.org/hello_world_py:latest" ...
DEBU[0000] Trying "quay.io/hello_world_py:latest" ...
DEBU[0000] Trying "registry.access.redhat.com/hello_world_py:latest" ...
DEBU[0000] Trying "registry.centos.org/hello_world_py:latest" ...
DEBU[0000] Trying "docker.io/library/hello_world_py:latest" ...
DEBU[0000] Trying "hello_world_py" ...
DEBU[0000] Trying "docker.io/custom/hello_world_py:latest" ...
DEBU[0000] parsed reference into "[overlay@/mnt/internal_storage/podman/lib/containers/storage+/mnt/internal_storage/podman/run/containers/storage]@be4fc274aebce36f0c519b2f942e918793796d52de4b8f81071017ffbd661255"
DEBU[0000] Found image "hello_world_py" as "docker.io/custom/hello_world_py:latest" in local containers storage
DEBU[0000] Found image "hello_world_py" as "docker.io/custom/hello_world_py:latest" in local containers storage ([overlay@/mnt/internal_storage/podman/lib/containers/storage+/mnt/internal_storage/podman/run/containers/storage]@be4fc274aebce36f0c519b2f942e918793796d52de4b8f81071017ffbd661255)
DEBU[0000] exporting opaque data as blob "sha256:be4fc274aebce36f0c519b2f942e918793796d52de4b8f81071017ffbd661255"
DEBU[0000] Inspecting image be4fc274aebce36f0c519b2f942e918793796d52de4b8f81071017ffbd661255
DEBU[0000] exporting opaque data as blob "sha256:be4fc274aebce36f0c519b2f942e918793796d52de4b8f81071017ffbd661255"
DEBU[0000] Inspecting image be4fc274aebce36f0c519b2f942e918793796d52de4b8f81071017ffbd661255
DEBU[0000] Inspecting image be4fc274aebce36f0c519b2f942e918793796d52de4b8f81071017ffbd661255
DEBU[0000] Inspecting image be4fc274aebce36f0c519b2f942e918793796d52de4b8f81071017ffbd661255
DEBU[0000] Inspecting image be4fc274aebce36f0c519b2f942e918793796d52de4b8f81071017ffbd661255
DEBU[0000] using systemd mode: false
DEBU[0000] No hostname set; container's hostname will default to runtime default
DEBU[0000] Loading default seccomp profile
DEBU[0000] Allocated lock 1 for container c74ad897ce050f0a00611b7469612b43053a8581b271fd78510fa88438d5a0b8
DEBU[0000] parsed reference into "[overlay@/mnt/internal_storage/podman/lib/containers/storage+/mnt/internal_storage/podman/run/containers/storage]@be4fc274aebce36f0c519b2f942e918793796d52de4b8f81071017ffbd661255"
DEBU[0000] exporting opaque data as blob "sha256:be4fc274aebce36f0c519b2f942e918793796d52de4b8f81071017ffbd661255"
DEBU[0000] Cached value indicated that idmapped mounts for overlay are not supported
DEBU[0000] Check for idmapped mounts support
DEBU[0000] Created container "c74ad897ce050f0a00611b7469612b43053a8581b271fd78510fa88438d5a0b8"
DEBU[0000] Container "c74ad897ce050f0a00611b7469612b43053a8581b271fd78510fa88438d5a0b8" has work directory "/mnt/internal_storage/podman/lib/containers/storage/overlay-containers/c74ad897ce050f0a00611b7469612b43053a8581b271fd78510fa88438d5a0b8/userdata"
DEBU[0000] Container "c74ad897ce050f0a00611b7469612b43053a8581b271fd78510fa88438d5a0b8" has run directory "/mnt/internal_storage/podman/run/containers/storage/overlay-containers/c74ad897ce050f0a00611b7469612b43053a8581b271fd78510fa88438d5a0b8/userdata"
DEBU[0000] Handling terminal attach
INFO[0000] Received shutdown.Stop(), terminating! PID=1629
DEBU[0000] Enabling signal proxying
DEBU[0000] overlay: mount_data=lowerdir=/mnt/internal_storage/podman/lib/containers/storage/overlay/l/Y5NARJP77ENBSCPOQ5A6CFVBZK:/mnt/internal_storage/podman/lib/containers/storage/overlay/l/FZ2WWB4RC7VOOBDHZ2NFUQB3MS:/mnt/internal_storage/podman/lib/containers/storage/overlay/l/YCB5QE52WKM3PKPRTJAL5C7JYO:/mnt/internal_storage/podman/lib/containers/storage/overlay/l/MNTZI3SXWFG44CFMFLL5JLJNO6,upperdir=/mnt/internal_storage/podman/lib/containers/storage/overlay/5348e734723000ba186307d7b17d1df0ca7b0086a9884f1c2fd4e0848cae18ec/diff,workdir=/mnt/internal_storage/podman/lib/containers/storage/overlay/5348e734723000ba186307d7b17d1df0ca7b0086a9884f1c2fd4e0848cae18ec/work,
DEBU[0000] Mounted container "c74ad897ce050f0a00611b7469612b43053a8581b271fd78510fa88438d5a0b8" at "/mnt/internal_storage/podman/lib/containers/storage/overlay/5348e734723000ba186307d7b17d1df0ca7b0086a9884f1c2fd4e0848cae18ec/merged"
DEBU[0000] Created root filesystem for container c74ad897ce050f0a00611b7469612b43053a8581b271fd78510fa88438d5a0b8 at /mnt/internal_storage/podman/lib/containers/storage/overlay/5348e734723000ba186307d7b17d1df0ca7b0086a9884f1c2fd4e0848cae18ec/merged
DEBU[0000] Made network namespace at /run/netns/netns-86e266ba-b56c-3e7e-bf14-663a1285af04 for container c74ad897ce050f0a00611b7469612b43053a8581b271fd78510fa88438d5a0b8
DEBU[0000] cni result for container c74ad897ce050f0a00611b7469612b43053a8581b271fd78510fa88438d5a0b8 network podman: &{0.4.0 [{Name:cni-podman0 Mac:3e:d8:fc:ae:33:99 Sandbox:} {Name:veth84aaebd9 Mac:86:4b:a6:4b:16:f9 Sandbox:} {Name:eth0 Mac:b6:f7:34:42:10:7c Sandbox:/run/netns/netns-86e266ba-b56c-3e7e-bf14-663a1285af04}] [{Version:4 Interface:0x40003f3278 Address:{IP:10.88.0.7 Mask:ffff0000} Gateway:10.88.0.1}] [{Dst:{IP:0.0.0.0 Mask:00000000} GW:}] {[] [] []}}
DEBU[0000] Adding nameserver(s) from network status of '[]'
INFO[0000] No non-localhost DNS nameservers are left in resolv.conf. Using default external servers: [nameserver 8.8.8.8 nameserver 8.8.4.4]
DEBU[0000] /etc/system-fips does not exist on host, not mounting FIPS mode subscription
DEBU[0000] Setting Cgroups for container c74ad897ce050f0a00611b7469612b43053a8581b271fd78510fa88438d5a0b8 to machine.slice:libpod:c74ad897ce050f0a00611b7469612b43053a8581b271fd78510fa88438d5a0b8
DEBU[0000] reading hooks from /usr/share/containers/oci/hooks.d
DEBU[0000] Workdir "/" resolved to host path "/mnt/internal_storage/podman/lib/containers/storage/overlay/5348e734723000ba186307d7b17d1df0ca7b0086a9884f1c2fd4e0848cae18ec/merged"
DEBU[0000] Created OCI spec for container c74ad897ce050f0a00611b7469612b43053a8581b271fd78510fa88438d5a0b8 at /mnt/internal_storage/podman/lib/containers/storage/overlay-containers/c74ad897ce050f0a00611b7469612b43053a8581b271fd78510fa88438d5a0b8/userdata/config.json
DEBU[0000] /usr/bin/conmon messages will be logged to syslog
DEBU[0000] running conmon: /usr/bin/conmon args="[--api-version 1 -c c74ad897ce050f0a00611b7469612b43053a8581b271fd78510fa88438d5a0b8 -u c74ad897ce050f0a00611b7469612b43053a8581b271fd78510fa88438d5a0b8 -r /usr/bin/runc -b /mnt/internal_storage/podman/lib/containers/storage/overlay-containers/c74ad897ce050f0a00611b7469612b43053a8581b271fd78510fa88438d5a0b8/userdata -p /mnt/internal_storage/podman/run/containers/storage/overlay-containers/c74ad897ce050f0a00611b7469612b43053a8581b271fd78510fa88438d5a0b8/userdata/pidfile -n adoring_jones --exit-dir /run/libpod/exits --full-attach -s -l journald --log-level debug --syslog -t --conmon-pidfile /mnt/internal_storage/podman/run/containers/storage/overlay-containers/c74ad897ce050f0a00611b7469612b43053a8581b271fd78510fa88438d5a0b8/userdata/conmon.pid --exit-command /usr/bin/podman --exit-command-arg --root --exit-command-arg /mnt/internal_storage/podman/lib/containers/storage --exit-command-arg --runroot --exit-command-arg /mnt/internal_storage/podman/run/containers/storage --exit-command-arg --log-level --exit-command-arg debug --exit-command-arg --cgroup-manager --exit-command-arg systemd --exit-command-arg --tmpdir --exit-command-arg /run/libpod --exit-command-arg --network-config-dir --exit-command-arg --exit-command-arg --network-backend --exit-command-arg cni --exit-command-arg --volumepath --exit-command-arg /mnt/internal_storage/podman/lib/containers/storage/volumes --exit-command-arg --transient-store=false --exit-command-arg --runtime --exit-command-arg runc --exit-command-arg --storage-driver --exit-command-arg overlay --exit-command-arg --events-backend --exit-command-arg journald --exit-command-arg --syslog --exit-command-arg container --exit-command-arg cleanup --exit-command-arg c74ad897ce050f0a00611b7469612b43053a8581b271fd78510fa88438d5a0b8]"
INFO[0000] Running conmon under slice machine.slice and unitName libpod-conmon-c74ad897ce050f0a00611b7469612b43053a8581b271fd78510fa88438d5a0b8.scope
DEBU[0000] Received: 1714
INFO[0000] Got Conmon PID as 1703
DEBU[0000] Created container c74ad897ce050f0a00611b7469612b43053a8581b271fd78510fa88438d5a0b8 in OCI runtime
DEBU[0000] Attaching to container c74ad897ce050f0a00611b7469612b43053a8581b271fd78510fa88438d5a0b8
DEBU[0000] Received a resize event: {Width:142 Height:36}
DEBU[0000] Starting container c74ad897ce050f0a00611b7469612b43053a8581b271fd78510fa88438d5a0b8 with command [/bin/sh -c /usr/bin/python3 /usr/local/bin/hello_world.py]
DEBU[0000] Started container c74ad897ce050f0a00611b7469612b43053a8581b271fd78510fa88438d5a0b8
DEBU[0000] Notify sent successfully
hello world
DEBU[0020] Exceeded conmon timeout waiting for container c74ad897ce050f0a00611b7469612b43053a8581b271fd78510fa88438d5a0b8 to exit
DEBU[0020] Called run.PersistentPostRunE(podman run -ti --log-level debug hello_world_py)
DEBU[0020] Shutting down engines
Describe the results you expected
it should not take extra time after the execution of the container is completed.
Issue Description
I encountered a situation that required an upgrade of Podman from version 4.1, which comes with Kirkstone by default, to version 4.4.4 in order to utilize the Quadlet feature. To test this, I created a Python-based "Hello World" image. Initially, when I tried to load and run the image, the container executed successfully without any delays.
However, after upgrading Podman to 4.4.4 to use Quadlet, I encountered an issue where, upon completion of the container's execution, it took an additional 20 seconds, and I received the following error:
Exceeded conmon timeout waiting for container e6ac3bb749659a82023fbe35db7b972be8af7d2cc755a32095df04b2beb58e5f to exit
To resolve this, I upgraded not only the Conmon package but also the runc packages and libseccomp. Despite these updates, the issue persisted.
Steps to reproduce the issue
Steps to reproduce the issue
Describe the results you received
root:# podman run -ti --log-level debug hello_world_py
INFO[0000] podman filtering at log level debug
DEBU[0000] Called run.PersistentPreRunE(podman run -ti --log-level debug hello_world_py)
DEBU[0000] Using conmon: "/usr/bin/conmon"
DEBU[0000] Initializing boltdb state at /mnt/internal_storage/podman/lib/containers/storage/libpod/bolt_state.db
DEBU[0000] Using graph driver overlay
DEBU[0000] Using graph root /mnt/internal_storage/podman/lib/containers/storage
DEBU[0000] Using run root /mnt/internal_storage/podman/run/containers/storage
DEBU[0000] Using static dir /mnt/internal_storage/podman/lib/containers/storage/libpod
DEBU[0000] Using tmp dir /run/libpod
DEBU[0000] Using volume path /mnt/internal_storage/podman/lib/containers/storage/volumes
DEBU[0000] Using transient store: false
DEBU[0000] Set libpod namespace to ""
DEBU[0000] [graphdriver] trying provided driver "overlay"
DEBU[0000] Cached value indicated that overlay is supported
DEBU[0000] Cached value indicated that overlay is supported
DEBU[0000] Cached value indicated that metacopy is not being used
DEBU[0000] Cached value indicated that native-diff is usable
DEBU[0000] backingFs=extfs, projectQuotaSupported=false, useNativeDiff=true, usingMetacopy=false
DEBU[0000] Initializing event backend journald
DEBU[0000] Configured OCI runtime runj initialization failed: no valid executable found for OCI runtime runj: invalid argument
DEBU[0000] Configured OCI runtime kata initialization failed: no valid executable found for OCI runtime kata: invalid argument
DEBU[0000] Configured OCI runtime runsc initialization failed: no valid executable found for OCI runtime runsc: invalid argument
DEBU[0000] Configured OCI runtime krun initialization failed: no valid executable found for OCI runtime krun: invalid argument
DEBU[0000] Configured OCI runtime ocijail initialization failed: no valid executable found for OCI runtime ocijail: invalid argument
DEBU[0000] Configured OCI runtime crun-wasm initialization failed: no valid executable found for OCI runtime crun-wasm: invalid argument
DEBU[0000] Configured OCI runtime youki initialization failed: no valid executable found for OCI runtime youki: invalid argument
DEBU[0000] Configured OCI runtime crun initialization failed: no valid executable found for OCI runtime crun: invalid argument
DEBU[0000] Using OCI runtime "/usr/bin/runc"
INFO[0000] Setting parallel job count to 13
DEBU[0000] Successfully loaded 1 networks
DEBU[0000] Pulling image hello_world_py (policy: missing)
DEBU[0000] Looking up image "hello_world_py" in local containers storage
DEBU[0000] Normalized platform linux/arm64 to {arm64 linux [] }
DEBU[0000] Loading registries configuration "/etc/containers/registries.conf"
DEBU[0000] Trying "localhost/hello_world_py:latest" ...
DEBU[0000] Trying "docker.io/library/hello_world_py:latest" ...
DEBU[0000] Trying "registry.fedoraproject.org/hello_world_py:latest" ...
DEBU[0000] Trying "quay.io/hello_world_py:latest" ...
DEBU[0000] Trying "registry.access.redhat.com/hello_world_py:latest" ...
DEBU[0000] Trying "registry.centos.org/hello_world_py:latest" ...
DEBU[0000] Trying "docker.io/library/hello_world_py:latest" ...
DEBU[0000] Trying "hello_world_py" ...
DEBU[0000] Trying "docker.io/custom/hello_world_py:latest" ...
DEBU[0000] parsed reference into "[overlay@/mnt/internal_storage/podman/lib/containers/storage+/mnt/internal_storage/podman/run/containers/storage]@be4fc274aebce36f0c519b2f942e918793796d52de4b8f81071017ffbd661255"
DEBU[0000] Found image "hello_world_py" as "docker.io/custom/hello_world_py:latest" in local containers storage
DEBU[0000] Found image "hello_world_py" as "docker.io/custom/hello_world_py:latest" in local containers storage ([overlay@/mnt/internal_storage/podman/lib/containers/storage+/mnt/internal_storage/podman/run/containers/storage]@be4fc274aebce36f0c519b2f942e918793796d52de4b8f81071017ffbd661255)
DEBU[0000] exporting opaque data as blob "sha256:be4fc274aebce36f0c519b2f942e918793796d52de4b8f81071017ffbd661255"
DEBU[0000] Looking up image "docker.io/custom/hello_world_py:latest" in local containers storage
DEBU[0000] Normalized platform linux/arm64 to {arm64 linux [] }
DEBU[0000] Trying "docker.io/custom/hello_world_py:latest" ...
DEBU[0000] parsed reference into "[overlay@/mnt/internal_storage/podman/lib/containers/storage+/mnt/internal_storage/podman/run/containers/storage]@be4fc274aebce36f0c519b2f942e918793796d52de4b8f81071017ffbd661255"
DEBU[0000] Found image "docker.io/custom/hello_world_py:latest" as "docker.io/custom/hello_world_py:latest" in local containers storage
DEBU[0000] Found image "docker.io/custom/hello_world_py:latest" as "docker.io/custom/hello_world_py:latest" in local containers storage ([overlay@/mnt/internal_storage/podman/lib/containers/storage+/mnt/internal_storage/podman/run/containers/storage]@be4fc274aebce36f0c519b2f942e918793796d52de4b8f81071017ffbd661255)
DEBU[0000] exporting opaque data as blob "sha256:be4fc274aebce36f0c519b2f942e918793796d52de4b8f81071017ffbd661255"
DEBU[0000] Looking up image "hello_world_py" in local containers storage
DEBU[0000] Normalized platform linux/arm64 to {arm64 linux [] }
DEBU[0000] Trying "localhost/hello_world_py:latest" ...
DEBU[0000] Trying "docker.io/library/hello_world_py:latest" ...
DEBU[0000] Trying "registry.fedoraproject.org/hello_world_py:latest" ...
DEBU[0000] Trying "quay.io/hello_world_py:latest" ...
DEBU[0000] Trying "registry.access.redhat.com/hello_world_py:latest" ...
DEBU[0000] Trying "registry.centos.org/hello_world_py:latest" ...
DEBU[0000] Trying "docker.io/library/hello_world_py:latest" ...
DEBU[0000] Trying "hello_world_py" ...
DEBU[0000] Trying "docker.io/custom/hello_world_py:latest" ...
DEBU[0000] parsed reference into "[overlay@/mnt/internal_storage/podman/lib/containers/storage+/mnt/internal_storage/podman/run/containers/storage]@be4fc274aebce36f0c519b2f942e918793796d52de4b8f81071017ffbd661255"
DEBU[0000] Found image "hello_world_py" as "docker.io/custom/hello_world_py:latest" in local containers storage
DEBU[0000] Found image "hello_world_py" as "docker.io/custom/hello_world_py:latest" in local containers storage ([overlay@/mnt/internal_storage/podman/lib/containers/storage+/mnt/internal_storage/podman/run/containers/storage]@be4fc274aebce36f0c519b2f942e918793796d52de4b8f81071017ffbd661255)
DEBU[0000] exporting opaque data as blob "sha256:be4fc274aebce36f0c519b2f942e918793796d52de4b8f81071017ffbd661255"
DEBU[0000] Inspecting image be4fc274aebce36f0c519b2f942e918793796d52de4b8f81071017ffbd661255
DEBU[0000] exporting opaque data as blob "sha256:be4fc274aebce36f0c519b2f942e918793796d52de4b8f81071017ffbd661255"
DEBU[0000] Inspecting image be4fc274aebce36f0c519b2f942e918793796d52de4b8f81071017ffbd661255
DEBU[0000] Inspecting image be4fc274aebce36f0c519b2f942e918793796d52de4b8f81071017ffbd661255
DEBU[0000] Inspecting image be4fc274aebce36f0c519b2f942e918793796d52de4b8f81071017ffbd661255
DEBU[0000] Inspecting image be4fc274aebce36f0c519b2f942e918793796d52de4b8f81071017ffbd661255
DEBU[0000] using systemd mode: false
DEBU[0000] No hostname set; container's hostname will default to runtime default
DEBU[0000] Loading default seccomp profile
DEBU[0000] Allocated lock 1 for container c74ad897ce050f0a00611b7469612b43053a8581b271fd78510fa88438d5a0b8
DEBU[0000] parsed reference into "[overlay@/mnt/internal_storage/podman/lib/containers/storage+/mnt/internal_storage/podman/run/containers/storage]@be4fc274aebce36f0c519b2f942e918793796d52de4b8f81071017ffbd661255"
DEBU[0000] exporting opaque data as blob "sha256:be4fc274aebce36f0c519b2f942e918793796d52de4b8f81071017ffbd661255"
DEBU[0000] Cached value indicated that idmapped mounts for overlay are not supported
DEBU[0000] Check for idmapped mounts support
DEBU[0000] Created container "c74ad897ce050f0a00611b7469612b43053a8581b271fd78510fa88438d5a0b8"
DEBU[0000] Container "c74ad897ce050f0a00611b7469612b43053a8581b271fd78510fa88438d5a0b8" has work directory "/mnt/internal_storage/podman/lib/containers/storage/overlay-containers/c74ad897ce050f0a00611b7469612b43053a8581b271fd78510fa88438d5a0b8/userdata"
DEBU[0000] Container "c74ad897ce050f0a00611b7469612b43053a8581b271fd78510fa88438d5a0b8" has run directory "/mnt/internal_storage/podman/run/containers/storage/overlay-containers/c74ad897ce050f0a00611b7469612b43053a8581b271fd78510fa88438d5a0b8/userdata"
DEBU[0000] Handling terminal attach
INFO[0000] Received shutdown.Stop(), terminating! PID=1629
DEBU[0000] Enabling signal proxying
DEBU[0000] overlay: mount_data=lowerdir=/mnt/internal_storage/podman/lib/containers/storage/overlay/l/Y5NARJP77ENBSCPOQ5A6CFVBZK:/mnt/internal_storage/podman/lib/containers/storage/overlay/l/FZ2WWB4RC7VOOBDHZ2NFUQB3MS:/mnt/internal_storage/podman/lib/containers/storage/overlay/l/YCB5QE52WKM3PKPRTJAL5C7JYO:/mnt/internal_storage/podman/lib/containers/storage/overlay/l/MNTZI3SXWFG44CFMFLL5JLJNO6,upperdir=/mnt/internal_storage/podman/lib/containers/storage/overlay/5348e734723000ba186307d7b17d1df0ca7b0086a9884f1c2fd4e0848cae18ec/diff,workdir=/mnt/internal_storage/podman/lib/containers/storage/overlay/5348e734723000ba186307d7b17d1df0ca7b0086a9884f1c2fd4e0848cae18ec/work,
DEBU[0000] Mounted container "c74ad897ce050f0a00611b7469612b43053a8581b271fd78510fa88438d5a0b8" at "/mnt/internal_storage/podman/lib/containers/storage/overlay/5348e734723000ba186307d7b17d1df0ca7b0086a9884f1c2fd4e0848cae18ec/merged"
DEBU[0000] Created root filesystem for container c74ad897ce050f0a00611b7469612b43053a8581b271fd78510fa88438d5a0b8 at /mnt/internal_storage/podman/lib/containers/storage/overlay/5348e734723000ba186307d7b17d1df0ca7b0086a9884f1c2fd4e0848cae18ec/merged
DEBU[0000] Made network namespace at /run/netns/netns-86e266ba-b56c-3e7e-bf14-663a1285af04 for container c74ad897ce050f0a00611b7469612b43053a8581b271fd78510fa88438d5a0b8
DEBU[0000] cni result for container c74ad897ce050f0a00611b7469612b43053a8581b271fd78510fa88438d5a0b8 network podman: &{0.4.0 [{Name:cni-podman0 Mac:3e:d8:fc:ae:33:99 Sandbox:} {Name:veth84aaebd9 Mac:86:4b:a6:4b:16:f9 Sandbox:} {Name:eth0 Mac:b6:f7:34:42:10:7c Sandbox:/run/netns/netns-86e266ba-b56c-3e7e-bf14-663a1285af04}] [{Version:4 Interface:0x40003f3278 Address:{IP:10.88.0.7 Mask:ffff0000} Gateway:10.88.0.1}] [{Dst:{IP:0.0.0.0 Mask:00000000} GW:}] {[] [] []}}
DEBU[0000] Adding nameserver(s) from network status of '[]'
INFO[0000] No non-localhost DNS nameservers are left in resolv.conf. Using default external servers: [nameserver 8.8.8.8 nameserver 8.8.4.4]
DEBU[0000] /etc/system-fips does not exist on host, not mounting FIPS mode subscription
DEBU[0000] Setting Cgroups for container c74ad897ce050f0a00611b7469612b43053a8581b271fd78510fa88438d5a0b8 to machine.slice:libpod:c74ad897ce050f0a00611b7469612b43053a8581b271fd78510fa88438d5a0b8
DEBU[0000] reading hooks from /usr/share/containers/oci/hooks.d
DEBU[0000] Workdir "/" resolved to host path "/mnt/internal_storage/podman/lib/containers/storage/overlay/5348e734723000ba186307d7b17d1df0ca7b0086a9884f1c2fd4e0848cae18ec/merged"
DEBU[0000] Created OCI spec for container c74ad897ce050f0a00611b7469612b43053a8581b271fd78510fa88438d5a0b8 at /mnt/internal_storage/podman/lib/containers/storage/overlay-containers/c74ad897ce050f0a00611b7469612b43053a8581b271fd78510fa88438d5a0b8/userdata/config.json
DEBU[0000] /usr/bin/conmon messages will be logged to syslog
DEBU[0000] running conmon: /usr/bin/conmon args="[--api-version 1 -c c74ad897ce050f0a00611b7469612b43053a8581b271fd78510fa88438d5a0b8 -u c74ad897ce050f0a00611b7469612b43053a8581b271fd78510fa88438d5a0b8 -r /usr/bin/runc -b /mnt/internal_storage/podman/lib/containers/storage/overlay-containers/c74ad897ce050f0a00611b7469612b43053a8581b271fd78510fa88438d5a0b8/userdata -p /mnt/internal_storage/podman/run/containers/storage/overlay-containers/c74ad897ce050f0a00611b7469612b43053a8581b271fd78510fa88438d5a0b8/userdata/pidfile -n adoring_jones --exit-dir /run/libpod/exits --full-attach -s -l journald --log-level debug --syslog -t --conmon-pidfile /mnt/internal_storage/podman/run/containers/storage/overlay-containers/c74ad897ce050f0a00611b7469612b43053a8581b271fd78510fa88438d5a0b8/userdata/conmon.pid --exit-command /usr/bin/podman --exit-command-arg --root --exit-command-arg /mnt/internal_storage/podman/lib/containers/storage --exit-command-arg --runroot --exit-command-arg /mnt/internal_storage/podman/run/containers/storage --exit-command-arg --log-level --exit-command-arg debug --exit-command-arg --cgroup-manager --exit-command-arg systemd --exit-command-arg --tmpdir --exit-command-arg /run/libpod --exit-command-arg --network-config-dir --exit-command-arg --exit-command-arg --network-backend --exit-command-arg cni --exit-command-arg --volumepath --exit-command-arg /mnt/internal_storage/podman/lib/containers/storage/volumes --exit-command-arg --transient-store=false --exit-command-arg --runtime --exit-command-arg runc --exit-command-arg --storage-driver --exit-command-arg overlay --exit-command-arg --events-backend --exit-command-arg journald --exit-command-arg --syslog --exit-command-arg container --exit-command-arg cleanup --exit-command-arg c74ad897ce050f0a00611b7469612b43053a8581b271fd78510fa88438d5a0b8]"
INFO[0000] Running conmon under slice machine.slice and unitName libpod-conmon-c74ad897ce050f0a00611b7469612b43053a8581b271fd78510fa88438d5a0b8.scope
DEBU[0000] Received: 1714
INFO[0000] Got Conmon PID as 1703
DEBU[0000] Created container c74ad897ce050f0a00611b7469612b43053a8581b271fd78510fa88438d5a0b8 in OCI runtime
DEBU[0000] Attaching to container c74ad897ce050f0a00611b7469612b43053a8581b271fd78510fa88438d5a0b8
DEBU[0000] Received a resize event: {Width:142 Height:36}
DEBU[0000] Starting container c74ad897ce050f0a00611b7469612b43053a8581b271fd78510fa88438d5a0b8 with command [/bin/sh -c /usr/bin/python3 /usr/local/bin/hello_world.py]
DEBU[0000] Started container c74ad897ce050f0a00611b7469612b43053a8581b271fd78510fa88438d5a0b8
DEBU[0000] Notify sent successfully
hello world
DEBU[0020] Exceeded conmon timeout waiting for container c74ad897ce050f0a00611b7469612b43053a8581b271fd78510fa88438d5a0b8 to exit
DEBU[0020] Called run.PersistentPostRunE(podman run -ti --log-level debug hello_world_py)
DEBU[0020] Shutting down engines
Describe the results you expected
it should not take extra time after the execution of the container is completed.
podman info output
Podman in a container
No
Privileged Or Rootless
None
Upstream Latest Release
No
The text was updated successfully, but these errors were encountered: