Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

containers-rootlessport dumps core #6018

Closed
marusak opened this issue Apr 28, 2020 · 6 comments · Fixed by #6036
Closed

containers-rootlessport dumps core #6018

marusak opened this issue Apr 28, 2020 · 6 comments · Fixed by #6036
Labels
kind/bug Categorizes issue or PR as related to a bug. locked - please file new issue/PR Assist humans wanting to comment on an old issue or PR with locked comments.

Comments

@marusak
Copy link
Contributor

marusak commented Apr 28, 2020

Is this a BUG REPORT or FEATURE REQUEST? (leave only one on its own line)

/kind bug

Description
In Cockpit tests we noticed that podman sometimes dumps core. I don't actually have reproducer (yet) as it does not happen always, but around in 1/3 of runs so slow to debug.

Describe the results you received:
I see podman coredumping:

  Command Line: containers-rootlessport                                         
    Executable: /usr/bin/podman                                                 
 Control Group: /user.slice/user-1000.slice/[email protected]/io.podman.service 
          Unit: [email protected]                                               
     User Unit: io.podman.service                                               
         Slice: user-1000.slice                                                 
     Owner UID: 1000 (admin)                                                    
       Boot ID: 4c93653f502c43e8ac08d914ce6dfad3                                
    Machine ID: 5ea7bfb5b3fa4a668686703decf4954a                                
      Hostname: localhost.localdomain                                           
       Storage: /var/lib/systemd/coredump/core.exe.1000.4c93653f502c43e8ac08d914ce6dfad3.7436.1588078680000000000000.lz4
       Message: Process 7436 (exe) of user 1000 dumped core.                    
@@@@@@@@@@@@@@@@                                                                
                Stack trace of thread 7444:                                     
                #0  0x0000562fce482251 runtime.raise (podman)                   
                #1  0x0000562fce467c6b runtime.dieFromSignal (podman)           
                #2  0x0000562fce4681fd runtime.sigfwdgo (podman)                
                #3  0x0000562fce467340 runtime.sigtrampgo (podman)              
                #4  0x0000562fce482543 runtime.sigtramp (podman)                
                #5  0x00007f191f830b20 __restore_rt (libpthread.so.0)           
                #6  0x0000562fce482251 runtime.raise (podman)                   
                #7  0x0000562fce467c6b runtime.dieFromSignal (podman)           
                #8  0x0000562fce467e0a runtime.crash (podman)                   
                #9  0x0000562fce452496 runtime.fatalpanic (podman)              
                #10 0x0000562fce451e56 runtime.gopanic (podman)                 
                #11 0x0000562fce428741 runtime.chansend (podman)                
                #12 0x0000562fce428145 runtime.chansend1 (podman)               
                #13 0x0000562fcf3ea3af github.com/containers/libpod/pkg/rootlessport.parent.func1 (podman)
                #14 0x0000562fcf3e98b8 github.com/containers/libpod/pkg/rootlessport.parent (podman)
                #15 0x0000562fcf3ea228 github.com/containers/libpod/pkg/rootlessport.init.0.func1 (podman)
                #16 0x0000562fcf73cacb main.main (podman)                       
                #17 0x0000562fce453bc6 runtime.main (podman)                    
                #18 0x0000562fce480841 runtime.goexit (podman)                  
@@@@@@@@@@@@@@@@                                                                
                Stack trace of thread 7436:                                     
                #0  0x0000562fce4827a3 runtime.futex (podman)                   
                #1  0x0000562fce44f6a6 runtime.futexsleep (podman)              
                #2  0x0000562fce42e353 runtime.notesleep (podman)               
                #3  0x0000562fce45844c runtime.stoplockedm (podman)             
                #4  0x0000562fce459d99 runtime.schedule (podman)                
                #5  0x0000562fce459f13 runtime.park_m (podman)                  
                #6  0x0000562fce47e733 runtime.mcall (podman)                   
                #7  0x0000562fce47e658 runtime.rt0_go (podman)                  
                #8  0x0000562fcf78af30 n/a (podman) 

I am attaching the whole backtrace from gdb: bt.txt

If this backtrace is useless and you cannot figure out what is actually happening, I'll spend more to try to find reproducer.

Output of podman version:

[root@localhost ~]# podman version
Version:            1.9.0
RemoteAPI Version:  1
Go Version:         go1.13.9
OS/Arch:            linux/amd64
@openshift-ci-robot openshift-ci-robot added the kind/bug Categorizes issue or PR as related to a bug. label Apr 28, 2020
@baude
Copy link
Member

baude commented Apr 28, 2020

@giuseppe PTAL ...

@baude
Copy link
Member

baude commented Apr 28, 2020

@marusak did you get a core file from this? any chance you saved it?

@marusak
Copy link
Contributor Author

marusak commented Apr 28, 2020

did you get a core file from this? any chance you saved it?

Our bots save those as well, one can be found here. I also have some local ones if these would end up being unusable.
(It is the 'exe' one, the other one is for #6021)

giuseppe added a commit to giuseppe/libpod that referenced this issue Apr 29, 2020
The same channel is written to by two different goroutines.

Use a different channel for each of them so to avoid writing to a
closed channel.

Closes: containers#6018

Signed-off-by: Giuseppe Scrivano <[email protected]>
@giuseppe
Copy link
Member

I've opened a PR here: #6036

giuseppe added a commit to giuseppe/libpod that referenced this issue Apr 29, 2020
The same channel is written to by two different goroutines.

Use a different channel for each of them so to avoid writing to a
closed channel.

Closes: containers#6018

Signed-off-by: Giuseppe Scrivano <[email protected]>
(cherry picked from commit 6d545bb)
@matzew
Copy link

matzew commented May 14, 2020

How can I get this fix, in 1.9.2 ?

I do not seem to get an update on my Fedora 32 ...

sudo dnf update podman --enablerepo updates-testing

gives me nothing to update 😞

@rhatdan
Copy link
Member

rhatdan commented May 14, 2020

It must not have gotten out to the mirrors yet.
You can grab the package from:
https://koji.fedoraproject.org/koji/buildinfo?buildID=1506922

snj33v pushed a commit to snj33v/libpod that referenced this issue May 31, 2020
The same channel is written to by two different goroutines.

Use a different channel for each of them so to avoid writing to a
closed channel.

Closes: containers#6018

Signed-off-by: Giuseppe Scrivano <[email protected]>
@github-actions github-actions bot added the locked - please file new issue/PR Assist humans wanting to comment on an old issue or PR with locked comments. label Sep 23, 2023
@github-actions github-actions bot locked as resolved and limited conversation to collaborators Sep 23, 2023
Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
kind/bug Categorizes issue or PR as related to a bug. locked - please file new issue/PR Assist humans wanting to comment on an old issue or PR with locked comments.
Projects
None yet
Development

Successfully merging a pull request may close this issue.

6 participants