$ kubectl get pod -o wide | awk '!/Running|Completed/ {n=split($8, host, "."); print $4, host[1], $2, $1}' RESTARTS NOMINATED READY NAME 0 <none> 0/1 signatures.sigprobs-5glgs 0 <none> 0/1 signatures.sigprobs-cfsnl 0 <none> 0/1 signatures.sigprobs-hdvqx 0 <none> 0/1 signatures.sigprobs-ltbnj 0 <none> 0/1 signatures.sigprobs-nggqc $ kubectl logs signatures.sigprobs-5glgs Traceback (most recent call last): File "/data/project/signatures/signatures/src/sigprobs.py", line 461, in <module> handle_args() File "/data/project/signatures/signatures/src/sigprobs.py", line 430, in handle_args with output_file(output, hostname, args.overwrite) as f: File "/data/project/signatures/signatures/src/sigprobs.py", line 453, in output_file return out_file.open("w") if overwrite else out_file.open("x") File "/usr/lib/python3.7/pathlib.py", line 1186, in open opener=self._opener) File "/usr/lib/python3.7/pathlib.py", line 1039, in _opener return self._accessor.open(self, flags, mode) FileExistsError: [Errno 17] File exists: '/data/project/signatures/signatures/data/en.wikipedia.org.json'
It looks like the job description that fired off these pods has been deleted, so maybe there is nothing else to do here? I'm not sure at what point Kubernetes will clean up the failed pods however.