I have the following pods:
NAME READY STATUS RESTARTS AGE
xxx-myactivities-79f49cdfb4-nwg22 1/1 Terminating 0 10h
xxx-mysearch-55864b5c59-6bnwl 1/1 Terminating 0 1d
xxx-mysearch-55864b5c59-rpn48 1/1 Terminating 0 13h
xxx-mysearch-6ff9bbb7cb-9qgbb 1/1 Terminating 0 3d
I am running the following code to forcefully delete those pods:
#
# Clean up dying pods
#
pods=$( kubectl get pods | grep -v Running | tail -n +2 | awk -F " " '{print $1}' )
for pod in $pods;
do
kubectl delete pod $pod --force
done
Here is the output:
pod "xxx-myactivities-79f49cdfb4-nwg22" deleted
pod "xxx-mysearch-55864b5c59-6bnwl" deleted
pod "xxx-mysearch-55864b5c59-rpn48" deleted
pod "xxx-mysearch-6ff9bbb7cb-9qgbb" deleted
After cleaning up, those pods still hang around.
NAME READY STATUS RESTARTS AGE
xxx-myactivities-79f49cdfb4-nwg22 1/1 Terminating 0 10h
xxx-mysearch-55864b5c59-6bnwl 1/1 Terminating 0 1d
xxx-mysearch-55864b5c59-rpn48 1/1 Terminating 0 13h
xxx-mysearch-6ff9bbb7cb-9qgbb 1/1 Terminating 0 3d
How do I clean up those pods?
Delete the deployment
<NAME>-xxxx
like this:For example to delete
tutorial-web-56fbccc56b-wbwjq
do:Then all corresponded pods of
tutorial-xxxx
will terminate by itself.To delete all pods in terminating state with one command do:
You have these alternatives:
Or
SSH into the node the stuck pod was scheduled on Running
docker ps | grep {pod name}
to get the Docker Container ID Runningdocker rm -f {container id}
Or