I have a Cronjob in Kubernetes that runs every 3 minutes. It seems to be running the job fine as shown below however the generated pod immediately deletes itself and I cannot look at any details as to why it gets deleted.
The cronjob skeleton is below,
apiVersion: batch/v1beta1
kind: CronJob
...
spec:
schedule: "*/3 * * * *"
successfulJobsHistoryLimit: 1
failedJobsHistoryLimit: 3
concurrencyPolicy: Forbid
startingDeadlineSeconds: 120
jobTemplate:
spec:
backoffLimit: 2
template:
spec:
...This generates the cronjob as below,
NAME SCHEDULE SUSPEND ACTIVE LAST SCHEDULE AGE
test */3 * * * * False 0 1m 51mThe job generated by this,
NAME DESIRED SUCCESSFUL AGE
test-1552177080 1 0 8m
test-1552177260 1 0 5m
test-1552177440 1 0 2mLooking at the details into one of these jobs i can see,
Name: test-1552177440
Namespace: storage
...
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Normal SuccessfulCreate 2m57s job-controller Created pod: test-1552177440-b5d6g
Normal SuccessfulDelete 2m40s job-controller Deleted pod: test-1552177440-b5d6g
Warning BackoffLimitExceeded 2m40s (x2 over 2m40s) job-controller Job has reached the specified backoff limitAs you can see the pod is deleted immediately with SuccessfulDelete. Is there anyway to stop this from happening ? Ultimately, id like to look at any logs or any details as to why the pod doesnt start.