Performed inline steps to verify the issue and checks against the test steps are passed. $ oc get clusterversion NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.11.0-0.nightly-2022-03-29-152521 True False 3h38m Cluster version is 4.11.0-0.nightly-2022-03-29-152521 Get the clusterip pattern & pods $ oc get svc NAME TYPE CLUSTER-IP EXTERNAL-IP PORT(S) AGE kubernetes ClusterIP 172.30.0.1 <none> 443/TCP 4h5m openshift ExternalName <none> kubernetes.default.svc.cluster.local <none> 4h ruby-hello-world ClusterIP 172.30.240.32 <none> 8080/TCP 31s $ oc get pods -o wide NAME READY STATUS RESTARTS AGE IP NODE NOMINATED NODE READINESS GATES ruby-hello-world-1-build 1/1 Running 0 43s 10.128.2.12 jmekkatt-fel-7676m-worker-northcentralus-5kpvm <none> <none> Get the master nodes $ oc get nodes | grep master jmekkatt-fel-7676m-master-0 Ready master 4h3m v1.23.3+54654d2 jmekkatt-fel-7676m-master-1 Ready master 4h2m v1.23.3+54654d2 jmekkatt-fel-7676m-master-2 Ready master 4h3m v1.23.3+54654d2 Remove node to pod networking route entries from one of the master. $ oc debug node/jmekkatt-fel-7676m-master-0 Starting pod/jmekkatt-fel-7676m-master-0-debug ... sh-4.4# chroot /host sh-4.4# route -n Kernel IP routing table Destination Gateway Genmask Flags Metric Ref Use Iface 0.0.0.0 10.0.0.1 0.0.0.0 UG 100 0 0 eth0 10.0.0.0 0.0.0.0 255.255.128.0 U 100 0 0 eth0 10.128.0.0 0.0.0.0 255.252.0.0 U 0 0 0 tun0 168.63.129.16 10.0.0.1 255.255.255.255 UGH 100 0 0 eth0 169.254.169.254 10.0.0.1 255.255.255.255 UGH 100 0 0 eth0 172.30.0.0 0.0.0.0 255.255.0.0 U 0 0 0 tun0 sh-4.4# route del -net 10.128.0.0 gw 0.0.0.0 netmask 255.252.0.0 tun0 sh-4.4# route del -net 172.30.0.0 gw 0.0.0.0 netmask 255.255.0.0 tun0 see the definition of job object $ cat job.yaml apiVersion: batch/v1 kind: Job metadata: name: pi spec: parallelism: 1 completions: 1 activeDeadlineSeconds: 1800 backoffLimit: 6 template: metadata: name: pi spec: containers: - name: pi image: perl command: ["perl", "-Mbignum=bpi", "-wle", "print bpi(2000)"] restartPolicy: OnFailure Create the job and see the status of job and pods - both are completed/successful. $ oc create -f job.yaml job.batch/pi created $ oc get jobs NAME COMPLETIONS DURATION AGE pi 1/1 31s 43s $ oc get pods NAME READY STATUS RESTARTS AGE pi-7fnb7 0/1 Completed 0 49s ruby-hello-world-1-build 0/1 Completed 0 4m8s ruby-hello-world-6856ff6f59-prff2 1/1 Running 0 2m55s Moved the bug to verified as the sanity tests works as expected.
Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory (Important: OpenShift Container Platform 4.11.0 bug fix and security update), and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHSA-2022:5069