Time Namespace Component RelatedObject Reason Message

openstack

glance-27086-default-external-api-0

Scheduled

Successfully assigned openstack/glance-27086-default-external-api-0 to master-0

openstack

bootstrap-dataplane-step-1-edpm-a-54cs7

Scheduled

Successfully assigned openstack/bootstrap-dataplane-step-1-edpm-a-54cs7 to master-0

openshift-marketplace

925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5kp5jf

Scheduled

Successfully assigned openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5kp5jf to master-0

openshift-marketplace

93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726rptjh

Scheduled

Successfully assigned openshift-marketplace/93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726rptjh to master-0

openshift-marketplace

certified-operators-cc955

Scheduled

Successfully assigned openshift-marketplace/certified-operators-cc955 to master-0

openshift-route-controller-manager

route-controller-manager-56f686584b-fdcx5

Scheduled

Successfully assigned openshift-route-controller-manager/route-controller-manager-56f686584b-fdcx5 to master-0

openshift-marketplace

7f6062bfcf66f08711c4d599873349559e66916847a22b4b74a32f97d4gkzrg

Scheduled

Successfully assigned openshift-marketplace/7f6062bfcf66f08711c4d599873349559e66916847a22b4b74a32f97d4gkzrg to master-0

openstack-operators

nova-operator-controller-manager-5d488d59fb-5rvds

Scheduled

Successfully assigned openstack-operators/nova-operator-controller-manager-5d488d59fb-5rvds to master-0

openshift-marketplace

2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1v54q9

Scheduled

Successfully assigned openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1v54q9 to master-0

cert-manager

cert-manager-545d4d4674-6ncn2

Scheduled

Successfully assigned cert-manager/cert-manager-545d4d4674-6ncn2 to master-0

openshift-multus

cni-sysctl-allowlist-ds-x2bwv

Scheduled

Successfully assigned openshift-multus/cni-sysctl-allowlist-ds-x2bwv to master-0

openshift-monitoring

thanos-querier-5dc898db58-5m6qg

Scheduled

Successfully assigned openshift-monitoring/thanos-querier-5dc898db58-5m6qg to master-0

openshift-monitoring

telemeter-client-54684fb888-p62jk

Scheduled

Successfully assigned openshift-monitoring/telemeter-client-54684fb888-p62jk to master-0

openshift-monitoring

prometheus-operator-6c8df6d4b-5m857

Scheduled

Successfully assigned openshift-monitoring/prometheus-operator-6c8df6d4b-5m857 to master-0

openshift-monitoring

prometheus-k8s-0

Scheduled

Successfully assigned openshift-monitoring/prometheus-k8s-0 to master-0

openshift-monitoring

prometheus-k8s-0

Scheduled

Successfully assigned openshift-monitoring/prometheus-k8s-0 to master-0

cert-manager

cert-manager-cainjector-5545bd876-mtx97

Scheduled

Successfully assigned cert-manager/cert-manager-cainjector-5545bd876-mtx97 to master-0

openshift-monitoring

openshift-state-metrics-5dc6c74576-j8tmv

Scheduled

Successfully assigned openshift-monitoring/openshift-state-metrics-5dc6c74576-j8tmv to master-0

openshift-monitoring

node-exporter-lb4t5

Scheduled

Successfully assigned openshift-monitoring/node-exporter-lb4t5 to master-0

openshift-monitoring

monitoring-plugin-b7d455597-jgpnw

Scheduled

Successfully assigned openshift-monitoring/monitoring-plugin-b7d455597-jgpnw to master-0

openshift-monitoring

metrics-server-64c67d44c4-s7vfs

Scheduled

Successfully assigned openshift-monitoring/metrics-server-64c67d44c4-s7vfs to master-0

openshift-monitoring

metrics-server-646d59cb8b-622bl

Scheduled

Successfully assigned openshift-monitoring/metrics-server-646d59cb8b-622bl to master-0

openshift-monitoring

kube-state-metrics-7bbc969446-qh6vq

Scheduled

Successfully assigned openshift-monitoring/kube-state-metrics-7bbc969446-qh6vq to master-0

openstack

nova-cell1-cell-mapping-4h8z9

Scheduled

Successfully assigned openstack/nova-cell1-cell-mapping-4h8z9 to master-0

openshift-marketplace

1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874g29cm

Scheduled

Successfully assigned openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874g29cm to master-0

cert-manager

cert-manager-webhook-6888856db4-kkc9m

Scheduled

Successfully assigned cert-manager/cert-manager-webhook-6888856db4-kkc9m to master-0

openstack

nova-cell1-3995-account-create-update-7pfxc

Scheduled

Successfully assigned openstack/nova-cell1-3995-account-create-update-7pfxc to master-0

openstack

nova-cell0-db-create-th2rs

Scheduled

Successfully assigned openstack/nova-cell0-db-create-th2rs to master-0

openstack

dnsmasq-dns-685c76cf85-qwz94

Scheduled

Successfully assigned openstack/dnsmasq-dns-685c76cf85-qwz94 to master-0

openstack

dnsmasq-dns-65db7fd8ff-r6h25

Scheduled

Successfully assigned openstack/dnsmasq-dns-65db7fd8ff-r6h25 to master-0

openstack

dnsmasq-dns-6c8447fc9-wqvmw

Scheduled

Successfully assigned openstack/dnsmasq-dns-6c8447fc9-wqvmw to master-0

openstack-operators

ovn-operator-controller-manager-884679f54-t875d

Scheduled

Successfully assigned openstack-operators/ovn-operator-controller-manager-884679f54-t875d to master-0

openshift-monitoring

alertmanager-main-0

Scheduled

Successfully assigned openshift-monitoring/alertmanager-main-0 to master-0

openshift-monitoring

alertmanager-main-0

Scheduled

Successfully assigned openshift-monitoring/alertmanager-main-0 to master-0

openshift-machine-api

metal3-image-customization-7dd5d8c865-5ppkq

Scheduled

Successfully assigned openshift-machine-api/metal3-image-customization-7dd5d8c865-5ppkq to master-0

openshift-machine-api

metal3-c957566cd-lgms8

Scheduled

Successfully assigned openshift-machine-api/metal3-c957566cd-lgms8 to master-0

openshift-image-registry

node-ca-8ffsg

Scheduled

Successfully assigned openshift-image-registry/node-ca-8ffsg to master-0

openshift-ingress-canary

ingress-canary-cst2b

Scheduled

Successfully assigned openshift-ingress-canary/ingress-canary-cst2b to master-0

openshift-machine-api

metal3-baremetal-operator-78474bdc48-ndj7c

Scheduled

Successfully assigned openshift-machine-api/metal3-baremetal-operator-78474bdc48-ndj7c to master-0

openshift-machine-api

ironic-proxy-28zbq

Scheduled

Successfully assigned openshift-machine-api/ironic-proxy-28zbq to master-0

openstack

nova-cell1-conductor-db-sync-bcpbl

Scheduled

Successfully assigned openstack/nova-cell1-conductor-db-sync-bcpbl to master-0

openshift-cloud-controller-manager-operator

cluster-cloud-controller-manager-operator-7dff898856-7vxxr

Scheduled

Successfully assigned openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-7dff898856-7vxxr to master-0

openstack

ovn-dataplane-step-2-edpm-b-tm4wb

Scheduled

Successfully assigned openstack/ovn-dataplane-step-2-edpm-b-tm4wb to master-0

openstack-operators

placement-operator-controller-manager-5784578c99-9wvcj

Scheduled

Successfully assigned openstack-operators/placement-operator-controller-manager-5784578c99-9wvcj to master-0

openstack-operators

rabbitmq-cluster-operator-manager-668c99d594-s2dqt

Scheduled

Successfully assigned openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-s2dqt to master-0

openstack

ovn-northd-0

Scheduled

Successfully assigned openstack/ovn-northd-0 to master-0

openstack-operators

swift-operator-controller-manager-c674c5965-srzmq

Scheduled

Successfully assigned openstack-operators/swift-operator-controller-manager-c674c5965-srzmq to master-0

openstack-operators

telemetry-operator-controller-manager-d6b694c5-lbvvg

Scheduled

Successfully assigned openstack-operators/telemetry-operator-controller-manager-d6b694c5-lbvvg to master-0

openstack-operators

test-operator-controller-manager-5c5cb9c4d7-nf9vn

Scheduled

Successfully assigned openstack-operators/test-operator-controller-manager-5c5cb9c4d7-nf9vn to master-0

openstack

nova-cell1-db-create-fh7jr

Scheduled

Successfully assigned openstack/nova-cell1-db-create-fh7jr to master-0

openshift-operators

observability-operator-6dd7dd855f-m27sp

Scheduled

Successfully assigned openshift-operators/observability-operator-6dd7dd855f-m27sp to master-0

openstack-operators

watcher-operator-controller-manager-6c4d75f7f9-kp7cg

Scheduled

Successfully assigned openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-kp7cg to master-0

sushy-emulator

nova-console-poller-557b677b49-vslxq

Scheduled

Successfully assigned sushy-emulator/nova-console-poller-557b677b49-vslxq to master-0

openstack-operators

keystone-operator-controller-manager-768b96df4c-2zdk8

Scheduled

Successfully assigned openstack-operators/keystone-operator-controller-manager-768b96df4c-2zdk8 to master-0

sushy-emulator

nova-console-recorder-6766d57b69-2wgj4

Scheduled

Successfully assigned sushy-emulator/nova-console-recorder-6766d57b69-2wgj4 to master-0

openstack

nova-cell1-novncproxy-0

Scheduled

Successfully assigned openstack/nova-cell1-novncproxy-0 to master-0

openstack

nova-cell1-novncproxy-0

Scheduled

Successfully assigned openstack/nova-cell1-novncproxy-0 to master-0

openshift-operators

obo-prometheus-operator-admission-webhook-6976bc9f87-xwq6f

Scheduled

Successfully assigned openshift-operators/obo-prometheus-operator-admission-webhook-6976bc9f87-xwq6f to master-0

openshift-route-controller-manager

route-controller-manager-ffdd4b479-rhmfx

FailedScheduling

0/1 nodes are available: 1 node(s) didn't match pod anti-affinity rules. preemption: 0/1 nodes are available: 1 node(s) didn't match pod anti-affinity rules.

openshift-operators

obo-prometheus-operator-admission-webhook-6976bc9f87-ngfr6

Scheduled

Successfully assigned openshift-operators/obo-prometheus-operator-admission-webhook-6976bc9f87-ngfr6 to master-0

openstack

ovsdbserver-nb-0

Scheduled

Successfully assigned openstack/ovsdbserver-nb-0 to master-0

openstack-operators

ironic-operator-controller-manager-6f787dddc9-dq5f2

Scheduled

Successfully assigned openstack-operators/ironic-operator-controller-manager-6f787dddc9-dq5f2 to master-0

openstack

ovsdbserver-sb-0

Scheduled

Successfully assigned openstack/ovsdbserver-sb-0 to master-0

openstack

placement-7754cdbcb6-lblgz

Scheduled

Successfully assigned openstack/placement-7754cdbcb6-lblgz to master-0

openstack

placement-8b47bf878-2ftwn

Scheduled

Successfully assigned openstack/placement-8b47bf878-2ftwn to master-0

openstack

placement-db-create-dqlc9

Scheduled

Successfully assigned openstack/placement-db-create-dqlc9 to master-0

openstack

placement-db-sync-8jqss

Scheduled

Successfully assigned openstack/placement-db-sync-8jqss to master-0

openstack

nova-metadata-0

Scheduled

Successfully assigned openstack/nova-metadata-0 to master-0

openshift-monitoring

kube-state-metrics-7bbc969446-qh6vq

Scheduled

Successfully assigned openshift-monitoring/kube-state-metrics-7bbc969446-qh6vq to master-0

openshift-operators

obo-prometheus-operator-8ff7d675-wvbr6

Scheduled

Successfully assigned openshift-operators/obo-prometheus-operator-8ff7d675-wvbr6 to master-0

openstack

nova-metadata-0

Scheduled

Successfully assigned openstack/nova-metadata-0 to master-0

openstack

nova-metadata-0

Scheduled

Successfully assigned openstack/nova-metadata-0 to master-0

openshift-monitoring

metrics-server-646d59cb8b-622bl

Scheduled

Successfully assigned openshift-monitoring/metrics-server-646d59cb8b-622bl to master-0

openshift-monitoring

metrics-server-64c67d44c4-s7vfs

Scheduled

Successfully assigned openshift-monitoring/metrics-server-64c67d44c4-s7vfs to master-0

openshift-route-controller-manager

route-controller-manager-ffdd4b479-rhmfx

Scheduled

Successfully assigned openshift-route-controller-manager/route-controller-manager-ffdd4b479-rhmfx to master-0

openshift-monitoring

monitoring-plugin-b7d455597-jgpnw

Scheduled

Successfully assigned openshift-monitoring/monitoring-plugin-b7d455597-jgpnw to master-0

openshift-monitoring

node-exporter-lb4t5

Scheduled

Successfully assigned openshift-monitoring/node-exporter-lb4t5 to master-0

openstack

nova-scheduler-0

Scheduled

Successfully assigned openstack/nova-scheduler-0 to master-0

openshift-controller-manager

controller-manager-bfcf4df58-l6xz7

FailedScheduling

0/1 nodes are available: 1 node(s) didn't match pod anti-affinity rules. preemption: 0/1 nodes are available: 1 node(s) didn't match pod anti-affinity rules.

openstack

nova-scheduler-0

Scheduled

Successfully assigned openstack/nova-scheduler-0 to master-0

openshift-monitoring

openshift-state-metrics-5dc6c74576-j8tmv

Scheduled

Successfully assigned openshift-monitoring/openshift-state-metrics-5dc6c74576-j8tmv to master-0

openshift-monitoring

prometheus-k8s-0

Scheduled

Successfully assigned openshift-monitoring/prometheus-k8s-0 to master-0

openshift-console-operator

console-operator-76b6568d85-8b8gv

Scheduled

Successfully assigned openshift-console-operator/console-operator-76b6568d85-8b8gv to master-0

openstack

placement-ffa9-account-create-update-22wx5

Scheduled

Successfully assigned openstack/placement-ffa9-account-create-update-22wx5 to master-0

openstack

rabbitmq-cell1-server-0

Scheduled

Successfully assigned openstack/rabbitmq-cell1-server-0 to master-0

openstack

rabbitmq-server-0

Scheduled

Successfully assigned openstack/rabbitmq-server-0 to master-0

openstack

reboot-os-dataplane-step-1-edpm-a-rwc7j

Scheduled

Successfully assigned openstack/reboot-os-dataplane-step-1-edpm-a-rwc7j to master-0

openstack

reboot-os-dataplane-step-1-edpm-b-9dw2p

Scheduled

Successfully assigned openstack/reboot-os-dataplane-step-1-edpm-b-9dw2p to master-0

openstack

dnsmasq-dns-6ff8fd9d5c-47t6r

Scheduled

Successfully assigned openstack/dnsmasq-dns-6ff8fd9d5c-47t6r to master-0

openstack

dnsmasq-dns-74db7c7d5f-t4vkw

Scheduled

Successfully assigned openstack/dnsmasq-dns-74db7c7d5f-t4vkw to master-0

openstack

root-account-create-update-54pxn

Scheduled

Successfully assigned openstack/root-account-create-update-54pxn to master-0

openstack

dnsmasq-dns-75597d7cbf-d9w8g

Scheduled

Successfully assigned openstack/dnsmasq-dns-75597d7cbf-d9w8g to master-0

openstack

root-account-create-update-nc4r5

Scheduled

Successfully assigned openstack/root-account-create-update-nc4r5 to master-0

openstack

root-account-create-update-ndh4m

Scheduled

Successfully assigned openstack/root-account-create-update-ndh4m to master-0

openstack

run-os-dataplane-step-1-edpm-a-mtbhq

Scheduled

Successfully assigned openstack/run-os-dataplane-step-1-edpm-a-mtbhq to master-0

openstack

run-os-dataplane-step-1-edpm-b-4xvvh

Scheduled

Successfully assigned openstack/run-os-dataplane-step-1-edpm-b-4xvvh to master-0

openstack

ssh-known-hosts-dataplane-step-1-kt96m

Scheduled

Successfully assigned openstack/ssh-known-hosts-dataplane-step-1-kt96m to master-0

openshift-console

downloads-66b8ffb895-dgq8g

Scheduled

Successfully assigned openshift-console/downloads-66b8ffb895-dgq8g to master-0

openstack

swift-proxy-84fcf8cbb-q9wtd

Scheduled

Successfully assigned openstack/swift-proxy-84fcf8cbb-q9wtd to master-0

openstack

swift-ring-rebalance-2zgwx

Scheduled

Successfully assigned openstack/swift-ring-rebalance-2zgwx to master-0

openshift-console

console-ffb45cdb5-7r4lm

Scheduled

Successfully assigned openshift-console/console-ffb45cdb5-7r4lm to master-0

openstack

dnsmasq-dns-76f498f559-52t76

Scheduled

Successfully assigned openstack/dnsmasq-dns-76f498f559-52t76 to master-0

openstack

dnsmasq-dns-7bcb7479f5-z74kg

Scheduled

Successfully assigned openstack/dnsmasq-dns-7bcb7479f5-z74kg to master-0

openstack

dnsmasq-dns-7cd5f68b57-cbttm

Scheduled

Successfully assigned openstack/dnsmasq-dns-7cd5f68b57-cbttm to master-0

openstack

dnsmasq-dns-7d58447bc7-xn5wl

Scheduled

Successfully assigned openstack/dnsmasq-dns-7d58447bc7-xn5wl to master-0

openstack-operators

placement-operator-controller-manager-5784578c99-9wvcj

Scheduled

Successfully assigned openstack-operators/placement-operator-controller-manager-5784578c99-9wvcj to master-0

openstack

dnsmasq-dns-8476fd89bc-24d8n

Scheduled

Successfully assigned openstack/dnsmasq-dns-8476fd89bc-24d8n to master-0

openstack

dnsmasq-dns-5d96bd4f7c-7qqmd

Scheduled

Successfully assigned openstack/dnsmasq-dns-5d96bd4f7c-7qqmd to master-0

openshift-nmstate

nmstate-metrics-9b8c8685d-p28fn

Scheduled

Successfully assigned openshift-nmstate/nmstate-metrics-9b8c8685d-p28fn to master-0

openshift-monitoring

prometheus-k8s-0

Scheduled

Successfully assigned openshift-monitoring/prometheus-k8s-0 to master-0

openshift-monitoring

prometheus-operator-6c8df6d4b-5m857

Scheduled

Successfully assigned openshift-monitoring/prometheus-operator-6c8df6d4b-5m857 to master-0

openshift-monitoring

telemeter-client-54684fb888-p62jk

Scheduled

Successfully assigned openshift-monitoring/telemeter-client-54684fb888-p62jk to master-0

openstack

nova-scheduler-0

Scheduled

Successfully assigned openstack/nova-scheduler-0 to master-0

openstack

openstack-cell1-galera-0

Scheduled

Successfully assigned openstack/openstack-cell1-galera-0 to master-0

openshift-monitoring

thanos-querier-5dc898db58-5m6qg

Scheduled

Successfully assigned openshift-monitoring/thanos-querier-5dc898db58-5m6qg to master-0

openshift-multus

cni-sysctl-allowlist-ds-x2bwv

Scheduled

Successfully assigned openshift-multus/cni-sysctl-allowlist-ds-x2bwv to master-0

openstack

openstack-galera-0

Scheduled

Successfully assigned openstack/openstack-galera-0 to master-0

openstack

nova-cell0-conductor-db-sync-zzsdf

Scheduled

Successfully assigned openstack/nova-cell0-conductor-db-sync-zzsdf to master-0

openstack

nova-cell0-conductor-0

Scheduled

Successfully assigned openstack/nova-cell0-conductor-0 to master-0

openstack

nova-cell0-cell-mapping-czjjp

Scheduled

Successfully assigned openstack/nova-cell0-cell-mapping-czjjp to master-0

openstack

nova-cell0-11a2-account-create-update-gkhvg

Scheduled

Successfully assigned openstack/nova-cell0-11a2-account-create-update-gkhvg to master-0

openstack

nova-api-db-create-hdkgz

Scheduled

Successfully assigned openstack/nova-api-db-create-hdkgz to master-0

openstack

nova-api-af59-account-create-update-tk2tl

Scheduled

Successfully assigned openstack/nova-api-af59-account-create-update-tk2tl to master-0

openstack

nova-api-0

Scheduled

Successfully assigned openstack/nova-api-0 to master-0

openstack

nova-api-0

Scheduled

Successfully assigned openstack/nova-api-0 to master-0

openstack

nova-api-0

Scheduled

Successfully assigned openstack/nova-api-0 to master-0

openstack

openstackclient

Scheduled

Successfully assigned openstack/openstackclient to master-0

openstack

ovn-controller-metrics-p4sw4

Scheduled

Successfully assigned openstack/ovn-controller-metrics-p4sw4 to master-0

openstack

ovn-controller-ovs-hz4j4

Scheduled

Successfully assigned openstack/ovn-controller-ovs-hz4j4 to master-0

openstack

ovn-controller-q5t48

Scheduled

Successfully assigned openstack/ovn-controller-q5t48 to master-0

openstack

ovn-dataplane-step-2-edpm-a-r2hdg

Scheduled

Successfully assigned openstack/ovn-dataplane-step-2-edpm-a-r2hdg to master-0

openstack

nova-api-0

Scheduled

Successfully assigned openstack/nova-api-0 to master-0

openstack

neutron-metadata-dataplane-step-2-edpm-b-xbtq9

Scheduled

Successfully assigned openstack/neutron-metadata-dataplane-step-2-edpm-b-xbtq9 to master-0

openstack

neutron-metadata-dataplane-step-2-edpm-b-qvqxd

Scheduled

Successfully assigned openstack/neutron-metadata-dataplane-step-2-edpm-b-qvqxd to master-0

openshift-controller-manager

controller-manager-bfcf4df58-l6xz7

Scheduled

Successfully assigned openshift-controller-manager/controller-manager-bfcf4df58-l6xz7 to master-0

openshift-marketplace

community-operators-dtqgc

Scheduled

Successfully assigned openshift-marketplace/community-operators-dtqgc to master-0

openstack-operators

manila-operator-controller-manager-55f864c847-d49s9

Scheduled

Successfully assigned openstack-operators/manila-operator-controller-manager-55f864c847-d49s9 to master-0

openshift-controller-manager

controller-manager-fc56bb77c-qd4sn

Scheduled

Successfully assigned openshift-controller-manager/controller-manager-fc56bb77c-qd4sn to master-0

openshift-machine-api

metal3-image-customization-7dd5d8c865-5ppkq

Scheduled

Successfully assigned openshift-machine-api/metal3-image-customization-7dd5d8c865-5ppkq to master-0

openstack

neutron-metadata-dataplane-step-2-edpm-b-nzczt

Scheduled

Successfully assigned openstack/neutron-metadata-dataplane-step-2-edpm-b-nzczt to master-0

openstack

neutron-metadata-dataplane-step-2-edpm-b-fq5nx

Scheduled

Successfully assigned openstack/neutron-metadata-dataplane-step-2-edpm-b-fq5nx to master-0

openshift-machine-api

metal3-c957566cd-lgms8

Scheduled

Successfully assigned openshift-machine-api/metal3-c957566cd-lgms8 to master-0

openstack

neutron-metadata-dataplane-step-2-edpm-b-cm55j

Scheduled

Successfully assigned openstack/neutron-metadata-dataplane-step-2-edpm-b-cm55j to master-0

openstack

neutron-metadata-dataplane-step-2-edpm-b-5g8zh

Scheduled

Successfully assigned openstack/neutron-metadata-dataplane-step-2-edpm-b-5g8zh to master-0

openshift-machine-api

metal3-baremetal-operator-78474bdc48-ndj7c

Scheduled

Successfully assigned openshift-machine-api/metal3-baremetal-operator-78474bdc48-ndj7c to master-0

openshift-machine-api

ironic-proxy-28zbq

Scheduled

Successfully assigned openshift-machine-api/ironic-proxy-28zbq to master-0

openstack

neutron-metadata-dataplane-step-2-edpm-b-4vq4r

Scheduled

Successfully assigned openstack/neutron-metadata-dataplane-step-2-edpm-b-4vq4r to master-0

openstack

neutron-metadata-dataplane-step-2-edpm-a-tk7z9

Scheduled

Successfully assigned openstack/neutron-metadata-dataplane-step-2-edpm-a-tk7z9 to master-0

openstack

neutron-metadata-dataplane-step-2-edpm-a-kwdqb

Scheduled

Successfully assigned openstack/neutron-metadata-dataplane-step-2-edpm-a-kwdqb to master-0

openshift-nmstate

nmstate-console-plugin-86f58fcf4-5wfxh

Scheduled

Successfully assigned openshift-nmstate/nmstate-console-plugin-86f58fcf4-5wfxh to master-0

openshift-nmstate

nmstate-handler-8ffsp

Scheduled

Successfully assigned openshift-nmstate/nmstate-handler-8ffsp to master-0

openstack

neutron-metadata-dataplane-step-2-edpm-a-kf242

Scheduled

Successfully assigned openstack/neutron-metadata-dataplane-step-2-edpm-a-kf242 to master-0

openstack

neutron-metadata-dataplane-step-2-edpm-a-jnw8j

Scheduled

Successfully assigned openstack/neutron-metadata-dataplane-step-2-edpm-a-jnw8j to master-0

openstack

neutron-metadata-dataplane-step-2-edpm-a-9lz8x

Scheduled

Successfully assigned openstack/neutron-metadata-dataplane-step-2-edpm-a-9lz8x to master-0

openshift-nmstate

nmstate-metrics-9b8c8685d-p28fn

Scheduled

Successfully assigned openshift-nmstate/nmstate-metrics-9b8c8685d-p28fn to master-0

openshift-authentication

oauth-openshift-dcb9594d9-wlht7

Scheduled

Successfully assigned openshift-authentication/oauth-openshift-dcb9594d9-wlht7 to master-0

openshift-authentication

oauth-openshift-dcb9594d9-wlht7

FailedScheduling

0/1 nodes are available: 1 node(s) didn't match pod anti-affinity rules. preemption: 0/1 nodes are available: 1 node(s) didn't match pod anti-affinity rules.

openstack

neutron-metadata-dataplane-step-2-edpm-a-8999k

Scheduled

Successfully assigned openstack/neutron-metadata-dataplane-step-2-edpm-a-8999k to master-0

openstack

neutron-metadata-dataplane-step-2-edpm-a-48qbs

Scheduled

Successfully assigned openstack/neutron-metadata-dataplane-step-2-edpm-a-48qbs to master-0

openshift-nmstate

nmstate-operator-796d4cfff4-q2cgp

Scheduled

Successfully assigned openshift-nmstate/nmstate-operator-796d4cfff4-q2cgp to master-0

openshift-cluster-machine-approver

machine-approver-5c6485487f-qb94j

Scheduled

Successfully assigned openshift-cluster-machine-approver/machine-approver-5c6485487f-qb94j to master-0

openshift-nmstate

nmstate-webhook-5f558f5558-w8sq6

Scheduled

Successfully assigned openshift-nmstate/nmstate-webhook-5f558f5558-w8sq6 to master-0

openstack

neutron-db-sync-q6j55

Scheduled

Successfully assigned openstack/neutron-db-sync-q6j55 to master-0

openstack

neutron-db-create-bbvbz

Scheduled

Successfully assigned openstack/neutron-db-create-bbvbz to master-0

openstack

neutron-6bc775ffc9-jcrc5

Scheduled

Successfully assigned openstack/neutron-6bc775ffc9-jcrc5 to master-0

openstack

dnsmasq-dns-b5b845b79-xn624

Scheduled

Successfully assigned openstack/dnsmasq-dns-b5b845b79-xn624 to master-0

openstack

neutron-555d58bf7b-nplsv

Scheduled

Successfully assigned openstack/neutron-555d58bf7b-nplsv to master-0

openshift-authentication

oauth-openshift-86f9447d7-znl8c

Scheduled

Successfully assigned openshift-authentication/oauth-openshift-86f9447d7-znl8c to master-0

openstack

edpm-a-provisionserver-checksum-discovery-49xmz

Scheduled

Successfully assigned openstack/edpm-a-provisionserver-checksum-discovery-49xmz to master-0

openstack

neutron-3e52-account-create-update-qlzvf

Scheduled

Successfully assigned openstack/neutron-3e52-account-create-update-qlzvf to master-0

openstack

edpm-a-provisionserver-openstackprovisionserver-78fd94b97-tkt7j

Scheduled

Successfully assigned openstack/edpm-a-provisionserver-openstackprovisionserver-78fd94b97-tkt7j to master-0

openstack

edpm-b-provisionserver-checksum-discovery-mf2jz

Scheduled

Successfully assigned openstack/edpm-b-provisionserver-checksum-discovery-mf2jz to master-0

openstack-operators

mariadb-operator-controller-manager-67ccfc9778-vxvcl

Scheduled

Successfully assigned openstack-operators/mariadb-operator-controller-manager-67ccfc9778-vxvcl to master-0

openstack

memcached-0

Scheduled

Successfully assigned openstack/memcached-0 to master-0

openstack

keystone-ec41-account-create-update-wlbg9

Scheduled

Successfully assigned openstack/keystone-ec41-account-create-update-wlbg9 to master-0

metallb-system

speaker-6cndp

Scheduled

Successfully assigned metallb-system/speaker-6cndp to master-0

metallb-system

metallb-operator-webhook-server-6dc87c8bf8-srnwn

Scheduled

Successfully assigned metallb-system/metallb-operator-webhook-server-6dc87c8bf8-srnwn to master-0

openshift-operators

perses-operator-777f7bbbbb-d5qjm

Scheduled

Successfully assigned openshift-operators/perses-operator-777f7bbbbb-d5qjm to master-0

openstack

keystone-db-sync-jg5hc

Scheduled

Successfully assigned openstack/keystone-db-sync-jg5hc to master-0

openshift-authentication

oauth-openshift-8684f7dbf-tqtg9

Scheduled

Successfully assigned openshift-authentication/oauth-openshift-8684f7dbf-tqtg9 to master-0

openstack

dnsmasq-dns-5cc86b957c-nqjxf

Scheduled

Successfully assigned openstack/dnsmasq-dns-5cc86b957c-nqjxf to master-0

openstack

keystone-db-create-cslxs

Scheduled

Successfully assigned openstack/keystone-db-create-cslxs to master-0

metallb-system

metallb-operator-controller-manager-64ff997757-xgs48

Scheduled

Successfully assigned metallb-system/metallb-operator-controller-manager-64ff997757-xgs48 to master-0

metallb-system

controller-7bb4cc7c98-77wvd

Scheduled

Successfully assigned metallb-system/controller-7bb4cc7c98-77wvd to master-0

metallb-system

frr-k8s-webhook-server-bcc4b6f68-lrv7p

Scheduled

Successfully assigned metallb-system/frr-k8s-webhook-server-bcc4b6f68-lrv7p to master-0

metallb-system

frr-k8s-pp5nf

Scheduled

Successfully assigned metallb-system/frr-k8s-pp5nf to master-0

metallb-system

controller-7bb4cc7c98-77wvd

Scheduled

Successfully assigned metallb-system/controller-7bb4cc7c98-77wvd to master-0

cert-manager

cert-manager-webhook-6888856db4-kkc9m

Scheduled

Successfully assigned cert-manager/cert-manager-webhook-6888856db4-kkc9m to master-0

cert-manager

cert-manager-cainjector-5545bd876-mtx97

Scheduled

Successfully assigned cert-manager/cert-manager-cainjector-5545bd876-mtx97 to master-0

cert-manager

cert-manager-545d4d4674-6ncn2

Scheduled

Successfully assigned cert-manager/cert-manager-545d4d4674-6ncn2 to master-0

openshift-marketplace

redhat-marketplace-hqqrk

Scheduled

Successfully assigned openshift-marketplace/redhat-marketplace-hqqrk to master-0

openshift-marketplace

redhat-operators-jstrn

Scheduled

Successfully assigned openshift-marketplace/redhat-operators-jstrn to master-0

openshift-monitoring

alertmanager-main-0

Scheduled

Successfully assigned openshift-monitoring/alertmanager-main-0 to master-0

openstack

keystone-cron-29566681-szrxn

Scheduled

Successfully assigned openstack/keystone-cron-29566681-szrxn to master-0

openshift-monitoring

alertmanager-main-0

Scheduled

Successfully assigned openshift-monitoring/alertmanager-main-0 to master-0

metallb-system

frr-k8s-pp5nf

Scheduled

Successfully assigned metallb-system/frr-k8s-pp5nf to master-0

openstack

keystone-bootstrap-xz2l7

Scheduled

Successfully assigned openstack/keystone-bootstrap-xz2l7 to master-0

openstack

keystone-bootstrap-x7bjf

Scheduled

Successfully assigned openstack/keystone-bootstrap-x7bjf to master-0

openstack

keystone-5987d8f9fd-q4m4b

Scheduled

Successfully assigned openstack/keystone-5987d8f9fd-q4m4b to master-0

openstack

install-os-dataplane-step-1-edpm-b-6gmw5

Scheduled

Successfully assigned openstack/install-os-dataplane-step-1-edpm-b-6gmw5 to master-0

openshift-storage

lvms-operator-dd6b59969-zzdqx

Scheduled

Successfully assigned openshift-storage/lvms-operator-dd6b59969-zzdqx to master-0

openshift-storage

vg-manager-rpqkf

Scheduled

Successfully assigned openshift-storage/vg-manager-rpqkf to master-0

openstack-operators

openstack-operator-index-5g65w

Scheduled

Successfully assigned openstack-operators/openstack-operator-index-5g65w to master-0

openstack-operators

7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514ckhsqj

Scheduled

Successfully assigned openstack-operators/7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514ckhsqj to master-0

openstack-operators

openstack-operator-controller-manager-86bd8996f6-cfkmp

Scheduled

Successfully assigned openstack-operators/openstack-operator-controller-manager-86bd8996f6-cfkmp to master-0

openstack-operators

barbican-operator-controller-manager-59bc569d95-vds6m

Scheduled

Successfully assigned openstack-operators/barbican-operator-controller-manager-59bc569d95-vds6m to master-0

openstack-operators

openstack-operator-controller-init-b85c4d696-9zm7k

Scheduled

Successfully assigned openstack-operators/openstack-operator-controller-init-b85c4d696-9zm7k to master-0

openstack-operators

openstack-baremetal-operator-controller-manager-74c4796899s6vn4

Scheduled

Successfully assigned openstack-operators/openstack-baremetal-operator-controller-manager-74c4796899s6vn4 to master-0

openstack-operators

octavia-operator-controller-manager-5b9f45d989-fr295

Scheduled

Successfully assigned openstack-operators/octavia-operator-controller-manager-5b9f45d989-fr295 to master-0

openstack-operators

nova-operator-controller-manager-5d488d59fb-5rvds

Scheduled

Successfully assigned openstack-operators/nova-operator-controller-manager-5d488d59fb-5rvds to master-0

openstack-operators

watcher-operator-controller-manager-6c4d75f7f9-kp7cg

Scheduled

Successfully assigned openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-kp7cg to master-0

openstack-operators

neutron-operator-controller-manager-767865f676-94tfj

Scheduled

Successfully assigned openstack-operators/neutron-operator-controller-manager-767865f676-94tfj to master-0

openshift-console

console-58857c5dc9-9wf6k

Scheduled

Successfully assigned openshift-console/console-58857c5dc9-9wf6k to master-0

openstack-operators

mariadb-operator-controller-manager-67ccfc9778-vxvcl

Scheduled

Successfully assigned openstack-operators/mariadb-operator-controller-manager-67ccfc9778-vxvcl to master-0

openstack-operators

cinder-operator-controller-manager-8d58dc466-sfshp

Scheduled

Successfully assigned openstack-operators/cinder-operator-controller-manager-8d58dc466-sfshp to master-0

openstack-operators

manila-operator-controller-manager-55f864c847-d49s9

Scheduled

Successfully assigned openstack-operators/manila-operator-controller-manager-55f864c847-d49s9 to master-0

openshift-console

console-5df6fc5897-d9ws4

Scheduled

Successfully assigned openshift-console/console-5df6fc5897-d9ws4 to master-0

openshift-network-console

networking-console-plugin-7c6b76c555-zbpk4

Scheduled

Successfully assigned openshift-network-console/networking-console-plugin-7c6b76c555-zbpk4 to master-0

openstack

install-os-dataplane-step-1-edpm-a-4zvvk

Scheduled

Successfully assigned openstack/install-os-dataplane-step-1-edpm-a-4zvvk to master-0

openstack-operators

keystone-operator-controller-manager-768b96df4c-2zdk8

Scheduled

Successfully assigned openstack-operators/keystone-operator-controller-manager-768b96df4c-2zdk8 to master-0

openstack-operators

ironic-operator-controller-manager-6f787dddc9-dq5f2

Scheduled

Successfully assigned openstack-operators/ironic-operator-controller-manager-6f787dddc9-dq5f2 to master-0

metallb-system

frr-k8s-webhook-server-bcc4b6f68-lrv7p

Scheduled

Successfully assigned metallb-system/frr-k8s-webhook-server-bcc4b6f68-lrv7p to master-0

openstack-operators

designate-operator-controller-manager-588d4d986b-bghwc

Scheduled

Successfully assigned openstack-operators/designate-operator-controller-manager-588d4d986b-bghwc to master-0

openstack-operators

infra-operator-controller-manager-7dd6bb94c9-7rp5f

Scheduled

Successfully assigned openstack-operators/infra-operator-controller-manager-7dd6bb94c9-7rp5f to master-0

openstack-operators

horizon-operator-controller-manager-8464cc45fb-jbfpt

Scheduled

Successfully assigned openstack-operators/horizon-operator-controller-manager-8464cc45fb-jbfpt to master-0

openstack

install-certs-dataplane-step-2-edpm-b-w5v2r

Scheduled

Successfully assigned openstack/install-certs-dataplane-step-2-edpm-b-w5v2r to master-0

openstack-operators

test-operator-controller-manager-5c5cb9c4d7-nf9vn

Scheduled

Successfully assigned openstack-operators/test-operator-controller-manager-5c5cb9c4d7-nf9vn to master-0

openshift-authentication

oauth-openshift-5bc4f8b544-czdmk

Scheduled

Successfully assigned openshift-authentication/oauth-openshift-5bc4f8b544-czdmk to master-0

openshift-multus

multus-admission-controller-649577484c-p72cd

Scheduled

Successfully assigned openshift-multus/multus-admission-controller-649577484c-p72cd to master-0

openshift-multus

multus-admission-controller-69c949f678-v7wb6

Scheduled

Successfully assigned openshift-multus/multus-admission-controller-69c949f678-v7wb6 to master-0

openstack-operators

telemetry-operator-controller-manager-d6b694c5-lbvvg

Scheduled

Successfully assigned openstack-operators/telemetry-operator-controller-manager-d6b694c5-lbvvg to master-0

metallb-system

metallb-operator-controller-manager-64ff997757-xgs48

Scheduled

Successfully assigned metallb-system/metallb-operator-controller-manager-64ff997757-xgs48 to master-0

openstack-operators

heat-operator-controller-manager-67dd5f86f5-5h2n9

Scheduled

Successfully assigned openstack-operators/heat-operator-controller-manager-67dd5f86f5-5h2n9 to master-0

openstack-operators

glance-operator-controller-manager-79df6bcc97-nt245

Scheduled

Successfully assigned openstack-operators/glance-operator-controller-manager-79df6bcc97-nt245 to master-0

openstack-operators

glance-operator-controller-manager-79df6bcc97-nt245

Scheduled

Successfully assigned openstack-operators/glance-operator-controller-manager-79df6bcc97-nt245 to master-0

openstack-operators

heat-operator-controller-manager-67dd5f86f5-5h2n9

Scheduled

Successfully assigned openstack-operators/heat-operator-controller-manager-67dd5f86f5-5h2n9 to master-0

openstack-operators

designate-operator-controller-manager-588d4d986b-bghwc

Scheduled

Successfully assigned openstack-operators/designate-operator-controller-manager-588d4d986b-bghwc to master-0

openshift-authentication

oauth-openshift-5bc4f8b544-czdmk

FailedScheduling

0/1 nodes are available: 1 node(s) didn't match pod anti-affinity rules. preemption: 0/1 nodes are available: 1 node(s) didn't match pod anti-affinity rules.

openstack-operators

swift-operator-controller-manager-c674c5965-srzmq

Scheduled

Successfully assigned openstack-operators/swift-operator-controller-manager-c674c5965-srzmq to master-0

openstack-operators

cinder-operator-controller-manager-8d58dc466-sfshp

Scheduled

Successfully assigned openstack-operators/cinder-operator-controller-manager-8d58dc466-sfshp to master-0

metallb-system

metallb-operator-webhook-server-6dc87c8bf8-srnwn

Scheduled

Successfully assigned metallb-system/metallb-operator-webhook-server-6dc87c8bf8-srnwn to master-0

openstack-operators

rabbitmq-cluster-operator-manager-668c99d594-s2dqt

Scheduled

Successfully assigned openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-s2dqt to master-0

openshift-nmstate

nmstate-console-plugin-86f58fcf4-5wfxh

Scheduled

Successfully assigned openshift-nmstate/nmstate-console-plugin-86f58fcf4-5wfxh to master-0

openshift-nmstate

nmstate-handler-8ffsp

Scheduled

Successfully assigned openshift-nmstate/nmstate-handler-8ffsp to master-0

openstack

install-certs-dataplane-step-2-edpm-a-5xzqr

Scheduled

Successfully assigned openstack/install-certs-dataplane-step-2-edpm-a-5xzqr to master-0

openstack

glance-fb00-account-create-update-9vcp7

Scheduled

Successfully assigned openstack/glance-fb00-account-create-update-9vcp7 to master-0

openstack

glance-27086-default-external-api-0

Scheduled

Successfully assigned openstack/glance-27086-default-external-api-0 to master-0

openstack-operators

ovn-operator-controller-manager-884679f54-t875d

Scheduled

Successfully assigned openstack-operators/ovn-operator-controller-manager-884679f54-t875d to master-0

openshift-nmstate

nmstate-webhook-5f558f5558-w8sq6

Scheduled

Successfully assigned openshift-nmstate/nmstate-webhook-5f558f5558-w8sq6 to master-0

openshift-operators

obo-prometheus-operator-8ff7d675-wvbr6

Scheduled

Successfully assigned openshift-operators/obo-prometheus-operator-8ff7d675-wvbr6 to master-0

openstack-operators

openstack-operator-index-5g65w

Scheduled

Successfully assigned openstack-operators/openstack-operator-index-5g65w to master-0

sushy-emulator

sushy-emulator-59477995f9-mk4kq

Scheduled

Successfully assigned sushy-emulator/sushy-emulator-59477995f9-mk4kq to master-0

openshift-operators

obo-prometheus-operator-admission-webhook-6976bc9f87-ngfr6

Scheduled

Successfully assigned openshift-operators/obo-prometheus-operator-admission-webhook-6976bc9f87-ngfr6 to master-0

openstack-operators

openstack-operator-controller-manager-86bd8996f6-cfkmp

Scheduled

Successfully assigned openstack-operators/openstack-operator-controller-manager-86bd8996f6-cfkmp to master-0

openstack-operators

openstack-operator-controller-init-b85c4d696-9zm7k

Scheduled

Successfully assigned openstack-operators/openstack-operator-controller-init-b85c4d696-9zm7k to master-0

openshift-operators

obo-prometheus-operator-admission-webhook-6976bc9f87-xwq6f

Scheduled

Successfully assigned openshift-operators/obo-prometheus-operator-admission-webhook-6976bc9f87-xwq6f to master-0

openshift-operators

observability-operator-6dd7dd855f-m27sp

Scheduled

Successfully assigned openshift-operators/observability-operator-6dd7dd855f-m27sp to master-0

openshift-operators

perses-operator-777f7bbbbb-d5qjm

Scheduled

Successfully assigned openshift-operators/perses-operator-777f7bbbbb-d5qjm to master-0

openshift-storage

lvms-operator-dd6b59969-zzdqx

Scheduled

Successfully assigned openshift-storage/lvms-operator-dd6b59969-zzdqx to master-0

metallb-system

speaker-6cndp

Scheduled

Successfully assigned metallb-system/speaker-6cndp to master-0

openstack-operators

openstack-baremetal-operator-controller-manager-74c4796899s6vn4

Scheduled

Successfully assigned openstack-operators/openstack-baremetal-operator-controller-manager-74c4796899s6vn4 to master-0

openshift-storage

vg-manager-rpqkf

Scheduled

Successfully assigned openshift-storage/vg-manager-rpqkf to master-0

openstack

dnsmasq-dns-6974f6bb85-z8znc

Scheduled

Successfully assigned openstack/dnsmasq-dns-6974f6bb85-z8znc to master-0

openstack

bootstrap-dataplane-step-1-edpm-b-k2lgn

Scheduled

Successfully assigned openstack/bootstrap-dataplane-step-1-edpm-b-k2lgn to master-0

openstack

cinder-3574-account-create-update-mw5s6

Scheduled

Successfully assigned openstack/cinder-3574-account-create-update-mw5s6 to master-0

openstack

cinder-c920a-api-0

Scheduled

Successfully assigned openstack/cinder-c920a-api-0 to master-0

openstack

cinder-c920a-api-0

Scheduled

Successfully assigned openstack/cinder-c920a-api-0 to master-0

openstack-operators

octavia-operator-controller-manager-5b9f45d989-fr295

Scheduled

Successfully assigned openstack-operators/octavia-operator-controller-manager-5b9f45d989-fr295 to master-0

openstack

cinder-c920a-backup-0

Scheduled

Successfully assigned openstack/cinder-c920a-backup-0 to master-0

openstack-operators

horizon-operator-controller-manager-8464cc45fb-jbfpt

Scheduled

Successfully assigned openstack-operators/horizon-operator-controller-manager-8464cc45fb-jbfpt to master-0

openstack

cinder-c920a-backup-0

Scheduled

Successfully assigned openstack/cinder-c920a-backup-0 to master-0

openstack

cinder-c920a-db-sync-5sx9b

Scheduled

Successfully assigned openstack/cinder-c920a-db-sync-5sx9b to master-0

openstack

cinder-c920a-scheduler-0

Scheduled

Successfully assigned openstack/cinder-c920a-scheduler-0 to master-0

openstack

cinder-c920a-scheduler-0

Scheduled

Successfully assigned openstack/cinder-c920a-scheduler-0 to master-0

openstack

cinder-c920a-volume-lvm-iscsi-0

Scheduled

Successfully assigned openstack/cinder-c920a-volume-lvm-iscsi-0 to master-0

openstack

cinder-c920a-volume-lvm-iscsi-0

Scheduled

Successfully assigned openstack/cinder-c920a-volume-lvm-iscsi-0 to master-0

openstack

cinder-db-create-vtvmv

Scheduled

Successfully assigned openstack/cinder-db-create-vtvmv to master-0

openstack

configure-network-dataplane-step-1-edpm-a-4hslh

Scheduled

Successfully assigned openstack/configure-network-dataplane-step-1-edpm-a-4hslh to master-0

openstack

configure-network-dataplane-step-1-edpm-b-5xnfd

Scheduled

Successfully assigned openstack/configure-network-dataplane-step-1-edpm-b-5xnfd to master-0

openstack

configure-os-dataplane-step-1-edpm-a-sgpm2

Scheduled

Successfully assigned openstack/configure-os-dataplane-step-1-edpm-a-sgpm2 to master-0

openstack

configure-os-dataplane-step-1-edpm-b-jdmpl

Scheduled

Successfully assigned openstack/configure-os-dataplane-step-1-edpm-b-jdmpl to master-0

openstack

dnsmasq-dns-586dbdbb8c-28nq4

Scheduled

Successfully assigned openstack/dnsmasq-dns-586dbdbb8c-28nq4 to master-0

openshift-console

console-d4699dbdf-6z95l

Scheduled

Successfully assigned openshift-console/console-d4699dbdf-6z95l to master-0

openstack-operators

barbican-operator-controller-manager-59bc569d95-vds6m

Scheduled

Successfully assigned openstack-operators/barbican-operator-controller-manager-59bc569d95-vds6m to master-0

openstack-operators

infra-operator-controller-manager-7dd6bb94c9-7rp5f

Scheduled

Successfully assigned openstack-operators/infra-operator-controller-manager-7dd6bb94c9-7rp5f to master-0

openstack-operators

7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514ckhsqj

Scheduled

Successfully assigned openstack-operators/7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514ckhsqj to master-0

openstack

validate-network-dataplane-step-1-edpm-b-x6qvv

Scheduled

Successfully assigned openstack/validate-network-dataplane-step-1-edpm-b-x6qvv to master-0

openshift-multus

multus-admission-controller-69c949f678-v7wb6

Scheduled

Successfully assigned openshift-multus/multus-admission-controller-69c949f678-v7wb6 to master-0

openstack

validate-network-dataplane-step-1-edpm-a-b7h26

Scheduled

Successfully assigned openstack/validate-network-dataplane-step-1-edpm-a-b7h26 to master-0

openshift-multus

multus-admission-controller-649577484c-p72cd

Scheduled

Successfully assigned openshift-multus/multus-admission-controller-649577484c-p72cd to master-0

openstack

glance-db-sync-662t6

Scheduled

Successfully assigned openstack/glance-db-sync-662t6 to master-0

openstack

glance-db-create-l87pv

Scheduled

Successfully assigned openstack/glance-db-create-l87pv to master-0

openstack

glance-27086-default-internal-api-0

Scheduled

Successfully assigned openstack/glance-27086-default-internal-api-0 to master-0

openstack

swift-storage-0

Scheduled

Successfully assigned openstack/swift-storage-0 to master-0

openstack

glance-27086-default-internal-api-0

Scheduled

Successfully assigned openstack/glance-27086-default-internal-api-0 to master-0

openstack

glance-27086-default-internal-api-0

Scheduled

Successfully assigned openstack/glance-27086-default-internal-api-0 to master-0

openstack

swift-ring-rebalance-p9cff

Scheduled

Successfully assigned openstack/swift-ring-rebalance-p9cff to master-0

openstack-operators

neutron-operator-controller-manager-767865f676-94tfj

Scheduled

Successfully assigned openstack-operators/neutron-operator-controller-manager-767865f676-94tfj to master-0

openshift-nmstate

nmstate-operator-796d4cfff4-q2cgp

Scheduled

Successfully assigned openshift-nmstate/nmstate-operator-796d4cfff4-q2cgp to master-0

openstack

nova-cell1-conductor-0

Scheduled

Successfully assigned openstack/nova-cell1-conductor-0 to master-0

openstack

glance-27086-default-external-api-0

Scheduled

Successfully assigned openstack/glance-27086-default-external-api-0 to master-0

openstack

edpm-b-provisionserver-openstackprovisionserver-6bf75656cftnmfr

Scheduled

Successfully assigned openstack/edpm-b-provisionserver-openstackprovisionserver-6bf75656cftnmfr to master-0

openstack

dnsmasq-dns-5bd59cc7c7-jcbnm

Scheduled

Successfully assigned openstack/dnsmasq-dns-5bd59cc7c7-jcbnm to master-0

openstack

dnsmasq-dns-5bf8b865dc-dhd96

Scheduled

Successfully assigned openstack/dnsmasq-dns-5bf8b865dc-dhd96 to master-0

default

apiserver

openshift-kube-apiserver

ShutdownInitiated

Received signal to terminate, becoming unready, but keeping serving

default

apiserver

openshift-kube-apiserver

TerminationPreShutdownHooksFinished

All pre-shutdown hooks have been finished

default

apiserver

openshift-kube-apiserver

AfterShutdownDelayDuration

The minimal shutdown duration of 0s finished

default

apiserver

openshift-kube-apiserver

InFlightRequestsDrained

All non long-running request(s) in-flight have drained

default

apiserver

openshift-kube-apiserver

HTTPServerStoppedListening

HTTP Server has stopped listening

default

apiserver

openshift-kube-apiserver

TerminationGracefulTerminationFinished

All pending requests processed

kube-system

default-scheduler

kube-scheduler

LeaderElection

master-0_c5825f02-edf4-476b-9048-b0368db953dc became leader

kube-system

kube-controller-manager

kube-controller-manager

LeaderElection

master-0_e3ca6aae-e687-4e9f-b1b7-6f353b878341 became leader

kube-system

Required control plane pods have been created

kube-system

cluster-policy-controller

cluster-policy-controller-lock

LeaderElection

master-0_9530582b-fcf7-4fc7-a534-5508cc3eecd6 became leader

kube-system

cluster-policy-controller

bootstrap-kube-controller-manager-master-0

ControlPlaneTopology

unable to get control plane topology, using HA cluster values for leader election: the server could not find the requested resource (get infrastructures.config.openshift.io cluster)

default

apiserver

openshift-kube-apiserver

KubeAPIReadyz

readyz=true

kube-system

cluster-policy-controller-namespace-security-allocation-controller

bootstrap-kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openshift-kube-apiserver namespace

kube-system

cluster-policy-controller-namespace-security-allocation-controller

bootstrap-kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for default namespace

kube-system

cluster-policy-controller-namespace-security-allocation-controller

bootstrap-kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for kube-node-lease namespace

kube-system

cluster-policy-controller-namespace-security-allocation-controller

bootstrap-kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for kube-public namespace

kube-system

cluster-policy-controller-namespace-security-allocation-controller

bootstrap-kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for kube-system namespace

kube-system

cluster-policy-controller-namespace-security-allocation-controller

bootstrap-kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openshift-cluster-version namespace

kube-system

cluster-policy-controller-namespace-security-allocation-controller

bootstrap-kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openshift-infra namespace

kube-system

cluster-policy-controller-namespace-security-allocation-controller

bootstrap-kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openshift-kube-controller-manager-operator namespace

kube-system

cluster-policy-controller-namespace-security-allocation-controller

bootstrap-kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openshift-kube-controller-manager namespace

kube-system

cluster-policy-controller-namespace-security-allocation-controller

bootstrap-kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openshift-kube-apiserver-operator namespace

kube-system

cluster-policy-controller-namespace-security-allocation-controller

bootstrap-kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openshift-etcd namespace

kube-system

cluster-policy-controller-namespace-security-allocation-controller

bootstrap-kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openshift-kube-scheduler namespace

kube-system

cluster-policy-controller-namespace-security-allocation-controller

bootstrap-kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for assisted-installer namespace

kube-system

kube-controller-manager

kube-controller-manager

LeaderElection

master-0_564deec6-a4da-454b-a70e-9e3a21dae407 became leader

kube-system

cluster-policy-controller-namespace-security-allocation-controller

bootstrap-kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openshift-cloud-credential-operator namespace

assisted-installer

job-controller

assisted-installer-controller

SuccessfulCreate

Created pod: assisted-installer-controller-w2zwp

kube-system

cluster-policy-controller-namespace-security-allocation-controller

bootstrap-kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openshift-ingress-operator namespace

openshift-cluster-version

openshift-cluster-version

version

LeaderElection

master-0_71eb15f7-75ad-4007-8d7e-9956d8cadc9d became leader

openshift-cluster-version

openshift-cluster-version

version

LeaderElection

master-0_bb039d46-f059-4c00-a96f-e0a579b47241 became leader

openshift-cluster-version

deployment-controller

cluster-version-operator

ScalingReplicaSet

Scaled up replica set cluster-version-operator-56d8475767 to 1

openshift-cluster-version

openshift-cluster-version

version

LeaderElection

master-0_0edab607-7000-4203-b83b-4d8fffd8beb0 became leader

openshift-cluster-version

openshift-cluster-version

version

LoadPayload

Loading payload version="4.18.35" image="quay.io/openshift-release-dev/ocp-release@sha256:59727c4b3fef19e5149675cf3350735bbfe2c6588a57654b2e4552dd719f58b1"

openshift-cluster-version

openshift-cluster-version

version

RetrievePayload

Retrieving and verifying payload version="4.18.35" image="quay.io/openshift-release-dev/ocp-release@sha256:59727c4b3fef19e5149675cf3350735bbfe2c6588a57654b2e4552dd719f58b1"

kube-system

cluster-policy-controller-namespace-security-allocation-controller

bootstrap-kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openshift-cloud-network-config-controller namespace

kube-system

cluster-policy-controller-namespace-security-allocation-controller

bootstrap-kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openshift-cluster-storage-operator namespace

openshift-cluster-version

openshift-cluster-version

version

PayloadLoaded

Payload loaded version="4.18.35" image="quay.io/openshift-release-dev/ocp-release@sha256:59727c4b3fef19e5149675cf3350735bbfe2c6588a57654b2e4552dd719f58b1" architecture="amd64"

kube-system

cluster-policy-controller-namespace-security-allocation-controller

bootstrap-kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openshift-config-operator namespace

kube-system

cluster-policy-controller-namespace-security-allocation-controller

bootstrap-kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openshift-controller-manager-operator namespace

kube-system

cluster-policy-controller-namespace-security-allocation-controller

bootstrap-kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openshift-kube-scheduler-operator namespace

kube-system

cluster-policy-controller-namespace-security-allocation-controller

bootstrap-kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openshift-network-operator namespace

kube-system

cluster-policy-controller-namespace-security-allocation-controller

bootstrap-kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openshift-marketplace namespace

kube-system

cluster-policy-controller-namespace-security-allocation-controller

bootstrap-kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openshift-etcd-operator namespace

kube-system

cluster-policy-controller-namespace-security-allocation-controller

bootstrap-kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openshift-cluster-machine-approver namespace

kube-system

cluster-policy-controller-namespace-security-allocation-controller

bootstrap-kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openshift-cloud-controller-manager-operator namespace

kube-system

cluster-policy-controller-namespace-security-allocation-controller

bootstrap-kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openshift-authentication-operator namespace

kube-system

cluster-policy-controller-namespace-security-allocation-controller

bootstrap-kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openshift-cluster-node-tuning-operator namespace

kube-system

cluster-policy-controller-namespace-security-allocation-controller

bootstrap-kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openshift-cluster-csi-drivers namespace

kube-system

cluster-policy-controller-namespace-security-allocation-controller

bootstrap-kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openshift-insights namespace

kube-system

cluster-policy-controller-namespace-security-allocation-controller

bootstrap-kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openshift-apiserver-operator namespace

kube-system

cluster-policy-controller-namespace-security-allocation-controller

bootstrap-kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openshift-cloud-controller-manager namespace

kube-system

cluster-policy-controller-namespace-security-allocation-controller

bootstrap-kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openshift-dns-operator namespace

kube-system

cluster-policy-controller-namespace-security-allocation-controller

bootstrap-kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openshift-service-ca-operator namespace

kube-system

cluster-policy-controller-namespace-security-allocation-controller

bootstrap-kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openshift-image-registry namespace

kube-system

cluster-policy-controller-namespace-security-allocation-controller

bootstrap-kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openshift-kube-storage-version-migrator-operator namespace

kube-system

cluster-policy-controller-namespace-security-allocation-controller

bootstrap-kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openshift-cluster-samples-operator namespace

kube-system

cluster-policy-controller-namespace-security-allocation-controller

bootstrap-kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openshift-machine-config-operator namespace

kube-system

cluster-policy-controller-namespace-security-allocation-controller

bootstrap-kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openshift-cluster-olm-operator namespace

kube-system

cluster-policy-controller-namespace-security-allocation-controller

bootstrap-kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openshift-openstack-infra namespace

kube-system

cluster-policy-controller-namespace-security-allocation-controller

bootstrap-kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openshift-kni-infra namespace

kube-system

cluster-policy-controller-namespace-security-allocation-controller

bootstrap-kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openshift-operator-lifecycle-manager namespace

kube-system

cluster-policy-controller-namespace-security-allocation-controller

bootstrap-kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openshift-ovirt-infra namespace

kube-system

cluster-policy-controller-namespace-security-allocation-controller

bootstrap-kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openshift-operators namespace

kube-system

cluster-policy-controller-namespace-security-allocation-controller

bootstrap-kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openshift-vsphere-infra namespace

openshift-kube-controller-manager-operator

deployment-controller

kube-controller-manager-operator

ScalingReplicaSet

Scaled up replica set kube-controller-manager-operator-ff989d6cc to 1

kube-system

cluster-policy-controller-namespace-security-allocation-controller

bootstrap-kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openshift-nutanix-infra namespace

openshift-cluster-olm-operator

deployment-controller

cluster-olm-operator

ScalingReplicaSet

Scaled up replica set cluster-olm-operator-67dcd4998 to 1

openshift-kube-storage-version-migrator-operator

deployment-controller

kube-storage-version-migrator-operator

ScalingReplicaSet

Scaled up replica set kube-storage-version-migrator-operator-6bb5bfb6fd to 1

openshift-service-ca-operator

deployment-controller

service-ca-operator

ScalingReplicaSet

Scaled up replica set service-ca-operator-b865698dc to 1

openshift-kube-scheduler-operator

deployment-controller

openshift-kube-scheduler-operator

ScalingReplicaSet

Scaled up replica set openshift-kube-scheduler-operator-dddff6458 to 1

kube-system

cluster-policy-controller-namespace-security-allocation-controller

bootstrap-kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openshift-cloud-platform-infra namespace

openshift-dns-operator

deployment-controller

dns-operator

ScalingReplicaSet

Scaled up replica set dns-operator-9c5679d8f to 1

openshift-etcd-operator

deployment-controller

etcd-operator

ScalingReplicaSet

Scaled up replica set etcd-operator-8544cbcf9c to 1

openshift-network-operator

deployment-controller

network-operator

ScalingReplicaSet

Scaled up replica set network-operator-7bd846bfc4 to 1

openshift-controller-manager-operator

deployment-controller

openshift-controller-manager-operator

ScalingReplicaSet

Scaled up replica set openshift-controller-manager-operator-8c94f4649 to 1

openshift-apiserver-operator

deployment-controller

openshift-apiserver-operator

ScalingReplicaSet

Scaled up replica set openshift-apiserver-operator-d65958b8 to 1
(x2)

openshift-operator-lifecycle-manager

controllermanager

packageserver-pdb

NoPods

No matching pods found

openshift-marketplace

deployment-controller

marketplace-operator

ScalingReplicaSet

Scaled up replica set marketplace-operator-89ccd998f to 1

kube-system

cluster-policy-controller-namespace-security-allocation-controller

bootstrap-kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openshift-monitoring namespace

openshift-authentication-operator

deployment-controller

authentication-operator

ScalingReplicaSet

Scaled up replica set authentication-operator-5885bfd7f4 to 1

kube-system

cluster-policy-controller-namespace-security-allocation-controller

bootstrap-kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openshift-user-workload-monitoring namespace
(x14)

openshift-cluster-version

replicaset-controller

cluster-version-operator-56d8475767

FailedCreate

Error creating: pods "cluster-version-operator-56d8475767-" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes
(x12)

openshift-cluster-olm-operator

replicaset-controller

cluster-olm-operator-67dcd4998

FailedCreate

Error creating: pods "cluster-olm-operator-67dcd4998-" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes

kube-system

cluster-policy-controller-namespace-security-allocation-controller

bootstrap-kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openshift-machine-api namespace

kube-system

cluster-policy-controller-namespace-security-allocation-controller

bootstrap-kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openshift-config-managed namespace
(x12)

openshift-kube-storage-version-migrator-operator

replicaset-controller

kube-storage-version-migrator-operator-6bb5bfb6fd

FailedCreate

Error creating: pods "kube-storage-version-migrator-operator-6bb5bfb6fd-" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes
(x12)

openshift-kube-scheduler-operator

replicaset-controller

openshift-kube-scheduler-operator-dddff6458

FailedCreate

Error creating: pods "openshift-kube-scheduler-operator-dddff6458-" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes

kube-system

cluster-policy-controller-namespace-security-allocation-controller

bootstrap-kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openshift-config namespace
(x12)

openshift-service-ca-operator

replicaset-controller

service-ca-operator-b865698dc

FailedCreate

Error creating: pods "service-ca-operator-b865698dc-" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes
(x12)

openshift-network-operator

replicaset-controller

network-operator-7bd846bfc4

FailedCreate

Error creating: pods "network-operator-7bd846bfc4-" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes

openshift-cluster-node-tuning-operator

deployment-controller

cluster-node-tuning-operator

ScalingReplicaSet

Scaled up replica set cluster-node-tuning-operator-598fbc5f8f to 1
(x12)

openshift-etcd-operator

replicaset-controller

etcd-operator-8544cbcf9c

FailedCreate

Error creating: pods "etcd-operator-8544cbcf9c-" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes
(x12)

openshift-apiserver-operator

replicaset-controller

openshift-apiserver-operator-d65958b8

FailedCreate

Error creating: pods "openshift-apiserver-operator-d65958b8-" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes

openshift-cluster-node-tuning-operator

deployment-controller

cluster-node-tuning-operator

ScalingReplicaSet

Scaled up replica set cluster-node-tuning-operator-598fbc5f8f to 1
(x12)

openshift-dns-operator

replicaset-controller

dns-operator-9c5679d8f

FailedCreate

Error creating: pods "dns-operator-9c5679d8f-" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes
(x12)

openshift-controller-manager-operator

replicaset-controller

openshift-controller-manager-operator-8c94f4649

FailedCreate

Error creating: pods "openshift-controller-manager-operator-8c94f4649-" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes
(x12)

openshift-marketplace

replicaset-controller

marketplace-operator-89ccd998f

FailedCreate

Error creating: pods "marketplace-operator-89ccd998f-" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes

openshift-cluster-storage-operator

deployment-controller

csi-snapshot-controller-operator

ScalingReplicaSet

Scaled up replica set csi-snapshot-controller-operator-5f5d689c6b to 1

openshift-operator-lifecycle-manager

deployment-controller

package-server-manager

ScalingReplicaSet

Scaled up replica set package-server-manager-7b95f86987 to 1
(x12)

openshift-authentication-operator

replicaset-controller

authentication-operator-5885bfd7f4

FailedCreate

Error creating: pods "authentication-operator-5885bfd7f4-" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes

openshift-monitoring

deployment-controller

cluster-monitoring-operator

ScalingReplicaSet

Scaled up replica set cluster-monitoring-operator-58845fbb57 to 1

openshift-monitoring

deployment-controller

cluster-monitoring-operator

ScalingReplicaSet

Scaled up replica set cluster-monitoring-operator-58845fbb57 to 1

openshift-operator-lifecycle-manager

deployment-controller

olm-operator

ScalingReplicaSet

Scaled up replica set olm-operator-5c9796789 to 1

openshift-image-registry

deployment-controller

cluster-image-registry-operator

ScalingReplicaSet

Scaled up replica set cluster-image-registry-operator-5549dc66cb to 1
(x10)

openshift-monitoring

replicaset-controller

cluster-monitoring-operator-58845fbb57

FailedCreate

Error creating: pods "cluster-monitoring-operator-58845fbb57-" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes

openshift-kube-apiserver-operator

deployment-controller

kube-apiserver-operator

ScalingReplicaSet

Scaled up replica set kube-apiserver-operator-8b68b9d9b to 1

openshift-ingress-operator

deployment-controller

ingress-operator

ScalingReplicaSet

Scaled up replica set ingress-operator-66b84d69b to 1

openshift-operator-lifecycle-manager

deployment-controller

catalog-operator

ScalingReplicaSet

Scaled up replica set catalog-operator-68f85b4d6c to 1
(x10)

openshift-monitoring

replicaset-controller

cluster-monitoring-operator-58845fbb57

FailedCreate

Error creating: pods "cluster-monitoring-operator-58845fbb57-" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes
(x11)

openshift-cluster-node-tuning-operator

replicaset-controller

cluster-node-tuning-operator-598fbc5f8f

FailedCreate

Error creating: pods "cluster-node-tuning-operator-598fbc5f8f-" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes
(x11)

openshift-cluster-node-tuning-operator

replicaset-controller

cluster-node-tuning-operator-598fbc5f8f

FailedCreate

Error creating: pods "cluster-node-tuning-operator-598fbc5f8f-" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes

openshift-config-operator

deployment-controller

openshift-config-operator

ScalingReplicaSet

Scaled up replica set openshift-config-operator-95bf4f4d to 1
(x10)

openshift-ingress-operator

replicaset-controller

ingress-operator-66b84d69b

FailedCreate

Error creating: pods "ingress-operator-66b84d69b-" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes

openshift-machine-api

deployment-controller

cluster-baremetal-operator

ScalingReplicaSet

Scaled up replica set cluster-baremetal-operator-6f69995874 to 1
(x10)

openshift-kube-apiserver-operator

replicaset-controller

kube-apiserver-operator-8b68b9d9b

FailedCreate

Error creating: pods "kube-apiserver-operator-8b68b9d9b-" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes
(x6)

openshift-machine-config-operator

replicaset-controller

machine-config-operator-84d549f6d5

FailedCreate

Error creating: pods "machine-config-operator-84d549f6d5-" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes
(x11)

openshift-cluster-storage-operator

replicaset-controller

csi-snapshot-controller-operator-5f5d689c6b

FailedCreate

Error creating: pods "csi-snapshot-controller-operator-5f5d689c6b-" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes

openshift-cluster-storage-operator

deployment-controller

cluster-storage-operator

ScalingReplicaSet

Scaled up replica set cluster-storage-operator-7d87854d6 to 1

kube-system

Required control plane pods have been created
(x11)

assisted-installer

default-scheduler

assisted-installer-controller-w2zwp

FailedScheduling

no nodes available to schedule pods
(x7)

openshift-cluster-storage-operator

replicaset-controller

cluster-storage-operator-7d87854d6

FailedCreate

Error creating: pods "cluster-storage-operator-7d87854d6-" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes
(x10)

openshift-operator-lifecycle-manager

replicaset-controller

olm-operator-5c9796789

FailedCreate

Error creating: pods "olm-operator-5c9796789-" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes
(x7)

openshift-machine-api

replicaset-controller

cluster-baremetal-operator-6f69995874

FailedCreate

Error creating: pods "cluster-baremetal-operator-6f69995874-" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes
(x10)

openshift-operator-lifecycle-manager

replicaset-controller

catalog-operator-68f85b4d6c

FailedCreate

Error creating: pods "catalog-operator-68f85b4d6c-" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes
(x11)

openshift-operator-lifecycle-manager

replicaset-controller

package-server-manager-7b95f86987

FailedCreate

Error creating: pods "package-server-manager-7b95f86987-" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes

openshift-machine-config-operator

deployment-controller

machine-config-operator

ScalingReplicaSet

Scaled up replica set machine-config-operator-84d549f6d5 to 1
(x7)

openshift-machine-api

replicaset-controller

cluster-baremetal-operator-6f69995874

FailedCreate

Error creating: pods "cluster-baremetal-operator-6f69995874-" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes
(x9)

openshift-config-operator

replicaset-controller

openshift-config-operator-95bf4f4d

FailedCreate

Error creating: pods "openshift-config-operator-95bf4f4d-" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes
(x10)

openshift-image-registry

replicaset-controller

cluster-image-registry-operator-5549dc66cb

FailedCreate

Error creating: pods "cluster-image-registry-operator-5549dc66cb-" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes

openshift-machine-api

deployment-controller

cluster-baremetal-operator

ScalingReplicaSet

Scaled up replica set cluster-baremetal-operator-6f69995874 to 1
(x13)

openshift-kube-controller-manager-operator

replicaset-controller

kube-controller-manager-operator-ff989d6cc

FailedCreate

Error creating: pods "kube-controller-manager-operator-ff989d6cc-" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes

default

apiserver

openshift-kube-apiserver

ShutdownInitiated

Received signal to terminate, becoming unready, but keeping serving

default

apiserver

openshift-kube-apiserver

HTTPServerStoppedListening

HTTP Server has stopped listening

default

apiserver

openshift-kube-apiserver

InFlightRequestsDrained

All non long-running request(s) in-flight have drained

default

apiserver

openshift-kube-apiserver

TerminationPreShutdownHooksFinished

All pre-shutdown hooks have been finished

default

apiserver

openshift-kube-apiserver

AfterShutdownDelayDuration

The minimal shutdown duration of 0s finished

kube-system

default-scheduler

kube-scheduler

LeaderElection

master-0_0ec4deda-5a86-4717-b20e-598c71939d2e became leader

kube-system

cluster-policy-controller

cluster-policy-controller-lock

LeaderElection

master-0_899d98b0-68da-446e-8967-d8731bd2d70f became leader

default

apiserver

openshift-kube-apiserver

KubeAPIReadyz

readyz=true
(x5)

assisted-installer

default-scheduler

assisted-installer-controller-w2zwp

FailedScheduling

no nodes available to schedule pods

openshift-operator-lifecycle-manager

controllermanager

packageserver-pdb

NoPods

No matching pods found

kube-system

kube-controller-manager

kube-controller-manager

LeaderElection

master-0_4cbbea7e-097c-4547-9eef-62ee607631de became leader
(x5)

openshift-network-operator

replicaset-controller

network-operator-7bd846bfc4

FailedCreate

Error creating: pods "network-operator-7bd846bfc4-" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes
(x6)

openshift-cluster-node-tuning-operator

replicaset-controller

cluster-node-tuning-operator-598fbc5f8f

FailedCreate

Error creating: pods "cluster-node-tuning-operator-598fbc5f8f-" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes
(x6)

openshift-config-operator

replicaset-controller

openshift-config-operator-95bf4f4d

FailedCreate

Error creating: pods "openshift-config-operator-95bf4f4d-" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes
(x5)

openshift-operator-lifecycle-manager

replicaset-controller

package-server-manager-7b95f86987

FailedCreate

Error creating: pods "package-server-manager-7b95f86987-" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes
(x6)

openshift-apiserver-operator

replicaset-controller

openshift-apiserver-operator-d65958b8

FailedCreate

Error creating: pods "openshift-apiserver-operator-d65958b8-" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes
(x6)

openshift-image-registry

replicaset-controller

cluster-image-registry-operator-5549dc66cb

FailedCreate

Error creating: pods "cluster-image-registry-operator-5549dc66cb-" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes
(x6)

openshift-cluster-node-tuning-operator

replicaset-controller

cluster-node-tuning-operator-598fbc5f8f

FailedCreate

Error creating: pods "cluster-node-tuning-operator-598fbc5f8f-" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes
(x6)

openshift-authentication-operator

replicaset-controller

authentication-operator-5885bfd7f4

FailedCreate

Error creating: pods "authentication-operator-5885bfd7f4-" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes
(x6)

openshift-cluster-olm-operator

replicaset-controller

cluster-olm-operator-67dcd4998

FailedCreate

Error creating: pods "cluster-olm-operator-67dcd4998-" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes
(x6)

openshift-controller-manager-operator

replicaset-controller

openshift-controller-manager-operator-8c94f4649

FailedCreate

Error creating: pods "openshift-controller-manager-operator-8c94f4649-" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes
(x5)

openshift-operator-lifecycle-manager

replicaset-controller

olm-operator-5c9796789

FailedCreate

Error creating: pods "olm-operator-5c9796789-" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes
(x6)

openshift-cluster-storage-operator

replicaset-controller

cluster-storage-operator-7d87854d6

FailedCreate

Error creating: pods "cluster-storage-operator-7d87854d6-" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes
(x6)

openshift-cluster-storage-operator

replicaset-controller

csi-snapshot-controller-operator-5f5d689c6b

FailedCreate

Error creating: pods "csi-snapshot-controller-operator-5f5d689c6b-" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes
(x4)

openshift-machine-config-operator

kubelet

kube-rbac-proxy-crio-master-0

BackOff

Back-off restarting failed container kube-rbac-proxy-crio in pod kube-rbac-proxy-crio-master-0_openshift-machine-config-operator(1249822f86f23526277d165c0d5d3c19)
(x3)

openshift-service-ca-operator

replicaset-controller

service-ca-operator-b865698dc

FailedCreate

Error creating: pods "service-ca-operator-b865698dc-" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes
(x6)

openshift-kube-apiserver-operator

replicaset-controller

kube-apiserver-operator-8b68b9d9b

FailedCreate

Error creating: pods "kube-apiserver-operator-8b68b9d9b-" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes
(x6)

openshift-machine-api

replicaset-controller

cluster-baremetal-operator-6f69995874

FailedCreate

Error creating: pods "cluster-baremetal-operator-6f69995874-" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes
(x6)

openshift-kube-storage-version-migrator-operator

replicaset-controller

kube-storage-version-migrator-operator-6bb5bfb6fd

FailedCreate

Error creating: pods "kube-storage-version-migrator-operator-6bb5bfb6fd-" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes
(x6)

openshift-etcd-operator

replicaset-controller

etcd-operator-8544cbcf9c

FailedCreate

Error creating: pods "etcd-operator-8544cbcf9c-" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes
(x6)

openshift-cluster-version

replicaset-controller

cluster-version-operator-56d8475767

FailedCreate

Error creating: pods "cluster-version-operator-56d8475767-" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes
(x6)

openshift-machine-api

replicaset-controller

cluster-baremetal-operator-6f69995874

FailedCreate

Error creating: pods "cluster-baremetal-operator-6f69995874-" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes
(x6)

openshift-ingress-operator

replicaset-controller

ingress-operator-66b84d69b

FailedCreate

Error creating: pods "ingress-operator-66b84d69b-" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes
(x6)

openshift-dns-operator

replicaset-controller

dns-operator-9c5679d8f

FailedCreate

Error creating: pods "dns-operator-9c5679d8f-" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes
(x6)

openshift-kube-scheduler-operator

replicaset-controller

openshift-kube-scheduler-operator-dddff6458

FailedCreate

Error creating: pods "openshift-kube-scheduler-operator-dddff6458-" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes
(x6)

openshift-kube-controller-manager-operator

replicaset-controller

kube-controller-manager-operator-ff989d6cc

FailedCreate

Error creating: pods "kube-controller-manager-operator-ff989d6cc-" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes

openshift-authentication-operator

default-scheduler

authentication-operator-5885bfd7f4-62zrx

FailedScheduling

0/1 nodes are available: 1 node(s) had untolerated taint {node.kubernetes.io/not-ready: }. preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.
(x6)

openshift-monitoring

replicaset-controller

cluster-monitoring-operator-58845fbb57

FailedCreate

Error creating: pods "cluster-monitoring-operator-58845fbb57-" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes

openshift-authentication-operator

replicaset-controller

authentication-operator-5885bfd7f4

SuccessfulCreate

Created pod: authentication-operator-5885bfd7f4-62zrx

openshift-operator-lifecycle-manager

default-scheduler

package-server-manager-7b95f86987-2pg77

FailedScheduling

0/1 nodes are available: 1 node(s) had untolerated taint {node.kubernetes.io/not-ready: }. preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.

openshift-operator-lifecycle-manager

replicaset-controller

package-server-manager-7b95f86987

SuccessfulCreate

Created pod: package-server-manager-7b95f86987-2pg77

openshift-network-operator

default-scheduler

network-operator-7bd846bfc4-mt454

Scheduled

Successfully assigned openshift-network-operator/network-operator-7bd846bfc4-mt454 to master-0

openshift-apiserver-operator

default-scheduler

openshift-apiserver-operator-d65958b8-th2vj

FailedScheduling

0/1 nodes are available: 1 node(s) had untolerated taint {node.kubernetes.io/not-ready: }. preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.

openshift-network-operator

replicaset-controller

network-operator-7bd846bfc4

SuccessfulCreate

Created pod: network-operator-7bd846bfc4-mt454
(x6)

openshift-marketplace

replicaset-controller

marketplace-operator-89ccd998f

FailedCreate

Error creating: pods "marketplace-operator-89ccd998f-" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes
(x6)

openshift-machine-config-operator

replicaset-controller

machine-config-operator-84d549f6d5

FailedCreate

Error creating: pods "machine-config-operator-84d549f6d5-" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes

openshift-apiserver-operator

replicaset-controller

openshift-apiserver-operator-d65958b8

SuccessfulCreate

Created pod: openshift-apiserver-operator-d65958b8-th2vj

openshift-cluster-olm-operator

default-scheduler

cluster-olm-operator-67dcd4998-c5742

FailedScheduling

0/1 nodes are available: 1 node(s) had untolerated taint {node.kubernetes.io/not-ready: }. preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.
(x6)

openshift-operator-lifecycle-manager

replicaset-controller

catalog-operator-68f85b4d6c

FailedCreate

Error creating: pods "catalog-operator-68f85b4d6c-" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes

openshift-operator-lifecycle-manager

default-scheduler

olm-operator-5c9796789-tjm9l

FailedScheduling

0/1 nodes are available: 1 node(s) had untolerated taint {node.kubernetes.io/not-ready: }. preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.

openshift-cluster-olm-operator

replicaset-controller

cluster-olm-operator-67dcd4998

SuccessfulCreate

Created pod: cluster-olm-operator-67dcd4998-c5742
(x6)

openshift-monitoring

replicaset-controller

cluster-monitoring-operator-58845fbb57

FailedCreate

Error creating: pods "cluster-monitoring-operator-58845fbb57-" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes

openshift-operator-lifecycle-manager

replicaset-controller

olm-operator-5c9796789

SuccessfulCreate

Created pod: olm-operator-5c9796789-tjm9l

openshift-cluster-node-tuning-operator

default-scheduler

cluster-node-tuning-operator-598fbc5f8f-vxzvg

FailedScheduling

0/1 nodes are available: 1 node(s) had untolerated taint {node.kubernetes.io/not-ready: }. preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.

openshift-cluster-storage-operator

default-scheduler

csi-snapshot-controller-operator-5f5d689c6b-4dqrh

FailedScheduling

0/1 nodes are available: 1 node(s) had untolerated taint {node.kubernetes.io/not-ready: }. preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.

openshift-cluster-node-tuning-operator

replicaset-controller

cluster-node-tuning-operator-598fbc5f8f

SuccessfulCreate

Created pod: cluster-node-tuning-operator-598fbc5f8f-vxzvg

openshift-kube-storage-version-migrator-operator

default-scheduler

kube-storage-version-migrator-operator-6bb5bfb6fd-64dg5

FailedScheduling

0/1 nodes are available: 1 node(s) had untolerated taint {node.kubernetes.io/not-ready: }. preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.

openshift-cluster-storage-operator

replicaset-controller

cluster-storage-operator-7d87854d6

SuccessfulCreate

Created pod: cluster-storage-operator-7d87854d6-vlq7h

openshift-cluster-storage-operator

default-scheduler

cluster-storage-operator-7d87854d6-vlq7h

FailedScheduling

0/1 nodes are available: 1 node(s) had untolerated taint {node.kubernetes.io/not-ready: }. preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.

openshift-kube-storage-version-migrator-operator

replicaset-controller

kube-storage-version-migrator-operator-6bb5bfb6fd

SuccessfulCreate

Created pod: kube-storage-version-migrator-operator-6bb5bfb6fd-64dg5

openshift-controller-manager-operator

default-scheduler

openshift-controller-manager-operator-8c94f4649-p7pt6

FailedScheduling

0/1 nodes are available: 1 node(s) had untolerated taint {node.kubernetes.io/not-ready: }. preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.

openshift-cluster-version

replicaset-controller

cluster-version-operator-56d8475767

SuccessfulCreate

Created pod: cluster-version-operator-56d8475767-9gfkg

openshift-cluster-version

default-scheduler

cluster-version-operator-56d8475767-9gfkg

Scheduled

Successfully assigned openshift-cluster-version/cluster-version-operator-56d8475767-9gfkg to master-0

openshift-service-ca-operator

default-scheduler

service-ca-operator-b865698dc-qzb2h

FailedScheduling

0/1 nodes are available: 1 node(s) had untolerated taint {node.kubernetes.io/not-ready: }. preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.

openshift-config-operator

default-scheduler

openshift-config-operator-95bf4f4d-25cml

FailedScheduling

0/1 nodes are available: 1 node(s) had untolerated taint {node.kubernetes.io/not-ready: }. preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.

openshift-image-registry

default-scheduler

cluster-image-registry-operator-5549dc66cb-fmhbq

FailedScheduling

0/1 nodes are available: 1 node(s) had untolerated taint {node.kubernetes.io/not-ready: }. preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.

openshift-image-registry

replicaset-controller

cluster-image-registry-operator-5549dc66cb

SuccessfulCreate

Created pod: cluster-image-registry-operator-5549dc66cb-fmhbq

openshift-cluster-node-tuning-operator

replicaset-controller

cluster-node-tuning-operator-598fbc5f8f

SuccessfulCreate

Created pod: cluster-node-tuning-operator-598fbc5f8f-vxzvg

openshift-cluster-storage-operator

replicaset-controller

csi-snapshot-controller-operator-5f5d689c6b

SuccessfulCreate

Created pod: csi-snapshot-controller-operator-5f5d689c6b-4dqrh

openshift-cluster-node-tuning-operator

default-scheduler

cluster-node-tuning-operator-598fbc5f8f-vxzvg

FailedScheduling

0/1 nodes are available: 1 node(s) had untolerated taint {node.kubernetes.io/not-ready: }. preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.

openshift-kube-apiserver-operator

replicaset-controller

kube-apiserver-operator-8b68b9d9b

SuccessfulCreate

Created pod: kube-apiserver-operator-8b68b9d9b-c6vkz

openshift-controller-manager-operator

replicaset-controller

openshift-controller-manager-operator-8c94f4649

SuccessfulCreate

Created pod: openshift-controller-manager-operator-8c94f4649-p7pt6

openshift-config-operator

replicaset-controller

openshift-config-operator-95bf4f4d

SuccessfulCreate

Created pod: openshift-config-operator-95bf4f4d-25cml

openshift-service-ca-operator

replicaset-controller

service-ca-operator-b865698dc

SuccessfulCreate

Created pod: service-ca-operator-b865698dc-qzb2h

openshift-kube-apiserver-operator

default-scheduler

kube-apiserver-operator-8b68b9d9b-c6vkz

FailedScheduling

0/1 nodes are available: 1 node(s) had untolerated taint {node.kubernetes.io/not-ready: }. preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.

openshift-machine-api

replicaset-controller

cluster-baremetal-operator-6f69995874

SuccessfulCreate

Created pod: cluster-baremetal-operator-6f69995874-dv6cd

openshift-machine-api

default-scheduler

cluster-baremetal-operator-6f69995874-dv6cd

FailedScheduling

0/1 nodes are available: 1 node(s) had untolerated taint {node.kubernetes.io/not-ready: }. preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.

openshift-kube-controller-manager-operator

replicaset-controller

kube-controller-manager-operator-ff989d6cc

SuccessfulCreate

Created pod: kube-controller-manager-operator-ff989d6cc-qcpb4

openshift-etcd-operator

default-scheduler

etcd-operator-8544cbcf9c-brhw4

FailedScheduling

0/1 nodes are available: 1 node(s) had untolerated taint {node.kubernetes.io/not-ready: }. preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.

openshift-ingress-operator

default-scheduler

ingress-operator-66b84d69b-gzg9m

FailedScheduling

0/1 nodes are available: 1 node(s) had untolerated taint {node.kubernetes.io/not-ready: }. preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.

openshift-monitoring

default-scheduler

cluster-monitoring-operator-58845fbb57-nljsr

FailedScheduling

0/1 nodes are available: 1 node(s) had untolerated taint {node.kubernetes.io/not-ready: }. preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.

default

node-controller

master-0

RegisteredNode

Node master-0 event: Registered Node master-0 in Controller

openshift-marketplace

replicaset-controller

marketplace-operator-89ccd998f

SuccessfulCreate

Created pod: marketplace-operator-89ccd998f-mvn4t

openshift-kube-controller-manager-operator

default-scheduler

kube-controller-manager-operator-ff989d6cc-qcpb4

FailedScheduling

0/1 nodes are available: 1 node(s) had untolerated taint {node.kubernetes.io/not-ready: }. preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.

openshift-machine-api

replicaset-controller

cluster-baremetal-operator-6f69995874

SuccessfulCreate

Created pod: cluster-baremetal-operator-6f69995874-dv6cd

openshift-etcd-operator

replicaset-controller

etcd-operator-8544cbcf9c

SuccessfulCreate

Created pod: etcd-operator-8544cbcf9c-brhw4

openshift-ingress-operator

replicaset-controller

ingress-operator-66b84d69b

SuccessfulCreate

Created pod: ingress-operator-66b84d69b-gzg9m

openshift-operator-lifecycle-manager

default-scheduler

catalog-operator-68f85b4d6c-fzm28

FailedScheduling

0/1 nodes are available: 1 node(s) had untolerated taint {node.kubernetes.io/not-ready: }. preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.

assisted-installer

default-scheduler

assisted-installer-controller-w2zwp

Scheduled

Successfully assigned assisted-installer/assisted-installer-controller-w2zwp to master-0

openshift-monitoring

replicaset-controller

cluster-monitoring-operator-58845fbb57

SuccessfulCreate

Created pod: cluster-monitoring-operator-58845fbb57-nljsr

openshift-kube-scheduler-operator

default-scheduler

openshift-kube-scheduler-operator-dddff6458-vx5d7

FailedScheduling

0/1 nodes are available: 1 node(s) had untolerated taint {node.kubernetes.io/not-ready: }. preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.

openshift-machine-config-operator

default-scheduler

machine-config-operator-84d549f6d5-gm4qr

FailedScheduling

0/1 nodes are available: 1 node(s) had untolerated taint {node.kubernetes.io/not-ready: }. preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.

openshift-marketplace

default-scheduler

marketplace-operator-89ccd998f-mvn4t

FailedScheduling

0/1 nodes are available: 1 node(s) had untolerated taint {node.kubernetes.io/not-ready: }. preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.

openshift-dns-operator

replicaset-controller

dns-operator-9c5679d8f

SuccessfulCreate

Created pod: dns-operator-9c5679d8f-r6dm8

openshift-machine-config-operator

replicaset-controller

machine-config-operator-84d549f6d5

SuccessfulCreate

Created pod: machine-config-operator-84d549f6d5-gm4qr

openshift-operator-lifecycle-manager

replicaset-controller

catalog-operator-68f85b4d6c

SuccessfulCreate

Created pod: catalog-operator-68f85b4d6c-fzm28

openshift-dns-operator

default-scheduler

dns-operator-9c5679d8f-r6dm8

FailedScheduling

0/1 nodes are available: 1 node(s) had untolerated taint {node.kubernetes.io/not-ready: }. preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.

openshift-machine-api

default-scheduler

cluster-baremetal-operator-6f69995874-dv6cd

FailedScheduling

0/1 nodes are available: 1 node(s) had untolerated taint {node.kubernetes.io/not-ready: }. preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.

openshift-kube-scheduler-operator

replicaset-controller

openshift-kube-scheduler-operator-dddff6458

SuccessfulCreate

Created pod: openshift-kube-scheduler-operator-dddff6458-vx5d7

openshift-monitoring

replicaset-controller

cluster-monitoring-operator-58845fbb57

SuccessfulCreate

Created pod: cluster-monitoring-operator-58845fbb57-nljsr

openshift-monitoring

default-scheduler

cluster-monitoring-operator-58845fbb57-nljsr

FailedScheduling

0/1 nodes are available: 1 node(s) had untolerated taint {node.kubernetes.io/not-ready: }. preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.
(x4)

openshift-machine-config-operator

kubelet

kube-rbac-proxy-crio-master-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine
(x4)

openshift-machine-config-operator

kubelet

kube-rbac-proxy-crio-master-0

Started

Started container kube-rbac-proxy-crio
(x4)

openshift-machine-config-operator

kubelet

kube-rbac-proxy-crio-master-0

Created

Created container: kube-rbac-proxy-crio

openshift-network-operator

kubelet

network-operator-7bd846bfc4-mt454

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ec8fd46dfb35ed10e8f98933166f69ce579c2f35b8db03d21e4c34fc544553e4"

assisted-installer

kubelet

assisted-installer-controller-w2zwp

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1e1faad2d9167d84e23585c1cea5962301845548043cf09578f943f79ca98016"

assisted-installer

kubelet

assisted-installer-controller-w2zwp

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1e1faad2d9167d84e23585c1cea5962301845548043cf09578f943f79ca98016" in 6.623s (6.623s including waiting). Image size: 687949580 bytes.

assisted-installer

kubelet

assisted-installer-controller-w2zwp

Created

Created container: assisted-installer-controller

assisted-installer

kubelet

assisted-installer-controller-w2zwp

Started

Started container assisted-installer-controller

openshift-network-operator

kubelet

network-operator-7bd846bfc4-mt454

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ec8fd46dfb35ed10e8f98933166f69ce579c2f35b8db03d21e4c34fc544553e4" in 6.609s (6.609s including waiting). Image size: 621648710 bytes.

openshift-network-operator

kubelet

network-operator-7bd846bfc4-mt454

Created

Created container: network-operator

openshift-network-operator

kubelet

network-operator-7bd846bfc4-mt454

Started

Started container network-operator

openshift-network-operator

cluster-network-operator

network-operator

FeatureGatesInitialized

FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{"AWSEFSDriverVolumeMetrics", "AdminNetworkPolicy", "AlibabaPlatform", "AzureWorkloadIdentity", "BareMetalLoadBalancer", "BuildCSIVolumes", "ChunkSizeMiB", "CloudDualStackNodeIPs", "DisableKubeletCloudCredentialProviders", "GCPLabelsTags", "HardwareSpeed", "IngressControllerLBSubnetsAWS", "KMSv1", "ManagedBootImages", "ManagedBootImagesAWS", "MultiArchInstallAWS", "MultiArchInstallGCP", "NetworkDiagnosticsConfig", "NetworkLiveMigration", "NetworkSegmentation", "NewOLM", "NodeDisruptionPolicy", "OnClusterBuild", "PersistentIPsForVirtualization", "PrivateHostedZoneAWS", "SetEIPForNLBIngressController", "StreamingCollectionEncodingToJSON", "StreamingCollectionEncodingToProtobuf", "VSphereControlPlaneMachineSet", "VSphereDriverConfiguration", "VSphereMultiVCenters", "VSphereStaticIPs", "ValidatingAdmissionPolicy"}, Disabled:[]v1.FeatureGateName{"AWSClusterHostedDNS", "AdditionalRoutingCapabilities", "AutomatedEtcdBackup", "BootcNodeManagement", "CSIDriverSharedResource", "ClusterAPIInstall", "ClusterAPIInstallIBMCloud", "ClusterMonitoringConfig", "ConsolePluginContentSecurityPolicy", "DNSNameResolver", "DynamicResourceAllocation", "EtcdBackendQuota", "EventedPLEG", "Example", "ExternalOIDC", "ExternalOIDCWithUIDAndExtraClaimMappings", "GCPClusterHostedDNS", "GatewayAPI", "ImageStreamImportMode", "IngressControllerDynamicConfigurationManager", "InsightsConfig", "InsightsConfigAPI", "InsightsOnDemandDataGather", "InsightsRuntimeExtractor", "MachineAPIMigration", "MachineAPIOperatorDisableMachineHealthCheckController", "MachineAPIProviderOpenStack", "MachineConfigNodes", "MaxUnavailableStatefulSet", "MetricsCollectionProfiles", "MinimumKubeletVersion", "MixedCPUsAllocation", "MultiArchInstallAzure", "NodeSwap", "NutanixMultiSubnets", "OVNObservability", "OpenShiftPodSecurityAdmission", "PinnedImages", "PlatformOperators", "ProcMountType", "RouteAdvertisements", "RouteExternalCertificate", "ServiceAccountTokenNodeBinding", "SignatureStores", "SigstoreImageVerification", "TranslateStreamCloseWebsocketRequests", "UpgradeStatus", "UserNamespacesPodSecurityStandards", "UserNamespacesSupport", "VSphereMultiNetworks", "VolumeAttributesClass", "VolumeGroupSnapshot"}}

openshift-network-operator

network-operator

network-operator-lock

LeaderElection

master-0_57d904a2-afc8-4de5-9b66-b3cabdc9230e became leader

assisted-installer

job-controller

assisted-installer-controller

Completed

Job completed

openshift-network-operator

default-scheduler

mtu-prober-fwvpv

Scheduled

Successfully assigned openshift-network-operator/mtu-prober-fwvpv to master-0

openshift-network-operator

kubelet

mtu-prober-fwvpv

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ec8fd46dfb35ed10e8f98933166f69ce579c2f35b8db03d21e4c34fc544553e4" already present on machine

openshift-network-operator

job-controller

mtu-prober

SuccessfulCreate

Created pod: mtu-prober-fwvpv

openshift-network-operator

kubelet

mtu-prober-fwvpv

Started

Started container prober

openshift-network-operator

kubelet

mtu-prober-fwvpv

Created

Created container: prober

openshift-network-operator

job-controller

mtu-prober

Completed

Job completed

kube-system

cluster-policy-controller-namespace-security-allocation-controller

bootstrap-kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openshift-multus namespace

openshift-multus

default-scheduler

multus-additional-cni-plugins-rpbcn

Scheduled

Successfully assigned openshift-multus/multus-additional-cni-plugins-rpbcn to master-0

openshift-multus

kubelet

multus-2fp4b

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5a55ec7ec64efd0f595d8084377b7e463a1807829b7617e5d4a9092dcd924c36"

openshift-multus

daemonset-controller

multus-additional-cni-plugins

SuccessfulCreate

Created pod: multus-additional-cni-plugins-rpbcn

openshift-multus

daemonset-controller

multus

SuccessfulCreate

Created pod: multus-2fp4b

openshift-multus

daemonset-controller

multus-additional-cni-plugins

SuccessfulCreate

Created pod: multus-additional-cni-plugins-rpbcn

openshift-multus

default-scheduler

multus-additional-cni-plugins-rpbcn

Scheduled

Successfully assigned openshift-multus/multus-additional-cni-plugins-rpbcn to master-0

openshift-multus

daemonset-controller

multus

SuccessfulCreate

Created pod: multus-2fp4b

openshift-multus

kubelet

multus-additional-cni-plugins-rpbcn

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:21b4c9cf268bb7abef7af187cd775d3f74d0bd33626250095428d53b705ee946"

openshift-multus

default-scheduler

multus-2fp4b

Scheduled

Successfully assigned openshift-multus/multus-2fp4b to master-0

openshift-multus

kubelet

multus-additional-cni-plugins-rpbcn

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:21b4c9cf268bb7abef7af187cd775d3f74d0bd33626250095428d53b705ee946"

openshift-multus

default-scheduler

multus-2fp4b

Scheduled

Successfully assigned openshift-multus/multus-2fp4b to master-0

openshift-multus

kubelet

multus-2fp4b

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5a55ec7ec64efd0f595d8084377b7e463a1807829b7617e5d4a9092dcd924c36"

openshift-multus

daemonset-controller

network-metrics-daemon

SuccessfulCreate

Created pod: network-metrics-daemon-srdjm

openshift-multus

default-scheduler

network-metrics-daemon-srdjm

Scheduled

Successfully assigned openshift-multus/network-metrics-daemon-srdjm to master-0

openshift-multus

daemonset-controller

network-metrics-daemon

SuccessfulCreate

Created pod: network-metrics-daemon-srdjm

openshift-multus

default-scheduler

network-metrics-daemon-srdjm

Scheduled

Successfully assigned openshift-multus/network-metrics-daemon-srdjm to master-0

openshift-multus

default-scheduler

multus-admission-controller-5dbbb8b86f-5rrrh

FailedScheduling

0/1 nodes are available: 1 node(s) had untolerated taint {node.kubernetes.io/not-ready: }. preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.

openshift-multus

kubelet

multus-additional-cni-plugins-rpbcn

Started

Started container egress-router-binary-copy

openshift-multus

replicaset-controller

multus-admission-controller-5dbbb8b86f

SuccessfulCreate

Created pod: multus-admission-controller-5dbbb8b86f-5rrrh

openshift-multus

deployment-controller

multus-admission-controller

ScalingReplicaSet

Scaled up replica set multus-admission-controller-5dbbb8b86f to 1

openshift-multus

kubelet

multus-additional-cni-plugins-rpbcn

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:21b4c9cf268bb7abef7af187cd775d3f74d0bd33626250095428d53b705ee946" in 2.659s (2.659s including waiting). Image size: 528956487 bytes.

openshift-multus

kubelet

multus-additional-cni-plugins-rpbcn

Created

Created container: egress-router-binary-copy

openshift-multus

kubelet

multus-additional-cni-plugins-rpbcn

Started

Started container egress-router-binary-copy

openshift-multus

kubelet

multus-additional-cni-plugins-rpbcn

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:21b4c9cf268bb7abef7af187cd775d3f74d0bd33626250095428d53b705ee946" in 2.659s (2.659s including waiting). Image size: 528956487 bytes.

openshift-multus

kubelet

multus-additional-cni-plugins-rpbcn

Created

Created container: egress-router-binary-copy

openshift-multus

deployment-controller

multus-admission-controller

ScalingReplicaSet

Scaled up replica set multus-admission-controller-5dbbb8b86f to 1

openshift-multus

replicaset-controller

multus-admission-controller-5dbbb8b86f

SuccessfulCreate

Created pod: multus-admission-controller-5dbbb8b86f-5rrrh

openshift-multus

default-scheduler

multus-admission-controller-5dbbb8b86f-5rrrh

FailedScheduling

0/1 nodes are available: 1 node(s) had untolerated taint {node.kubernetes.io/not-ready: }. preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.

openshift-multus

kubelet

multus-additional-cni-plugins-rpbcn

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5e12e4dc52214d3ada5ba5106caebe079eac1d9292c2571a5fe83411ce8e900d"

kube-system

cluster-policy-controller-namespace-security-allocation-controller

bootstrap-kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openshift-ovn-kubernetes namespace

openshift-multus

kubelet

multus-additional-cni-plugins-rpbcn

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5e12e4dc52214d3ada5ba5106caebe079eac1d9292c2571a5fe83411ce8e900d"

openshift-ovn-kubernetes

deployment-controller

ovnkube-control-plane

ScalingReplicaSet

Scaled up replica set ovnkube-control-plane-57f769d897 to 1

openshift-ovn-kubernetes

replicaset-controller

ovnkube-control-plane-57f769d897

SuccessfulCreate

Created pod: ovnkube-control-plane-57f769d897-z2zpj

kube-system

cluster-policy-controller-namespace-security-allocation-controller

bootstrap-kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openshift-host-network namespace

openshift-ovn-kubernetes

default-scheduler

ovnkube-control-plane-57f769d897-z2zpj

Scheduled

Successfully assigned openshift-ovn-kubernetes/ovnkube-control-plane-57f769d897-z2zpj to master-0

openshift-ovn-kubernetes

default-scheduler

ovnkube-node-nsmzc

Scheduled

Successfully assigned openshift-ovn-kubernetes/ovnkube-node-nsmzc to master-0

kube-system

cluster-policy-controller-namespace-security-allocation-controller

bootstrap-kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openshift-network-diagnostics namespace

openshift-ovn-kubernetes

daemonset-controller

ovnkube-node

SuccessfulCreate

Created pod: ovnkube-node-nsmzc

openshift-network-diagnostics

default-scheduler

network-check-source-b4bf74f6-fhvg6

FailedScheduling

0/1 nodes are available: 1 node(s) had untolerated taint {node-role.kubernetes.io/master: }. preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.

openshift-network-diagnostics

replicaset-controller

network-check-source-b4bf74f6

SuccessfulCreate

Created pod: network-check-source-b4bf74f6-fhvg6

openshift-network-diagnostics

deployment-controller

network-check-source

ScalingReplicaSet

Scaled up replica set network-check-source-b4bf74f6 to 1

openshift-network-diagnostics

daemonset-controller

network-check-target

SuccessfulCreate

Created pod: network-check-target-xnrw6

openshift-network-diagnostics

default-scheduler

network-check-target-xnrw6

Scheduled

Successfully assigned openshift-network-diagnostics/network-check-target-xnrw6 to master-0

kube-system

cluster-policy-controller-namespace-security-allocation-controller

bootstrap-kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openshift-network-node-identity namespace

openshift-ovn-kubernetes

kubelet

ovnkube-control-plane-57f769d897-z2zpj

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-multus

kubelet

multus-additional-cni-plugins-rpbcn

Created

Created container: cni-plugins

openshift-multus

kubelet

multus-2fp4b

Started

Started container kube-multus

openshift-multus

kubelet

multus-2fp4b

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5a55ec7ec64efd0f595d8084377b7e463a1807829b7617e5d4a9092dcd924c36" in 16.325s (16.325s including waiting). Image size: 1238100502 bytes.

openshift-ovn-kubernetes

kubelet

ovnkube-control-plane-57f769d897-z2zpj

Started

Started container kube-rbac-proxy

openshift-ovn-kubernetes

kubelet

ovnkube-control-plane-57f769d897-z2zpj

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:2abc1fd79e7781634ed5ed9e8f2b98b9094ea51f40ac3a773c5e5224607bf3d7"

openshift-ovn-kubernetes

kubelet

ovnkube-node-nsmzc

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:2abc1fd79e7781634ed5ed9e8f2b98b9094ea51f40ac3a773c5e5224607bf3d7"

openshift-multus

kubelet

multus-2fp4b

Created

Created container: kube-multus

openshift-multus

kubelet

multus-additional-cni-plugins-rpbcn

Started

Started container cni-plugins

openshift-ovn-kubernetes

kubelet

ovnkube-control-plane-57f769d897-z2zpj

Created

Created container: kube-rbac-proxy

openshift-multus

kubelet

multus-2fp4b

Created

Created container: kube-multus

openshift-multus

kubelet

multus-additional-cni-plugins-rpbcn

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5e12e4dc52214d3ada5ba5106caebe079eac1d9292c2571a5fe83411ce8e900d" in 12.948s (12.948s including waiting). Image size: 683195416 bytes.

openshift-multus

kubelet

multus-additional-cni-plugins-rpbcn

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5e12e4dc52214d3ada5ba5106caebe079eac1d9292c2571a5fe83411ce8e900d" in 12.948s (12.948s including waiting). Image size: 683195416 bytes.

openshift-multus

kubelet

multus-additional-cni-plugins-rpbcn

Created

Created container: cni-plugins

openshift-multus

kubelet

multus-2fp4b

Started

Started container kube-multus

openshift-multus

kubelet

multus-2fp4b

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5a55ec7ec64efd0f595d8084377b7e463a1807829b7617e5d4a9092dcd924c36" in 16.325s (16.325s including waiting). Image size: 1238100502 bytes.

openshift-multus

kubelet

multus-additional-cni-plugins-rpbcn

Started

Started container cni-plugins

openshift-multus

kubelet

multus-additional-cni-plugins-rpbcn

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:0e66fd50be6f83ce321a566dfb76f3725b597374077d5af13813b928f6b1267e"

openshift-network-node-identity

daemonset-controller

network-node-identity

SuccessfulCreate

Created pod: network-node-identity-6t5vb

openshift-multus

kubelet

multus-additional-cni-plugins-rpbcn

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:0e66fd50be6f83ce321a566dfb76f3725b597374077d5af13813b928f6b1267e"

openshift-network-node-identity

default-scheduler

network-node-identity-6t5vb

Scheduled

Successfully assigned openshift-network-node-identity/network-node-identity-6t5vb to master-0

openshift-network-node-identity

kubelet

network-node-identity-6t5vb

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:2abc1fd79e7781634ed5ed9e8f2b98b9094ea51f40ac3a773c5e5224607bf3d7"

openshift-multus

kubelet

multus-additional-cni-plugins-rpbcn

Started

Started container bond-cni-plugin

openshift-multus

kubelet

multus-additional-cni-plugins-rpbcn

Created

Created container: bond-cni-plugin

openshift-multus

kubelet

multus-additional-cni-plugins-rpbcn

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:0e66fd50be6f83ce321a566dfb76f3725b597374077d5af13813b928f6b1267e" in 2.484s (2.484s including waiting). Image size: 411587146 bytes.

openshift-multus

kubelet

multus-additional-cni-plugins-rpbcn

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:0e66fd50be6f83ce321a566dfb76f3725b597374077d5af13813b928f6b1267e" in 2.484s (2.484s including waiting). Image size: 411587146 bytes.

openshift-multus

kubelet

multus-additional-cni-plugins-rpbcn

Created

Created container: bond-cni-plugin

openshift-multus

kubelet

multus-additional-cni-plugins-rpbcn

Started

Started container bond-cni-plugin

openshift-multus

kubelet

multus-additional-cni-plugins-rpbcn

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e3a494212f1ba17f0f0980eef583218330eccb56eadf6b8cb0548c76d99b5014"

openshift-multus

kubelet

multus-additional-cni-plugins-rpbcn

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e3a494212f1ba17f0f0980eef583218330eccb56eadf6b8cb0548c76d99b5014"

openshift-multus

kubelet

multus-additional-cni-plugins-rpbcn

Started

Started container routeoverride-cni

openshift-multus

kubelet

multus-additional-cni-plugins-rpbcn

Started

Started container routeoverride-cni

openshift-multus

kubelet

multus-additional-cni-plugins-rpbcn

Created

Created container: routeoverride-cni

openshift-multus

kubelet

multus-additional-cni-plugins-rpbcn

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e3a494212f1ba17f0f0980eef583218330eccb56eadf6b8cb0548c76d99b5014" in 2.446s (2.446s including waiting). Image size: 407347125 bytes.

openshift-multus

kubelet

multus-additional-cni-plugins-rpbcn

Created

Created container: routeoverride-cni

openshift-multus

kubelet

multus-additional-cni-plugins-rpbcn

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e3a494212f1ba17f0f0980eef583218330eccb56eadf6b8cb0548c76d99b5014" in 2.446s (2.446s including waiting). Image size: 407347125 bytes.

openshift-multus

kubelet

multus-additional-cni-plugins-rpbcn

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:0a09f5a3ba4f60cce0145769509bab92553c8075d210af4ac058965d2ae11efa"

openshift-multus

kubelet

multus-additional-cni-plugins-rpbcn

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:0a09f5a3ba4f60cce0145769509bab92553c8075d210af4ac058965d2ae11efa"
(x7)

openshift-multus

kubelet

network-metrics-daemon-srdjm

FailedMount

MountVolume.SetUp failed for volume "metrics-certs" : object "openshift-multus"/"metrics-daemon-secret" not registered
(x7)

openshift-multus

kubelet

network-metrics-daemon-srdjm

FailedMount

MountVolume.SetUp failed for volume "metrics-certs" : object "openshift-multus"/"metrics-daemon-secret" not registered
(x18)

openshift-multus

kubelet

network-metrics-daemon-srdjm

NetworkNotReady

network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?
(x18)

openshift-multus

kubelet

network-metrics-daemon-srdjm

NetworkNotReady

network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?

openshift-ovn-kubernetes

kubelet

ovnkube-control-plane-57f769d897-z2zpj

Created

Created container: ovnkube-cluster-manager

openshift-network-node-identity

kubelet

network-node-identity-6t5vb

Started

Started container approver

openshift-multus

kubelet

multus-additional-cni-plugins-rpbcn

Created

Created container: whereabouts-cni-bincopy

openshift-multus

kubelet

multus-additional-cni-plugins-rpbcn

Started

Started container whereabouts-cni-bincopy

openshift-ovn-kubernetes

kubelet

ovnkube-node-nsmzc

Created

Created container: kubecfg-setup

openshift-ovn-kubernetes

kubelet

ovnkube-node-nsmzc

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:2abc1fd79e7781634ed5ed9e8f2b98b9094ea51f40ac3a773c5e5224607bf3d7" in 21.404s (21.404s including waiting). Image size: 1637455533 bytes.

openshift-ovn-kubernetes

kubelet

ovnkube-control-plane-57f769d897-z2zpj

Started

Started container ovnkube-cluster-manager

openshift-ovn-kubernetes

kubelet

ovnkube-control-plane-57f769d897-z2zpj

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:2abc1fd79e7781634ed5ed9e8f2b98b9094ea51f40ac3a773c5e5224607bf3d7" in 21.176s (21.176s including waiting). Image size: 1637455533 bytes.

openshift-multus

kubelet

multus-additional-cni-plugins-rpbcn

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:0a09f5a3ba4f60cce0145769509bab92553c8075d210af4ac058965d2ae11efa" in 14.124s (14.124s including waiting). Image size: 876160834 bytes.

openshift-network-node-identity

master-0_e002d2b0-ed7c-4024-9011-bbb4409da84a

ovnkube-identity

LeaderElection

master-0_e002d2b0-ed7c-4024-9011-bbb4409da84a became leader

openshift-multus

kubelet

multus-additional-cni-plugins-rpbcn

Started

Started container whereabouts-cni-bincopy

openshift-ovn-kubernetes

ovnk-controlplane

ovn-kubernetes-master

LeaderElection

ovnkube-control-plane-57f769d897-z2zpj became leader

openshift-multus

kubelet

multus-additional-cni-plugins-rpbcn

Created

Created container: whereabouts-cni-bincopy

openshift-multus

kubelet

multus-additional-cni-plugins-rpbcn

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:0a09f5a3ba4f60cce0145769509bab92553c8075d210af4ac058965d2ae11efa" in 14.124s (14.124s including waiting). Image size: 876160834 bytes.

openshift-network-node-identity

kubelet

network-node-identity-6t5vb

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:2abc1fd79e7781634ed5ed9e8f2b98b9094ea51f40ac3a773c5e5224607bf3d7" in 19.266s (19.266s including waiting). Image size: 1637455533 bytes.

openshift-network-node-identity

kubelet

network-node-identity-6t5vb

Created

Created container: webhook

openshift-network-node-identity

kubelet

network-node-identity-6t5vb

Started

Started container webhook

openshift-network-node-identity

kubelet

network-node-identity-6t5vb

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:2abc1fd79e7781634ed5ed9e8f2b98b9094ea51f40ac3a773c5e5224607bf3d7" already present on machine

openshift-network-node-identity

kubelet

network-node-identity-6t5vb

Created

Created container: approver

openshift-ovn-kubernetes

kubelet

ovnkube-node-nsmzc

Started

Started container kubecfg-setup

openshift-ovn-kubernetes

daemonset-controller

ovnkube-node

SuccessfulDelete

Deleted pod: ovnkube-node-nsmzc

openshift-multus

kubelet

multus-additional-cni-plugins-rpbcn

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:0a09f5a3ba4f60cce0145769509bab92553c8075d210af4ac058965d2ae11efa" already present on machine

openshift-multus

kubelet

multus-additional-cni-plugins-rpbcn

Created

Created container: whereabouts-cni

openshift-multus

kubelet

multus-additional-cni-plugins-rpbcn

Started

Started container whereabouts-cni

openshift-multus

kubelet

multus-additional-cni-plugins-rpbcn

Started

Started container whereabouts-cni

openshift-multus

kubelet

multus-additional-cni-plugins-rpbcn

Created

Created container: whereabouts-cni

openshift-multus

kubelet

multus-additional-cni-plugins-rpbcn

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:0a09f5a3ba4f60cce0145769509bab92553c8075d210af4ac058965d2ae11efa" already present on machine

openshift-ovn-kubernetes

kubelet

ovnkube-node-nsmzc

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:2abc1fd79e7781634ed5ed9e8f2b98b9094ea51f40ac3a773c5e5224607bf3d7" already present on machine

openshift-ovn-kubernetes

kubelet

ovnkube-node-nsmzc

Started

Started container kube-rbac-proxy-ovn-metrics

openshift-ovn-kubernetes

kubelet

ovnkube-node-nsmzc

Created

Created container: kube-rbac-proxy-ovn-metrics

openshift-ovn-kubernetes

kubelet

ovnkube-node-nsmzc

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-ovn-kubernetes

kubelet

ovnkube-node-nsmzc

Started

Started container kube-rbac-proxy-node

openshift-ovn-kubernetes

kubelet

ovnkube-node-nsmzc

Created

Created container: kube-rbac-proxy-node

openshift-ovn-kubernetes

kubelet

ovnkube-node-nsmzc

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-ovn-kubernetes

kubelet

ovnkube-node-nsmzc

Started

Started container ovn-acl-logging

openshift-ovn-kubernetes

kubelet

ovnkube-node-nsmzc

Created

Created container: ovn-acl-logging

openshift-ovn-kubernetes

kubelet

ovnkube-node-nsmzc

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:2abc1fd79e7781634ed5ed9e8f2b98b9094ea51f40ac3a773c5e5224607bf3d7" already present on machine

openshift-ovn-kubernetes

kubelet

ovnkube-node-nsmzc

Started

Started container ovn-controller

openshift-ovn-kubernetes

kubelet

ovnkube-node-nsmzc

Created

Created container: ovn-controller

openshift-ovn-kubernetes

kubelet

ovnkube-node-nsmzc

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:2abc1fd79e7781634ed5ed9e8f2b98b9094ea51f40ac3a773c5e5224607bf3d7" already present on machine

openshift-ovn-kubernetes

kubelet

ovnkube-node-nsmzc

Created

Created container: nbdb

openshift-ovn-kubernetes

kubelet

ovnkube-node-nsmzc

Created

Created container: northd

openshift-ovn-kubernetes

kubelet

ovnkube-node-nsmzc

Started

Started container northd

openshift-multus

kubelet

multus-additional-cni-plugins-rpbcn

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5a55ec7ec64efd0f595d8084377b7e463a1807829b7617e5d4a9092dcd924c36" already present on machine

openshift-multus

kubelet

multus-additional-cni-plugins-rpbcn

Created

Created container: kube-multus-additional-cni-plugins

openshift-multus

kubelet

multus-additional-cni-plugins-rpbcn

Created

Created container: kube-multus-additional-cni-plugins

openshift-multus

kubelet

multus-additional-cni-plugins-rpbcn

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5a55ec7ec64efd0f595d8084377b7e463a1807829b7617e5d4a9092dcd924c36" already present on machine

openshift-ovn-kubernetes

kubelet

ovnkube-node-nsmzc

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:2abc1fd79e7781634ed5ed9e8f2b98b9094ea51f40ac3a773c5e5224607bf3d7" already present on machine

openshift-ovn-kubernetes

kubelet

ovnkube-node-nsmzc

Started

Started container nbdb

openshift-ovn-kubernetes

kubelet

ovnkube-node-nsmzc

Started

Started container sbdb

openshift-ovn-kubernetes

kubelet

ovnkube-node-nsmzc

Created

Created container: sbdb

openshift-ovn-kubernetes

kubelet

ovnkube-node-nsmzc

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:2abc1fd79e7781634ed5ed9e8f2b98b9094ea51f40ac3a773c5e5224607bf3d7" already present on machine
(x8)

openshift-cluster-version

kubelet

cluster-version-operator-56d8475767-9gfkg

FailedMount

MountVolume.SetUp failed for volume "serving-cert" : secret "cluster-version-operator-serving-cert" not found

default

ovnkube-csr-approver-controller

csr-bzzq8

CSRApproved

CSR "csr-bzzq8" has been approved

openshift-ovn-kubernetes

kubelet

ovnkube-node-rxdwp

Started

Started container kubecfg-setup

openshift-ovn-kubernetes

kubelet

ovnkube-node-rxdwp

Created

Created container: kubecfg-setup

openshift-ovn-kubernetes

kubelet

ovnkube-node-rxdwp

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:2abc1fd79e7781634ed5ed9e8f2b98b9094ea51f40ac3a773c5e5224607bf3d7" already present on machine

openshift-ovn-kubernetes

default-scheduler

ovnkube-node-rxdwp

Scheduled

Successfully assigned openshift-ovn-kubernetes/ovnkube-node-rxdwp to master-0

openshift-ovn-kubernetes

daemonset-controller

ovnkube-node

SuccessfulCreate

Created pod: ovnkube-node-rxdwp

openshift-ovn-kubernetes

kubelet

ovnkube-node-rxdwp

Started

Started container ovn-acl-logging

openshift-ovn-kubernetes

kubelet

ovnkube-node-rxdwp

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:2abc1fd79e7781634ed5ed9e8f2b98b9094ea51f40ac3a773c5e5224607bf3d7" already present on machine

openshift-ovn-kubernetes

kubelet

ovnkube-node-rxdwp

Started

Started container kube-rbac-proxy-ovn-metrics

openshift-ovn-kubernetes

kubelet

ovnkube-node-rxdwp

Created

Created container: kube-rbac-proxy-ovn-metrics

openshift-ovn-kubernetes

kubelet

ovnkube-node-rxdwp

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-ovn-kubernetes

kubelet

ovnkube-node-rxdwp

Started

Started container kube-rbac-proxy-node

openshift-ovn-kubernetes

kubelet

ovnkube-node-rxdwp

Created

Created container: kube-rbac-proxy-node

openshift-ovn-kubernetes

kubelet

ovnkube-node-rxdwp

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-ovn-kubernetes

kubelet

ovnkube-node-rxdwp

Created

Created container: ovn-acl-logging

openshift-ovn-kubernetes

kubelet

ovnkube-node-rxdwp

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:2abc1fd79e7781634ed5ed9e8f2b98b9094ea51f40ac3a773c5e5224607bf3d7" already present on machine

openshift-ovn-kubernetes

kubelet

ovnkube-node-rxdwp

Created

Created container: ovn-controller

openshift-ovn-kubernetes

kubelet

ovnkube-node-rxdwp

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:2abc1fd79e7781634ed5ed9e8f2b98b9094ea51f40ac3a773c5e5224607bf3d7" already present on machine

openshift-ovn-kubernetes

kubelet

ovnkube-node-rxdwp

Started

Started container ovn-controller
(x7)

openshift-network-diagnostics

kubelet

network-check-target-xnrw6

FailedMount

MountVolume.SetUp failed for volume "kube-api-access-w2zzd" : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered]
(x18)

openshift-network-diagnostics

kubelet

network-check-target-xnrw6

NetworkNotReady

network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?

openshift-ovn-kubernetes

kubelet

ovnkube-node-rxdwp

Created

Created container: northd

openshift-ovn-kubernetes

kubelet

ovnkube-node-rxdwp

Started

Started container nbdb

openshift-ovn-kubernetes

kubelet

ovnkube-node-rxdwp

Created

Created container: nbdb

openshift-ovn-kubernetes

kubelet

ovnkube-node-rxdwp

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:2abc1fd79e7781634ed5ed9e8f2b98b9094ea51f40ac3a773c5e5224607bf3d7" already present on machine

openshift-ovn-kubernetes

kubelet

ovnkube-node-rxdwp

Started

Started container northd

openshift-ovn-kubernetes

kubelet

ovnkube-node-rxdwp

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:2abc1fd79e7781634ed5ed9e8f2b98b9094ea51f40ac3a773c5e5224607bf3d7" already present on machine

openshift-ovn-kubernetes

kubelet

ovnkube-node-rxdwp

Created

Created container: sbdb

openshift-ovn-kubernetes

kubelet

ovnkube-node-rxdwp

Started

Started container sbdb

openshift-ovn-kubernetes

kubelet

ovnkube-node-rxdwp

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:2abc1fd79e7781634ed5ed9e8f2b98b9094ea51f40ac3a773c5e5224607bf3d7" already present on machine

default

ovnk-controlplane

master-0

ErrorAddingResource

[k8s.ovn.org/node-chassis-id annotation not found for node master-0, error getting gateway config for node master-0: k8s.ovn.org/l3-gateway-config annotation not found for node "master-0", failed to update chassis to local for local node master-0, error: failed to parse node chassis-id for node - master-0, error: k8s.ovn.org/node-chassis-id annotation not found for node master-0]

default

ovnkube-csr-approver-controller

csr-qpzlb

CSRApproved

CSR "csr-qpzlb" has been approved

openshift-machine-api

default-scheduler

cluster-baremetal-operator-6f69995874-dv6cd

Scheduled

Successfully assigned openshift-machine-api/cluster-baremetal-operator-6f69995874-dv6cd to master-0

openshift-controller-manager-operator

default-scheduler

openshift-controller-manager-operator-8c94f4649-p7pt6

Scheduled

Successfully assigned openshift-controller-manager-operator/openshift-controller-manager-operator-8c94f4649-p7pt6 to master-0

openshift-kube-apiserver-operator

default-scheduler

kube-apiserver-operator-8b68b9d9b-c6vkz

Scheduled

Successfully assigned openshift-kube-apiserver-operator/kube-apiserver-operator-8b68b9d9b-c6vkz to master-0

openshift-operator-lifecycle-manager

default-scheduler

package-server-manager-7b95f86987-2pg77

Scheduled

Successfully assigned openshift-operator-lifecycle-manager/package-server-manager-7b95f86987-2pg77 to master-0

openshift-monitoring

default-scheduler

cluster-monitoring-operator-58845fbb57-nljsr

Scheduled

Successfully assigned openshift-monitoring/cluster-monitoring-operator-58845fbb57-nljsr to master-0

openshift-config-operator

default-scheduler

openshift-config-operator-95bf4f4d-25cml

Scheduled

Successfully assigned openshift-config-operator/openshift-config-operator-95bf4f4d-25cml to master-0

openshift-monitoring

default-scheduler

cluster-monitoring-operator-58845fbb57-nljsr

Scheduled

Successfully assigned openshift-monitoring/cluster-monitoring-operator-58845fbb57-nljsr to master-0

openshift-cluster-node-tuning-operator

default-scheduler

cluster-node-tuning-operator-598fbc5f8f-vxzvg

Scheduled

Successfully assigned openshift-cluster-node-tuning-operator/cluster-node-tuning-operator-598fbc5f8f-vxzvg to master-0

openshift-multus

default-scheduler

multus-admission-controller-5dbbb8b86f-5rrrh

Scheduled

Successfully assigned openshift-multus/multus-admission-controller-5dbbb8b86f-5rrrh to master-0

openshift-operator-lifecycle-manager

default-scheduler

olm-operator-5c9796789-tjm9l

Scheduled

Successfully assigned openshift-operator-lifecycle-manager/olm-operator-5c9796789-tjm9l to master-0

openshift-kube-storage-version-migrator-operator

default-scheduler

kube-storage-version-migrator-operator-6bb5bfb6fd-64dg5

Scheduled

Successfully assigned openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6bb5bfb6fd-64dg5 to master-0

openshift-dns-operator

default-scheduler

dns-operator-9c5679d8f-r6dm8

Scheduled

Successfully assigned openshift-dns-operator/dns-operator-9c5679d8f-r6dm8 to master-0

openshift-cluster-olm-operator

default-scheduler

cluster-olm-operator-67dcd4998-c5742

Scheduled

Successfully assigned openshift-cluster-olm-operator/cluster-olm-operator-67dcd4998-c5742 to master-0

openshift-network-operator

daemonset-controller

iptables-alerter

SuccessfulCreate

Created pod: iptables-alerter-dd9wv

openshift-cluster-node-tuning-operator

default-scheduler

cluster-node-tuning-operator-598fbc5f8f-vxzvg

Scheduled

Successfully assigned openshift-cluster-node-tuning-operator/cluster-node-tuning-operator-598fbc5f8f-vxzvg to master-0

openshift-network-operator

default-scheduler

iptables-alerter-dd9wv

Scheduled

Successfully assigned openshift-network-operator/iptables-alerter-dd9wv to master-0

openshift-apiserver-operator

default-scheduler

openshift-apiserver-operator-d65958b8-th2vj

Scheduled

Successfully assigned openshift-apiserver-operator/openshift-apiserver-operator-d65958b8-th2vj to master-0

openshift-kube-scheduler-operator

default-scheduler

openshift-kube-scheduler-operator-dddff6458-vx5d7

Scheduled

Successfully assigned openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-dddff6458-vx5d7 to master-0

openshift-etcd-operator

default-scheduler

etcd-operator-8544cbcf9c-brhw4

Scheduled

Successfully assigned openshift-etcd-operator/etcd-operator-8544cbcf9c-brhw4 to master-0

openshift-operator-lifecycle-manager

default-scheduler

catalog-operator-68f85b4d6c-fzm28

Scheduled

Successfully assigned openshift-operator-lifecycle-manager/catalog-operator-68f85b4d6c-fzm28 to master-0

openshift-cluster-storage-operator

default-scheduler

cluster-storage-operator-7d87854d6-vlq7h

Scheduled

Successfully assigned openshift-cluster-storage-operator/cluster-storage-operator-7d87854d6-vlq7h to master-0

openshift-ingress-operator

default-scheduler

ingress-operator-66b84d69b-gzg9m

Scheduled

Successfully assigned openshift-ingress-operator/ingress-operator-66b84d69b-gzg9m to master-0

openshift-service-ca-operator

default-scheduler

service-ca-operator-b865698dc-qzb2h

Scheduled

Successfully assigned openshift-service-ca-operator/service-ca-operator-b865698dc-qzb2h to master-0

openshift-machine-api

default-scheduler

cluster-baremetal-operator-6f69995874-dv6cd

Scheduled

Successfully assigned openshift-machine-api/cluster-baremetal-operator-6f69995874-dv6cd to master-0

openshift-multus

default-scheduler

multus-admission-controller-5dbbb8b86f-5rrrh

Scheduled

Successfully assigned openshift-multus/multus-admission-controller-5dbbb8b86f-5rrrh to master-0

openshift-authentication-operator

default-scheduler

authentication-operator-5885bfd7f4-62zrx

Scheduled

Successfully assigned openshift-authentication-operator/authentication-operator-5885bfd7f4-62zrx to master-0

openshift-machine-config-operator

default-scheduler

machine-config-operator-84d549f6d5-gm4qr

Scheduled

Successfully assigned openshift-machine-config-operator/machine-config-operator-84d549f6d5-gm4qr to master-0

openshift-kube-controller-manager-operator

default-scheduler

kube-controller-manager-operator-ff989d6cc-qcpb4

Scheduled

Successfully assigned openshift-kube-controller-manager-operator/kube-controller-manager-operator-ff989d6cc-qcpb4 to master-0

openshift-image-registry

default-scheduler

cluster-image-registry-operator-5549dc66cb-fmhbq

Scheduled

Successfully assigned openshift-image-registry/cluster-image-registry-operator-5549dc66cb-fmhbq to master-0

openshift-cluster-storage-operator

default-scheduler

csi-snapshot-controller-operator-5f5d689c6b-4dqrh

Scheduled

Successfully assigned openshift-cluster-storage-operator/csi-snapshot-controller-operator-5f5d689c6b-4dqrh to master-0

openshift-marketplace

default-scheduler

marketplace-operator-89ccd998f-mvn4t

Scheduled

Successfully assigned openshift-marketplace/marketplace-operator-89ccd998f-mvn4t to master-0

openshift-cluster-olm-operator

kubelet

cluster-olm-operator-67dcd4998-c5742

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3062f6485aec4770e60852b535c69a42527b305161fe856499c8658ead6d1e85"

openshift-cluster-olm-operator

multus

cluster-olm-operator-67dcd4998-c5742

AddedInterface

Add eth0 [10.128.0.6/23] from ovn-kubernetes

openshift-config-operator

multus

openshift-config-operator-95bf4f4d-25cml

AddedInterface

Add eth0 [10.128.0.17/23] from ovn-kubernetes

openshift-kube-controller-manager-operator

kubelet

kube-controller-manager-operator-ff989d6cc-qcpb4

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c032f87ae61d6f757ff3ce52620a70a43516591987731f25da77aba152f17458"

openshift-config-operator

kubelet

openshift-config-operator-95bf4f4d-25cml

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:292560e2d80b460468bb19fe0ddf289767c655027b03a76ee6c40c91ffe4c483"

openshift-kube-storage-version-migrator-operator

kubelet

kube-storage-version-migrator-operator-6bb5bfb6fd-64dg5

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5c983016b9ceed0fca1f51bd49c2653243c7e5af91cbf2f478b091db6e028252"

openshift-kube-apiserver-operator

multus

kube-apiserver-operator-8b68b9d9b-c6vkz

AddedInterface

Add eth0 [10.128.0.13/23] from ovn-kubernetes

openshift-kube-apiserver-operator

kubelet

kube-apiserver-operator-8b68b9d9b-c6vkz

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c5ce3d1134d6500e2b8528516c1889d7bbc6259aba4981c6983395b0e9eeff65" already present on machine

openshift-kube-storage-version-migrator-operator

multus

kube-storage-version-migrator-operator-6bb5bfb6fd-64dg5

AddedInterface

Add eth0 [10.128.0.14/23] from ovn-kubernetes

openshift-cluster-storage-operator

multus

csi-snapshot-controller-operator-5f5d689c6b-4dqrh

AddedInterface

Add eth0 [10.128.0.5/23] from ovn-kubernetes

openshift-cluster-storage-operator

kubelet

csi-snapshot-controller-operator-5f5d689c6b-4dqrh

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c527b4e8239a1f4f4e0a851113e7dd633b7dcb9d75b0e7b21c23d26304abcb3"

openshift-service-ca-operator

multus

service-ca-operator-b865698dc-qzb2h

AddedInterface

Add eth0 [10.128.0.10/23] from ovn-kubernetes

openshift-cluster-storage-operator

kubelet

cluster-storage-operator-7d87854d6-vlq7h

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:30a2f97d7785ce8b0ea5115e67c4554b64adefbc7856bcf6f4fe6cc7e938a310"

openshift-kube-apiserver-operator

kubelet

kube-apiserver-operator-8b68b9d9b-c6vkz

Created

Created container: kube-apiserver-operator

openshift-kube-apiserver-operator

kubelet

kube-apiserver-operator-8b68b9d9b-c6vkz

Started

Started container kube-apiserver-operator

openshift-service-ca-operator

kubelet

service-ca-operator-b865698dc-qzb2h

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:812819a9d712b9e345ef5f1404b242c281e2518ad724baebc393ec0fd3b3d263"

openshift-kube-controller-manager-operator

multus

kube-controller-manager-operator-ff989d6cc-qcpb4

AddedInterface

Add eth0 [10.128.0.18/23] from ovn-kubernetes

openshift-kube-scheduler-operator

kubelet

openshift-kube-scheduler-operator-dddff6458-vx5d7

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ef199844317b7b012879ed8d29f9b6bc37fad8a6fdb336103cbd5cabc74c4302"

openshift-cluster-storage-operator

multus

cluster-storage-operator-7d87854d6-vlq7h

AddedInterface

Add eth0 [10.128.0.21/23] from ovn-kubernetes

openshift-kube-scheduler-operator

multus

openshift-kube-scheduler-operator-dddff6458-vx5d7

AddedInterface

Add eth0 [10.128.0.27/23] from ovn-kubernetes

openshift-network-operator

kubelet

iptables-alerter-dd9wv

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c6a4383333a1fd6d05c3f60ec793913f7937ee3d77f002d85e6c61e20507bf55"

openshift-etcd-operator

kubelet

etcd-operator-8544cbcf9c-brhw4

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d5a971d5889f167cfe61a64c366424b87c17a6dc141ffcc43406cdcbb50cae2a"

openshift-etcd-operator

multus

etcd-operator-8544cbcf9c-brhw4

AddedInterface

Add eth0 [10.128.0.28/23] from ovn-kubernetes

openshift-authentication-operator

kubelet

authentication-operator-5885bfd7f4-62zrx

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bfe394b58ec6195de8b8420e781b7630d85a412b9112d892fea903f92b783427"

openshift-apiserver-operator

kubelet

openshift-apiserver-operator-d65958b8-th2vj

Failed

Error: ErrImagePull

openshift-controller-manager-operator

kubelet

openshift-controller-manager-operator-8c94f4649-p7pt6

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:313d1d8ca85e65236a59f058a3316c49436dde691b3a3930d5bc5e3b4b8c8a71"

openshift-apiserver-operator

kubelet

openshift-apiserver-operator-d65958b8-th2vj

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1f23bac0a2a6cfd638e4af679dc787a8790d99c391f6e2ade8087dc477ff765e"

openshift-kube-apiserver-operator

kube-apiserver-operator

kube-apiserver-operator

FeatureGatesInitialized

FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{"AWSEFSDriverVolumeMetrics", "AdminNetworkPolicy", "AlibabaPlatform", "AzureWorkloadIdentity", "BareMetalLoadBalancer", "BuildCSIVolumes", "ChunkSizeMiB", "CloudDualStackNodeIPs", "DisableKubeletCloudCredentialProviders", "GCPLabelsTags", "HardwareSpeed", "IngressControllerLBSubnetsAWS", "KMSv1", "ManagedBootImages", "ManagedBootImagesAWS", "MultiArchInstallAWS", "MultiArchInstallGCP", "NetworkDiagnosticsConfig", "NetworkLiveMigration", "NetworkSegmentation", "NewOLM", "NodeDisruptionPolicy", "OnClusterBuild", "PersistentIPsForVirtualization", "PrivateHostedZoneAWS", "SetEIPForNLBIngressController", "StreamingCollectionEncodingToJSON", "StreamingCollectionEncodingToProtobuf", "VSphereControlPlaneMachineSet", "VSphereDriverConfiguration", "VSphereMultiVCenters", "VSphereStaticIPs", "ValidatingAdmissionPolicy"}, Disabled:[]v1.FeatureGateName{"AWSClusterHostedDNS", "AdditionalRoutingCapabilities", "AutomatedEtcdBackup", "BootcNodeManagement", "CSIDriverSharedResource", "ClusterAPIInstall", "ClusterAPIInstallIBMCloud", "ClusterMonitoringConfig", "ConsolePluginContentSecurityPolicy", "DNSNameResolver", "DynamicResourceAllocation", "EtcdBackendQuota", "EventedPLEG", "Example", "ExternalOIDC", "ExternalOIDCWithUIDAndExtraClaimMappings", "GCPClusterHostedDNS", "GatewayAPI", "ImageStreamImportMode", "IngressControllerDynamicConfigurationManager", "InsightsConfig", "InsightsConfigAPI", "InsightsOnDemandDataGather", "InsightsRuntimeExtractor", "MachineAPIMigration", "MachineAPIOperatorDisableMachineHealthCheckController", "MachineAPIProviderOpenStack", "MachineConfigNodes", "MaxUnavailableStatefulSet", "MetricsCollectionProfiles", "MinimumKubeletVersion", "MixedCPUsAllocation", "MultiArchInstallAzure", "NodeSwap", "NutanixMultiSubnets", "OVNObservability", "OpenShiftPodSecurityAdmission", "PinnedImages", "PlatformOperators", "ProcMountType", "RouteAdvertisements", "RouteExternalCertificate", "ServiceAccountTokenNodeBinding", "SignatureStores", "SigstoreImageVerification", "TranslateStreamCloseWebsocketRequests", "UpgradeStatus", "UserNamespacesPodSecurityStandards", "UserNamespacesSupport", "VSphereMultiNetworks", "VolumeAttributesClass", "VolumeGroupSnapshot"}}

openshift-authentication-operator

multus

authentication-operator-5885bfd7f4-62zrx

AddedInterface

Add eth0 [10.128.0.25/23] from ovn-kubernetes

openshift-apiserver-operator

multus

openshift-apiserver-operator-d65958b8-th2vj

AddedInterface

Add eth0 [10.128.0.8/23] from ovn-kubernetes

openshift-controller-manager-operator

multus

openshift-controller-manager-operator-8c94f4649-p7pt6

AddedInterface

Add eth0 [10.128.0.23/23] from ovn-kubernetes

openshift-kube-apiserver-operator

kube-apiserver-operator

kube-apiserver-operator-lock

LeaderElection

kube-apiserver-operator-8b68b9d9b-c6vkz_2bd0786e-628d-4975-b853-0103ed13667e became leader

openshift-apiserver-operator

kubelet

openshift-apiserver-operator-d65958b8-th2vj

Failed

Failed to pull image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1f23bac0a2a6cfd638e4af679dc787a8790d99c391f6e2ade8087dc477ff765e": pull QPS exceeded
(x2)

openshift-kube-apiserver-operator

kube-apiserver-operator-status-controller-statussyncer_kube-apiserver

kube-apiserver-operator

OperatorVersionChanged

clusteroperator/kube-apiserver version "raw-internal" changed from "" to "4.18.35"

openshift-kube-apiserver-operator

kube-apiserver-operator-status-controller-statussyncer_kube-apiserver

kube-apiserver-operator

OperatorStatusChanged

Status for clusteroperator/kube-apiserver changed: Degraded set to Unknown (""),Progressing set to Unknown (""),Available set to Unknown (""),Upgradeable set to Unknown (""),EvaluationConditionsDetected set to Unknown (""),status.relatedObjects changed from [{"operator.openshift.io" "kubeapiservers" "" "cluster"} {"apiextensions.k8s.io" "customresourcedefinitions" "" ""} {"security.openshift.io" "securitycontextconstraints" "" ""} {"" "namespaces" "" "openshift-config"} {"" "namespaces" "" "openshift-config-managed"} {"" "namespaces" "" "openshift-kube-apiserver-operator"} {"" "namespaces" "" "openshift-kube-apiserver"} {"admissionregistration.k8s.io" "mutatingwebhookconfigurations" "" ""} {"admissionregistration.k8s.io" "validatingwebhookconfigurations" "" ""} {"controlplane.operator.openshift.io" "podnetworkconnectivitychecks" "openshift-kube-apiserver" ""} {"apiserver.openshift.io" "apirequestcounts" "" ""}] to [{"operator.openshift.io" "kubeapiservers" "" "cluster"} {"apiextensions.k8s.io" "customresourcedefinitions" "" ""} {"security.openshift.io" "securitycontextconstraints" "" ""} {"" "namespaces" "" "openshift-config"} {"" "namespaces" "" "openshift-config-managed"} {"" "namespaces" "" "openshift-kube-apiserver-operator"} {"" "namespaces" "" "openshift-kube-apiserver"} {"admissionregistration.k8s.io" "mutatingwebhookconfigurations" "" ""} {"admissionregistration.k8s.io" "validatingwebhookconfigurations" "" ""} {"controlplane.operator.openshift.io" "podnetworkconnectivitychecks" "openshift-kube-apiserver" ""} {"apiserver.openshift.io" "apirequestcounts" "" ""} {"config.openshift.io" "nodes" "" "cluster"}],status.versions changed from [] to [{"raw-internal" "4.18.35"}]
(x2)

openshift-apiserver-operator

kubelet

openshift-apiserver-operator-d65958b8-th2vj

BackOff

Back-off pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1f23bac0a2a6cfd638e4af679dc787a8790d99c391f6e2ade8087dc477ff765e"

openshift-kube-apiserver-operator

kube-apiserver-operator-high-cpu-usage-alert-controller-highcpuusagealertcontroller

kube-apiserver-operator

PrometheusRuleCreated

Created PrometheusRule.monitoring.coreos.com/cpu-utilization -n openshift-kube-apiserver because it was missing
(x2)

openshift-apiserver-operator

kubelet

openshift-apiserver-operator-d65958b8-th2vj

Failed

Error: ImagePullBackOff
(x4)

openshift-operator-lifecycle-manager

kubelet

package-server-manager-7b95f86987-2pg77

FailedMount

MountVolume.SetUp failed for volume "package-server-manager-serving-cert" : secret "package-server-manager-serving-cert" not found
(x4)

openshift-ingress-operator

kubelet

ingress-operator-66b84d69b-gzg9m

FailedMount

MountVolume.SetUp failed for volume "metrics-tls" : secret "metrics-tls" not found
(x4)

openshift-monitoring

kubelet

cluster-monitoring-operator-58845fbb57-nljsr

FailedMount

MountVolume.SetUp failed for volume "cluster-monitoring-operator-tls" : secret "cluster-monitoring-operator-tls" not found
(x4)

openshift-monitoring

kubelet

cluster-monitoring-operator-58845fbb57-nljsr

FailedMount

MountVolume.SetUp failed for volume "cluster-monitoring-operator-tls" : secret "cluster-monitoring-operator-tls" not found
(x4)

openshift-multus

kubelet

multus-admission-controller-5dbbb8b86f-5rrrh

FailedMount

MountVolume.SetUp failed for volume "webhook-certs" : secret "multus-admission-controller-secret" not found
(x4)

openshift-cluster-node-tuning-operator

kubelet

cluster-node-tuning-operator-598fbc5f8f-vxzvg

FailedMount

MountVolume.SetUp failed for volume "apiservice-cert" : secret "performance-addon-operator-webhook-cert" not found
(x4)

openshift-cluster-node-tuning-operator

kubelet

cluster-node-tuning-operator-598fbc5f8f-vxzvg

FailedMount

MountVolume.SetUp failed for volume "node-tuning-operator-tls" : secret "node-tuning-operator-tls" not found
(x4)

openshift-dns-operator

kubelet

dns-operator-9c5679d8f-r6dm8

FailedMount

MountVolume.SetUp failed for volume "metrics-tls" : secret "metrics-tls" not found
(x4)

openshift-cluster-node-tuning-operator

kubelet

cluster-node-tuning-operator-598fbc5f8f-vxzvg

FailedMount

MountVolume.SetUp failed for volume "node-tuning-operator-tls" : secret "node-tuning-operator-tls" not found
(x4)

openshift-image-registry

kubelet

cluster-image-registry-operator-5549dc66cb-fmhbq

FailedMount

MountVolume.SetUp failed for volume "image-registry-operator-tls" : secret "image-registry-operator-tls" not found
(x4)

openshift-marketplace

kubelet

marketplace-operator-89ccd998f-mvn4t

FailedMount

MountVolume.SetUp failed for volume "marketplace-operator-metrics" : secret "marketplace-operator-metrics" not found

openshift-kube-apiserver-operator

kube-apiserver-operator-status-controller-statussyncer_kube-apiserver

kube-apiserver-operator

OperatorStatusChanged

Status for clusteroperator/kube-apiserver changed: Upgradeable message changed from "All is well" to "KubeletMinorVersionUpgradeable: Kubelet and API server minor versions are synced."
(x4)

openshift-machine-api

kubelet

cluster-baremetal-operator-6f69995874-dv6cd

FailedMount

MountVolume.SetUp failed for volume "cluster-baremetal-operator-tls" : secret "cluster-baremetal-operator-tls" not found
(x4)

openshift-operator-lifecycle-manager

kubelet

olm-operator-5c9796789-tjm9l

FailedMount

MountVolume.SetUp failed for volume "srv-cert" : secret "olm-operator-serving-cert" not found
(x4)

openshift-machine-api

kubelet

cluster-baremetal-operator-6f69995874-dv6cd

FailedMount

MountVolume.SetUp failed for volume "cert" : secret "cluster-baremetal-webhook-server-cert" not found
(x4)

openshift-cluster-node-tuning-operator

kubelet

cluster-node-tuning-operator-598fbc5f8f-vxzvg

FailedMount

MountVolume.SetUp failed for volume "apiservice-cert" : secret "performance-addon-operator-webhook-cert" not found

openshift-kube-apiserver-operator

kube-apiserver-operator-kube-apiserver-node

kube-apiserver-operator

MasterNodesReadyChanged

All master nodes are ready

openshift-kube-apiserver-operator

kube-apiserver-operator-kube-apiserver-node

kube-apiserver-operator

MasterNodeObserved

Observed new master node master-0
(x4)

openshift-machine-api

kubelet

cluster-baremetal-operator-6f69995874-dv6cd

FailedMount

MountVolume.SetUp failed for volume "cluster-baremetal-operator-tls" : secret "cluster-baremetal-operator-tls" not found
(x4)

openshift-machine-api

kubelet

cluster-baremetal-operator-6f69995874-dv6cd

FailedMount

MountVolume.SetUp failed for volume "cert" : secret "cluster-baremetal-webhook-server-cert" not found
(x4)

openshift-multus

kubelet

multus-admission-controller-5dbbb8b86f-5rrrh

FailedMount

MountVolume.SetUp failed for volume "webhook-certs" : secret "multus-admission-controller-secret" not found
(x4)

openshift-operator-lifecycle-manager

kubelet

catalog-operator-68f85b4d6c-fzm28

FailedMount

MountVolume.SetUp failed for volume "srv-cert" : secret "catalog-operator-serving-cert" not found

openshift-kube-apiserver-operator

kube-apiserver-operator-status-controller-statussyncer_kube-apiserver

kube-apiserver-operator

OperatorStatusChanged

Status for clusteroperator/kube-apiserver changed: Degraded changed from Unknown to False ("All is well")
(x4)

openshift-machine-config-operator

kubelet

machine-config-operator-84d549f6d5-gm4qr

FailedMount

MountVolume.SetUp failed for volume "proxy-tls" : secret "mco-proxy-tls" not found

openshift-kube-apiserver-operator

kube-apiserver-operator-status-controller-statussyncer_kube-apiserver

kube-apiserver-operator

OperatorStatusChanged

Status for clusteroperator/kube-apiserver changed: Upgradeable changed from Unknown to True ("All is well"),EvaluationConditionsDetected changed from Unknown to False ("All is well")

openshift-kube-apiserver-operator

kube-apiserver-operator-serviceaccountissuercontroller

kube-apiserver-operator

ServiceAccountIssuer

Issuer set to default value "https://kubernetes.default.svc"

openshift-kube-apiserver-operator

kube-apiserver-operator-status-controller-statussyncer_kube-apiserver

kube-apiserver-operator

OperatorStatusChanged

Status for clusteroperator/kube-apiserver changed: Degraded message changed from "NodeControllerDegraded: All master nodes are ready\nNodeKubeconfigControllerDegraded: \"secret/node-kubeconfigs\": secret \"node-system-admin-client\" not found\nRevisionControllerDegraded: configmap \"kube-apiserver-pod\" not found" to "NodeControllerDegraded: All master nodes are ready\nNodeKubeconfigControllerDegraded: \"secret/node-kubeconfigs\": secret \"node-system-admin-client\" not found\nRevisionControllerDegraded: configmap \"kube-apiserver-pod\" not found\nInstallerControllerDegraded: missing required resources: [configmaps: bound-sa-token-signing-certs-0,config-0,etcd-serving-ca-0,kube-apiserver-audit-policies-0,kube-apiserver-cert-syncer-kubeconfig-0,kube-apiserver-pod-0,kubelet-serving-ca-0,sa-token-signing-certs-0, secrets: etcd-client-0,localhost-recovery-client-token-0,localhost-recovery-serving-certkey-0]",Progressing message changed from "All is well" to "NodeInstallerProgressing: 1 node is at revision 0",Available changed from Unknown to False ("StaticPodsAvailable: 0 nodes are active; 1 node is at revision 0")

openshift-kube-apiserver-operator

kube-apiserver-operator-status-controller-statussyncer_kube-apiserver

kube-apiserver-operator

OperatorStatusChanged

Status for clusteroperator/kube-apiserver changed: Degraded message changed from "NodeControllerDegraded: All master nodes are ready" to "NodeControllerDegraded: All master nodes are ready\nNodeKubeconfigControllerDegraded: \"secret/node-kubeconfigs\": secret \"node-system-admin-client\" not found"

openshift-kube-apiserver-operator

kube-apiserver-operator-audit-policy-controller-auditpolicycontroller

kube-apiserver-operator

FastControllerResync

Controller "auditPolicyController" resync interval is set to 10s which might lead to client request throttling

openshift-kube-apiserver-operator

kube-apiserver-operator-cert-rotation-controller

kube-apiserver-operator

SignerUpdateRequired

"node-system-admin-signer" in "openshift-kube-apiserver-operator" requires a new signing cert/key pair: secret doesn't exist

openshift-kube-apiserver-operator

kube-apiserver-operator-status-controller-statussyncer_kube-apiserver

kube-apiserver-operator

OperatorStatusChanged

Status for clusteroperator/kube-apiserver changed: Degraded message changed from "All is well" to "NodeControllerDegraded: All master nodes are ready"
(x2)

openshift-kube-apiserver-operator

kube-apiserver-operator-cert-rotation-controller

kube-apiserver-operator

CABundleUpdateRequired

"loadbalancer-serving-ca" in "openshift-kube-apiserver-operator" requires a new cert: configmap doesn't exist

openshift-kube-apiserver-operator

kube-apiserver-operator-status-controller-statussyncer_kube-apiserver

kube-apiserver-operator

OperatorStatusChanged

Status for clusteroperator/kube-apiserver changed: Degraded message changed from "NodeControllerDegraded: All master nodes are ready\nNodeKubeconfigControllerDegraded: \"secret/node-kubeconfigs\": secret \"node-system-admin-client\" not found" to "NodeControllerDegraded: All master nodes are ready\nNodeKubeconfigControllerDegraded: \"secret/node-kubeconfigs\": secret \"node-system-admin-client\" not found\nRevisionControllerDegraded: configmap \"kube-apiserver-pod\" not found"

openshift-kube-apiserver-operator

kube-apiserver-operator-status-controller-statussyncer_kube-apiserver

kube-apiserver-operator

OperatorStatusChanged

Status for clusteroperator/kube-apiserver changed: Progressing changed from Unknown to False ("All is well")

openshift-kube-apiserver-operator

kube-apiserver-operator-cert-rotation-controller

kube-apiserver-operator

SignerUpdateRequired

"localhost-recovery-serving-signer" in "openshift-kube-apiserver-operator" requires a new signing cert/key pair: secret doesn't exist

openshift-kube-apiserver-operator

kube-apiserver-operator-cert-rotation-controller

kube-apiserver-operator

CABundleUpdateRequired

"service-network-serving-ca" in "openshift-kube-apiserver-operator" requires a new cert: configmap doesn't exist

openshift-kube-apiserver-operator

kube-apiserver-operator-cert-rotation-controller

kube-apiserver-operator

CABundleUpdateRequired

"localhost-serving-ca" in "openshift-kube-apiserver-operator" requires a new cert: configmap doesn't exist

openshift-kube-apiserver-operator

kube-apiserver-operator-cert-rotation-controller

kube-apiserver-operator

CABundleUpdateRequired

"kube-apiserver-aggregator-client-ca" in "openshift-config-managed" requires a new cert: configmap doesn't exist

openshift-kube-apiserver-operator

kube-apiserver-operator-cert-rotation-controller

kube-apiserver-operator

SecretUpdated

Updated Secret/aggregator-client-signer -n openshift-kube-apiserver-operator because it changed

openshift-kube-apiserver-operator

kube-apiserver-operator-cert-rotation-controller

kube-apiserver-operator

CABundleUpdateRequired

"node-system-admin-ca" in "openshift-kube-apiserver-operator" requires a new cert: configmap doesn't exist

openshift-kube-apiserver-operator

kube-apiserver-operator-cert-rotation-controller

kube-apiserver-operator

SecretCreated

Created Secret/node-system-admin-signer -n openshift-kube-apiserver-operator because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-cert-rotation-controller

kube-apiserver-operator

SecretUpdated

Updated Secret/kube-control-plane-signer -n openshift-kube-apiserver-operator because it changed

openshift-kube-apiserver-operator

kube-apiserver-operator-cert-rotation-controller

kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/loadbalancer-serving-ca -n openshift-kube-apiserver-operator because it was missing

default

kubelet

master-0

NodeAllocatableEnforced

Updated Node Allocatable limit across pods

default

kubelet

master-0

NodeHasSufficientPID

Node master-0 status is now: NodeHasSufficientPID

openshift-kube-apiserver-operator

kube-apiserver-operator-status-controller-statussyncer_kube-apiserver

kube-apiserver-operator

OperatorStatusChanged

Status for clusteroperator/kube-apiserver changed: Degraded message changed from "NodeControllerDegraded: All master nodes are ready\nNodeKubeconfigControllerDegraded: \"secret/node-kubeconfigs\": secret \"node-system-admin-client\" not found\nRevisionControllerDegraded: configmap \"kube-apiserver-pod\" not found\nInstallerControllerDegraded: missing required resources: [configmaps: bound-sa-token-signing-certs-0,config-0,etcd-serving-ca-0,kube-apiserver-audit-policies-0,kube-apiserver-cert-syncer-kubeconfig-0,kube-apiserver-pod-0,kubelet-serving-ca-0,sa-token-signing-certs-0, secrets: etcd-client-0,localhost-recovery-client-token-0,localhost-recovery-serving-certkey-0]" to "NodeControllerDegraded: All master nodes are ready\nNodeKubeconfigControllerDegraded: \"secret/node-kubeconfigs\": secret \"node-system-admin-client\" not found\nRevisionControllerDegraded: configmap \"kube-apiserver-pod\" not found\nInstallerControllerDegraded: missing required resources: [configmaps: bound-sa-token-signing-certs-0,config-0,etcd-serving-ca-0,kube-apiserver-audit-policies-0,kube-apiserver-cert-syncer-kubeconfig-0,kube-apiserver-pod-0,kubelet-serving-ca-0,sa-token-signing-certs-0, secrets: etcd-client-0,localhost-recovery-client-token-0,localhost-recovery-serving-certkey-0]\nCertRotation_ExternalLoadBalancerServing_Degraded: configmaps \"loadbalancer-serving-ca\" already exists"

openshift-kube-apiserver-operator

kube-apiserver-operator-cert-rotation-controller

kube-apiserver-operator

TargetUpdateRequired

"external-loadbalancer-serving-certkey" in "openshift-kube-apiserver" requires a new target cert/key pair: secret doesn't exist

openshift-kube-apiserver-operator

kube-apiserver-operator-cert-rotation-controller

kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/localhost-serving-ca -n openshift-kube-apiserver-operator because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-cert-rotation-controller

kube-apiserver-operator

TargetUpdateRequired

"localhost-serving-cert-certkey" in "openshift-kube-apiserver" requires a new target cert/key pair: secret doesn't exist

default

kubelet

master-0

NodeHasNoDiskPressure

Node master-0 status is now: NodeHasNoDiskPressure

openshift-kube-apiserver-operator

kube-apiserver-operator-cert-rotation-controller

kube-apiserver-operator

TargetUpdateRequired

"internal-loadbalancer-serving-certkey" in "openshift-kube-apiserver" requires a new target cert/key pair: secret doesn't exist

openshift-kube-apiserver-operator

kube-apiserver-operator-cert-rotation-controller

kube-apiserver-operator

ConfigMapCreateFailed

Failed to create ConfigMap/: configmaps "loadbalancer-serving-ca" already exists

openshift-kube-apiserver-operator

kube-apiserver-operator-boundsatokensignercontroller

kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/bound-sa-token-signing-certs -n openshift-kube-apiserver because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-cert-rotation-controller

kube-apiserver-operator

CABundleUpdateRequired

"kube-apiserver-to-kubelet-client-ca" in "openshift-kube-apiserver-operator" requires a new cert: configmap doesn't exist

default

kubelet

master-0

Starting

Starting kubelet.

openshift-kube-apiserver-operator

kube-apiserver-operator-cert-rotation-controller

kube-apiserver-operator

SecretUpdated

Updated Secret/kube-apiserver-to-kubelet-signer -n openshift-kube-apiserver-operator because it changed

openshift-kube-apiserver-operator

kube-apiserver-operator-cert-rotation-controller-cert-rotation-controller-ExternalLoadBalancerServing-certrotationcontroller

kube-apiserver-operator

RotationError

configmaps "loadbalancer-serving-ca" already exists

default

kubelet

master-0

NodeHasSufficientMemory

Node master-0 status is now: NodeHasSufficientMemory

openshift-kube-apiserver-operator

kube-apiserver-operator-cert-rotation-controller

kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/service-network-serving-ca -n openshift-kube-apiserver-operator because it was missing
(x2)

openshift-kube-apiserver-operator

kube-apiserver-operator-cert-rotation-controller

kube-apiserver-operator

CABundleUpdateRequired

"kube-control-plane-signer-ca" in "openshift-kube-apiserver-operator" requires a new cert: configmap doesn't exist

openshift-kube-apiserver-operator

kube-apiserver-operator-backingresourcecontroller-backingresourcecontroller-staticresources

kube-apiserver-operator

ServiceAccountCreated

Created ServiceAccount/installer-sa -n openshift-kube-apiserver because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-cert-rotation-controller

kube-apiserver-operator

TargetUpdateRequired

"service-network-serving-certkey" in "openshift-kube-apiserver" requires a new target cert/key pair: secret doesn't exist

openshift-kube-apiserver-operator

kube-apiserver-operator-resource-sync-controller-resourcesynccontroller

kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/sa-token-signing-certs -n openshift-kube-apiserver because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-backingresourcecontroller-backingresourcecontroller-staticresources

kube-apiserver-operator

ClusterRoleBindingCreated

Created ClusterRoleBinding.rbac.authorization.k8s.io/system:openshift:operator:openshift-kube-apiserver-installer because it was missing

openshift-kube-storage-version-migrator-operator

kubelet

kube-storage-version-migrator-operator-6bb5bfb6fd-64dg5

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5c983016b9ceed0fca1f51bd49c2653243c7e5af91cbf2f478b091db6e028252"

openshift-kube-apiserver-operator

kube-apiserver-operator-cert-rotation-controller

kube-apiserver-operator

SecretCreated

Created Secret/localhost-recovery-serving-signer -n openshift-kube-apiserver-operator because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-cert-rotation-controller

kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/kube-control-plane-signer-ca -n openshift-kube-apiserver-operator because it was missing

openshift-service-ca-operator

kubelet

service-ca-operator-b865698dc-qzb2h

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:812819a9d712b9e345ef5f1404b242c281e2518ad724baebc393ec0fd3b3d263"

openshift-kube-apiserver-operator

kube-apiserver-operator-cert-rotation-controller

kube-apiserver-operator

TargetUpdateRequired

"kube-scheduler-client-cert-key" in "openshift-config-managed" requires a new target cert/key pair: secret doesn't exist

openshift-cluster-storage-operator

kubelet

csi-snapshot-controller-operator-5f5d689c6b-4dqrh

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c527b4e8239a1f4f4e0a851113e7dd633b7dcb9d75b0e7b21c23d26304abcb3"

openshift-kube-apiserver-operator

kube-apiserver-operator-cert-rotation-controller

kube-apiserver-operator

CABundleUpdateRequired

"localhost-recovery-serving-ca" in "openshift-kube-apiserver-operator" requires a new cert: configmap doesn't exist

openshift-apiserver-operator

kubelet

openshift-apiserver-operator-d65958b8-th2vj

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1f23bac0a2a6cfd638e4af679dc787a8790d99c391f6e2ade8087dc477ff765e"

openshift-controller-manager-operator

kubelet

openshift-controller-manager-operator-8c94f4649-p7pt6

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:313d1d8ca85e65236a59f058a3316c49436dde691b3a3930d5bc5e3b4b8c8a71"

openshift-authentication-operator

kubelet

authentication-operator-5885bfd7f4-62zrx

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bfe394b58ec6195de8b8420e781b7630d85a412b9112d892fea903f92b783427"

openshift-network-operator

kubelet

iptables-alerter-dd9wv

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c6a4383333a1fd6d05c3f60ec793913f7937ee3d77f002d85e6c61e20507bf55"

openshift-kube-apiserver-operator

kube-apiserver-operator-cert-rotation-controller

kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/node-system-admin-ca -n openshift-kube-apiserver-operator because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-audit-policy-controller-auditpolicycontroller

kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/kube-apiserver-audit-policies -n openshift-kube-apiserver because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-cert-rotation-controller

kube-apiserver-operator

TargetUpdateRequired

"node-system-admin-client" in "openshift-kube-apiserver-operator" requires a new target cert/key pair: secret doesn't exist

openshift-kube-controller-manager-operator

kubelet

kube-controller-manager-operator-ff989d6cc-qcpb4

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c032f87ae61d6f757ff3ce52620a70a43516591987731f25da77aba152f17458"

openshift-kube-apiserver-operator

kube-apiserver-operator-status-controller-statussyncer_kube-apiserver

kube-apiserver-operator

OperatorStatusChanged

Status for clusteroperator/kube-apiserver changed: Degraded message changed from "NodeControllerDegraded: All master nodes are ready\nNodeKubeconfigControllerDegraded: \"secret/node-kubeconfigs\": secret \"node-system-admin-client\" not found\nRevisionControllerDegraded: configmap \"kube-apiserver-pod\" not found\nInstallerControllerDegraded: missing required resources: [configmaps: bound-sa-token-signing-certs-0,config-0,etcd-serving-ca-0,kube-apiserver-audit-policies-0,kube-apiserver-cert-syncer-kubeconfig-0,kube-apiserver-pod-0,kubelet-serving-ca-0,sa-token-signing-certs-0, secrets: etcd-client-0,localhost-recovery-client-token-0,localhost-recovery-serving-certkey-0]\nCertRotation_ExternalLoadBalancerServing_Degraded: configmaps \"loadbalancer-serving-ca\" already exists" to "NodeControllerDegraded: All master nodes are ready\nNodeKubeconfigControllerDegraded: \"secret/node-kubeconfigs\": secret \"node-system-admin-client\" not found\nRevisionControllerDegraded: configmap \"kube-apiserver-pod\" not found\nInstallerControllerDegraded: missing required resources: [configmaps: bound-sa-token-signing-certs-0,config-0,etcd-serving-ca-0,kube-apiserver-audit-policies-0,kube-apiserver-cert-syncer-kubeconfig-0,kube-apiserver-pod-0,kubelet-serving-ca-0,sa-token-signing-certs-0, secrets: etcd-client-0,localhost-recovery-client-token-0,localhost-recovery-serving-certkey-0]\nCertRotation_ExternalLoadBalancerServing_Degraded: configmaps \"loadbalancer-serving-ca\" already exists\nConfigObservationDegraded: error writing updated observed config: Operation cannot be fulfilled on kubeapiservers.operator.openshift.io \"cluster\": the object has been modified; please apply your changes to the latest version and try again"

openshift-kube-apiserver-operator

kube-apiserver-operator-cert-rotation-controller

kube-apiserver-operator

SecretCreated

Created Secret/localhost-serving-cert-certkey -n openshift-kube-apiserver because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-cert-rotation-controller

kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/kube-apiserver-aggregator-client-ca -n openshift-config-managed because it was missing

openshift-kube-scheduler-operator

kubelet

openshift-kube-scheduler-operator-dddff6458-vx5d7

Failed

Error: ErrImagePull

openshift-kube-scheduler-operator

kubelet

openshift-kube-scheduler-operator-dddff6458-vx5d7

Failed

Failed to pull image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ef199844317b7b012879ed8d29f9b6bc37fad8a6fdb336103cbd5cabc74c4302": rpc error: code = Canceled desc = copying config: context canceled

openshift-kube-apiserver-operator

kube-apiserver-operator-boundsatokensignercontroller

kube-apiserver-operator

SecretCreated

Created Secret/bound-service-account-signing-key -n openshift-kube-apiserver because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-cert-rotation-controller

kube-apiserver-operator

TargetUpdateRequired

"kube-controller-manager-client-cert-key" in "openshift-config-managed" requires a new target cert/key pair: secret doesn't exist

openshift-kube-apiserver-operator

kube-apiserver-operator-cert-rotation-controller

kube-apiserver-operator

SecretCreated

Created Secret/internal-loadbalancer-serving-certkey -n openshift-kube-apiserver because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-cert-rotation-controller

kube-apiserver-operator

TargetUpdateRequired

"aggregator-client" in "openshift-kube-apiserver" requires a new target cert/key pair: secret doesn't exist

openshift-service-ca-operator

kubelet

service-ca-operator-b865698dc-qzb2h

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:812819a9d712b9e345ef5f1404b242c281e2518ad724baebc393ec0fd3b3d263" in 2.082s (2.082s including waiting). Image size: 508544745 bytes.

openshift-cluster-olm-operator

kubelet

cluster-olm-operator-67dcd4998-c5742

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3062f6485aec4770e60852b535c69a42527b305161fe856499c8658ead6d1e85" already present on machine

openshift-kube-storage-version-migrator-operator

kubelet

kube-storage-version-migrator-operator-6bb5bfb6fd-64dg5

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5c983016b9ceed0fca1f51bd49c2653243c7e5af91cbf2f478b091db6e028252" in 2.386s (2.386s including waiting). Image size: 504625081 bytes.

openshift-network-diagnostics

multus

network-check-target-xnrw6

AddedInterface

Add eth0 [10.128.0.3/23] from ovn-kubernetes

openshift-config-operator

kubelet

openshift-config-operator-95bf4f4d-25cml

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:292560e2d80b460468bb19fe0ddf289767c655027b03a76ee6c40c91ffe4c483" already present on machine

openshift-config-operator

kubelet

openshift-config-operator-95bf4f4d-25cml

Created

Created container: openshift-api

openshift-config-operator

kubelet

openshift-config-operator-95bf4f4d-25cml

Started

Started container openshift-api

openshift-etcd-operator

kubelet

etcd-operator-8544cbcf9c-brhw4

Failed

Failed to pull image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d5a971d5889f167cfe61a64c366424b87c17a6dc141ffcc43406cdcbb50cae2a": rpc error: code = Canceled desc = copying config: context canceled

openshift-cluster-storage-operator

kubelet

cluster-storage-operator-7d87854d6-vlq7h

Failed

Error: ErrImagePull

openshift-cluster-storage-operator

kubelet

cluster-storage-operator-7d87854d6-vlq7h

Failed

Failed to pull image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:30a2f97d7785ce8b0ea5115e67c4554b64adefbc7856bcf6f4fe6cc7e938a310": rpc error: code = Canceled desc = copying config: context canceled

openshift-controller-manager-operator

kubelet

openshift-controller-manager-operator-8c94f4649-p7pt6

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:313d1d8ca85e65236a59f058a3316c49436dde691b3a3930d5bc5e3b4b8c8a71" in 2.694s (2.694s including waiting). Image size: 507972093 bytes.

openshift-kube-apiserver-operator

kube-apiserver-operator-cert-rotation-controller

kube-apiserver-operator

TargetUpdateRequired

"kubelet-client" in "openshift-kube-apiserver" requires a new target cert/key pair: secret doesn't exist

openshift-etcd-operator

kubelet

etcd-operator-8544cbcf9c-brhw4

Failed

Error: ErrImagePull

openshift-kube-controller-manager-operator

kubelet

kube-controller-manager-operator-ff989d6cc-qcpb4

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c032f87ae61d6f757ff3ce52620a70a43516591987731f25da77aba152f17458" in 2.39s (2.39s including waiting). Image size: 508888171 bytes.

openshift-network-operator

kubelet

iptables-alerter-dd9wv

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c6a4383333a1fd6d05c3f60ec793913f7937ee3d77f002d85e6c61e20507bf55" in 2.074s (2.074s including waiting). Image size: 582154903 bytes.

openshift-authentication-operator

kubelet

authentication-operator-5885bfd7f4-62zrx

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bfe394b58ec6195de8b8420e781b7630d85a412b9112d892fea903f92b783427" in 2.698s (2.698s including waiting). Image size: 513221333 bytes.

openshift-cluster-storage-operator

kubelet

csi-snapshot-controller-operator-5f5d689c6b-4dqrh

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c527b4e8239a1f4f4e0a851113e7dd633b7dcb9d75b0e7b21c23d26304abcb3" in 2.691s (2.691s including waiting). Image size: 506480167 bytes.

openshift-kube-apiserver-operator

kube-apiserver-operator-cert-rotation-controller

kube-apiserver-operator

TargetUpdateRequired

"control-plane-node-admin-client-cert-key" in "openshift-kube-apiserver" requires a new target cert/key pair: secret doesn't exist

openshift-kube-apiserver-operator

kube-apiserver-operator-cert-rotation-controller

kube-apiserver-operator

ConfigMapCreateFailed

Failed to create ConfigMap/: configmaps "kube-control-plane-signer-ca" already exists

openshift-kube-apiserver-operator

kube-apiserver-operator-cert-rotation-controller

kube-apiserver-operator

SecretCreated

Created Secret/service-network-serving-certkey -n openshift-kube-apiserver because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-cert-rotation-controller

kube-apiserver-operator

SecretCreated

Created Secret/external-loadbalancer-serving-certkey -n openshift-kube-apiserver because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-cert-rotation-controller

kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/kube-apiserver-to-kubelet-client-ca -n openshift-kube-apiserver-operator because it was missing

openshift-cluster-storage-operator

kubelet

csi-snapshot-controller-operator-5f5d689c6b-4dqrh

Started

Started container csi-snapshot-controller-operator

openshift-config-operator

kubelet

openshift-config-operator-95bf4f4d-25cml

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:712d334b7752d95580571059aae2c50e111d879af4fd8ea7cc3dbaf1a8e7dc69"

openshift-network-diagnostics

kubelet

network-check-target-xnrw6

Created

Created container: network-check-target-container

openshift-cluster-storage-operator

kubelet

csi-snapshot-controller-operator-5f5d689c6b-4dqrh

Created

Created container: csi-snapshot-controller-operator

openshift-cluster-olm-operator

kubelet

cluster-olm-operator-67dcd4998-c5742

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:4b5ea1ef4e09b673a0c68c8848ca162ab11d9ac373a377daa52dea702ffa3023"

openshift-kube-apiserver-operator

kube-apiserver-operator-cert-rotation-controller

kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/localhost-recovery-serving-ca -n openshift-kube-apiserver-operator because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-cert-rotation-controller

kube-apiserver-operator

TargetUpdateRequired

"localhost-recovery-serving-certkey" in "openshift-kube-apiserver" requires a new target cert/key pair: secret doesn't exist

openshift-kube-apiserver-operator

kube-apiserver-operator-cert-rotation-controller

kube-apiserver-operator

SecretCreated

Created Secret/kube-scheduler-client-cert-key -n openshift-config-managed because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-cert-rotation-controller

kube-apiserver-operator

SecretCreated

Created Secret/node-system-admin-client -n openshift-kube-apiserver-operator because it was missing

openshift-network-diagnostics

kubelet

network-check-target-xnrw6

Started

Started container network-check-target-container

openshift-cluster-olm-operator

kubelet

cluster-olm-operator-67dcd4998-c5742

Started

Started container copy-catalogd-manifests

openshift-cluster-olm-operator

kubelet

cluster-olm-operator-67dcd4998-c5742

Created

Created container: copy-catalogd-manifests

openshift-kube-storage-version-migrator-operator

openshift-kube-storage-version-migrator-operator

openshift-kube-storage-version-migrator-operator-lock

LeaderElection

kube-storage-version-migrator-operator-6bb5bfb6fd-64dg5_af974ef0-7f81-4ece-b30c-e092c4122294 became leader

openshift-service-ca-operator

service-ca-operator

service-ca-operator-lock

LeaderElection

service-ca-operator-b865698dc-qzb2h_f3b56674-14f0-4807-9adb-41bf192511fe became leader

openshift-authentication-operator

cluster-authentication-operator

cluster-authentication-operator-lock

LeaderElection

authentication-operator-5885bfd7f4-62zrx_7f42e53a-05c1-46ee-93f8-77b6bf5733e5 became leader

openshift-network-diagnostics

kubelet

network-check-target-xnrw6

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ec8fd46dfb35ed10e8f98933166f69ce579c2f35b8db03d21e4c34fc544553e4" already present on machine

openshift-kube-storage-version-migrator-operator

openshift-kube-storage-version-migrator-operator-status-controller-statussyncer_kube-storage-version-migrator

kube-storage-version-migrator-operator

OperatorVersionChanged

clusteroperator/kube-storage-version-migrator version "operator" changed from "" to "4.18.35"

openshift-kube-storage-version-migrator-operator

openshift-kube-storage-version-migrator-operator-status-controller-statussyncer_kube-storage-version-migrator

kube-storage-version-migrator-operator

OperatorStatusChanged

Status for clusteroperator/kube-storage-version-migrator changed: Degraded set to Unknown (""),Progressing set to Unknown (""),Available set to Unknown (""),Upgradeable set to Unknown (""),EvaluationConditionsDetected set to Unknown (""),status.versions changed from [] to [{"operator" "4.18.35"}]

openshift-controller-manager-operator

openshift-controller-manager-operator

openshift-controller-manager-operator

ConfigMapCreateFailed

Failed to create ConfigMap/config -n openshift-controller-manager: namespaces "openshift-controller-manager" not found

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Degraded set to False ("All is well"),Progressing set to Unknown (""),Available set to False ("OAuthServerServiceEndpointAccessibleControllerAvailable: service \"oauth-openshift\" not found\nOAuthServerServiceEndpointsEndpointAccessibleControllerAvailable: endpoints \"oauth-openshift\" not found"),Upgradeable set to True ("All is well"),EvaluationConditionsDetected set to Unknown (""),status.versions changed from [] to [{"operator" "4.18.35"}]

openshift-controller-manager-operator

openshift-controller-manager-operator-openshiftcontrollermanagerstaticresources-openshiftcontrollermanagerstaticresources-staticresources

openshift-controller-manager-operator

RoleCreated

Created Role.rbac.authorization.k8s.io/prometheus-k8s -n openshift-route-controller-manager because it was missing

openshift-controller-manager-operator

openshift-controller-manager-operator-openshiftcontrollermanagerstaticresources-openshiftcontrollermanagerstaticresources-staticresources

openshift-controller-manager-operator

RoleBindingCreated

Created RoleBinding.rbac.authorization.k8s.io/system:openshift:sa-creating-route-controller-manager -n openshift-infra because it was missing

kube-system

cluster-policy-controller-namespace-security-allocation-controller

bootstrap-kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openshift-kube-storage-version-migrator namespace

kube-system

cluster-policy-controller-namespace-security-allocation-controller

bootstrap-kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openshift-controller-manager namespace

kube-system

cluster-policy-controller-namespace-security-allocation-controller

bootstrap-kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openshift-service-ca namespace

kube-system

cluster-policy-controller-namespace-security-allocation-controller

bootstrap-kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openshift-route-controller-manager namespace

openshift-controller-manager-operator

openshift-controller-manager-operator-openshiftcontrollermanagerstaticresources-openshiftcontrollermanagerstaticresources-staticresources

openshift-controller-manager-operator

RoleCreated

Created Role.rbac.authorization.k8s.io/system:openshift:sa-creating-route-controller-manager -n openshift-infra because it was missing

openshift-controller-manager-operator

openshift-controller-manager-operator-openshiftcontrollermanagerstaticresources-openshiftcontrollermanagerstaticresources-staticresources

openshift-controller-manager-operator

ServiceAccountCreated

Created ServiceAccount/route-controller-manager-sa -n openshift-route-controller-manager because it was missing

openshift-cluster-storage-operator

deployment-controller

csi-snapshot-controller

ScalingReplicaSet

Scaled up replica set csi-snapshot-controller-64854d9cff to 1

openshift-controller-manager-operator

openshift-controller-manager-operator-openshiftcontrollermanagerstaticresources-openshiftcontrollermanagerstaticresources-staticresources

openshift-controller-manager-operator

NamespaceCreated

Created Namespace/openshift-route-controller-manager because it was missing

openshift-controller-manager-operator

openshift-controller-manager-operator-openshiftcontrollermanagerstaticresources-openshiftcontrollermanagerstaticresources-staticresources

openshift-controller-manager-operator

RoleBindingCreateFailed

Failed to create RoleBinding.rbac.authorization.k8s.io/system:openshift:leader-locking-openshift-route-controller-manager -n openshift-route-controller-manager: namespaces "openshift-route-controller-manager" not found

openshift-controller-manager-operator

openshift-controller-manager-operator

openshift-controller-manager-operator

DeploymentCreated

Created Deployment.apps/controller-manager -n openshift-controller-manager because it was missing

openshift-controller-manager-operator

openshift-controller-manager-operator-openshiftcontrollermanagerstaticresources-openshiftcontrollermanagerstaticresources-staticresources

openshift-controller-manager-operator

RoleCreateFailed

Failed to create Role.rbac.authorization.k8s.io/system:openshift:leader-locking-openshift-route-controller-manager -n openshift-route-controller-manager: namespaces "openshift-route-controller-manager" not found

openshift-cluster-storage-operator

csi-snapshot-controller-operator-status-controller-statussyncer_csi-snapshot-controller

csi-snapshot-controller-operator

OperatorStatusChanged

Status for clusteroperator/csi-snapshot-controller changed: Progressing message changed from "CSISnapshotControllerProgressing: Waiting for Deployment to act on changes" to "CSISnapshotControllerProgressing: Waiting for Deployment to deploy pods"

openshift-controller-manager-operator

openshift-controller-manager-operator

openshift-controller-manager-operator

ConfigMapCreateFailed

Failed to create ConfigMap/config -n openshift-route-controller-manager: namespaces "openshift-route-controller-manager" not found

openshift-controller-manager-operator

openshift-controller-manager-operator-openshiftcontrollermanagerstaticresources-openshiftcontrollermanagerstaticresources-staticresources

openshift-controller-manager-operator

ClusterRoleBindingCreated

Created ClusterRoleBinding.rbac.authorization.k8s.io/system:openshift:openshift-route-controller-manager because it was missing

openshift-cluster-storage-operator

csi-snapshot-controller-operator-csisnapshotgueststaticresourcecontroller-csisnapshotgueststaticresourcecontroller-staticresources

csi-snapshot-controller-operator

CustomResourceDefinitionCreated

Created CustomResourceDefinition.apiextensions.k8s.io/volumesnapshotclasses.snapshot.storage.k8s.io because it was missing

openshift-cluster-storage-operator

csi-snapshot-controller-operator-status-controller-statussyncer_csi-snapshot-controller

csi-snapshot-controller-operator

OperatorStatusChanged

Status for clusteroperator/csi-snapshot-controller changed: Progressing changed from Unknown to True ("CSISnapshotControllerProgressing: Waiting for Deployment to act on changes"),Available changed from Unknown to False ("CSISnapshotControllerAvailable: Waiting for Deployment")

openshift-cluster-storage-operator

csi-snapshot-controller-operator-csisnapshotgueststaticresourcecontroller-csisnapshotgueststaticresourcecontroller-staticresources

csi-snapshot-controller-operator

CustomResourceDefinitionCreated

Created CustomResourceDefinition.apiextensions.k8s.io/volumesnapshotcontents.snapshot.storage.k8s.io because it was missing

openshift-cluster-storage-operator

csi-snapshot-controller-operator-status-controller-statussyncer_csi-snapshot-controller

csi-snapshot-controller-operator

OperatorStatusChanged

Status for clusteroperator/csi-snapshot-controller changed: Upgradeable changed from Unknown to True ("All is well")

openshift-cluster-storage-operator

csi-snapshot-controller-operator-csisnapshotgueststaticresourcecontroller-csisnapshotgueststaticresourcecontroller-staticresources

csi-snapshot-controller-operator

CustomResourceDefinitionCreated

Created CustomResourceDefinition.apiextensions.k8s.io/volumesnapshots.snapshot.storage.k8s.io because it was missing

openshift-cluster-storage-operator

csi-snapshot-controller-operator-csisnapshotcontroller-deployment-controller--csisnapshotcontroller

csi-snapshot-controller-operator

DeploymentCreated

Created Deployment.apps/csi-snapshot-controller -n openshift-cluster-storage-operator because it was missing

openshift-cluster-storage-operator

csi-snapshot-controller-operator-status-controller-statussyncer_csi-snapshot-controller

csi-snapshot-controller-operator

OperatorStatusChanged

Status for clusteroperator/csi-snapshot-controller changed: Degraded set to False ("All is well"),Progressing set to Unknown (""),Available set to Unknown (""),Upgradeable set to Unknown (""),EvaluationConditionsDetected set to Unknown (""),status.relatedObjects changed from [] to [{"" "namespaces" "" "openshift-cluster-storage-operator"} {"operator.openshift.io" "csisnapshotcontrollers" "" "cluster"}]

openshift-cluster-storage-operator

csi-snapshot-controller-operator-csisnapshotstaticresourcecontroller-csisnapshotstaticresourcecontroller-staticresources

csi-snapshot-controller-operator

ServiceAccountCreated

Created ServiceAccount/csi-snapshot-controller -n openshift-cluster-storage-operator because it was missing

openshift-cluster-storage-operator

csi-snapshot-controller-operator

csi-snapshot-controller-operator

FeatureGatesInitialized

FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{"AWSEFSDriverVolumeMetrics", "AdminNetworkPolicy", "AlibabaPlatform", "AzureWorkloadIdentity", "BareMetalLoadBalancer", "BuildCSIVolumes", "ChunkSizeMiB", "CloudDualStackNodeIPs", "DisableKubeletCloudCredentialProviders", "GCPLabelsTags", "HardwareSpeed", "IngressControllerLBSubnetsAWS", "KMSv1", "ManagedBootImages", "ManagedBootImagesAWS", "MultiArchInstallAWS", "MultiArchInstallGCP", "NetworkDiagnosticsConfig", "NetworkLiveMigration", "NetworkSegmentation", "NewOLM", "NodeDisruptionPolicy", "OnClusterBuild", "PersistentIPsForVirtualization", "PrivateHostedZoneAWS", "SetEIPForNLBIngressController", "StreamingCollectionEncodingToJSON", "StreamingCollectionEncodingToProtobuf", "VSphereControlPlaneMachineSet", "VSphereDriverConfiguration", "VSphereMultiVCenters", "VSphereStaticIPs", "ValidatingAdmissionPolicy"}, Disabled:[]v1.FeatureGateName{"AWSClusterHostedDNS", "AdditionalRoutingCapabilities", "AutomatedEtcdBackup", "BootcNodeManagement", "CSIDriverSharedResource", "ClusterAPIInstall", "ClusterAPIInstallIBMCloud", "ClusterMonitoringConfig", "ConsolePluginContentSecurityPolicy", "DNSNameResolver", "DynamicResourceAllocation", "EtcdBackendQuota", "EventedPLEG", "Example", "ExternalOIDC", "ExternalOIDCWithUIDAndExtraClaimMappings", "GCPClusterHostedDNS", "GatewayAPI", "ImageStreamImportMode", "IngressControllerDynamicConfigurationManager", "InsightsConfig", "InsightsConfigAPI", "InsightsOnDemandDataGather", "InsightsRuntimeExtractor", "MachineAPIMigration", "MachineAPIOperatorDisableMachineHealthCheckController", "MachineAPIProviderOpenStack", "MachineConfigNodes", "MaxUnavailableStatefulSet", "MetricsCollectionProfiles", "MinimumKubeletVersion", "MixedCPUsAllocation", "MultiArchInstallAzure", "NodeSwap", "NutanixMultiSubnets", "OVNObservability", "OpenShiftPodSecurityAdmission", "PinnedImages", "PlatformOperators", "ProcMountType", "RouteAdvertisements", "RouteExternalCertificate", "ServiceAccountTokenNodeBinding", "SignatureStores", "SigstoreImageVerification", "TranslateStreamCloseWebsocketRequests", "UpgradeStatus", "UserNamespacesPodSecurityStandards", "UserNamespacesSupport", "VSphereMultiNetworks", "VolumeAttributesClass", "VolumeGroupSnapshot"}}

openshift-controller-manager-operator

openshift-controller-manager-operator

openshift-controller-manager-operator

ConfigMapCreated

Created ConfigMap/openshift-global-ca -n openshift-controller-manager because it was missing

openshift-controller-manager-operator

openshift-controller-manager-operator-openshiftcontrollermanagerstaticresources-openshiftcontrollermanagerstaticresources-staticresources

openshift-controller-manager-operator

ClusterRoleCreated

Created ClusterRole.rbac.authorization.k8s.io/system:openshift:openshift-route-controller-manager because it was missing

openshift-controller-manager-operator

openshift-controller-manager-operator-openshiftcontrollermanagerstaticresources-openshiftcontrollermanagerstaticresources-staticresources

openshift-controller-manager-operator

NamespaceCreated

Created Namespace/openshift-controller-manager because it was missing

openshift-apiserver-operator

kubelet

openshift-apiserver-operator-d65958b8-th2vj

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1f23bac0a2a6cfd638e4af679dc787a8790d99c391f6e2ade8087dc477ff765e" in 4.024s (4.024s including waiting). Image size: 512274055 bytes.

openshift-cluster-storage-operator

csi-snapshot-controller-operator

csi-snapshot-controller-operator-lock

LeaderElection

csi-snapshot-controller-operator-5f5d689c6b-4dqrh_9c48412b-d53f-411b-930b-c89edae89b25 became leader

openshift-controller-manager-operator

openshift-controller-manager-operator-openshiftcontrollermanagerstaticresources-openshiftcontrollermanagerstaticresources-staticresources

openshift-controller-manager-operator

RoleBindingCreateFailed

Failed to create RoleBinding.rbac.authorization.k8s.io/system:openshift:leader-locking-openshift-controller-manager -n openshift-controller-manager: namespaces "openshift-controller-manager" not found

openshift-cluster-storage-operator

replicaset-controller

csi-snapshot-controller-64854d9cff

SuccessfulCreate

Created pod: csi-snapshot-controller-64854d9cff-f44gr

openshift-controller-manager-operator

openshift-controller-manager-operator

openshift-controller-manager-operator

ConfigMapCreateFailed

Failed to create configmap/openshift-service-ca-n openshift-controller-manager: namespaces "openshift-controller-manager" not found

openshift-controller-manager-operator

openshift-controller-manager-operator-openshiftcontrollermanagerstaticresources-openshiftcontrollermanagerstaticresources-staticresources

openshift-controller-manager-operator

RoleCreateFailed

Failed to create Role.rbac.authorization.k8s.io/system:openshift:leader-locking-openshift-controller-manager -n openshift-controller-manager: namespaces "openshift-controller-manager" not found

openshift-controller-manager-operator

openshift-controller-manager-operator-status-controller-statussyncer_openshift-controller-manager

openshift-controller-manager-operator

OperatorStatusChanged

Status for clusteroperator/openshift-controller-manager changed: Degraded changed from Unknown to False ("All is well")

openshift-controller-manager-operator

openshift-controller-manager-operator-openshiftcontrollermanagerstaticresources-openshiftcontrollermanagerstaticresources-staticresources

openshift-controller-manager-operator

ClusterRoleBindingCreated

Created ClusterRoleBinding.rbac.authorization.k8s.io/system:openshift:tokenreview-openshift-controller-manager because it was missing

openshift-controller-manager-operator

openshift-controller-manager-operator-status-controller-statussyncer_openshift-controller-manager

openshift-controller-manager-operator

OperatorStatusChanged

Status for clusteroperator/openshift-controller-manager changed: Degraded set to Unknown (""),Progressing set to Unknown (""),Available set to Unknown (""),Upgradeable set to Unknown (""),EvaluationConditionsDetected set to Unknown (""),status.relatedObjects changed from [] to [{"operator.openshift.io" "openshiftcontrollermanagers" "" "cluster"} {"" "namespaces" "" "openshift-config"} {"" "namespaces" "" "openshift-config-managed"} {"" "namespaces" "" "openshift-controller-manager-operator"} {"" "namespaces" "" "openshift-controller-manager"} {"" "namespaces" "" "openshift-route-controller-manager"}]

openshift-controller-manager-operator

openshift-controller-manager-operator-openshiftcontrollermanagerstaticresources-openshiftcontrollermanagerstaticresources-staticresources

openshift-controller-manager-operator

ClusterRoleCreated

Created ClusterRole.rbac.authorization.k8s.io/system:openshift:tokenreview-openshift-controller-manager because it was missing

openshift-controller-manager-operator

openshift-controller-manager-operator-config-observer-configobserver

openshift-controller-manager-operator

ObservedConfigChanged

Writing updated observed config: map[string]any{ + "build": map[string]any{ + "buildDefaults": map[string]any{"resources": map[string]any{}}, + "imageTemplateFormat": map[string]any{ + "format": string("quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e7030c5cce"...), + }, + }, + "controllers": []any{ + string("openshift.io/build"), string("openshift.io/build-config-change"), + string("openshift.io/builder-rolebindings"), + string("openshift.io/builder-serviceaccount"), + string("-openshift.io/default-rolebindings"), string("openshift.io/deployer"), + string("openshift.io/deployer-rolebindings"), + string("openshift.io/deployer-serviceaccount"), + string("openshift.io/deploymentconfig"), string("openshift.io/image-import"), + string("openshift.io/image-puller-rolebindings"), + string("openshift.io/image-signature-import"), + string("openshift.io/image-trigger"), string("openshift.io/ingress-ip"), + string("openshift.io/ingress-to-route"), + string("openshift.io/origin-namespace"), ..., + }, + "deployer": map[string]any{ + "imageTemplateFormat": map[string]any{ + "format": string("quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6e5004457a"...), + }, + }, + "featureGates": []any{string("BuildCSIVolumes=true")}, + "ingress": map[string]any{"ingressIPNetworkCIDR": string("")}, }

openshift-controller-manager-operator

openshift-controller-manager-operator-config-observer-configobserver

openshift-controller-manager-operator

ObserveFeatureFlagsUpdated

Updated featureGates to BuildCSIVolumes=true

openshift-controller-manager-operator

openshift-controller-manager-operator-openshiftcontrollermanagerstaticresources-openshiftcontrollermanagerstaticresources-staticresources

openshift-controller-manager-operator

ClusterRoleBindingCreated

Created ClusterRoleBinding.rbac.authorization.k8s.io/system:openshift:openshift-controller-manager because it was missing

openshift-controller-manager-operator

openshift-controller-manager-operator-openshiftcontrollermanagerstaticresources-openshiftcontrollermanagerstaticresources-staticresources

openshift-controller-manager-operator

ClusterRoleCreated

Created ClusterRole.rbac.authorization.k8s.io/system:openshift:openshift-controller-manager because it was missing

openshift-controller-manager-operator

openshift-controller-manager-operator

openshift-controller-manager-operator

FeatureGatesInitialized

FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{"AWSEFSDriverVolumeMetrics", "AdminNetworkPolicy", "AlibabaPlatform", "AzureWorkloadIdentity", "BareMetalLoadBalancer", "BuildCSIVolumes", "ChunkSizeMiB", "CloudDualStackNodeIPs", "DisableKubeletCloudCredentialProviders", "GCPLabelsTags", "HardwareSpeed", "IngressControllerLBSubnetsAWS", "KMSv1", "ManagedBootImages", "ManagedBootImagesAWS", "MultiArchInstallAWS", "MultiArchInstallGCP", "NetworkDiagnosticsConfig", "NetworkLiveMigration", "NetworkSegmentation", "NewOLM", "NodeDisruptionPolicy", "OnClusterBuild", "PersistentIPsForVirtualization", "PrivateHostedZoneAWS", "SetEIPForNLBIngressController", "StreamingCollectionEncodingToJSON", "StreamingCollectionEncodingToProtobuf", "VSphereControlPlaneMachineSet", "VSphereDriverConfiguration", "VSphereMultiVCenters", "VSphereStaticIPs", "ValidatingAdmissionPolicy"}, Disabled:[]v1.FeatureGateName{"AWSClusterHostedDNS", "AdditionalRoutingCapabilities", "AutomatedEtcdBackup", "BootcNodeManagement", "CSIDriverSharedResource", "ClusterAPIInstall", "ClusterAPIInstallIBMCloud", "ClusterMonitoringConfig", "ConsolePluginContentSecurityPolicy", "DNSNameResolver", "DynamicResourceAllocation", "EtcdBackendQuota", "EventedPLEG", "Example", "ExternalOIDC", "ExternalOIDCWithUIDAndExtraClaimMappings", "GCPClusterHostedDNS", "GatewayAPI", "ImageStreamImportMode", "IngressControllerDynamicConfigurationManager", "InsightsConfig", "InsightsConfigAPI", "InsightsOnDemandDataGather", "InsightsRuntimeExtractor", "MachineAPIMigration", "MachineAPIOperatorDisableMachineHealthCheckController", "MachineAPIProviderOpenStack", "MachineConfigNodes", "MaxUnavailableStatefulSet", "MetricsCollectionProfiles", "MinimumKubeletVersion", "MixedCPUsAllocation", "MultiArchInstallAzure", "NodeSwap", "NutanixMultiSubnets", "OVNObservability", "OpenShiftPodSecurityAdmission", "PinnedImages", "PlatformOperators", "ProcMountType", "RouteAdvertisements", "RouteExternalCertificate", "ServiceAccountTokenNodeBinding", "SignatureStores", "SigstoreImageVerification", "TranslateStreamCloseWebsocketRequests", "UpgradeStatus", "UserNamespacesPodSecurityStandards", "UserNamespacesSupport", "VSphereMultiNetworks", "VolumeAttributesClass", "VolumeGroupSnapshot"}}

openshift-controller-manager-operator

openshift-controller-manager-operator

openshift-controller-manager-operator-lock

LeaderElection

openshift-controller-manager-operator-8c94f4649-p7pt6_bc43d1d2-5cbb-4b6d-8675-0d97f3f10354 became leader

openshift-kube-storage-version-migrator-operator

openshift-kube-storage-version-migrator-operator-kubestorageversionmigratorstaticresources-kubestorageversionmigratorstaticresources-staticresources

kube-storage-version-migrator-operator

NamespaceCreated

Created Namespace/openshift-kube-storage-version-migrator because it was missing

openshift-kube-storage-version-migrator-operator

openshift-kube-storage-version-migrator-operator-status-controller-statussyncer_kube-storage-version-migrator

kube-storage-version-migrator-operator

OperatorStatusChanged

Status for clusteroperator/kube-storage-version-migrator changed: Upgradeable changed from Unknown to True ("All is well")

openshift-kube-storage-version-migrator-operator

openshift-kube-storage-version-migrator-operator-kubestorageversionmigratorstaticresources-kubestorageversionmigratorstaticresources-staticresources

kube-storage-version-migrator-operator

ServiceAccountCreated

Created ServiceAccount/kube-storage-version-migrator-sa -n openshift-kube-storage-version-migrator because it was missing

openshift-kube-storage-version-migrator-operator

openshift-kube-storage-version-migrator-operator-kubestorageversionmigratorstaticresources-kubestorageversionmigratorstaticresources-staticresources

kube-storage-version-migrator-operator

ClusterRoleBindingCreated

Created ClusterRoleBinding.rbac.authorization.k8s.io/storage-version-migration-migrator because it was missing

openshift-cluster-storage-operator

default-scheduler

csi-snapshot-controller-64854d9cff-f44gr

Scheduled

Successfully assigned openshift-cluster-storage-operator/csi-snapshot-controller-64854d9cff-f44gr to master-0

openshift-kube-storage-version-migrator-operator

openshift-kube-storage-version-migrator-operator-kubestorageversionmigrator-deployment-controller--kubestorageversionmigrator

kube-storage-version-migrator-operator

DeploymentCreated

Created Deployment.apps/migrator -n openshift-kube-storage-version-migrator because it was missing

openshift-service-ca-operator

service-ca-operator

service-ca-operator

RoleBindingCreated

Created RoleBinding.rbac.authorization.k8s.io/system:openshift:controller:service-ca -n openshift-service-ca because it was missing

openshift-service-ca-operator

service-ca-operator

service-ca-operator

RoleCreated

Created Role.rbac.authorization.k8s.io/system:openshift:controller:service-ca -n openshift-service-ca because it was missing

openshift-service-ca-operator

service-ca-operator

service-ca-operator

ClusterRoleBindingCreated

Created ClusterRoleBinding.rbac.authorization.k8s.io/system:openshift:controller:service-ca because it was missing

openshift-service-ca-operator

service-ca-operator

service-ca-operator

ClusterRoleCreated

Created ClusterRole.rbac.authorization.k8s.io/system:openshift:controller:service-ca because it was missing

openshift-service-ca-operator

service-ca-operator

service-ca-operator

NamespaceCreated

Created Namespace/openshift-service-ca because it was missing

openshift-kube-storage-version-migrator-operator

openshift-kube-storage-version-migrator-operator-status-controller-statussyncer_kube-storage-version-migrator

kube-storage-version-migrator-operator

OperatorStatusChanged

Status for clusteroperator/kube-storage-version-migrator changed: Degraded changed from Unknown to False ("All is well")

openshift-kube-storage-version-migrator-operator

openshift-kube-storage-version-migrator-operator-status-controller-statussyncer_kube-storage-version-migrator

kube-storage-version-migrator-operator

OperatorStatusChanged

Status for clusteroperator/kube-storage-version-migrator changed: Progressing changed from Unknown to True ("KubeStorageVersionMigratorProgressing: Waiting for Deployment to act on changes"),Available changed from Unknown to False ("KubeStorageVersionMigratorAvailable: Waiting for Deployment")

openshift-kube-storage-version-migrator-operator

openshift-kube-storage-version-migrator-operator-status-controller-statussyncer_kube-storage-version-migrator

kube-storage-version-migrator-operator

OperatorStatusChanged

Status for clusteroperator/kube-storage-version-migrator changed: Progressing message changed from "KubeStorageVersionMigratorProgressing: Waiting for Deployment to act on changes" to "KubeStorageVersionMigratorProgressing: Waiting for Deployment to deploy pods"

openshift-service-ca-operator

service-ca-operator-status-controller-statussyncer_service-ca

service-ca-operator

OperatorStatusChanged

Status for clusteroperator/service-ca changed: Degraded changed from Unknown to False ("All is well")

openshift-service-ca-operator

service-ca-operator-status-controller-statussyncer_service-ca

service-ca-operator

OperatorStatusChanged

Status for clusteroperator/service-ca changed: Degraded set to Unknown (""),Progressing set to Unknown (""),Available set to Unknown (""),Upgradeable set to Unknown (""),EvaluationConditionsDetected set to Unknown (""),status.relatedObjects changed from [] to [{"operator.openshift.io" "servicecas" "" "cluster"} {"" "namespaces" "" "openshift-config"} {"" "namespaces" "" "openshift-config-managed"} {"" "namespaces" "" "openshift-service-ca-operator"} {"" "namespaces" "" "openshift-service-ca"}]

openshift-kube-storage-version-migrator

deployment-controller

migrator

ScalingReplicaSet

Scaled up replica set migrator-8487694857 to 1

openshift-controller-manager

deployment-controller

controller-manager

ScalingReplicaSet

Scaled up replica set controller-manager-797c5b47d8 to 1

openshift-kube-storage-version-migrator

replicaset-controller

migrator-8487694857

SuccessfulCreate

Created pod: migrator-8487694857-w5tlr

openshift-kube-storage-version-migrator

kubelet

migrator-8487694857-w5tlr

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:951ecfeba9b2da4b653034d09275f925396a79c2d8461b8a7c71c776fee67ba0"

openshift-kube-storage-version-migrator

multus

migrator-8487694857-w5tlr

AddedInterface

Add eth0 [10.128.0.30/23] from ovn-kubernetes

openshift-kube-apiserver-operator

kube-apiserver-operator-kubeapiserverstaticresources-kubeapiserverstaticresources-staticresources

kube-apiserver-operator

ServiceCreated

Created Service/apiserver -n openshift-kube-apiserver because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-status-controller-statussyncer_kube-apiserver

kube-apiserver-operator

OperatorStatusChanged

Status for clusteroperator/kube-apiserver changed: Degraded message changed from "NodeControllerDegraded: All master nodes are ready\nNodeKubeconfigControllerDegraded: \"secret/node-kubeconfigs\": secret \"node-system-admin-client\" not found\nRevisionControllerDegraded: configmap \"kube-apiserver-pod\" not found\nInstallerControllerDegraded: missing required resources: [configmaps: bound-sa-token-signing-certs-0,config-0,etcd-serving-ca-0,kube-apiserver-audit-policies-0,kube-apiserver-cert-syncer-kubeconfig-0,kube-apiserver-pod-0,kubelet-serving-ca-0,sa-token-signing-certs-0, secrets: etcd-client-0,localhost-recovery-client-token-0,localhost-recovery-serving-certkey-0]\nCertRotation_ExternalLoadBalancerServing_Degraded: configmaps \"loadbalancer-serving-ca\" already exists\nConfigObservationDegraded: error writing updated observed config: Operation cannot be fulfilled on kubeapiservers.operator.openshift.io \"cluster\": the object has been modified; please apply your changes to the latest version and try again" to "NodeControllerDegraded: All master nodes are ready\nNodeKubeconfigControllerDegraded: \"secret/node-kubeconfigs\": secret \"node-system-admin-client\" not found\nRevisionControllerDegraded: configmap \"kube-apiserver-pod\" not found\nInstallerControllerDegraded: missing required resources: [configmaps: bound-sa-token-signing-certs-0,config-0,etcd-serving-ca-0,kube-apiserver-audit-policies-0,kube-apiserver-cert-syncer-kubeconfig-0,kube-apiserver-pod-0,kubelet-serving-ca-0,sa-token-signing-certs-0, secrets: etcd-client-0,localhost-recovery-client-token-0,localhost-recovery-serving-certkey-0]\nConfigObservationDegraded: error writing updated observed config: Operation cannot be fulfilled on kubeapiservers.operator.openshift.io \"cluster\": the object has been modified; please apply your changes to the latest version and try again"

openshift-kube-apiserver-operator

kube-apiserver-operator-cert-rotation-controller

kube-apiserver-operator

SecretCreated

Created Secret/aggregator-client -n openshift-kube-apiserver because it was missing

openshift-kube-storage-version-migrator

default-scheduler

migrator-8487694857-w5tlr

Scheduled

Successfully assigned openshift-kube-storage-version-migrator/migrator-8487694857-w5tlr to master-0
(x2)

openshift-kube-controller-manager-operator

kube-controller-manager-operator-kube-controller-manager-node

kube-controller-manager-operator

MasterNodesReadyChanged

All master nodes are ready

openshift-kube-controller-manager-operator

kube-controller-manager-operator-status-controller-statussyncer_kube-controller-manager

kube-controller-manager-operator

OperatorStatusChanged

Status for clusteroperator/kube-controller-manager changed: Degraded set to Unknown (""),Progressing set to Unknown (""),Available set to Unknown (""),Upgradeable set to Unknown (""),EvaluationConditionsDetected set to Unknown (""),status.relatedObjects changed from [{"operator.openshift.io" "kubecontrollermanagers" "" "cluster"} {"" "namespaces" "" "openshift-config"} {"" "namespaces" "" "openshift-config-managed"} {"" "namespaces" "" "openshift-kube-controller-manager"} {"" "namespaces" "" "openshift-kube-controller-manager-operator"} {"" "namespaces" "" "kube-system"} {"" "nodes" "" ""} {"certificates.k8s.io" "certificatesigningrequests" "" ""}] to [{"operator.openshift.io" "kubecontrollermanagers" "" "cluster"} {"" "namespaces" "" "openshift-config"} {"" "namespaces" "" "openshift-config-managed"} {"" "namespaces" "" "openshift-kube-controller-manager"} {"" "namespaces" "" "openshift-kube-controller-manager-operator"} {"" "namespaces" "" "kube-system"} {"certificates.k8s.io" "certificatesigningrequests" "" ""} {"" "nodes" "" ""} {"config.openshift.io" "nodes" "" "cluster"}],status.versions changed from [] to [{"raw-internal" "4.18.35"}]
(x2)

openshift-kube-controller-manager-operator

kube-controller-manager-operator-status-controller-statussyncer_kube-controller-manager

kube-controller-manager-operator

OperatorVersionChanged

clusteroperator/kube-controller-manager version "raw-internal" changed from "" to "4.18.35"

openshift-kube-controller-manager-operator

kube-controller-manager-operator

kube-controller-manager-operator

CABundleUpdateRequired

"csr-controller-signer-ca" in "openshift-kube-controller-manager-operator" requires a new cert: configmap doesn't exist
(x2)

openshift-kube-controller-manager-operator

kube-controller-manager-operator-kube-controller-manager-node

kube-controller-manager-operator

MasterNodeObserved

Observed new master node master-0

openshift-kube-controller-manager-operator

kube-controller-manager-operator

kube-controller-manager-operator

FeatureGatesInitialized

FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{"AWSEFSDriverVolumeMetrics", "AdminNetworkPolicy", "AlibabaPlatform", "AzureWorkloadIdentity", "BareMetalLoadBalancer", "BuildCSIVolumes", "ChunkSizeMiB", "CloudDualStackNodeIPs", "DisableKubeletCloudCredentialProviders", "GCPLabelsTags", "HardwareSpeed", "IngressControllerLBSubnetsAWS", "KMSv1", "ManagedBootImages", "ManagedBootImagesAWS", "MultiArchInstallAWS", "MultiArchInstallGCP", "NetworkDiagnosticsConfig", "NetworkLiveMigration", "NetworkSegmentation", "NewOLM", "NodeDisruptionPolicy", "OnClusterBuild", "PersistentIPsForVirtualization", "PrivateHostedZoneAWS", "SetEIPForNLBIngressController", "StreamingCollectionEncodingToJSON", "StreamingCollectionEncodingToProtobuf", "VSphereControlPlaneMachineSet", "VSphereDriverConfiguration", "VSphereMultiVCenters", "VSphereStaticIPs", "ValidatingAdmissionPolicy"}, Disabled:[]v1.FeatureGateName{"AWSClusterHostedDNS", "AdditionalRoutingCapabilities", "AutomatedEtcdBackup", "BootcNodeManagement", "CSIDriverSharedResource", "ClusterAPIInstall", "ClusterAPIInstallIBMCloud", "ClusterMonitoringConfig", "ConsolePluginContentSecurityPolicy", "DNSNameResolver", "DynamicResourceAllocation", "EtcdBackendQuota", "EventedPLEG", "Example", "ExternalOIDC", "ExternalOIDCWithUIDAndExtraClaimMappings", "GCPClusterHostedDNS", "GatewayAPI", "ImageStreamImportMode", "IngressControllerDynamicConfigurationManager", "InsightsConfig", "InsightsConfigAPI", "InsightsOnDemandDataGather", "InsightsRuntimeExtractor", "MachineAPIMigration", "MachineAPIOperatorDisableMachineHealthCheckController", "MachineAPIProviderOpenStack", "MachineConfigNodes", "MaxUnavailableStatefulSet", "MetricsCollectionProfiles", "MinimumKubeletVersion", "MixedCPUsAllocation", "MultiArchInstallAzure", "NodeSwap", "NutanixMultiSubnets", "OVNObservability", "OpenShiftPodSecurityAdmission", "PinnedImages", "PlatformOperators", "ProcMountType", "RouteAdvertisements", "RouteExternalCertificate", "ServiceAccountTokenNodeBinding", "SignatureStores", "SigstoreImageVerification", "TranslateStreamCloseWebsocketRequests", "UpgradeStatus", "UserNamespacesPodSecurityStandards", "UserNamespacesSupport", "VSphereMultiNetworks", "VolumeAttributesClass", "VolumeGroupSnapshot"}}

openshift-kube-controller-manager-operator

kube-controller-manager-operator

kube-controller-manager-operator-lock

LeaderElection

kube-controller-manager-operator-ff989d6cc-qcpb4_2978297d-0752-40c0-8095-d515a08a5ada became leader

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Degraded message changed from "All is well" to "OAuthServiceDegraded: Unable to get oauth server service: service \"oauth-openshift\" not found"

openshift-authentication-operator

oauth-apiserver-audit-policy-controller-auditpolicycontroller

authentication-operator

FastControllerResync

Controller "auditPolicyController" resync interval is set to 10s which might lead to client request throttling

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorVersionChanged

clusteroperator/authentication version "operator" changed from "" to "4.18.35"

openshift-kube-apiserver-operator

kube-apiserver-operator-cert-rotation-controller

kube-apiserver-operator

SecretCreated

Created Secret/kube-controller-manager-client-cert-key -n openshift-config-managed because it was missing

openshift-controller-manager-operator

openshift-controller-manager-operator-openshiftcontrollermanagerstaticresources-openshiftcontrollermanagerstaticresources-staticresources

openshift-controller-manager-operator

ClusterRoleCreated

Created ClusterRole.rbac.authorization.k8s.io/system:openshift:openshift-controller-manager:ingress-to-route-controller because it was missing

openshift-controller-manager-operator

openshift-controller-manager-operator

openshift-controller-manager-operator

DeploymentCreated

Created Deployment.apps/route-controller-manager -n openshift-route-controller-manager because it was missing

openshift-controller-manager-operator

openshift-controller-manager-operator-openshiftcontrollermanagerstaticresources-openshiftcontrollermanagerstaticresources-staticresources

openshift-controller-manager-operator

ClusterRoleBindingCreated

Created ClusterRoleBinding.rbac.authorization.k8s.io/system:openshift:openshift-controller-manager:image-trigger-controller because it was missing

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Degraded message changed from "OAuthServiceDegraded: Unable to get oauth server service: service \"oauth-openshift\" not found" to "OAuthServiceDegraded: Unable to get oauth server service: service \"oauth-openshift\" not found\nRevisionControllerDegraded: configmap \"audit\" not found"

openshift-controller-manager-operator

openshift-controller-manager-operator-openshiftcontrollermanagerstaticresources-openshiftcontrollermanagerstaticresources-staticresources

openshift-controller-manager-operator

ClusterRoleCreated

Created ClusterRole.rbac.authorization.k8s.io/system:openshift:openshift-controller-manager:image-trigger-controller because it was missing

openshift-controller-manager-operator

openshift-controller-manager-operator-openshiftcontrollermanagerstaticresources-openshiftcontrollermanagerstaticresources-staticresources

openshift-controller-manager-operator

ClusterRoleBindingCreated

Created ClusterRoleBinding.rbac.authorization.k8s.io/system:deployer because it was missing

openshift-route-controller-manager

replicaset-controller

route-controller-manager-6cd6978d68

SuccessfulCreate

Created pod: route-controller-manager-6cd6978d68-4chhh

openshift-controller-manager

default-scheduler

controller-manager-797c5b47d8-xrgq6

Scheduled

Successfully assigned openshift-controller-manager/controller-manager-797c5b47d8-xrgq6 to master-0

openshift-controller-manager-operator

openshift-controller-manager-operator-openshiftcontrollermanagerstaticresources-openshiftcontrollermanagerstaticresources-staticresources

openshift-controller-manager-operator

ClusterRoleCreated

Created ClusterRole.rbac.authorization.k8s.io/system:deployer because it was missing

openshift-controller-manager-operator

openshift-controller-manager-operator-openshiftcontrollermanagerstaticresources-openshiftcontrollermanagerstaticresources-staticresources

openshift-controller-manager-operator

ClusterRoleBindingCreated

Created ClusterRoleBinding.rbac.authorization.k8s.io/system:openshift:openshift-controller-manager:update-buildconfig-status because it was missing

openshift-controller-manager-operator

openshift-controller-manager-operator-openshiftcontrollermanagerstaticresources-openshiftcontrollermanagerstaticresources-staticresources

openshift-controller-manager-operator

ClusterRoleCreated

Created ClusterRole.rbac.authorization.k8s.io/system:openshift:openshift-controller-manager:update-buildconfig-status because it was missing

openshift-controller-manager-operator

openshift-controller-manager-operator-openshiftcontrollermanagerstaticresources-openshiftcontrollermanagerstaticresources-staticresources

openshift-controller-manager-operator

RoleBindingCreated

Created RoleBinding.rbac.authorization.k8s.io/prometheus-k8s -n openshift-controller-manager because it was missing

openshift-controller-manager-operator

openshift-controller-manager-operator-openshiftcontrollermanagerstaticresources-openshiftcontrollermanagerstaticresources-staticresources

openshift-controller-manager-operator

RoleCreated

Created Role.rbac.authorization.k8s.io/prometheus-k8s -n openshift-controller-manager because it was missing

openshift-controller-manager-operator

openshift-controller-manager-operator-openshiftcontrollermanagerstaticresources-openshiftcontrollermanagerstaticresources-staticresources

openshift-controller-manager-operator

ServiceCreated

Created Service/controller-manager -n openshift-controller-manager because it was missing

openshift-controller-manager-operator

openshift-controller-manager-operator-openshiftcontrollermanagerstaticresources-openshiftcontrollermanagerstaticresources-staticresources

openshift-controller-manager-operator

ServiceAccountCreated

Created ServiceAccount/openshift-controller-manager-sa -n openshift-controller-manager because it was missing

openshift-controller-manager-operator

openshift-controller-manager-operator-openshiftcontrollermanagerstaticresources-openshiftcontrollermanagerstaticresources-staticresources

openshift-controller-manager-operator

RoleBindingCreated

Created RoleBinding.rbac.authorization.k8s.io/system:openshift:sa-creating-openshift-controller-manager -n openshift-infra because it was missing

openshift-controller-manager-operator

openshift-controller-manager-operator-openshiftcontrollermanagerstaticresources-openshiftcontrollermanagerstaticresources-staticresources

openshift-controller-manager-operator

RoleCreated

Created Role.rbac.authorization.k8s.io/system:openshift:sa-creating-openshift-controller-manager -n openshift-infra because it was missing

openshift-controller-manager-operator

openshift-controller-manager-operator-openshiftcontrollermanagerstaticresources-openshiftcontrollermanagerstaticresources-staticresources

openshift-controller-manager-operator

RoleBindingCreated

Created RoleBinding.rbac.authorization.k8s.io/system:openshift:leader-locking-openshift-controller-manager -n kube-system because it was missing

openshift-controller-manager-operator

openshift-controller-manager-operator-openshiftcontrollermanagerstaticresources-openshiftcontrollermanagerstaticresources-staticresources

openshift-controller-manager-operator

RoleCreated

Created Role.rbac.authorization.k8s.io/system:openshift:leader-locking-openshift-controller-manager -n kube-system because it was missing

openshift-controller-manager-operator

openshift-controller-manager-operator-openshiftcontrollermanagerstaticresources-openshiftcontrollermanagerstaticresources-staticresources

openshift-controller-manager-operator

ClusterRoleBindingCreated

Created ClusterRoleBinding.rbac.authorization.k8s.io/system:openshift:openshift-controller-manager:ingress-to-route-controller because it was missing

openshift-kube-controller-manager-operator

kube-controller-manager-operator-status-controller-statussyncer_kube-controller-manager

kube-controller-manager-operator

OperatorStatusChanged

Status for clusteroperator/kube-controller-manager changed: Progressing changed from Unknown to False ("All is well")

openshift-controller-manager-operator

openshift-controller-manager-operator-openshiftcontrollermanagerstaticresources-openshiftcontrollermanagerstaticresources-staticresources

openshift-controller-manager-operator

ClusterRoleBindingCreated

Created ClusterRoleBinding.rbac.authorization.k8s.io/system:openshift:tokenreview-openshift-route-controller-manager because it was missing

openshift-controller-manager-operator

openshift-controller-manager-operator-openshiftcontrollermanagerstaticresources-openshiftcontrollermanagerstaticresources-staticresources

openshift-controller-manager-operator

ClusterRoleCreated

Created ClusterRole.rbac.authorization.k8s.io/system:openshift:tokenreview-openshift-route-controller-manager because it was missing

openshift-controller-manager-operator

openshift-controller-manager-operator-openshiftcontrollermanagerstaticresources-openshiftcontrollermanagerstaticresources-staticresources

openshift-controller-manager-operator

ServiceCreated

Created Service/route-controller-manager -n openshift-route-controller-manager because it was missing

openshift-controller-manager-operator

openshift-controller-manager-operator-openshiftcontrollermanagerstaticresources-openshiftcontrollermanagerstaticresources-staticresources

openshift-controller-manager-operator

RoleBindingCreated

Created RoleBinding.rbac.authorization.k8s.io/prometheus-k8s -n openshift-route-controller-manager because it was missing
(x7)

openshift-controller-manager

replicaset-controller

controller-manager-797c5b47d8

FailedCreate

Error creating: pods "controller-manager-797c5b47d8-" is forbidden: error looking up service account openshift-controller-manager/openshift-controller-manager-sa: serviceaccount "openshift-controller-manager-sa" not found

openshift-controller-manager

replicaset-controller

controller-manager-797c5b47d8

SuccessfulCreate

Created pod: controller-manager-797c5b47d8-xrgq6

openshift-route-controller-manager

deployment-controller

route-controller-manager

ScalingReplicaSet

Scaled up replica set route-controller-manager-6cd6978d68 to 1

openshift-service-ca-operator

service-ca-operator

service-ca-operator

ServiceAccountCreated

Created ServiceAccount/service-ca -n openshift-service-ca because it was missing

openshift-apiserver-operator

openshift-apiserver-operator

openshift-apiserver-operator-lock

LeaderElection

openshift-apiserver-operator-d65958b8-th2vj_fba48b0d-bec6-403d-8e66-13a019ff654c became leader

openshift-apiserver-operator

openshift-apiserver-operator

openshift-apiserver-operator

FeatureGatesInitialized

FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{"AWSEFSDriverVolumeMetrics", "AdminNetworkPolicy", "AlibabaPlatform", "AzureWorkloadIdentity", "BareMetalLoadBalancer", "BuildCSIVolumes", "ChunkSizeMiB", "CloudDualStackNodeIPs", "DisableKubeletCloudCredentialProviders", "GCPLabelsTags", "HardwareSpeed", "IngressControllerLBSubnetsAWS", "KMSv1", "ManagedBootImages", "ManagedBootImagesAWS", "MultiArchInstallAWS", "MultiArchInstallGCP", "NetworkDiagnosticsConfig", "NetworkLiveMigration", "NetworkSegmentation", "NewOLM", "NodeDisruptionPolicy", "OnClusterBuild", "PersistentIPsForVirtualization", "PrivateHostedZoneAWS", "SetEIPForNLBIngressController", "StreamingCollectionEncodingToJSON", "StreamingCollectionEncodingToProtobuf", "VSphereControlPlaneMachineSet", "VSphereDriverConfiguration", "VSphereMultiVCenters", "VSphereStaticIPs", "ValidatingAdmissionPolicy"}, Disabled:[]v1.FeatureGateName{"AWSClusterHostedDNS", "AdditionalRoutingCapabilities", "AutomatedEtcdBackup", "BootcNodeManagement", "CSIDriverSharedResource", "ClusterAPIInstall", "ClusterAPIInstallIBMCloud", "ClusterMonitoringConfig", "ConsolePluginContentSecurityPolicy", "DNSNameResolver", "DynamicResourceAllocation", "EtcdBackendQuota", "EventedPLEG", "Example", "ExternalOIDC", "ExternalOIDCWithUIDAndExtraClaimMappings", "GCPClusterHostedDNS", "GatewayAPI", "ImageStreamImportMode", "IngressControllerDynamicConfigurationManager", "InsightsConfig", "InsightsConfigAPI", "InsightsOnDemandDataGather", "InsightsRuntimeExtractor", "MachineAPIMigration", "MachineAPIOperatorDisableMachineHealthCheckController", "MachineAPIProviderOpenStack", "MachineConfigNodes", "MaxUnavailableStatefulSet", "MetricsCollectionProfiles", "MinimumKubeletVersion", "MixedCPUsAllocation", "MultiArchInstallAzure", "NodeSwap", "NutanixMultiSubnets", "OVNObservability", "OpenShiftPodSecurityAdmission", "PinnedImages", "PlatformOperators", "ProcMountType", "RouteAdvertisements", "RouteExternalCertificate", "ServiceAccountTokenNodeBinding", "SignatureStores", "SigstoreImageVerification", "TranslateStreamCloseWebsocketRequests", "UpgradeStatus", "UserNamespacesPodSecurityStandards", "UserNamespacesSupport", "VSphereMultiNetworks", "VolumeAttributesClass", "VolumeGroupSnapshot"}}

openshift-apiserver-operator

openshift-apiserver-operator-status-controller-statussyncer_openshift-apiserver

openshift-apiserver-operator

OperatorVersionChanged

clusteroperator/openshift-apiserver version "operator" changed from "" to "4.18.35"

openshift-service-ca-operator

service-ca-operator

service-ca-operator

SecretCreated

Created Secret/signing-key -n openshift-service-ca because it was missing

openshift-kube-controller-manager-operator

kube-controller-manager-operator-status-controller-statussyncer_kube-controller-manager

kube-controller-manager-operator

OperatorStatusChanged

Status for clusteroperator/kube-controller-manager changed: Degraded message changed from "NodeControllerDegraded: All master nodes are ready\nRevisionControllerDegraded: configmap \"kube-controller-manager-pod\" not found" to "NodeControllerDegraded: All master nodes are ready\nRevisionControllerDegraded: configmap \"kube-controller-manager-pod\" not found\nInstallerControllerDegraded: missing required resources: [configmaps: cluster-policy-controller-config-0,config-0,controller-manager-kubeconfig-0,kube-controller-cert-syncer-kubeconfig-0,kube-controller-manager-pod-0,recycler-config-0,service-ca-0,serviceaccount-ca-0, secrets: localhost-recovery-client-token-0,service-account-private-key-0]",Progressing message changed from "All is well" to "NodeInstallerProgressing: 1 node is at revision 0",Available changed from Unknown to False ("StaticPodsAvailable: 0 nodes are active; 1 node is at revision 0")

openshift-route-controller-manager

default-scheduler

route-controller-manager-6cd6978d68-4chhh

Scheduled

Successfully assigned openshift-route-controller-manager/route-controller-manager-6cd6978d68-4chhh to master-0

openshift-kube-controller-manager-operator

kube-controller-manager-operator-status-controller-statussyncer_kube-controller-manager

kube-controller-manager-operator

OperatorStatusChanged

Status for clusteroperator/kube-controller-manager changed: Degraded message changed from "NodeControllerDegraded: All master nodes are ready" to "NodeControllerDegraded: All master nodes are ready\nRevisionControllerDegraded: configmap \"kube-controller-manager-pod\" not found"

openshift-kube-controller-manager-operator

kube-controller-manager-operator-status-controller-statussyncer_kube-controller-manager

kube-controller-manager-operator

OperatorStatusChanged

Status for clusteroperator/kube-controller-manager changed: Degraded changed from Unknown to False ("NodeControllerDegraded: All master nodes are ready"),Upgradeable changed from Unknown to True ("All is well")

openshift-kube-apiserver-operator

kube-apiserver-operator-cert-rotation-controller

kube-apiserver-operator

SecretCreated

Created Secret/localhost-recovery-serving-certkey -n openshift-kube-apiserver because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-cert-rotation-controller

kube-apiserver-operator

SecretCreated

Created Secret/control-plane-node-admin-client-cert-key -n openshift-kube-apiserver because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-status-controller-statussyncer_kube-apiserver

kube-apiserver-operator

OperatorStatusChanged

Status for clusteroperator/kube-apiserver changed: Degraded message changed from "NodeControllerDegraded: All master nodes are ready\nNodeKubeconfigControllerDegraded: \"secret/node-kubeconfigs\": secret \"node-system-admin-client\" not found\nRevisionControllerDegraded: configmap \"kube-apiserver-pod\" not found\nInstallerControllerDegraded: missing required resources: [configmaps: bound-sa-token-signing-certs-0,config-0,etcd-serving-ca-0,kube-apiserver-audit-policies-0,kube-apiserver-cert-syncer-kubeconfig-0,kube-apiserver-pod-0,kubelet-serving-ca-0,sa-token-signing-certs-0, secrets: etcd-client-0,localhost-recovery-client-token-0,localhost-recovery-serving-certkey-0]\nConfigObservationDegraded: error writing updated observed config: Operation cannot be fulfilled on kubeapiservers.operator.openshift.io \"cluster\": the object has been modified; please apply your changes to the latest version and try again" to "NodeControllerDegraded: All master nodes are ready\nNodeKubeconfigControllerDegraded: \"secret/node-kubeconfigs\": configmap \"kube-apiserver-server-ca\" not found\nRevisionControllerDegraded: configmap \"kube-apiserver-pod\" not found\nInstallerControllerDegraded: missing required resources: [configmaps: bound-sa-token-signing-certs-0,config-0,etcd-serving-ca-0,kube-apiserver-audit-policies-0,kube-apiserver-cert-syncer-kubeconfig-0,kube-apiserver-pod-0,kubelet-serving-ca-0,sa-token-signing-certs-0, secrets: etcd-client-0,localhost-recovery-client-token-0,localhost-recovery-serving-certkey-0]\nConfigObservationDegraded: error writing updated observed config: Operation cannot be fulfilled on kubeapiservers.operator.openshift.io \"cluster\": the object has been modified; please apply your changes to the latest version and try again"

openshift-kube-apiserver-operator

kube-apiserver-operator-cert-rotation-controller

kube-apiserver-operator

SecretCreated

Created Secret/kubelet-client -n openshift-kube-apiserver because it was missing

openshift-cluster-storage-operator

kubelet

csi-snapshot-controller-64854d9cff-f44gr

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9609c00207cc4db97f0fd6162eb429d7f81654137f020a677e30cba26a887a24"

openshift-cluster-storage-operator

multus

csi-snapshot-controller-64854d9cff-f44gr

AddedInterface

Add eth0 [10.128.0.31/23] from ovn-kubernetes

openshift-apiserver-operator

openshift-apiserver-operator-status-controller-statussyncer_openshift-apiserver

openshift-apiserver-operator

OperatorStatusChanged

Status for clusteroperator/openshift-apiserver changed: Upgradeable changed from Unknown to True ("All is well")

openshift-service-ca-operator

service-ca-operator-status-controller-statussyncer_service-ca

service-ca-operator

OperatorStatusChanged

Status for clusteroperator/service-ca changed: Progressing changed from Unknown to True ("Progressing: \nProgressing: service-ca does not have available replicas"),Available changed from Unknown to True ("All is well"),Upgradeable changed from Unknown to True ("All is well")

openshift-service-ca-operator

service-ca-operator

service-ca-operator

DeploymentCreated

Created Deployment.apps/service-ca -n openshift-service-ca because it was missing

openshift-service-ca-operator

service-ca-operator

service-ca-operator

ConfigMapCreated

Created ConfigMap/signing-cabundle -n openshift-service-ca because it was missing

openshift-service-ca-operator

service-ca-operator-resource-sync-controller-resourcesynccontroller

service-ca-operator

ConfigMapCreated

Created ConfigMap/service-ca -n openshift-config-managed because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-kubeapiserverstaticresources-kubeapiserverstaticresources-staticresources

kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/kube-apiserver-cert-syncer-kubeconfig -n openshift-kube-apiserver because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-kubeapiserverstaticresources-kubeapiserverstaticresources-staticresources

kube-apiserver-operator

ClusterRoleCreated

Created ClusterRole.rbac.authorization.k8s.io/system:openshift:controller:check-endpoints because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-kubeapiserverstaticresources-kubeapiserverstaticresources-staticresources

kube-apiserver-operator

ClusterRoleCreated

Created ClusterRole.rbac.authorization.k8s.io/system:openshift:controller:check-endpoints-node-reader because it was missing

openshift-kube-controller-manager-operator

kube-controller-manager-operator-config-observer-configobserver

kube-controller-manager-operator

ObserveFeatureFlagsUpdated

Updated extendedArguments.feature-gates to AWSEFSDriverVolumeMetrics=true,AdminNetworkPolicy=true,AlibabaPlatform=true,AzureWorkloadIdentity=true,BareMetalLoadBalancer=true,BuildCSIVolumes=true,ChunkSizeMiB=true,CloudDualStackNodeIPs=true,DisableKubeletCloudCredentialProviders=true,GCPLabelsTags=true,HardwareSpeed=true,IngressControllerLBSubnetsAWS=true,KMSv1=true,ManagedBootImages=true,ManagedBootImagesAWS=true,MultiArchInstallAWS=true,MultiArchInstallGCP=true,NetworkDiagnosticsConfig=true,NetworkLiveMigration=true,NetworkSegmentation=true,NewOLM=true,NodeDisruptionPolicy=true,OnClusterBuild=true,PersistentIPsForVirtualization=true,PrivateHostedZoneAWS=true,SetEIPForNLBIngressController=true,StreamingCollectionEncodingToJSON=true,StreamingCollectionEncodingToProtobuf=true,VSphereControlPlaneMachineSet=true,VSphereDriverConfiguration=true,VSphereMultiVCenters=true,VSphereStaticIPs=true,ValidatingAdmissionPolicy=true,AWSClusterHostedDNS=false,AdditionalRoutingCapabilities=false,AutomatedEtcdBackup=false,BootcNodeManagement=false,CSIDriverSharedResource=false,ClusterAPIInstall=false,ClusterAPIInstallIBMCloud=false,ClusterMonitoringConfig=false,ConsolePluginContentSecurityPolicy=false,DNSNameResolver=false,DynamicResourceAllocation=false,EtcdBackendQuota=false,EventedPLEG=false,Example=false,ExternalOIDC=false,ExternalOIDCWithUIDAndExtraClaimMappings=false,GCPClusterHostedDNS=false,GatewayAPI=false,ImageStreamImportMode=false,IngressControllerDynamicConfigurationManager=false,InsightsConfig=false,InsightsConfigAPI=false,InsightsOnDemandDataGather=false,InsightsRuntimeExtractor=false,MachineAPIMigration=false,MachineAPIOperatorDisableMachineHealthCheckController=false,MachineAPIProviderOpenStack=false,MachineConfigNodes=false,MaxUnavailableStatefulSet=false,MetricsCollectionProfiles=false,MinimumKubeletVersion=false,MixedCPUsAllocation=false,MultiArchInstallAzure=false,NodeSwap=false,NutanixMultiSubnets=false,OVNObservability=false,OpenShiftPodSecurityAdmission=false,PinnedImages=false,PlatformOperators=false,ProcMountType=false,RouteAdvertisements=false,RouteExternalCertificate=false,ServiceAccountTokenNodeBinding=false,SignatureStores=false,SigstoreImageVerification=false,TranslateStreamCloseWebsocketRequests=false,UpgradeStatus=false,UserNamespacesPodSecurityStandards=false,UserNamespacesSupport=false,VSphereMultiNetworks=false,VolumeAttributesClass=false,VolumeGroupSnapshot=false

openshift-kube-controller-manager-operator

kube-controller-manager-operator-config-observer-configobserver

kube-controller-manager-operator

ObserveTLSSecurityProfile

minTLSVersion changed to VersionTLS12

openshift-kube-controller-manager-operator

kube-controller-manager-operator-config-observer-configobserver

kube-controller-manager-operator

ObserveTLSSecurityProfile

cipherSuites changed to ["TLS_AES_128_GCM_SHA256" "TLS_AES_256_GCM_SHA384" "TLS_CHACHA20_POLY1305_SHA256" "TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256" "TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256" "TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384" "TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384" "TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256" "TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256"]

openshift-kube-controller-manager-operator

kube-controller-manager-operator-config-observer-configobserver

kube-controller-manager-operator

ObserveFeatureFlagsUpdated

Updated featureGates to AWSEFSDriverVolumeMetrics=true,AdminNetworkPolicy=true,AlibabaPlatform=true,AzureWorkloadIdentity=true,BareMetalLoadBalancer=true,BuildCSIVolumes=true,ChunkSizeMiB=true,CloudDualStackNodeIPs=true,DisableKubeletCloudCredentialProviders=true,GCPLabelsTags=true,HardwareSpeed=true,IngressControllerLBSubnetsAWS=true,KMSv1=true,ManagedBootImages=true,ManagedBootImagesAWS=true,MultiArchInstallAWS=true,MultiArchInstallGCP=true,NetworkDiagnosticsConfig=true,NetworkLiveMigration=true,NetworkSegmentation=true,NewOLM=true,NodeDisruptionPolicy=true,OnClusterBuild=true,PersistentIPsForVirtualization=true,PrivateHostedZoneAWS=true,SetEIPForNLBIngressController=true,StreamingCollectionEncodingToJSON=true,StreamingCollectionEncodingToProtobuf=true,VSphereControlPlaneMachineSet=true,VSphereDriverConfiguration=true,VSphereMultiVCenters=true,VSphereStaticIPs=true,ValidatingAdmissionPolicy=true,AWSClusterHostedDNS=false,AdditionalRoutingCapabilities=false,AutomatedEtcdBackup=false,BootcNodeManagement=false,CSIDriverSharedResource=false,ClusterAPIInstall=false,ClusterAPIInstallIBMCloud=false,ClusterMonitoringConfig=false,ConsolePluginContentSecurityPolicy=false,DNSNameResolver=false,DynamicResourceAllocation=false,EtcdBackendQuota=false,EventedPLEG=false,Example=false,ExternalOIDC=false,ExternalOIDCWithUIDAndExtraClaimMappings=false,GCPClusterHostedDNS=false,GatewayAPI=false,ImageStreamImportMode=false,IngressControllerDynamicConfigurationManager=false,InsightsConfig=false,InsightsConfigAPI=false,InsightsOnDemandDataGather=false,InsightsRuntimeExtractor=false,MachineAPIMigration=false,MachineAPIOperatorDisableMachineHealthCheckController=false,MachineAPIProviderOpenStack=false,MachineConfigNodes=false,MaxUnavailableStatefulSet=false,MetricsCollectionProfiles=false,MinimumKubeletVersion=false,MixedCPUsAllocation=false,MultiArchInstallAzure=false,NodeSwap=false,NutanixMultiSubnets=false,OVNObservability=false,OpenShiftPodSecurityAdmission=false,PinnedImages=false,PlatformOperators=false,ProcMountType=false,RouteAdvertisements=false,RouteExternalCertificate=false,ServiceAccountTokenNodeBinding=false,SignatureStores=false,SigstoreImageVerification=false,TranslateStreamCloseWebsocketRequests=false,UpgradeStatus=false,UserNamespacesPodSecurityStandards=false,UserNamespacesSupport=false,VSphereMultiNetworks=false,VolumeAttributesClass=false,VolumeGroupSnapshot=false

openshift-kube-controller-manager-operator

kube-controller-manager-operator-config-observer-configobserver

kube-controller-manager-operator

ObservedConfigChanged

Writing updated observed config: map[string]any{ + "extendedArguments": map[string]any{ + "cluster-cidr": []any{string("10.128.0.0/16")}, + "cluster-name": []any{string("sno-bhfwr")}, + "feature-gates": []any{ + string("AWSEFSDriverVolumeMetrics=true"), string("AdminNetworkPolicy=true"), + string("AlibabaPlatform=true"), string("AzureWorkloadIdentity=true"), + string("BareMetalLoadBalancer=true"), string("BuildCSIVolumes=true"), + string("ChunkSizeMiB=true"), string("CloudDualStackNodeIPs=true"), ..., + }, + "service-cluster-ip-range": []any{string("172.30.0.0/16")}, + }, + "featureGates": []any{ + string("AWSEFSDriverVolumeMetrics=true"), string("AdminNetworkPolicy=true"), + string("AlibabaPlatform=true"), string("AzureWorkloadIdentity=true"), + string("BareMetalLoadBalancer=true"), string("BuildCSIVolumes=true"), + string("ChunkSizeMiB=true"), string("CloudDualStackNodeIPs=true"), + string("DisableKubeletCloudCredentialProviders=true"), + string("GCPLabelsTags=true"), string("HardwareSpeed=true"), + string("IngressControllerLBSubnetsAWS=true"), string("KMSv1=true"), + string("ManagedBootImages=true"), string("ManagedBootImagesAWS=true"), + string("MultiArchInstallAWS=true"), ..., + }, + "servingInfo": map[string]any{ + "cipherSuites": []any{ + string("TLS_AES_128_GCM_SHA256"), string("TLS_AES_256_GCM_SHA384"), + string("TLS_CHACHA20_POLY1305_SHA256"), + string("TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256"), + string("TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256"), + string("TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384"), + string("TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384"), + string("TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256"), ..., + }, + "minTLSVersion": string("VersionTLS12"), + }, }

openshift-apiserver-operator

openshift-apiserver-operator-status-controller-statussyncer_openshift-apiserver

openshift-apiserver-operator

OperatorStatusChanged

Status for clusteroperator/openshift-apiserver changed: Degraded set to Unknown (""),Progressing set to Unknown (""),Available set to Unknown (""),Upgradeable set to Unknown (""),EvaluationConditionsDetected set to Unknown (""),status.relatedObjects changed from [] to [{"operator.openshift.io" "openshiftapiservers" "" "cluster"} {"" "namespaces" "" "openshift-config"} {"" "namespaces" "" "openshift-config-managed"} {"" "namespaces" "" "openshift-apiserver-operator"} {"" "namespaces" "" "openshift-apiserver"} {"" "namespaces" "" "openshift-etcd-operator"} {"" "endpoints" "openshift-etcd" "host-etcd-2"} {"controlplane.operator.openshift.io" "podnetworkconnectivitychecks" "openshift-apiserver" ""} {"apiregistration.k8s.io" "apiservices" "" "v1.apps.openshift.io"} {"apiregistration.k8s.io" "apiservices" "" "v1.authorization.openshift.io"} {"apiregistration.k8s.io" "apiservices" "" "v1.build.openshift.io"} {"apiregistration.k8s.io" "apiservices" "" "v1.image.openshift.io"} {"apiregistration.k8s.io" "apiservices" "" "v1.project.openshift.io"} {"apiregistration.k8s.io" "apiservices" "" "v1.quota.openshift.io"} {"apiregistration.k8s.io" "apiservices" "" "v1.route.openshift.io"} {"apiregistration.k8s.io" "apiservices" "" "v1.security.openshift.io"} {"apiregistration.k8s.io" "apiservices" "" "v1.template.openshift.io"}],status.versions changed from [] to [{"operator" "4.18.35"}]

openshift-network-operator

kubelet

iptables-alerter-dd9wv

Created

Created container: iptables-alerter

openshift-service-ca

deployment-controller

service-ca

ScalingReplicaSet

Scaled up replica set service-ca-79bc6b8d76 to 1

openshift-service-ca

replicaset-controller

service-ca-79bc6b8d76

SuccessfulCreate

Created pod: service-ca-79bc6b8d76-72j8t

openshift-kube-apiserver-operator

kube-apiserver-operator-kubeapiserverstaticresources-kubeapiserverstaticresources-staticresources

kube-apiserver-operator

ClusterRoleBindingCreated

Created ClusterRoleBinding.rbac.authorization.k8s.io/system:openshift:controller:kube-apiserver-check-endpoints-auth-delegator because it was missing

openshift-service-ca

default-scheduler

service-ca-79bc6b8d76-72j8t

Scheduled

Successfully assigned openshift-service-ca/service-ca-79bc6b8d76-72j8t to master-0

openshift-network-operator

kubelet

iptables-alerter-dd9wv

Started

Started container iptables-alerter

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Degraded message changed from "OAuthServiceDegraded: Unable to get oauth server service: service \"oauth-openshift\" not found\nRevisionControllerDegraded: configmap \"audit\" not found" to "OAuthServiceDegraded: Unable to get oauth server service: service \"oauth-openshift\" not found\nRevisionControllerDegraded: configmap \"audit\" not found\nWebhookAuthenticatorControllerDegraded: failed to retrieve service openshift-oauth-apiserver/api: service \"api\" not found"

openshift-kube-apiserver-operator

kube-apiserver-operator-kubeapiserverstaticresources-kubeapiserverstaticresources-staticresources

kube-apiserver-operator

ClusterRoleCreated

Created ClusterRole.rbac.authorization.k8s.io/system:openshift:controller:check-endpoints-crd-reader because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-kubeapiserverstaticresources-kubeapiserverstaticresources-staticresources

kube-apiserver-operator

ClusterRoleBindingCreated

Created ClusterRoleBinding.rbac.authorization.k8s.io/system:openshift:controller:kube-apiserver-check-endpoints-crd-reader because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-kubeapiserverstaticresources-kubeapiserverstaticresources-staticresources

kube-apiserver-operator

ClusterRoleBindingCreated

Created ClusterRoleBinding.rbac.authorization.k8s.io/system:openshift:controller:kube-apiserver-check-endpoints-node-reader because it was missing
(x2)

openshift-route-controller-manager

kubelet

route-controller-manager-6cd6978d68-4chhh

FailedMount

MountVolume.SetUp failed for volume "config" : configmap "config" not found
(x5)

openshift-ingress-operator

kubelet

ingress-operator-66b84d69b-gzg9m

FailedMount

MountVolume.SetUp failed for volume "metrics-tls" : secret "metrics-tls" not found

openshift-apiserver-operator

openshift-apiserver-operator-config-observer-configobserver

openshift-apiserver-operator

ObserveStorageUpdated

Updated storage urls to https://192.168.32.10:2379

openshift-route-controller-manager

replicaset-controller

route-controller-manager-5ff6bdb8cf

SuccessfulCreate

Created pod: route-controller-manager-5ff6bdb8cf-b6499

openshift-controller-manager-operator

openshift-controller-manager-operator-openshiftcontrollermanagerstaticresources-openshiftcontrollermanagerstaticresources-staticresources

openshift-controller-manager-operator

RoleBindingCreated

Created RoleBinding.rbac.authorization.k8s.io/system:openshift:leader-locking-openshift-route-controller-manager -n openshift-route-controller-manager because it was missing

openshift-controller-manager-operator

openshift-controller-manager-operator-openshiftcontrollermanagerstaticresources-openshiftcontrollermanagerstaticresources-staticresources

openshift-controller-manager-operator

RoleCreated

Created Role.rbac.authorization.k8s.io/system:openshift:leader-locking-openshift-route-controller-manager -n openshift-route-controller-manager because it was missing
(x3)

openshift-route-controller-manager

kubelet

route-controller-manager-6cd6978d68-4chhh

FailedMount

MountVolume.SetUp failed for volume "client-ca" : configmap "client-ca" not found

openshift-controller-manager-operator

openshift-controller-manager-operator-openshiftcontrollermanagerstaticresources-openshiftcontrollermanagerstaticresources-staticresources

openshift-controller-manager-operator

RoleBindingCreated

Created RoleBinding.rbac.authorization.k8s.io/system:openshift:leader-locking-openshift-controller-manager -n openshift-controller-manager because it was missing
(x3)

openshift-route-controller-manager

kubelet

route-controller-manager-6cd6978d68-4chhh

FailedMount

MountVolume.SetUp failed for volume "serving-cert" : secret "serving-cert" not found

openshift-controller-manager-operator

openshift-controller-manager-operator-openshiftcontrollermanagerstaticresources-openshiftcontrollermanagerstaticresources-staticresources

openshift-controller-manager-operator

RoleCreated

Created Role.rbac.authorization.k8s.io/system:openshift:leader-locking-openshift-controller-manager -n openshift-controller-manager because it was missing

openshift-route-controller-manager

replicaset-controller

route-controller-manager-6cd6978d68

SuccessfulDelete

Deleted pod: route-controller-manager-6cd6978d68-4chhh

openshift-kube-apiserver-operator

kube-apiserver-operator-resource-sync-controller-resourcesynccontroller

kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/aggregator-client-ca -n openshift-kube-apiserver because it was missing
(x5)

openshift-image-registry

kubelet

cluster-image-registry-operator-5549dc66cb-fmhbq

FailedMount

MountVolume.SetUp failed for volume "image-registry-operator-tls" : secret "image-registry-operator-tls" not found

openshift-kube-apiserver-operator

kube-apiserver-operator-cert-rotation-controller-cert-rotation-controller-CheckEndpointsClient-certrotationcontroller

kube-apiserver-operator

RotationError

configmaps "kube-control-plane-signer-ca" already exists
(x5)

openshift-cluster-node-tuning-operator

kubelet

cluster-node-tuning-operator-598fbc5f8f-vxzvg

FailedMount

MountVolume.SetUp failed for volume "apiservice-cert" : secret "performance-addon-operator-webhook-cert" not found
(x3)

openshift-controller-manager

kubelet

controller-manager-797c5b47d8-xrgq6

FailedMount

MountVolume.SetUp failed for volume "client-ca" : configmap "client-ca" not found
(x3)

openshift-controller-manager

kubelet

controller-manager-797c5b47d8-xrgq6

FailedMount

MountVolume.SetUp failed for volume "serving-cert" : secret "serving-cert" not found
(x3)

openshift-controller-manager

kubelet

controller-manager-797c5b47d8-xrgq6

FailedMount

MountVolume.SetUp failed for volume "config" : configmap "config" not found
(x5)

openshift-cluster-node-tuning-operator

kubelet

cluster-node-tuning-operator-598fbc5f8f-vxzvg

FailedMount

MountVolume.SetUp failed for volume "node-tuning-operator-tls" : secret "node-tuning-operator-tls" not found

openshift-kube-controller-manager-operator

kube-controller-manager-operator

kube-controller-manager-operator

ConfigMapCreated

Created ConfigMap/csr-controller-signer-ca -n openshift-kube-controller-manager-operator because it was missing

openshift-controller-manager

replicaset-controller

controller-manager-797c5b47d8

SuccessfulDelete

Deleted pod: controller-manager-797c5b47d8-xrgq6

openshift-kube-apiserver-operator

kube-apiserver-operator-cert-rotation-controller

kube-apiserver-operator

TargetUpdateRequired

"check-endpoints-client-cert-key" in "openshift-kube-apiserver" requires a new target cert/key pair: secret doesn't exist

openshift-route-controller-manager

deployment-controller

route-controller-manager

ScalingReplicaSet

Scaled down replica set route-controller-manager-6cd6978d68 to 0 from 1

openshift-route-controller-manager

deployment-controller

route-controller-manager

ScalingReplicaSet

Scaled up replica set route-controller-manager-5ff6bdb8cf to 1 from 0

openshift-kube-apiserver-operator

kube-apiserver-operator-status-controller-statussyncer_kube-apiserver

kube-apiserver-operator

OperatorStatusChanged

Status for clusteroperator/kube-apiserver changed: Degraded message changed from "NodeControllerDegraded: All master nodes are ready\nNodeKubeconfigControllerDegraded: \"secret/node-kubeconfigs\": configmap \"kube-apiserver-server-ca\" not found\nRevisionControllerDegraded: configmap \"kube-apiserver-pod\" not found\nInstallerControllerDegraded: missing required resources: [configmaps: bound-sa-token-signing-certs-0,config-0,etcd-serving-ca-0,kube-apiserver-audit-policies-0,kube-apiserver-cert-syncer-kubeconfig-0,kube-apiserver-pod-0,kubelet-serving-ca-0,sa-token-signing-certs-0, secrets: etcd-client-0,localhost-recovery-client-token-0,localhost-recovery-serving-certkey-0]\nConfigObservationDegraded: error writing updated observed config: Operation cannot be fulfilled on kubeapiservers.operator.openshift.io \"cluster\": the object has been modified; please apply your changes to the latest version and try again" to "NodeControllerDegraded: All master nodes are ready\nNodeKubeconfigControllerDegraded: \"secret/node-kubeconfigs\": configmap \"kube-apiserver-server-ca\" not found\nRevisionControllerDegraded: configmap \"kube-apiserver-pod\" not found\nInstallerControllerDegraded: missing required resources: [configmaps: bound-sa-token-signing-certs-0,config-0,etcd-serving-ca-0,kube-apiserver-audit-policies-0,kube-apiserver-cert-syncer-kubeconfig-0,kube-apiserver-pod-0,kubelet-serving-ca-0,sa-token-signing-certs-0, secrets: etcd-client-0,localhost-recovery-client-token-0,localhost-recovery-serving-certkey-0]\nCertRotation_CheckEndpointsClient_Degraded: configmaps \"kube-control-plane-signer-ca\" already exists\nConfigObservationDegraded: error writing updated observed config: Operation cannot be fulfilled on kubeapiservers.operator.openshift.io \"cluster\": the object has been modified; please apply your changes to the latest version and try again"
(x5)

openshift-dns-operator

kubelet

dns-operator-9c5679d8f-r6dm8

FailedMount

MountVolume.SetUp failed for volume "metrics-tls" : secret "metrics-tls" not found
(x5)

openshift-cluster-version

kubelet

cluster-version-operator-56d8475767-9gfkg

FailedMount

MountVolume.SetUp failed for volume "serving-cert" : secret "cluster-version-operator-serving-cert" not found

openshift-apiserver-operator

openshift-apiserver-operator-audit-policy-controller-auditpolicycontroller

openshift-apiserver-operator

FastControllerResync

Controller "auditPolicyController" resync interval is set to 10s which might lead to client request throttling

openshift-controller-manager

default-scheduler

controller-manager-f9d74d9cf-ks2l5

FailedScheduling

0/1 nodes are available: 1 node(s) didn't match pod anti-affinity rules. preemption: 0/1 nodes are available: 1 node(s) didn't match pod anti-affinity rules.

openshift-apiserver-operator

openshift-apiserver-operator-config-observer-configobserver

openshift-apiserver-operator

ObserveFeatureFlagsUpdated

Updated apiServerArguments.feature-gates to AWSEFSDriverVolumeMetrics=true,AdminNetworkPolicy=true,AlibabaPlatform=true,AzureWorkloadIdentity=true,BareMetalLoadBalancer=true,BuildCSIVolumes=true,ChunkSizeMiB=true,CloudDualStackNodeIPs=true,DisableKubeletCloudCredentialProviders=true,GCPLabelsTags=true,HardwareSpeed=true,IngressControllerLBSubnetsAWS=true,KMSv1=true,ManagedBootImages=true,ManagedBootImagesAWS=true,MultiArchInstallAWS=true,MultiArchInstallGCP=true,NetworkDiagnosticsConfig=true,NetworkLiveMigration=true,NetworkSegmentation=true,NewOLM=true,NodeDisruptionPolicy=true,OnClusterBuild=true,PersistentIPsForVirtualization=true,PrivateHostedZoneAWS=true,SetEIPForNLBIngressController=true,StreamingCollectionEncodingToJSON=true,StreamingCollectionEncodingToProtobuf=true,VSphereControlPlaneMachineSet=true,VSphereDriverConfiguration=true,VSphereMultiVCenters=true,VSphereStaticIPs=true,ValidatingAdmissionPolicy=true,AWSClusterHostedDNS=false,AdditionalRoutingCapabilities=false,AutomatedEtcdBackup=false,BootcNodeManagement=false,CSIDriverSharedResource=false,ClusterAPIInstall=false,ClusterAPIInstallIBMCloud=false,ClusterMonitoringConfig=false,ConsolePluginContentSecurityPolicy=false,DNSNameResolver=false,DynamicResourceAllocation=false,EtcdBackendQuota=false,EventedPLEG=false,Example=false,ExternalOIDC=false,ExternalOIDCWithUIDAndExtraClaimMappings=false,GCPClusterHostedDNS=false,GatewayAPI=false,ImageStreamImportMode=false,IngressControllerDynamicConfigurationManager=false,InsightsConfig=false,InsightsConfigAPI=false,InsightsOnDemandDataGather=false,InsightsRuntimeExtractor=false,MachineAPIMigration=false,MachineAPIOperatorDisableMachineHealthCheckController=false,MachineAPIProviderOpenStack=false,MachineConfigNodes=false,MaxUnavailableStatefulSet=false,MetricsCollectionProfiles=false,MinimumKubeletVersion=false,MixedCPUsAllocation=false,MultiArchInstallAzure=false,NodeSwap=false,NutanixMultiSubnets=false,OVNObservability=false,OpenShiftPodSecurityAdmission=false,PinnedImages=false,PlatformOperators=false,ProcMountType=false,RouteAdvertisements=false,RouteExternalCertificate=false,ServiceAccountTokenNodeBinding=false,SignatureStores=false,SigstoreImageVerification=false,TranslateStreamCloseWebsocketRequests=false,UpgradeStatus=false,UserNamespacesPodSecurityStandards=false,UserNamespacesSupport=false,VSphereMultiNetworks=false,VolumeAttributesClass=false,VolumeGroupSnapshot=false

openshift-controller-manager-operator

openshift-controller-manager-operator

openshift-controller-manager-operator

ConfigMapCreated

Created ConfigMap/config -n openshift-controller-manager because it was missing
(x5)

openshift-cluster-node-tuning-operator

kubelet

cluster-node-tuning-operator-598fbc5f8f-vxzvg

FailedMount

MountVolume.SetUp failed for volume "apiservice-cert" : secret "performance-addon-operator-webhook-cert" not found

openshift-controller-manager

replicaset-controller

controller-manager-f9d74d9cf

SuccessfulCreate

Created pod: controller-manager-f9d74d9cf-ks2l5

openshift-kube-controller-manager-operator

kube-controller-manager-operator

kube-controller-manager-operator

TargetUpdateRequired

"csr-signer" in "openshift-kube-controller-manager-operator" requires a new target cert/key pair: secret doesn't exist

openshift-apiserver-operator

openshift-apiserver-operator-config-observer-configobserver

openshift-apiserver-operator

RoutingConfigSubdomainChanged

Domain changed from "" to "apps.sno.openstack.lab"

openshift-apiserver-operator

openshift-apiserver-operator-config-observer-configobserver

openshift-apiserver-operator

ObserveTLSSecurityProfile

minTLSVersion changed to VersionTLS12

openshift-apiserver-operator

openshift-apiserver-operator-config-observer-configobserver

openshift-apiserver-operator

ObserveTLSSecurityProfile

cipherSuites changed to ["TLS_AES_128_GCM_SHA256" "TLS_AES_256_GCM_SHA384" "TLS_CHACHA20_POLY1305_SHA256" "TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256" "TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256" "TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384" "TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384" "TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256" "TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256"]
(x5)

openshift-cluster-node-tuning-operator

kubelet

cluster-node-tuning-operator-598fbc5f8f-vxzvg

FailedMount

MountVolume.SetUp failed for volume "node-tuning-operator-tls" : secret "node-tuning-operator-tls" not found

openshift-apiserver-operator

openshift-apiserver-operator-config-observer-configobserver

openshift-apiserver-operator

ObservedConfigChanged

Writing updated observed config:   map[string]any{ +  "apiServerArguments": map[string]any{ +  "feature-gates": []any{ +  string("AWSEFSDriverVolumeMetrics=true"), string("AdminNetworkPolicy=true"), +  string("AlibabaPlatform=true"), string("AzureWorkloadIdentity=true"), +  string("BareMetalLoadBalancer=true"), string("BuildCSIVolumes=true"), +  string("ChunkSizeMiB=true"), string("CloudDualStackNodeIPs=true"), ..., +  }, +  }, +  "projectConfig": map[string]any{"projectRequestMessage": string("")}, +  "routingConfig": map[string]any{"subdomain": string("apps.sno.openstack.lab")}, +  "servingInfo": map[string]any{ +  "cipherSuites": []any{ +  string("TLS_AES_128_GCM_SHA256"), string("TLS_AES_256_GCM_SHA384"), +  string("TLS_CHACHA20_POLY1305_SHA256"), +  string("TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256"), +  string("TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256"), +  string("TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384"), +  string("TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384"), +  string("TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256"), ..., +  }, +  "minTLSVersion": string("VersionTLS12"), +  }, +  "storageConfig": map[string]any{"urls": []any{string("https://192.168.32.10:2379")}},   }

openshift-apiserver-operator

openshift-apiserver-operator-status-controller-statussyncer_openshift-apiserver

openshift-apiserver-operator

OperatorStatusChanged

Status for clusteroperator/openshift-apiserver changed: Degraded changed from Unknown to False ("RevisionControllerDegraded: configmap \"audit\" not found")

openshift-controller-manager-operator

openshift-controller-manager-operator

openshift-controller-manager-operator

ConfigMapCreated

Created configmap/openshift-service-ca-n openshift-controller-manager because it was missing

openshift-controller-manager-operator

openshift-controller-manager-operator

openshift-controller-manager-operator

ConfigMapCreated

Created ConfigMap/config -n openshift-route-controller-manager because it was missing

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Degraded message changed from "OAuthServiceDegraded: Unable to get oauth server service: service \"oauth-openshift\" not found\nRevisionControllerDegraded: configmap \"audit\" not found\nWebhookAuthenticatorControllerDegraded: failed to retrieve service openshift-oauth-apiserver/api: service \"api\" not found" to "OAuthServiceDegraded: Unable to get oauth server service: service \"oauth-openshift\" not found\nRevisionControllerDegraded: configmap \"audit\" not found\nWebhookAuthenticatorControllerDegraded: failed to retrieve service openshift-oauth-apiserver/api: service \"api\" not found\nAPIServerDeploymentDegraded: waiting for .status.latestAvailableRevision to be available\nAPIServerDeploymentDegraded: \nAPIServerWorkloadDegraded: waiting for .status.latestAvailableRevision to be available\nAPIServerWorkloadDegraded: "

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Available message changed from "APIServicesAvailable: endpoints \"api\" not found\nOAuthServerServiceEndpointAccessibleControllerAvailable: service \"oauth-openshift\" not found\nOAuthServerServiceEndpointsEndpointAccessibleControllerAvailable: endpoints \"oauth-openshift\" not found" to "APIServicesAvailable: endpoints \"api\" not found\nOAuthServerServiceEndpointAccessibleControllerAvailable: service \"oauth-openshift\" not found\nOAuthServerServiceEndpointsEndpointAccessibleControllerAvailable: endpoints \"oauth-openshift\" not found\nReadyIngressNodesAvailable: Authentication requires functional ingress which requires at least one schedulable and ready node. Got 0 worker nodes, 1 master nodes, 0 custom target nodes (none are schedulable or ready for ingress pods)."

openshift-controller-manager

deployment-controller

controller-manager

ScalingReplicaSet

Scaled up replica set controller-manager-f9d74d9cf to 1 from 0

openshift-authentication-operator

oauth-apiserver-apiservice-openshift-apiserver-controller-openshift-apiserver-apiservice

authentication-operator

FastControllerResync

Controller "openshift-apiserver-APIService" resync interval is set to 10s which might lead to client request throttling

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Available message changed from "OAuthServerServiceEndpointAccessibleControllerAvailable: service \"oauth-openshift\" not found\nOAuthServerServiceEndpointsEndpointAccessibleControllerAvailable: endpoints \"oauth-openshift\" not found" to "APIServicesAvailable: endpoints \"api\" not found\nOAuthServerServiceEndpointAccessibleControllerAvailable: service \"oauth-openshift\" not found\nOAuthServerServiceEndpointsEndpointAccessibleControllerAvailable: endpoints \"oauth-openshift\" not found"

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Progressing changed from Unknown to False ("All is well")

openshift-controller-manager

deployment-controller

controller-manager

ScalingReplicaSet

Scaled down replica set controller-manager-797c5b47d8 to 0 from 1

openshift-service-ca

kubelet

service-ca-79bc6b8d76-72j8t

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:812819a9d712b9e345ef5f1404b242c281e2518ad724baebc393ec0fd3b3d263" already present on machine

openshift-controller-manager-operator

openshift-controller-manager-operator-status-controller-statussyncer_openshift-controller-manager

openshift-controller-manager-operator

OperatorStatusChanged

Status for clusteroperator/openshift-controller-manager changed: Degraded message changed from "All is well" to "OpenshiftControllerManagerStaticResourcesDegraded: \"assets/openshift-controller-manager/leader-role.yaml\" (string): namespaces \"openshift-controller-manager\" not found\nOpenshiftControllerManagerStaticResourcesDegraded: \"assets/openshift-controller-manager/leader-rolebinding.yaml\" (string): namespaces \"openshift-controller-manager\" not found\nOpenshiftControllerManagerStaticResourcesDegraded: \"assets/openshift-controller-manager/route-controller-manager-leader-role.yaml\" (string): namespaces \"openshift-route-controller-manager\" not found\nOpenshiftControllerManagerStaticResourcesDegraded: \"assets/openshift-controller-manager/route-controller-manager-leader-rolebinding.yaml\" (string): namespaces \"openshift-route-controller-manager\" not found\nOpenshiftControllerManagerStaticResourcesDegraded: "

openshift-kube-storage-version-migrator

kubelet

migrator-8487694857-w5tlr

Started

Started container migrator

openshift-kube-apiserver-operator

kube-apiserver-operator-kubeapiserverstaticresources-kubeapiserverstaticresources-staticresources

kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/check-endpoints-kubeconfig -n openshift-kube-apiserver because it was missing

openshift-kube-controller-manager-operator

kube-controller-manager-operator-resource-sync-controller-resourcesynccontroller

kube-controller-manager-operator

ConfigMapCreated

Created ConfigMap/aggregator-client-ca -n openshift-kube-controller-manager because it was missing

openshift-controller-manager

default-scheduler

controller-manager-f9d74d9cf-ks2l5

Scheduled

Successfully assigned openshift-controller-manager/controller-manager-f9d74d9cf-ks2l5 to master-0

openshift-authentication-operator

cluster-authentication-operator-config-observer-configobserver

authentication-operator

ObserveTLSSecurityProfile

minTLSVersion changed to VersionTLS12

openshift-apiserver-operator

openshift-apiserver-operator-status-controller-statussyncer_openshift-apiserver

openshift-apiserver-operator

OperatorStatusChanged

Status for clusteroperator/openshift-apiserver changed: Progressing changed from Unknown to False ("All is well")

openshift-authentication-operator

cluster-authentication-operator-config-observer-configobserver

authentication-operator

ObserveTLSSecurityProfile

cipherSuites changed to ["TLS_AES_128_GCM_SHA256" "TLS_AES_256_GCM_SHA384" "TLS_CHACHA20_POLY1305_SHA256" "TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256" "TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256" "TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384" "TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384" "TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256" "TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256"]

openshift-apiserver-operator

openshift-apiserver-operator-status-controller-statussyncer_openshift-apiserver

openshift-apiserver-operator

OperatorStatusChanged

Status for clusteroperator/openshift-apiserver changed: Available changed from Unknown to False ("APIServicesAvailable: endpoints \"api\" not found")

openshift-apiserver-operator

openshift-apiserver-operator-apiservice-openshift-apiserver-controller-openshift-apiserver-apiservice

openshift-apiserver-operator

FastControllerResync

Controller "openshift-apiserver-APIService" resync interval is set to 10s which might lead to client request throttling

openshift-kube-storage-version-migrator

kubelet

migrator-8487694857-w5tlr

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:951ecfeba9b2da4b653034d09275f925396a79c2d8461b8a7c71c776fee67ba0" in 3.766s (3.766s including waiting). Image size: 443272037 bytes.

openshift-kube-apiserver-operator

kube-apiserver-operator-resource-sync-controller-resourcesynccontroller

kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/bound-sa-token-signing-certs -n openshift-config-managed because it was missing

openshift-cluster-storage-operator

kubelet

csi-snapshot-controller-64854d9cff-f44gr

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9609c00207cc4db97f0fd6162eb429d7f81654137f020a677e30cba26a887a24" in 3.612s (3.612s including waiting). Image size: 463705930 bytes.

openshift-kube-apiserver-operator

kube-apiserver-operator-kubeapiserverstaticresources-kubeapiserverstaticresources-staticresources

kube-apiserver-operator

RoleBindingCreated

Created RoleBinding.rbac.authorization.k8s.io/system:openshift:controller:check-endpoints -n openshift-kube-apiserver because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-kubeapiserverstaticresources-kubeapiserverstaticresources-staticresources

kube-apiserver-operator

RoleBindingCreated

Created RoleBinding.rbac.authorization.k8s.io/system:openshift:controller:kube-apiserver-check-endpoints -n kube-system because it was missing

openshift-cluster-storage-operator

kubelet

csi-snapshot-controller-64854d9cff-f44gr

Started

Started container snapshot-controller

openshift-kube-storage-version-migrator

kubelet

migrator-8487694857-w5tlr

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:951ecfeba9b2da4b653034d09275f925396a79c2d8461b8a7c71c776fee67ba0" already present on machine

openshift-kube-storage-version-migrator

kubelet

migrator-8487694857-w5tlr

Created

Created container: migrator

openshift-service-ca

multus

service-ca-79bc6b8d76-72j8t

AddedInterface

Add eth0 [10.128.0.34/23] from ovn-kubernetes

openshift-authentication-operator

cluster-authentication-operator-config-observer-configobserver

authentication-operator

ObserveAPIAudiences

service account issuer changed from to https://kubernetes.default.svc

openshift-authentication-operator

cluster-authentication-operator-config-observer-configobserver

authentication-operator

ObserveStorageUpdated

Updated storage urls to https://192.168.32.10:2379
(x2)

openshift-route-controller-manager

default-scheduler

route-controller-manager-5ff6bdb8cf-b6499

FailedScheduling

0/1 nodes are available: 1 node(s) didn't match pod anti-affinity rules. preemption: 0/1 nodes are available: 1 node(s) didn't match pod anti-affinity rules.

openshift-kube-controller-manager-operator

kube-controller-manager-operator

kube-controller-manager-operator

SecretCreated

Created Secret/csr-signer -n openshift-kube-controller-manager-operator because it was missing

openshift-config-operator

kubelet

openshift-config-operator-95bf4f4d-25cml

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:712d334b7752d95580571059aae2c50e111d879af4fd8ea7cc3dbaf1a8e7dc69" in 5.043s (5.043s including waiting). Image size: 495994673 bytes.

openshift-cluster-olm-operator

kubelet

cluster-olm-operator-67dcd4998-c5742

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:4b5ea1ef4e09b673a0c68c8848ca162ab11d9ac373a377daa52dea702ffa3023" in 5.014s (5.014s including waiting). Image size: 495065340 bytes.

openshift-cluster-olm-operator

kubelet

cluster-olm-operator-67dcd4998-c5742

Created

Created container: copy-operator-controller-manifests

openshift-kube-controller-manager-operator

kube-controller-manager-operator-backingresourcecontroller-backingresourcecontroller-staticresources

kube-controller-manager-operator

ServiceAccountCreated

Created ServiceAccount/installer-sa -n openshift-kube-controller-manager because it was missing

openshift-cluster-olm-operator

kubelet

cluster-olm-operator-67dcd4998-c5742

Started

Started container copy-operator-controller-manifests

openshift-kube-controller-manager-operator

kube-controller-manager-operator-backingresourcecontroller-backingresourcecontroller-staticresources

kube-controller-manager-operator

ClusterRoleBindingCreated

Created ClusterRoleBinding.rbac.authorization.k8s.io/system:openshift:operator:openshift-kube-controller-manager-installer because it was missing

openshift-authentication-operator

cluster-authentication-operator-config-observer-configobserver

authentication-operator

ObservedConfigChanged

Writing updated section ("oauthAPIServer") of observed config: " map[string]any(\n- \tnil,\n+ \t{\n+ \t\t\"apiServerArguments\": map[string]any{\n+ \t\t\t\"api-audiences\": []any{string(\"https://kubernetes.default.svc\")},\n+ \t\t\t\"cors-allowed-origins\": []any{string(`//127\\.0\\.0\\.1(:|$)`), string(\"//localhost(:|$)\")},\n+ \t\t\t\"etcd-servers\": []any{string(\"https://192.168.32.10:2379\")},\n+ \t\t\t\"tls-cipher-suites\": []any{\n+ \t\t\t\tstring(\"TLS_AES_128_GCM_SHA256\"), string(\"TLS_AES_256_GCM_SHA384\"),\n+ \t\t\t\tstring(\"TLS_CHACHA20_POLY1305_SHA256\"),\n+ \t\t\t\tstring(\"TLS_ECDHE_ECDSA_WITH_AES_128_GCM\"...), ...,\n+ \t\t\t},\n+ \t\t\t\"tls-min-version\": string(\"VersionTLS12\"),\n+ \t\t},\n+ \t},\n )\n"

openshift-authentication-operator

cluster-authentication-operator-config-observer-configobserver

authentication-operator

ObserveTemplates

templates changed to map["error":"/var/config/system/secrets/v4-0-config-system-ocp-branding-template/errors.html" "login":"/var/config/system/secrets/v4-0-config-system-ocp-branding-template/login.html" "providerSelection":"/var/config/system/secrets/v4-0-config-system-ocp-branding-template/providers.html"]

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Degraded message changed from "OAuthServiceDegraded: Unable to get oauth server service: service \"oauth-openshift\" not found\nRevisionControllerDegraded: configmap \"audit\" not found\nWebhookAuthenticatorControllerDegraded: failed to retrieve service openshift-oauth-apiserver/api: service \"api\" not found\nAPIServerDeploymentDegraded: waiting for .status.latestAvailableRevision to be available\nAPIServerDeploymentDegraded: \nAPIServerWorkloadDegraded: waiting for .status.latestAvailableRevision to be available\nAPIServerWorkloadDegraded: " to "OAuthServiceDegraded: Unable to get oauth server service: service \"oauth-openshift\" not found\nRevisionControllerDegraded: configmap \"audit\" not found\nWebhookAuthenticatorControllerDegraded: failed to retrieve service openshift-oauth-apiserver/api: service \"api\" not found\nAPIServerDeploymentDegraded: waiting for .status.latestAvailableRevision to be available\nAPIServerDeploymentDegraded: \nAPIServerWorkloadDegraded: waiting for .status.latestAvailableRevision to be available\nAPIServerWorkloadDegraded: \nIngressStateEndpointsDegraded: No endpoints found for oauth-server"

openshift-kube-apiserver-operator

kube-apiserver-operator-cert-rotation-controller

kube-apiserver-operator

SecretCreated

Created Secret/check-endpoints-client-cert-key -n openshift-kube-apiserver because it was missing

openshift-cluster-storage-operator

kubelet

csi-snapshot-controller-64854d9cff-f44gr

Created

Created container: snapshot-controller

openshift-service-ca

kubelet

service-ca-79bc6b8d76-72j8t

Created

Created container: service-ca-controller

openshift-authentication-operator

oauth-apiserver-webhook-authenticator-cert-approver-OpenShiftAuthenticator-webhookauthenticatorcertapprover_openshiftauthenticator

authentication-operator

CSRApproval

The CSR "system:openshift:openshift-authenticator-wjwxz" has been approved

openshift-cluster-olm-operator

kubelet

cluster-olm-operator-67dcd4998-c5742

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adb9f6f2fd701863c7caed747df43f83d3569ba9388cfa33ea7219ac6a606b11"

openshift-authentication-operator

oauth-apiserver-openshiftauthenticatorcertrequester

authentication-operator

NoValidCertificateFound

No valid client certificate for OpenShiftAuthenticatorCertRequester is found: unable to parse certificate: data does not contain any valid RSA or ECDSA certificates

openshift-authentication-operator

cluster-authentication-operator-openshiftauthenticationstaticresources-openshiftauthenticationstaticresources-staticresources

authentication-operator

ConfigMapCreateFailed

Failed to create ConfigMap/audit -n openshift-authentication: namespaces "openshift-authentication" not found

openshift-kube-controller-manager-operator

kube-controller-manager-operator-targetconfigcontroller

kube-controller-manager-operator

ConfigMapCreated

Created ConfigMap/config -n openshift-kube-controller-manager because it was missing

openshift-config-operator

config-operator-configoperatorcontroller

openshift-config-operator

FastControllerResync

Controller "ConfigOperatorController" resync interval is set to 10s which might lead to client request throttling

openshift-kube-controller-manager-operator

kube-controller-manager-operator-resource-sync-controller-resourcesynccontroller

kube-controller-manager-operator

TargetConfigDeleted

Deleted target configmap openshift-config-managed/csr-controller-ca because source config does not exist

openshift-kube-controller-manager-operator

kube-controller-manager-operator-kubecontrollermanagerstaticresources-kubecontrollermanagerstaticresources-staticresources

kube-controller-manager-operator

NamespaceUpdated

Updated Namespace/openshift-kube-controller-manager because it changed

openshift-service-ca

kubelet

service-ca-79bc6b8d76-72j8t

Started

Started container service-ca-controller

openshift-kube-storage-version-migrator

kubelet

migrator-8487694857-w5tlr

Created

Created container: graceful-termination

openshift-controller-manager-operator

openshift-controller-manager-operator-status-controller-statussyncer_openshift-controller-manager

openshift-controller-manager-operator

OperatorStatusChanged

Status for clusteroperator/openshift-controller-manager changed: Degraded message changed from "OpenshiftControllerManagerStaticResourcesDegraded: \"assets/openshift-controller-manager/leader-role.yaml\" (string): namespaces \"openshift-controller-manager\" not found\nOpenshiftControllerManagerStaticResourcesDegraded: \"assets/openshift-controller-manager/leader-rolebinding.yaml\" (string): namespaces \"openshift-controller-manager\" not found\nOpenshiftControllerManagerStaticResourcesDegraded: \"assets/openshift-controller-manager/route-controller-manager-leader-role.yaml\" (string): namespaces \"openshift-route-controller-manager\" not found\nOpenshiftControllerManagerStaticResourcesDegraded: \"assets/openshift-controller-manager/route-controller-manager-leader-rolebinding.yaml\" (string): namespaces \"openshift-route-controller-manager\" not found\nOpenshiftControllerManagerStaticResourcesDegraded: " to "All is well",Progressing changed from Unknown to True ("Progressing: deployment/controller-manager: observed generation is 1, desired generation is 2.\nProgressing: deployment/controller-manager: available replicas is 0, desired available replicas > 1\nProgressing: deployment/route-controller-manager: observed generation is 1, desired generation is 2.\nProgressing: deployment/route-controller-manager: available replicas is 0, desired available replicas > 1\nProgressing: openshiftcontrollermanagers.operator.openshift.io/cluster: observed generation is 0, desired generation is 2."),Available changed from Unknown to False ("Available: no pods available on any node."),Upgradeable changed from Unknown to True ("All is well")

openshift-kube-apiserver-operator

kube-apiserver-operator-kubeapiserverstaticresources-kubeapiserverstaticresources-staticresources

kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/control-plane-node-kubeconfig -n openshift-kube-apiserver because it was missing

openshift-config-operator

config-operator

config-operator-lock

LeaderElection

openshift-config-operator-95bf4f4d-25cml_caf79c8a-7124-4a01-82aa-08e5005c159d became leader

openshift-cluster-storage-operator

csi-snapshot-controller-operator-status-controller-statussyncer_csi-snapshot-controller

csi-snapshot-controller-operator

OperatorVersionChanged

clusteroperator/csi-snapshot-controller version "operator" changed from "" to "4.18.35"

openshift-cluster-storage-operator

csi-snapshot-controller-operator-status-controller-statussyncer_csi-snapshot-controller

csi-snapshot-controller-operator

OperatorStatusChanged

Status for clusteroperator/csi-snapshot-controller changed: Progressing changed from True to False ("All is well"),Available changed from False to True ("All is well"),status.versions changed from [] to [{"operator" "4.18.35"}]

openshift-cluster-storage-operator

csi-snapshot-controller-operator-status-controller-statussyncer_csi-snapshot-controller

csi-snapshot-controller-operator

OperatorVersionChanged

clusteroperator/csi-snapshot-controller version "csi-snapshot-controller" changed from "" to "4.18.35"

openshift-cluster-storage-operator

csi-snapshot-controller-operator-status-controller-statussyncer_csi-snapshot-controller

csi-snapshot-controller-operator

OperatorStatusChanged

Status for clusteroperator/csi-snapshot-controller changed: status.versions changed from [{"operator" "4.18.35"}] to [{"operator" "4.18.35"} {"csi-snapshot-controller" "4.18.35"}]

openshift-cluster-storage-operator

snapshot-controller-leader/csi-snapshot-controller-64854d9cff-f44gr

snapshot-controller-leader

LeaderElection

csi-snapshot-controller-64854d9cff-f44gr became leader

openshift-kube-apiserver-operator

kube-apiserver-operator-kubeapiserverstaticresources-kubeapiserverstaticresources-staticresources

kube-apiserver-operator

RoleBindingCreated

Created RoleBinding.rbac.authorization.k8s.io/authentication-reader-for-authenticated-users -n kube-system because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-kubeapiserverstaticresources-kubeapiserverstaticresources-staticresources

kube-apiserver-operator

ClusterRoleBindingCreated

Created ClusterRoleBinding.rbac.authorization.k8s.io/system:openshift:operator:kube-apiserver-recovery because it was missing

openshift-kube-storage-version-migrator

kubelet

migrator-8487694857-w5tlr

Started

Started container graceful-termination

openshift-kube-storage-version-migrator-operator

openshift-kube-storage-version-migrator-operator-status-controller-statussyncer_kube-storage-version-migrator

kube-storage-version-migrator-operator

OperatorStatusChanged

Status for clusteroperator/kube-storage-version-migrator changed: Progressing changed from True to False ("All is well"),Available changed from False to True ("All is well")

openshift-authentication-operator

oauth-apiserver-openshiftauthenticatorcertrequester

authentication-operator

CSRCreated

A csr "system:openshift:openshift-authenticator-wjwxz" is created for OpenShiftAuthenticatorCertRequester

openshift-apiserver-operator

openshift-apiserver-operator-apiserverstaticresources-apiserverstaticresources-staticresources

openshift-apiserver-operator

NamespaceCreated

Created Namespace/openshift-apiserver because it was missing

openshift-apiserver-operator

openshift-apiserver-operator-apiserverstaticresources-apiserverstaticresources-staticresources

openshift-apiserver-operator

ClusterRoleBindingCreated

Created ClusterRoleBinding.rbac.authorization.k8s.io/system:openshift:openshift-apiserver because it was missing

openshift-apiserver-operator

openshift-apiserver-operator-audit-policy-controller-auditpolicycontroller

openshift-apiserver-operator

ConfigMapCreated

Created ConfigMap/audit -n openshift-apiserver because it was missing

openshift-apiserver-operator

openshift-apiserver-operator-revisioncontroller

openshift-apiserver-operator

StartingNewRevision

new revision 1 triggered by "configmap \"audit-0\" not found"

openshift-config-operator

config-operator-status-controller-statussyncer_config-operator

openshift-config-operator

OperatorStatusChanged

Status for clusteroperator/config-operator changed: Degraded changed from Unknown to False ("All is well")

openshift-kube-apiserver-operator

kube-apiserver-operator-kubeapiserverstaticresources-kubeapiserverstaticresources-staticresources

kube-apiserver-operator

ServiceAccountCreated

Created ServiceAccount/localhost-recovery-client -n openshift-kube-apiserver because it was missing

openshift-config-operator

config-operator-status-controller-statussyncer_config-operator

openshift-config-operator

OperatorStatusChanged

Status for clusteroperator/config-operator changed: Degraded set to Unknown (""),Progressing set to False ("All is well"),Available set to True ("All is well"),Upgradeable set to True ("All is well"),EvaluationConditionsDetected set to Unknown (""),status.versions changed from [] to [{"feature-gates" "4.18.35"} {"operator" "4.18.35"}]
(x2)

openshift-config-operator

config-operator-status-controller-statussyncer_config-operator

openshift-config-operator

OperatorVersionChanged

clusteroperator/config-operator version "operator" changed from "" to "4.18.35"

kube-system

cluster-policy-controller-namespace-security-allocation-controller

bootstrap-kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openshift-apiserver namespace
(x2)

openshift-config-operator

config-operator-status-controller-statussyncer_config-operator

openshift-config-operator

OperatorVersionChanged

clusteroperator/config-operator version "feature-gates" changed from "" to "4.18.35"

openshift-service-ca

service-ca-controller

service-ca-controller-lock

LeaderElection

service-ca-79bc6b8d76-72j8t_9c62926e-104e-44fb-a165-55dabcc360fc became leader

openshift-service-ca-operator

service-ca-operator-status-controller-statussyncer_service-ca

service-ca-operator

OperatorVersionChanged

clusteroperator/service-ca version "operator" changed from "" to "4.18.35"

openshift-kube-controller-manager-operator

kube-controller-manager-operator-kubecontrollermanagerstaticresources-kubecontrollermanagerstaticresources-staticresources

kube-controller-manager-operator

ConfigMapCreated

Created ConfigMap/kube-controller-cert-syncer-kubeconfig -n openshift-kube-controller-manager because it was missing

openshift-apiserver-operator

openshift-apiserver-operator-openshiftapiserver-workloadworkloadcontroller

openshift-apiserver-operator

ConfigMapCreated

Created ConfigMap/config -n openshift-apiserver because it was missing

openshift-kube-controller-manager-operator

kube-controller-manager-operator-resource-sync-controller-resourcesynccontroller

kube-controller-manager-operator

ConfigMapCreated

Created ConfigMap/service-ca -n openshift-kube-controller-manager because it was missing

openshift-route-controller-manager

default-scheduler

route-controller-manager-5ff6bdb8cf-b6499

Scheduled

Successfully assigned openshift-route-controller-manager/route-controller-manager-5ff6bdb8cf-b6499 to master-0

openshift-kube-controller-manager-operator

kube-controller-manager-operator-satokensignercontroller

kube-controller-manager-operator

SecretCreated

Created Secret/service-account-private-key -n openshift-kube-controller-manager because it was missing

openshift-kube-controller-manager-operator

kube-controller-manager-operator-kubecontrollermanagerstaticresources-kubecontrollermanagerstaticresources-staticresources

kube-controller-manager-operator

RoleBindingCreated

Created RoleBinding.rbac.authorization.k8s.io/system:openshift:leader-election-lock-cluster-policy-controller -n openshift-kube-controller-manager because it was missing

openshift-kube-controller-manager-operator

kube-controller-manager-operator-kubecontrollermanagerstaticresources-kubecontrollermanagerstaticresources-staticresources

kube-controller-manager-operator

RoleCreated

Created Role.rbac.authorization.k8s.io/system:openshift:leader-election-lock-cluster-policy-controller -n openshift-kube-controller-manager because it was missing

openshift-service-ca-operator

service-ca-operator-status-controller-statussyncer_service-ca

service-ca-operator

OperatorStatusChanged

Status for clusteroperator/service-ca changed: status.versions changed from [] to [{"operator" "4.18.35"}]

openshift-config-operator

config-operator-configoperatorcontroller

openshift-config-operator

ConfigOperatorStatusChanged

Operator conditions defaulted: [{OperatorAvailable True 2026-03-20 08:36:25 +0000 UTC AsExpected } {OperatorProgressing False 2026-03-20 08:36:25 +0000 UTC AsExpected } {OperatorUpgradeable True 2026-03-20 08:36:25 +0000 UTC AsExpected }]

openshift-service-ca-operator

service-ca-operator-status-controller-statussyncer_service-ca

service-ca-operator

OperatorStatusChanged

Status for clusteroperator/service-ca changed: Progressing changed from True to False ("Progressing: All service-ca-operator deployments updated")

openshift-kube-controller-manager-operator

kube-controller-manager-operator-kubecontrollermanagerstaticresources-kubecontrollermanagerstaticresources-staticresources

kube-controller-manager-operator

RoleBindingCreated

Created RoleBinding.rbac.authorization.k8s.io/system:openshift:leader-locking-kube-controller-manager -n kube-system because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-kubeapiserverstaticresources-kubeapiserverstaticresources-staticresources

kube-apiserver-operator

StorageVersionMigrationCreated

Created StorageVersionMigration.migration.k8s.io/flowcontrol-flowschema-storage-version-migration-v1beta3 because it was missing

openshift-apiserver-operator

openshift-apiserver-operator-apiserverstaticresources-apiserverstaticresources-staticresources

openshift-apiserver-operator

ServiceCreated

Created Service/api -n openshift-apiserver because it was missing

openshift-authentication-operator

oauth-apiserver-apiserverstaticresources-apiserverstaticresources-staticresources

authentication-operator

NamespaceCreated

Created Namespace/openshift-oauth-apiserver because it was missing

openshift-authentication-operator

oauth-apiserver-apiserverstaticresources-apiserverstaticresources-staticresources

authentication-operator

ClusterRoleBindingCreated

Created ClusterRoleBinding.rbac.authorization.k8s.io/system:openshift:oauth-apiserver because it was missing

openshift-authentication-operator

oauth-apiserver-audit-policy-controller-auditpolicycontroller

authentication-operator

ConfigMapCreated

Created ConfigMap/audit -n openshift-oauth-apiserver because it was missing

openshift-authentication-operator

oauth-apiserver-revisioncontroller

authentication-operator

StartingNewRevision

new revision 1 triggered by "configmap \"audit-0\" not found"

openshift-kube-apiserver-operator

kube-apiserver-operator-resource-sync-controller-resourcesynccontroller

kube-apiserver-operator

TargetConfigDeleted

Deleted target configmap openshift-kube-apiserver/kubelet-serving-ca because source config does not exist

openshift-controller-manager-operator

openshift-controller-manager-operator-status-controller-statussyncer_openshift-controller-manager

openshift-controller-manager-operator

OperatorStatusChanged

Status for clusteroperator/openshift-controller-manager changed: Progressing message changed from "Progressing: deployment/controller-manager: observed generation is 1, desired generation is 2.\nProgressing: deployment/controller-manager: available replicas is 0, desired available replicas > 1\nProgressing: deployment/route-controller-manager: observed generation is 1, desired generation is 2.\nProgressing: deployment/route-controller-manager: available replicas is 0, desired available replicas > 1\nProgressing: openshiftcontrollermanagers.operator.openshift.io/cluster: observed generation is 0, desired generation is 2." to "Progressing: deployment/controller-manager: available replicas is 0, desired available replicas > 1\nProgressing: deployment/route-controller-manager: available replicas is 0, desired available replicas > 1"

kube-system

cluster-policy-controller-namespace-security-allocation-controller

bootstrap-kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openshift-oauth-apiserver namespace

openshift-kube-apiserver-operator

kube-apiserver-operator-kubeapiserverstaticresources-kubeapiserverstaticresources-staticresources

kube-apiserver-operator

PrometheusRuleCreated

Created PrometheusRule.monitoring.coreos.com/kube-apiserver-slos-basic -n openshift-kube-apiserver because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-kubeapiserverstaticresources-kubeapiserverstaticresources-staticresources

kube-apiserver-operator

PrometheusRuleCreated

Created PrometheusRule.monitoring.coreos.com/kube-apiserver-requests -n openshift-kube-apiserver because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-kubeapiserverstaticresources-kubeapiserverstaticresources-staticresources

kube-apiserver-operator

PrometheusRuleCreated

Created PrometheusRule.monitoring.coreos.com/audit-errors -n openshift-kube-apiserver because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-kubeapiserverstaticresources-kubeapiserverstaticresources-staticresources

kube-apiserver-operator

PrometheusRuleCreated

Created PrometheusRule.monitoring.coreos.com/api-usage -n openshift-kube-apiserver because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-kubeapiserverstaticresources-kubeapiserverstaticresources-staticresources

kube-apiserver-operator

StorageVersionMigrationCreated

Created StorageVersionMigration.migration.k8s.io/flowcontrol-prioritylevel-storage-version-migration-v1beta3 because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-kubeapiserverstaticresources-kubeapiserverstaticresources-staticresources

kube-apiserver-operator

StorageVersionMigrationCreated

Created StorageVersionMigration.migration.k8s.io/flowcontrol-prioritylevel-storage-version-migration because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-kubeapiserverstaticresources-kubeapiserverstaticresources-staticresources

kube-apiserver-operator

StorageVersionMigrationCreated

Created StorageVersionMigration.migration.k8s.io/flowcontrol-flowschema-storage-version-migration because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-kubeapiserverstaticresources-kubeapiserverstaticresources-staticresources

kube-apiserver-operator

CustomResourceDefinitionUpdated

Updated CustomResourceDefinition.apiextensions.k8s.io/apirequestcounts.apiserver.openshift.io because it changed

openshift-kube-apiserver-operator

kube-apiserver-operator-kubeapiserverstaticresources-kubeapiserverstaticresources-staticresources

kube-apiserver-operator

SecretCreated

Created Secret/localhost-recovery-client-token -n openshift-kube-apiserver because it was missing

openshift-apiserver-operator

openshift-apiserver-operator-status-controller-statussyncer_openshift-apiserver

openshift-apiserver-operator

OperatorStatusChanged

Status for clusteroperator/openshift-apiserver changed: Available message changed from "APIServicesAvailable: endpoints \"api\" not found" to "APIServicesAvailable: PreconditionNotReady"

openshift-kube-controller-manager-operator

kube-controller-manager-operator-resource-sync-controller-resourcesynccontroller

kube-controller-manager-operator

SecretCreated

Created Secret/kube-controller-manager-client-cert-key -n openshift-kube-controller-manager because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-kubeapiserverstaticresources-kubeapiserverstaticresources-staticresources

kube-apiserver-operator

PrometheusRuleCreated

Created PrometheusRule.monitoring.coreos.com/podsecurity -n openshift-kube-apiserver because it was missing

openshift-cluster-olm-operator

kubelet

cluster-olm-operator-67dcd4998-c5742

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adb9f6f2fd701863c7caed747df43f83d3569ba9388cfa33ea7219ac6a606b11" in 2.713s (2.713s including waiting). Image size: 511164375 bytes.

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Degraded message changed from "OAuthServiceDegraded: Unable to get oauth server service: service \"oauth-openshift\" not found\nRevisionControllerDegraded: configmap \"audit\" not found\nWebhookAuthenticatorControllerDegraded: failed to retrieve service openshift-oauth-apiserver/api: service \"api\" not found\nAPIServerDeploymentDegraded: waiting for .status.latestAvailableRevision to be available\nAPIServerDeploymentDegraded: \nAPIServerWorkloadDegraded: waiting for .status.latestAvailableRevision to be available\nAPIServerWorkloadDegraded: \nIngressStateEndpointsDegraded: No endpoints found for oauth-server" to "OAuthServiceDegraded: Unable to get oauth server service: service \"oauth-openshift\" not found\nRevisionControllerDegraded: configmap \"audit\" not found\nWebhookAuthenticatorControllerDegraded: failed to retrieve service openshift-oauth-apiserver/api: service \"api\" not found\nAPIServerDeploymentDegraded: waiting for .status.latestAvailableRevision to be available\nAPIServerDeploymentDegraded: \nAPIServerWorkloadDegraded: waiting for .status.latestAvailableRevision to be available\nAPIServerWorkloadDegraded: \nIngressStateEndpointsDegraded: No endpoints found for oauth-server\nRouterCertsDegraded: neither the custom secret/v4-0-config-system-router-certs -n openshift-authentication or default secret/v4-0-config-system-custom-router-certs -n openshift-authentication could be retrieved: secret \"v4-0-config-system-router-certs\" not found"

openshift-apiserver-operator

openshift-apiserver-operator-revisioncontroller

openshift-apiserver-operator

ConfigMapCreated

Created ConfigMap/audit-1 -n openshift-apiserver because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-resource-sync-controller-resourcesynccontroller

kube-apiserver-operator

TargetConfigDeleted

Deleted target configmap openshift-config-managed/kubelet-serving-ca because source config does not exist

openshift-kube-controller-manager-operator

kube-controller-manager-operator-config-observer-configobserver

kube-controller-manager-operator

ObservedConfigChanged

Writing updated observed config: map[string]any{ "extendedArguments": map[string]any{"cluster-cidr": []any{string("10.128.0.0/16")}, "cluster-name": []any{string("sno-bhfwr")}, "feature-gates": []any{string("AWSEFSDriverVolumeMetrics=true"), string("AdminNetworkPolicy=true"), string("AlibabaPlatform=true"), string("AzureWorkloadIdentity=true"), ...}, "service-cluster-ip-range": []any{string("172.30.0.0/16")}}, "featureGates": []any{string("AWSEFSDriverVolumeMetrics=true"), string("AdminNetworkPolicy=true"), string("AlibabaPlatform=true"), string("AzureWorkloadIdentity=true"), ...}, + "serviceServingCert": map[string]any{ + "certFile": string("/etc/kubernetes/static-pod-resources/configmaps/service-ca/ca-bundle.crt"), + }, "servingInfo": map[string]any{"cipherSuites": []any{string("TLS_AES_128_GCM_SHA256"), string("TLS_AES_256_GCM_SHA384"), string("TLS_CHACHA20_POLY1305_SHA256"), string("TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256"), ...}, "minTLSVersion": string("VersionTLS12")}, }

openshift-kube-controller-manager-operator

kube-controller-manager-operator-config-observer-configobserver

kube-controller-manager-operator

ObserveServiceCAConfigMap

observed change in config
(x4)

openshift-controller-manager

kubelet

controller-manager-f9d74d9cf-ks2l5

FailedMount

MountVolume.SetUp failed for volume "serving-cert" : secret "serving-cert" not found

openshift-authentication-operator

cluster-authentication-operator-config-observer-configobserver

authentication-operator

ObserveAuditProfile

AuditProfile changed from '%!s(<nil>)' to 'map[audit-log-format:[json] audit-log-maxbackup:[10] audit-log-maxsize:[100] audit-log-path:[/var/log/oauth-server/audit.log] audit-policy-file:[/var/run/configmaps/audit/audit.yaml]]'

openshift-authentication-operator

cluster-authentication-operator-config-observer-configobserver

authentication-operator

ObservedConfigChanged

Writing updated section ("oauthServer") of observed config: " map[string]any(\n- \tnil,\n+ \t{\n+ \t\t\"corsAllowedOrigins\": []any{string(`//127\\.0\\.0\\.1(:|$)`), string(\"//localhost(:|$)\")},\n+ \t\t\"oauthConfig\": map[string]any{\n+ \t\t\t\"assetPublicURL\": string(\"\"),\n+ \t\t\t\"loginURL\": string(\"https://api.sno.openstack.lab:6443\"),\n+ \t\t\t\"templates\": map[string]any{\n+ \t\t\t\t\"error\": string(\"/var/config/system/secrets/v4-0-\"...),\n+ \t\t\t\t\"login\": string(\"/var/config/system/secrets/v4-0-\"...),\n+ \t\t\t\t\"providerSelection\": string(\"/var/config/system/secrets/v4-0-\"...),\n+ \t\t\t},\n+ \t\t\t\"tokenConfig\": map[string]any{\n+ \t\t\t\t\"accessTokenMaxAgeSeconds\": float64(86400),\n+ \t\t\t\t\"authorizeTokenMaxAgeSeconds\": float64(300),\n+ \t\t\t},\n+ \t\t},\n+ \t\t\"serverArguments\": map[string]any{\n+ \t\t\t\"audit-log-format\": []any{string(\"json\")},\n+ \t\t\t\"audit-log-maxbackup\": []any{string(\"10\")},\n+ \t\t\t\"audit-log-maxsize\": []any{string(\"100\")},\n+ \t\t\t\"audit-log-path\": []any{string(\"/var/log/oauth-server/audit.log\")},\n+ \t\t\t\"audit-policy-file\": []any{string(\"/var/run/configmaps/audit/audit.\"...)},\n+ \t\t},\n+ \t\t\"servingInfo\": map[string]any{\n+ \t\t\t\"cipherSuites\": []any{\n+ \t\t\t\tstring(\"TLS_AES_128_GCM_SHA256\"), string(\"TLS_AES_256_GCM_SHA384\"),\n+ \t\t\t\tstring(\"TLS_CHACHA20_POLY1305_SHA256\"),\n+ \t\t\t\tstring(\"TLS_ECDHE_ECDSA_WITH_AES_128_GCM\"...), ...,\n+ \t\t\t},\n+ \t\t\t\"minTLSVersion\": string(\"VersionTLS12\"),\n+ \t\t},\n+ \t\t\"volumesToMount\": map[string]any{\"identityProviders\": string(\"{}\")},\n+ \t},\n )\n"

openshift-kube-controller-manager-operator

kube-controller-manager-operator-targetconfigcontroller

kube-controller-manager-operator

ConfigMapCreated

Created ConfigMap/cluster-policy-controller-config -n openshift-kube-controller-manager because it was missing

openshift-kube-controller-manager-operator

kube-controller-manager-operator-kubecontrollermanagerstaticresources-kubecontrollermanagerstaticresources-staticresources

kube-controller-manager-operator

ServiceCreated

Created Service/kube-controller-manager -n openshift-kube-controller-manager because it was missing

openshift-authentication-operator

cluster-authentication-operator-config-observer-configobserver

authentication-operator

ObserveTokenConfig

accessTokenMaxAgeSeconds changed from %!d(float64=0) to %!d(float64=86400)

openshift-authentication-operator

cluster-authentication-operator-config-observer-configobserver

authentication-operator

ObserveTLSSecurityProfile

cipherSuites changed to ["TLS_AES_128_GCM_SHA256" "TLS_AES_256_GCM_SHA384" "TLS_CHACHA20_POLY1305_SHA256" "TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256" "TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256" "TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384" "TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384" "TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256" "TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256"]

openshift-authentication-operator

cluster-authentication-operator-config-observer-configobserver

authentication-operator

ObserveTLSSecurityProfile

minTLSVersion changed to VersionTLS12

openshift-apiserver-operator

openshift-apiserver-operator-apiserverstaticresources-apiserverstaticresources-staticresources

openshift-apiserver-operator

ServiceAccountCreated

Created ServiceAccount/openshift-apiserver-sa -n openshift-apiserver because it was missing

openshift-apiserver-operator

openshift-apiserver-operator-openshiftapiserver-workloadworkloadcontroller

openshift-apiserver-operator

ConfigMapCreated

Created ConfigMap/image-import-ca -n openshift-apiserver because it was missing

openshift-authentication-operator

cluster-authentication-operator-config-observer-configobserver

authentication-operator

ObserveAPIServerURL

loginURL changed from to https://api.sno.openstack.lab:6443
(x2)

openshift-cluster-storage-operator

kubelet

cluster-storage-operator-7d87854d6-vlq7h

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:30a2f97d7785ce8b0ea5115e67c4554b64adefbc7856bcf6f4fe6cc7e938a310"

openshift-etcd-operator

kubelet

etcd-operator-8544cbcf9c-brhw4

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d5a971d5889f167cfe61a64c366424b87c17a6dc141ffcc43406cdcbb50cae2a" in 415ms (415ms including waiting). Image size: 518384969 bytes.
(x2)

openshift-cluster-olm-operator

kubelet

cluster-olm-operator-67dcd4998-c5742

Started

Started container cluster-olm-operator

kube-system

cluster-policy-controller-namespace-security-allocation-controller

bootstrap-kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openshift-authentication namespace

openshift-authentication-operator

cluster-authentication-operator-openshiftauthenticationstaticresources-openshiftauthenticationstaticresources-staticresources

authentication-operator

ClusterRoleBindingCreated

Created ClusterRoleBinding.rbac.authorization.k8s.io/system:openshift:openshift-authentication because it was missing

openshift-apiserver-operator

openshift-apiserver-operator-revisioncontroller

openshift-apiserver-operator

RevisionTriggered

new revision 1 triggered by "configmap \"audit-0\" not found"
(x2)

openshift-cluster-olm-operator

kubelet

cluster-olm-operator-67dcd4998-c5742

Created

Created container: cluster-olm-operator
(x2)

openshift-etcd-operator

kubelet

etcd-operator-8544cbcf9c-brhw4

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d5a971d5889f167cfe61a64c366424b87c17a6dc141ffcc43406cdcbb50cae2a"

openshift-cluster-olm-operator

kubelet

cluster-olm-operator-67dcd4998-c5742

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adb9f6f2fd701863c7caed747df43f83d3569ba9388cfa33ea7219ac6a606b11" already present on machine

openshift-apiserver-operator

openshift-apiserver-operator-status-controller-statussyncer_openshift-apiserver

openshift-apiserver-operator

OperatorStatusChanged

Status for clusteroperator/openshift-apiserver changed: Degraded message changed from "RevisionControllerDegraded: configmap \"audit\" not found" to "All is well"

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Degraded message changed from "OAuthServiceDegraded: Unable to get oauth server service: service \"oauth-openshift\" not found\nRevisionControllerDegraded: configmap \"audit\" not found\nWebhookAuthenticatorControllerDegraded: failed to retrieve service openshift-oauth-apiserver/api: service \"api\" not found\nAPIServerDeploymentDegraded: waiting for .status.latestAvailableRevision to be available\nAPIServerDeploymentDegraded: \nAPIServerWorkloadDegraded: waiting for .status.latestAvailableRevision to be available\nAPIServerWorkloadDegraded: \nIngressStateEndpointsDegraded: No endpoints found for oauth-server\nRouterCertsDegraded: neither the custom secret/v4-0-config-system-router-certs -n openshift-authentication or default secret/v4-0-config-system-custom-router-certs -n openshift-authentication could be retrieved: secret \"v4-0-config-system-router-certs\" not found" to "OAuthServiceDegraded: Unable to get oauth server service: service \"oauth-openshift\" not found\nRevisionControllerDegraded: configmap \"audit\" not found\nWebhookAuthenticatorControllerDegraded: failed to retrieve service openshift-oauth-apiserver/api: service \"api\" not found\nAPIServerDeploymentDegraded: waiting for .status.latestAvailableRevision to be available\nAPIServerDeploymentDegraded: \nAPIServerWorkloadDegraded: waiting for .status.latestAvailableRevision to be available\nAPIServerWorkloadDegraded: \nIngressStateEndpointsDegraded: No endpoints found for oauth-server\nRouterCertsDegraded: neither the custom secret/v4-0-config-system-router-certs -n openshift-authentication or default secret/v4-0-config-system-custom-router-certs -n openshift-authentication could be retrieved: secret \"v4-0-config-system-router-certs\" not found\nOAuthServerConfigObservationDegraded: secret \"v4-0-config-system-router-certs\" not found"

openshift-authentication-operator

cluster-authentication-operator-openshiftauthenticationstaticresources-openshiftauthenticationstaticresources-staticresources

authentication-operator

NamespaceCreated

Created Namespace/openshift-authentication because it was missing

openshift-apiserver-operator

openshift-apiserver-operator-apiserverstaticresources-apiserverstaticresources-staticresources

openshift-apiserver-operator

ConfigMapCreated

Created ConfigMap/trusted-ca-bundle -n openshift-apiserver because it was missing

openshift-cluster-node-tuning-operator

multus

cluster-node-tuning-operator-598fbc5f8f-vxzvg

AddedInterface

Add eth0 [10.128.0.9/23] from ovn-kubernetes

openshift-cluster-storage-operator

kubelet

cluster-storage-operator-7d87854d6-vlq7h

Created

Created container: cluster-storage-operator
(x6)

openshift-operator-lifecycle-manager

kubelet

package-server-manager-7b95f86987-2pg77

FailedMount

MountVolume.SetUp failed for volume "package-server-manager-serving-cert" : secret "package-server-manager-serving-cert" not found

openshift-image-registry

kubelet

cluster-image-registry-operator-5549dc66cb-fmhbq

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7af9f5c5af9d529840233ef4b519120cc0e3f14c4fe28cc43b0823f2c11d8f89"

openshift-image-registry

multus

cluster-image-registry-operator-5549dc66cb-fmhbq

AddedInterface

Add eth0 [10.128.0.12/23] from ovn-kubernetes

openshift-authentication-operator

oauth-apiserver-openshiftauthenticatorcertrequester

authentication-operator

ClientCertificateCreated

A new client certificate for OpenShiftAuthenticatorCertRequester is available
(x6)

openshift-operator-lifecycle-manager

kubelet

olm-operator-5c9796789-tjm9l

FailedMount

MountVolume.SetUp failed for volume "srv-cert" : secret "olm-operator-serving-cert" not found

openshift-dns-operator

kubelet

dns-operator-9c5679d8f-r6dm8

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ea5c8a93f30e0a4932da5697d22c0da7eda9a7035c0555eb006b6755e62bb2fc"

openshift-dns-operator

multus

dns-operator-9c5679d8f-r6dm8

AddedInterface

Add eth0 [10.128.0.15/23] from ovn-kubernetes

openshift-etcd-operator

openshift-cluster-etcd-operator

openshift-cluster-etcd-operator-lock

LeaderElection

etcd-operator-8544cbcf9c-brhw4_d2bf482f-20bc-4a90-b6b4-9ee766084981 became leader
(x6)

openshift-monitoring

kubelet

cluster-monitoring-operator-58845fbb57-nljsr

FailedMount

MountVolume.SetUp failed for volume "cluster-monitoring-operator-tls" : secret "cluster-monitoring-operator-tls" not found
(x6)

openshift-monitoring

kubelet

cluster-monitoring-operator-58845fbb57-nljsr

FailedMount

MountVolume.SetUp failed for volume "cluster-monitoring-operator-tls" : secret "cluster-monitoring-operator-tls" not found
(x6)

openshift-multus

kubelet

network-metrics-daemon-srdjm

FailedMount

MountVolume.SetUp failed for volume "metrics-certs" : secret "metrics-daemon-secret" not found
(x6)

openshift-multus

kubelet

multus-admission-controller-5dbbb8b86f-5rrrh

FailedMount

MountVolume.SetUp failed for volume "webhook-certs" : secret "multus-admission-controller-secret" not found
(x6)

openshift-machine-api

kubelet

cluster-baremetal-operator-6f69995874-dv6cd

FailedMount

MountVolume.SetUp failed for volume "cert" : secret "cluster-baremetal-webhook-server-cert" not found
(x6)

openshift-machine-api

kubelet

cluster-baremetal-operator-6f69995874-dv6cd

FailedMount

MountVolume.SetUp failed for volume "cluster-baremetal-operator-tls" : secret "cluster-baremetal-operator-tls" not found

openshift-cluster-version

kubelet

cluster-version-operator-56d8475767-9gfkg

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-release@sha256:59727c4b3fef19e5149675cf3350735bbfe2c6588a57654b2e4552dd719f58b1"

openshift-cluster-olm-operator

cluster-olm-operator

cluster-olm-operator-lock

LeaderElection

cluster-olm-operator-67dcd4998-c5742_b807bd40-fdd1-4bd5-ad35-669dbad40ea1 became leader
(x6)

openshift-operator-lifecycle-manager

kubelet

catalog-operator-68f85b4d6c-fzm28

FailedMount

MountVolume.SetUp failed for volume "srv-cert" : secret "catalog-operator-serving-cert" not found

openshift-ingress-operator

multus

ingress-operator-66b84d69b-gzg9m

AddedInterface

Add eth0 [10.128.0.26/23] from ovn-kubernetes

openshift-ingress-operator

kubelet

ingress-operator-66b84d69b-gzg9m

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:77fff570657d2fa0bfb709b2c8b6665bae0bf90a2be981d8dbca56c674715098"
(x2)

openshift-config-operator

kubelet

openshift-config-operator-95bf4f4d-25cml

Started

Started container openshift-config-operator
(x2)

openshift-config-operator

kubelet

openshift-config-operator-95bf4f4d-25cml

Created

Created container: openshift-config-operator
(x6)

openshift-machine-api

kubelet

cluster-baremetal-operator-6f69995874-dv6cd

FailedMount

MountVolume.SetUp failed for volume "cert" : secret "cluster-baremetal-webhook-server-cert" not found

openshift-cluster-node-tuning-operator

kubelet

cluster-node-tuning-operator-598fbc5f8f-vxzvg

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:aa5e782406f71c048b1ac3a4bf5d1227ff4be81111114083ad4c7a209c6bfb5a"

kube-system

cluster-policy-controller-namespace-security-allocation-controller

bootstrap-kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openshift-catalogd namespace

kube-system

cluster-policy-controller-namespace-security-allocation-controller

bootstrap-kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openshift-operator-controller namespace

openshift-config-operator

kubelet

openshift-config-operator-95bf4f4d-25cml

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:712d334b7752d95580571059aae2c50e111d879af4fd8ea7cc3dbaf1a8e7dc69" already present on machine

openshift-cluster-olm-operator

olm-status-controller-statussyncer_olm

cluster-olm-operator

OperatorVersionChanged

clusteroperator/olm version "operator" changed from "" to "4.18.35"
(x6)

openshift-machine-api

kubelet

cluster-baremetal-operator-6f69995874-dv6cd

FailedMount

MountVolume.SetUp failed for volume "cluster-baremetal-operator-tls" : secret "cluster-baremetal-operator-tls" not found

openshift-cluster-storage-operator

kubelet

cluster-storage-operator-7d87854d6-vlq7h

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:30a2f97d7785ce8b0ea5115e67c4554b64adefbc7856bcf6f4fe6cc7e938a310" in 514ms (514ms including waiting). Image size: 513582374 bytes.

openshift-cluster-olm-operator

OperatorControllerStaticResources-operatorcontrollerstaticresources-operatorcontrollerstaticresources-staticresources

cluster-olm-operator

NamespaceCreated

Created Namespace/openshift-operator-controller because it was missing

openshift-cluster-olm-operator

CatalogdStaticResources-catalogdstaticresources-catalogdstaticresources-staticresources

cluster-olm-operator

ClusterRoleCreated

Created ClusterRole.rbac.authorization.k8s.io/catalogd-manager-role because it was missing

openshift-cluster-olm-operator

OperatorControllerStaticResources-operatorcontrollerstaticresources-operatorcontrollerstaticresources-staticresources

cluster-olm-operator

RoleCreated

Created Role.rbac.authorization.k8s.io/operator-controller-leader-election-role -n openshift-operator-controller because it was missing

openshift-cluster-olm-operator

CatalogdStaticResources-catalogdstaticresources-catalogdstaticresources-staticresources

cluster-olm-operator

RoleCreated

Created Role.rbac.authorization.k8s.io/catalogd-manager-role -n openshift-config because it was missing

openshift-cluster-olm-operator

OperatorControllerStaticResources-operatorcontrollerstaticresources-operatorcontrollerstaticresources-staticresources

cluster-olm-operator

RoleCreated

Created Role.rbac.authorization.k8s.io/operator-controller-manager-role -n openshift-config because it was missing

openshift-cluster-olm-operator

CatalogdStaticResources-catalogdstaticresources-catalogdstaticresources-staticresources

cluster-olm-operator

RoleCreated

Created Role.rbac.authorization.k8s.io/catalogd-leader-election-role -n openshift-catalogd because it was missing

openshift-kube-controller-manager-operator

kube-controller-manager-operator-kubecontrollermanagerstaticresources-kubecontrollermanagerstaticresources-staticresources

kube-controller-manager-operator

ServiceAccountCreated

Created ServiceAccount/kube-controller-manager-sa -n openshift-kube-controller-manager because it was missing

openshift-kube-controller-manager-operator

kube-controller-manager-operator-targetconfigcontroller

kube-controller-manager-operator

ConfigMapCreated

Created ConfigMap/recycler-config -n openshift-kube-controller-manager because it was missing
(x6)

openshift-multus

kubelet

multus-admission-controller-5dbbb8b86f-5rrrh

FailedMount

MountVolume.SetUp failed for volume "webhook-certs" : secret "multus-admission-controller-secret" not found

openshift-cluster-storage-operator

kubelet

cluster-storage-operator-7d87854d6-vlq7h

Started

Started container cluster-storage-operator

openshift-cluster-storage-operator

cluster-storage-operator

cluster-storage-operator-lock

LeaderElection

cluster-storage-operator-7d87854d6-vlq7h_996c6839-e9f9-41c2-a2c1-9363c7c2814b became leader
(x6)

openshift-multus

kubelet

network-metrics-daemon-srdjm

FailedMount

MountVolume.SetUp failed for volume "metrics-certs" : secret "metrics-daemon-secret" not found

openshift-cluster-storage-operator

cluster-storage-operator

cluster-storage-operator

FeatureGatesInitialized

FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{"AWSEFSDriverVolumeMetrics", "AdminNetworkPolicy", "AlibabaPlatform", "AzureWorkloadIdentity", "BareMetalLoadBalancer", "BuildCSIVolumes", "ChunkSizeMiB", "CloudDualStackNodeIPs", "DisableKubeletCloudCredentialProviders", "GCPLabelsTags", "HardwareSpeed", "IngressControllerLBSubnetsAWS", "KMSv1", "ManagedBootImages", "ManagedBootImagesAWS", "MultiArchInstallAWS", "MultiArchInstallGCP", "NetworkDiagnosticsConfig", "NetworkLiveMigration", "NetworkSegmentation", "NewOLM", "NodeDisruptionPolicy", "OnClusterBuild", "PersistentIPsForVirtualization", "PrivateHostedZoneAWS", "SetEIPForNLBIngressController", "StreamingCollectionEncodingToJSON", "StreamingCollectionEncodingToProtobuf", "VSphereControlPlaneMachineSet", "VSphereDriverConfiguration", "VSphereMultiVCenters", "VSphereStaticIPs", "ValidatingAdmissionPolicy"}, Disabled:[]v1.FeatureGateName{"AWSClusterHostedDNS", "AdditionalRoutingCapabilities", "AutomatedEtcdBackup", "BootcNodeManagement", "CSIDriverSharedResource", "ClusterAPIInstall", "ClusterAPIInstallIBMCloud", "ClusterMonitoringConfig", "ConsolePluginContentSecurityPolicy", "DNSNameResolver", "DynamicResourceAllocation", "EtcdBackendQuota", "EventedPLEG", "Example", "ExternalOIDC", "ExternalOIDCWithUIDAndExtraClaimMappings", "GCPClusterHostedDNS", "GatewayAPI", "ImageStreamImportMode", "IngressControllerDynamicConfigurationManager", "InsightsConfig", "InsightsConfigAPI", "InsightsOnDemandDataGather", "InsightsRuntimeExtractor", "MachineAPIMigration", "MachineAPIOperatorDisableMachineHealthCheckController", "MachineAPIProviderOpenStack", "MachineConfigNodes", "MaxUnavailableStatefulSet", "MetricsCollectionProfiles", "MinimumKubeletVersion", "MixedCPUsAllocation", "MultiArchInstallAzure", "NodeSwap", "NutanixMultiSubnets", "OVNObservability", "OpenShiftPodSecurityAdmission", "PinnedImages", "PlatformOperators", "ProcMountType", "RouteAdvertisements", "RouteExternalCertificate", "ServiceAccountTokenNodeBinding", "SignatureStores", "SigstoreImageVerification", "TranslateStreamCloseWebsocketRequests", "UpgradeStatus", "UserNamespacesPodSecurityStandards", "UserNamespacesSupport", "VSphereMultiNetworks", "VolumeAttributesClass", "VolumeGroupSnapshot"}}

openshift-cluster-olm-operator

OperatorControllerStaticResources-operatorcontrollerstaticresources-operatorcontrollerstaticresources-staticresources

cluster-olm-operator

ServiceAccountCreated

Created ServiceAccount/operator-controller-controller-manager -n openshift-operator-controller because it was missing
(x6)

openshift-machine-config-operator

kubelet

machine-config-operator-84d549f6d5-gm4qr

FailedMount

MountVolume.SetUp failed for volume "proxy-tls" : secret "mco-proxy-tls" not found

openshift-cluster-olm-operator

olm-status-controller-statussyncer_olm

cluster-olm-operator

OperatorStatusChanged

Status for clusteroperator/olm changed: Degraded set to Unknown (""),Progressing set to Unknown (""),Available set to Unknown (""),Upgradeable set to Unknown (""),EvaluationConditionsDetected set to Unknown (""),status.relatedObjects changed from [{"" "namespaces" "" "openshift-cluster-olm-operator"} {"operator.openshift.io" "olms" "" "cluster"}] to [{"" "namespaces" "" "openshift-catalogd"} {"apiextensions.k8s.io" "customresourcedefinitions" "" "clustercatalogs.olm.operatorframework.io"} {"" "serviceaccounts" "openshift-catalogd" "catalogd-controller-manager"} {"rbac.authorization.k8s.io" "roles" "openshift-catalogd" "catalogd-leader-election-role"} {"rbac.authorization.k8s.io" "roles" "openshift-config" "catalogd-manager-role"} {"rbac.authorization.k8s.io" "clusterroles" "" "catalogd-manager-role"} {"rbac.authorization.k8s.io" "clusterroles" "" "catalogd-metrics-reader"} {"rbac.authorization.k8s.io" "clusterroles" "" "catalogd-proxy-role"} {"rbac.authorization.k8s.io" "rolebindings" "openshift-catalogd" "catalogd-leader-election-rolebinding"} {"rbac.authorization.k8s.io" "rolebindings" "openshift-config" "catalogd-manager-rolebinding"} {"rbac.authorization.k8s.io" "clusterrolebindings" "" "catalogd-manager-rolebinding"} {"rbac.authorization.k8s.io" "clusterrolebindings" "" "catalogd-proxy-rolebinding"} {"" "configmaps" "openshift-catalogd" "catalogd-trusted-ca-bundle"} {"" "services" "openshift-catalogd" "catalogd-service"} {"apps" "deployments" "openshift-catalogd" "catalogd-controller-manager"} {"olm.operatorframework.io" "clustercatalogs" "" "openshift-certified-operators"} {"olm.operatorframework.io" "clustercatalogs" "" "openshift-community-operators"} {"olm.operatorframework.io" "clustercatalogs" "" "openshift-redhat-marketplace"} {"olm.operatorframework.io" "clustercatalogs" "" "openshift-redhat-operators"} {"admissionregistration.k8s.io" "mutatingwebhookconfigurations" "" "catalogd-mutating-webhook-configuration"} {"" "namespaces" "" "openshift-operator-controller"} {"apiextensions.k8s.io" "customresourcedefinitions" "" "clusterextensions.olm.operatorframework.io"} {"" "serviceaccounts" "openshift-operator-controller" "operator-controller-controller-manager"} {"rbac.authorization.k8s.io" "roles" "openshift-config" "operator-controller-manager-role"} {"rbac.authorization.k8s.io" "roles" "openshift-operator-controller" "operator-controller-leader-election-role"} {"rbac.authorization.k8s.io" "roles" "openshift-operator-controller" "operator-controller-manager-role"} {"rbac.authorization.k8s.io" "clusterroles" "" "operator-controller-clusterextension-editor-role"} {"rbac.authorization.k8s.io" "clusterroles" "" "operator-controller-clusterextension-viewer-role"} {"rbac.authorization.k8s.io" "clusterroles" "" "operator-controller-extension-editor-role"} {"rbac.authorization.k8s.io" "clusterroles" "" "operator-controller-extension-viewer-role"} {"rbac.authorization.k8s.io" "clusterroles" "" "operator-controller-manager-role"} {"rbac.authorization.k8s.io" "clusterroles" "" "operator-controller-metrics-reader"} {"rbac.authorization.k8s.io" "clusterroles" "" "operator-controller-proxy-role"} {"rbac.authorization.k8s.io" "rolebindings" "openshift-config" "operator-controller-manager-rolebinding"} {"rbac.authorization.k8s.io" "rolebindings" "openshift-operator-controller" "operator-controller-leader-election-rolebinding"} {"rbac.authorization.k8s.io" "rolebindings" "openshift-operator-controller" "operator-controller-manager-rolebinding"} {"rbac.authorization.k8s.io" "clusterrolebindings" "" "operator-controller-manager-rolebinding"} {"rbac.authorization.k8s.io" "clusterrolebindings" "" "operator-controller-proxy-rolebinding"} {"" "configmaps" "openshift-operator-controller" "operator-controller-trusted-ca-bundle"} {"" "services" "openshift-operator-controller" "operator-controller-controller-manager-metrics-service"} {"apps" "deployments" "openshift-operator-controller" "operator-controller-controller-manager"} {"operator.openshift.io" "olms" "" "cluster"} {"" "namespaces" "" "openshift-cluster-olm-operator"}],status.versions changed from [] to [{"operator" "4.18.35"}]
(x2)

openshift-etcd-operator

openshift-cluster-etcd-operator-etcd-node

etcd-operator

MasterNodesReadyChanged

All master nodes are ready

openshift-etcd-operator

openshift-cluster-etcd-operator-status-controller-statussyncer_etcd

etcd-operator

OperatorStatusChanged

Status for clusteroperator/etcd changed: Upgradeable changed from Unknown to True ("All is well")

openshift-etcd-operator

openshift-cluster-etcd-operator-status-controller-statussyncer_etcd

etcd-operator

OperatorStatusChanged

Status for clusteroperator/etcd changed: Degraded set to False ("EtcdMembersControllerDegraded: getting cache client could not retrieve endpoints: configmaps lister not synced"),Progressing set to Unknown (""),Available set to Unknown (""),Upgradeable set to Unknown (""),EvaluationConditionsDetected set to Unknown (""),status.versions changed from [] to [{"raw-internal" "4.18.35"}]
(x2)

openshift-etcd-operator

openshift-cluster-etcd-operator-status-controller-statussyncer_etcd

etcd-operator

OperatorVersionChanged

clusteroperator/etcd version "raw-internal" changed from "" to "4.18.35"
(x2)

openshift-etcd-operator

openshift-cluster-etcd-operator-etcd-node

etcd-operator

MasterNodeObserved

Observed new master node master-0

openshift-etcd-operator

openshift-cluster-etcd-operator-member-observer-controller-etcdmemberscontroller

etcd-operator

ReportEtcdMembersErrorUpdatingStatus

etcds.operator.openshift.io "cluster" not found

openshift-etcd-operator

openshift-cluster-etcd-operator

etcd-operator

FeatureGatesInitialized

FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{"AWSEFSDriverVolumeMetrics", "AdminNetworkPolicy", "AlibabaPlatform", "AzureWorkloadIdentity", "BareMetalLoadBalancer", "BuildCSIVolumes", "ChunkSizeMiB", "CloudDualStackNodeIPs", "DisableKubeletCloudCredentialProviders", "GCPLabelsTags", "HardwareSpeed", "IngressControllerLBSubnetsAWS", "KMSv1", "ManagedBootImages", "ManagedBootImagesAWS", "MultiArchInstallAWS", "MultiArchInstallGCP", "NetworkDiagnosticsConfig", "NetworkLiveMigration", "NetworkSegmentation", "NewOLM", "NodeDisruptionPolicy", "OnClusterBuild", "PersistentIPsForVirtualization", "PrivateHostedZoneAWS", "SetEIPForNLBIngressController", "StreamingCollectionEncodingToJSON", "StreamingCollectionEncodingToProtobuf", "VSphereControlPlaneMachineSet", "VSphereDriverConfiguration", "VSphereMultiVCenters", "VSphereStaticIPs", "ValidatingAdmissionPolicy"}, Disabled:[]v1.FeatureGateName{"AWSClusterHostedDNS", "AdditionalRoutingCapabilities", "AutomatedEtcdBackup", "BootcNodeManagement", "CSIDriverSharedResource", "ClusterAPIInstall", "ClusterAPIInstallIBMCloud", "ClusterMonitoringConfig", "ConsolePluginContentSecurityPolicy", "DNSNameResolver", "DynamicResourceAllocation", "EtcdBackendQuota", "EventedPLEG", "Example", "ExternalOIDC", "ExternalOIDCWithUIDAndExtraClaimMappings", "GCPClusterHostedDNS", "GatewayAPI", "ImageStreamImportMode", "IngressControllerDynamicConfigurationManager", "InsightsConfig", "InsightsConfigAPI", "InsightsOnDemandDataGather", "InsightsRuntimeExtractor", "MachineAPIMigration", "MachineAPIOperatorDisableMachineHealthCheckController", "MachineAPIProviderOpenStack", "MachineConfigNodes", "MaxUnavailableStatefulSet", "MetricsCollectionProfiles", "MinimumKubeletVersion", "MixedCPUsAllocation", "MultiArchInstallAzure", "NodeSwap", "NutanixMultiSubnets", "OVNObservability", "OpenShiftPodSecurityAdmission", "PinnedImages", "PlatformOperators", "ProcMountType", "RouteAdvertisements", "RouteExternalCertificate", "ServiceAccountTokenNodeBinding", "SignatureStores", "SigstoreImageVerification", "TranslateStreamCloseWebsocketRequests", "UpgradeStatus", "UserNamespacesPodSecurityStandards", "UserNamespacesSupport", "VSphereMultiNetworks", "VolumeAttributesClass", "VolumeGroupSnapshot"}}

openshift-cluster-olm-operator

olm-status-controller-statussyncer_olm

cluster-olm-operator

OperatorStatusChanged

Status for clusteroperator/olm changed: Degraded changed from Unknown to False ("All is well")

openshift-cluster-olm-operator

CatalogdStaticResources-catalogdstaticresources-catalogdstaticresources-staticresources

cluster-olm-operator

ServiceAccountCreated

Created ServiceAccount/catalogd-controller-manager -n openshift-catalogd because it was missing

openshift-cluster-node-tuning-operator

multus

cluster-node-tuning-operator-598fbc5f8f-vxzvg

AddedInterface

Add eth0 [10.128.0.9/23] from ovn-kubernetes

openshift-cluster-node-tuning-operator

kubelet

cluster-node-tuning-operator-598fbc5f8f-vxzvg

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:aa5e782406f71c048b1ac3a4bf5d1227ff4be81111114083ad4c7a209c6bfb5a"

openshift-cluster-olm-operator

OperatorControllerStaticResources-operatorcontrollerstaticresources-operatorcontrollerstaticresources-staticresources

cluster-olm-operator

CustomResourceDefinitionCreated

Created CustomResourceDefinition.apiextensions.k8s.io/clusterextensions.olm.operatorframework.io because it was missing

openshift-cluster-olm-operator

CatalogdStaticResources-catalogdstaticresources-catalogdstaticresources-staticresources

cluster-olm-operator

CustomResourceDefinitionCreated

Created CustomResourceDefinition.apiextensions.k8s.io/clustercatalogs.olm.operatorframework.io because it was missing

openshift-cluster-olm-operator

olm-status-controller-statussyncer_olm

cluster-olm-operator

OperatorStatusChanged

Status for clusteroperator/olm changed: Upgradeable changed from Unknown to True ("All is well")
(x6)

openshift-marketplace

kubelet

marketplace-operator-89ccd998f-mvn4t

FailedMount

MountVolume.SetUp failed for volume "marketplace-operator-metrics" : secret "marketplace-operator-metrics" not found

openshift-cluster-olm-operator

CatalogdStaticResources-catalogdstaticresources-catalogdstaticresources-staticresources

cluster-olm-operator

NamespaceCreated

Created Namespace/openshift-catalogd because it was missing

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Degraded message changed from "OAuthServiceDegraded: Unable to get oauth server service: service \"oauth-openshift\" not found\nRevisionControllerDegraded: configmap \"audit\" not found\nWebhookAuthenticatorControllerDegraded: failed to retrieve service openshift-oauth-apiserver/api: service \"api\" not found\nAPIServerDeploymentDegraded: waiting for .status.latestAvailableRevision to be available\nAPIServerDeploymentDegraded: \nAPIServerWorkloadDegraded: waiting for .status.latestAvailableRevision to be available\nAPIServerWorkloadDegraded: \nIngressStateEndpointsDegraded: No endpoints found for oauth-server\nRouterCertsDegraded: neither the custom secret/v4-0-config-system-router-certs -n openshift-authentication or default secret/v4-0-config-system-custom-router-certs -n openshift-authentication could be retrieved: secret \"v4-0-config-system-router-certs\" not found\nOAuthServerConfigObservationDegraded: secret \"v4-0-config-system-router-certs\" not found" to "OAuthServiceDegraded: Unable to get oauth server service: service \"oauth-openshift\" not found\nRevisionControllerDegraded: configmap \"audit\" not found\nWebhookAuthenticatorControllerDegraded: failed to read service-ca crt bundle: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory\nAPIServerDeploymentDegraded: waiting for .status.latestAvailableRevision to be available\nAPIServerDeploymentDegraded: \nAPIServerWorkloadDegraded: waiting for .status.latestAvailableRevision to be available\nAPIServerWorkloadDegraded: \nIngressStateEndpointsDegraded: No endpoints found for oauth-server\nRouterCertsDegraded: neither the custom secret/v4-0-config-system-router-certs -n openshift-authentication or default secret/v4-0-config-system-custom-router-certs -n openshift-authentication could be retrieved: secret \"v4-0-config-system-router-certs\" not found\nOAuthServerConfigObservationDegraded: secret \"v4-0-config-system-router-certs\" not found"

openshift-etcd-operator

openshift-cluster-etcd-operator-status-controller-statussyncer_etcd

etcd-operator

OperatorStatusChanged

Status for clusteroperator/etcd changed: Degraded message changed from "EtcdMembersControllerDegraded: getting cache client could not retrieve endpoints: configmaps lister not synced" to "EtcdMembersControllerDegraded: getting cache client could not retrieve endpoints: configmaps lister not synced\nNodeControllerDegraded: All master nodes are ready"

openshift-config-operator

config-operator

config-operator-lock

LeaderElection

openshift-config-operator-95bf4f4d-25cml_45d9fa47-ab6d-4497-b38e-3171369cf138 became leader

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Available message changed from "APIServicesAvailable: endpoints \"api\" not found\nOAuthServerServiceEndpointAccessibleControllerAvailable: service \"oauth-openshift\" not found\nOAuthServerServiceEndpointsEndpointAccessibleControllerAvailable: endpoints \"oauth-openshift\" not found\nReadyIngressNodesAvailable: Authentication requires functional ingress which requires at least one schedulable and ready node. Got 0 worker nodes, 1 master nodes, 0 custom target nodes (none are schedulable or ready for ingress pods)." to "APIServicesAvailable: PreconditionNotReady\nOAuthServerServiceEndpointAccessibleControllerAvailable: service \"oauth-openshift\" not found\nOAuthServerServiceEndpointsEndpointAccessibleControllerAvailable: endpoints \"oauth-openshift\" not found\nReadyIngressNodesAvailable: Authentication requires functional ingress which requires at least one schedulable and ready node. Got 0 worker nodes, 1 master nodes, 0 custom target nodes (none are schedulable or ready for ingress pods)."
(x2)

openshift-cluster-storage-operator

cluster-storage-operator-status-controller-statussyncer_storage

cluster-storage-operator

OperatorVersionChanged

clusteroperator/storage version "operator" changed from "" to "4.18.35"
(x2)

openshift-kube-scheduler-operator

kubelet

openshift-kube-scheduler-operator-dddff6458-vx5d7

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ef199844317b7b012879ed8d29f9b6bc37fad8a6fdb336103cbd5cabc74c4302"

openshift-cluster-olm-operator

OperatorControllerStaticResources-operatorcontrollerstaticresources-operatorcontrollerstaticresources-staticresources

cluster-olm-operator

RoleCreated

Created Role.rbac.authorization.k8s.io/operator-controller-manager-role -n openshift-operator-controller because it was missing

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Progressing message changed from "All is well" to "AuthenticatorCertKeyProgressing: All is well"

openshift-cluster-storage-operator

cluster-storage-operator-status-controller-statussyncer_storage

cluster-storage-operator

OperatorStatusChanged

Status for clusteroperator/storage changed: Progressing changed from Unknown to False ("All is well"),Available changed from Unknown to True ("DefaultStorageClassControllerAvailable: No default StorageClass for this platform"),Upgradeable changed from Unknown to True ("All is well")

openshift-cluster-storage-operator

cluster-storage-operator-status-controller-statussyncer_storage

cluster-storage-operator

OperatorStatusChanged

Status for clusteroperator/storage changed: Degraded set to Unknown (""),Progressing set to Unknown (""),Available set to Unknown (""),Upgradeable set to Unknown (""),EvaluationConditionsDetected set to Unknown (""),status.relatedObjects changed from [] to [{"" "namespaces" "" "openshift-cluster-storage-operator"} {"" "namespaces" "" "openshift-cluster-csi-drivers"} {"operator.openshift.io" "storages" "" "cluster"} {"rbac.authorization.k8s.io" "clusterrolebindings" "" "cluster-storage-operator-role"}],status.versions changed from [] to [{"operator" "4.18.35"}]

openshift-cluster-olm-operator

CatalogdStaticResources-catalogdstaticresources-catalogdstaticresources-staticresources

cluster-olm-operator

ClusterRoleCreated

Created ClusterRole.rbac.authorization.k8s.io/catalogd-metrics-reader because it was missing

openshift-authentication-operator

oauth-apiserver-apiserverstaticresources-apiserverstaticresources-staticresources

authentication-operator

ServiceCreated

Created Service/api -n openshift-oauth-apiserver because it was missing

openshift-cluster-storage-operator

cluster-storage-operator-status-controller-statussyncer_storage

cluster-storage-operator

OperatorStatusChanged

Status for clusteroperator/storage changed: Degraded changed from Unknown to False ("All is well")

openshift-config-operator

config-operator-configoperatorcontroller

openshift-config-operator

FastControllerResync

Controller "ConfigOperatorController" resync interval is set to 10s which might lead to client request throttling

openshift-kube-controller-manager-operator

kube-controller-manager-operator-targetconfigcontroller

kube-controller-manager-operator

ConfigMapCreated

Created ConfigMap/csr-signer-ca -n openshift-kube-controller-manager-operator because it was missing

openshift-authentication-operator

cluster-authentication-operator-openshiftauthenticationstaticresources-openshiftauthenticationstaticresources-staticresources

authentication-operator

ConfigMapCreated

Created ConfigMap/v4-0-config-system-trusted-ca-bundle -n openshift-authentication because it was missing

openshift-etcd-operator

openshift-cluster-etcd-operator-config-observer-configobserver

etcd-operator

ObservedConfigChanged

Writing updated observed config:   map[string]any{ +  "controlPlane": map[string]any{"replicas": float64(1)}, +  "servingInfo": map[string]any{ +  "cipherSuites": []any{ +  string("TLS_AES_128_GCM_SHA256"), string("TLS_AES_256_GCM_SHA384"), +  string("TLS_CHACHA20_POLY1305_SHA256"), +  string("TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256"), +  string("TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256"), +  string("TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384"), +  string("TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384"), +  string("TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256"), ..., +  }, +  "minTLSVersion": string("VersionTLS12"), +  },   }

openshift-etcd-operator

openshift-cluster-etcd-operator-config-observer-configobserver

etcd-operator

ObserveTLSSecurityProfile

cipherSuites changed to ["TLS_AES_128_GCM_SHA256" "TLS_AES_256_GCM_SHA384" "TLS_CHACHA20_POLY1305_SHA256" "TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256" "TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256" "TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384" "TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384" "TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256" "TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256"]

openshift-kube-controller-manager-operator

kube-controller-manager-operator-kubecontrollermanagerstaticresources-kubecontrollermanagerstaticresources-staticresources

kube-controller-manager-operator

ServiceAccountCreated

Created ServiceAccount/pv-recycler-controller -n openshift-infra because it was missing

openshift-cluster-olm-operator

CatalogdStaticResources-catalogdstaticresources-catalogdstaticresources-staticresources

cluster-olm-operator

ClusterRoleCreated

Created ClusterRole.rbac.authorization.k8s.io/catalogd-proxy-role because it was missing

openshift-cluster-olm-operator

OperatorControllerStaticResources-operatorcontrollerstaticresources-operatorcontrollerstaticresources-staticresources

cluster-olm-operator

ClusterRoleCreated

Created ClusterRole.rbac.authorization.k8s.io/operator-controller-clusterextension-editor-role because it was missing

openshift-authentication-operator

oauth-apiserver-revisioncontroller

authentication-operator

ConfigMapCreated

Created ConfigMap/audit-1 -n openshift-oauth-apiserver because it was missing

openshift-etcd-operator

openshift-cluster-etcd-operator-resource-sync-controller-resourcesynccontroller

etcd-operator

ConfigMapCreated

Created ConfigMap/cluster-config-v1 -n openshift-etcd because it was missing

openshift-cluster-olm-operator

OperatorControllerStaticResources-operatorcontrollerstaticresources-operatorcontrollerstaticresources-staticresources

cluster-olm-operator

ClusterRoleCreated

Created ClusterRole.rbac.authorization.k8s.io/operator-controller-clusterextension-viewer-role because it was missing

openshift-etcd-operator

openshift-cluster-etcd-operator-script-controller-scriptcontroller

etcd-operator

ScriptControllerErrorUpdatingStatus

Operation cannot be fulfilled on etcds.operator.openshift.io "cluster": the object has been modified; please apply your changes to the latest version and try again
(x5)

openshift-route-controller-manager

kubelet

route-controller-manager-5ff6bdb8cf-b6499

FailedMount

MountVolume.SetUp failed for volume "serving-cert" : secret "serving-cert" not found
(x5)

openshift-route-controller-manager

kubelet

route-controller-manager-5ff6bdb8cf-b6499

FailedMount

MountVolume.SetUp failed for volume "client-ca" : configmap "client-ca" not found

openshift-kube-controller-manager-operator

kube-controller-manager-operator-resource-sync-controller-resourcesynccontroller

kube-controller-manager-operator

ConfigMapCreated

Created ConfigMap/csr-controller-ca -n openshift-config-managed because it was missing

openshift-kube-controller-manager-operator

kube-controller-manager-operator-targetconfigcontroller

kube-controller-manager-operator

ConfigMapCreated

Created ConfigMap/csr-controller-ca -n openshift-kube-controller-manager-operator because it was missing

openshift-kube-controller-manager-operator

kube-controller-manager-operator-kubecontrollermanagerstaticresources-kubecontrollermanagerstaticresources-staticresources

kube-controller-manager-operator

ClusterRoleBindingCreated

Created ClusterRoleBinding.rbac.authorization.k8s.io/system:openshift:operator:kube-controller-manager-recovery because it was missing

openshift-etcd-operator

openshift-cluster-etcd-operator-backingresourcecontroller-backingresourcecontroller-staticresources

etcd-operator

ServiceAccountCreated

Created ServiceAccount/installer-sa -n openshift-etcd because it was missing

openshift-apiserver-operator

openshift-apiserver-operator-openshiftapiserver-workloadworkloadcontroller

openshift-apiserver-operator

DeploymentCreated

Created Deployment.apps/apiserver -n openshift-apiserver because it was missing

openshift-apiserver

deployment-controller

apiserver

ScalingReplicaSet

Scaled up replica set apiserver-c765cd67b to 1

openshift-etcd-operator

openshift-cluster-etcd-operator-resource-sync-controller-resourcesynccontroller

etcd-operator

ConfigMapUpdated

Updated ConfigMap/etcd-ca-bundle -n openshift-etcd-operator: cause by changes in data.ca-bundle.crt

openshift-etcd-operator

openshift-cluster-etcd-operator-etcdstaticresources-etcdstaticresources-staticresources

etcd-operator

NamespaceUpdated

Updated Namespace/openshift-etcd because it changed
(x2)

openshift-kube-apiserver-operator

kube-apiserver-operator-resource-sync-controller-resourcesynccontroller

kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/kubelet-serving-ca -n openshift-kube-apiserver because it was missing

openshift-kube-controller-manager-operator

kube-controller-manager-operator-kubecontrollermanagerstaticresources-kubecontrollermanagerstaticresources-staticresources

kube-controller-manager-operator

ServiceAccountCreated

Created ServiceAccount/localhost-recovery-client -n openshift-kube-controller-manager because it was missing

openshift-cluster-olm-operator

OperatorControllerStaticResources-operatorcontrollerstaticresources-operatorcontrollerstaticresources-staticresources

cluster-olm-operator

ClusterRoleCreated

Created ClusterRole.rbac.authorization.k8s.io/operator-controller-extension-editor-role because it was missing

openshift-cluster-olm-operator

CatalogdStaticResources-catalogdstaticresources-catalogdstaticresources-staticresources

cluster-olm-operator

RoleBindingCreated

Created RoleBinding.rbac.authorization.k8s.io/catalogd-leader-election-rolebinding -n openshift-catalogd because it was missing

openshift-apiserver-operator

openshift-apiserver-operator-status-controller-statussyncer_openshift-apiserver

openshift-apiserver-operator

OperatorStatusChanged

Status for clusteroperator/openshift-apiserver changed: Progressing message changed from "OperatorConfigProgressing: openshiftapiserveroperatorconfigs/instance: observed generation is 0, desired generation is 2." to "APIServerDeploymentProgressing: deployment/apiserver.openshift-apiserver: observed generation is 0, desired generation is 1.\nOperatorConfigProgressing: openshiftapiserveroperatorconfigs/instance: observed generation is 0, desired generation is 2.",Available message changed from "APIServicesAvailable: PreconditionNotReady" to "APIServerDeploymentAvailable: no apiserver.openshift-apiserver pods available on any node.\nAPIServicesAvailable: PreconditionNotReady"

openshift-kube-apiserver-operator

kube-apiserver-operator-resource-sync-controller-resourcesynccontroller

kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/etcd-serving-ca -n openshift-kube-apiserver because it was missing

openshift-cluster-olm-operator

CatalogdStaticResources-catalogdstaticresources-catalogdstaticresources-staticresources

cluster-olm-operator

ClusterRoleBindingCreated

Created ClusterRoleBinding.rbac.authorization.k8s.io/catalogd-manager-rolebinding because it was missing
(x2)

openshift-kube-apiserver-operator

kube-apiserver-operator-resource-sync-controller-resourcesynccontroller

kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/kubelet-serving-ca -n openshift-config-managed because it was missing

openshift-apiserver

replicaset-controller

apiserver-c765cd67b

SuccessfulCreate

Created pod: apiserver-c765cd67b-xddmb

openshift-cluster-olm-operator

OperatorControllerStaticResources-operatorcontrollerstaticresources-operatorcontrollerstaticresources-staticresources

cluster-olm-operator

ClusterRoleCreated

Created ClusterRole.rbac.authorization.k8s.io/operator-controller-metrics-reader because it was missing

openshift-etcd-operator

openshift-cluster-etcd-operator-backingresourcecontroller-backingresourcecontroller-staticresources

etcd-operator

ClusterRoleBindingCreated

Created ClusterRoleBinding.rbac.authorization.k8s.io/system:openshift:operator:openshift-etcd-installer because it was missing

openshift-etcd-operator

openshift-cluster-etcd-operator-resource-sync-controller-resourcesynccontroller

etcd-operator

ConfigMapCreated

Created ConfigMap/etcd-ca-bundle -n openshift-config because it was missing

openshift-etcd-operator

openshift-cluster-etcd-operator-etcdstaticresources-etcdstaticresources-staticresources

etcd-operator

ServiceAccountCreated

Created ServiceAccount/etcd-sa -n openshift-etcd because it was missing

openshift-etcd-operator

openshift-cluster-etcd-operator-resource-sync-controller-resourcesynccontroller

etcd-operator

ConfigMapCreated

Created ConfigMap/etcd-serving-ca -n openshift-config because it was missing

openshift-cluster-olm-operator

OperatorControllerStaticResources-operatorcontrollerstaticresources-operatorcontrollerstaticresources-staticresources

cluster-olm-operator

ClusterRoleCreated

Created ClusterRole.rbac.authorization.k8s.io/operator-controller-manager-role because it was missing

openshift-cluster-olm-operator

CatalogdStaticResources-catalogdstaticresources-catalogdstaticresources-staticresources

cluster-olm-operator

RoleBindingCreated

Created RoleBinding.rbac.authorization.k8s.io/catalogd-manager-rolebinding -n openshift-config because it was missing

openshift-cluster-olm-operator

OperatorControllerStaticResources-operatorcontrollerstaticresources-operatorcontrollerstaticresources-staticresources

cluster-olm-operator

ClusterRoleCreated

Created ClusterRole.rbac.authorization.k8s.io/operator-controller-extension-viewer-role because it was missing

openshift-apiserver-operator

openshift-apiserver-operator-resource-sync-controller-resourcesynccontroller

openshift-apiserver-operator

ConfigMapCreated

Created ConfigMap/etcd-serving-ca -n openshift-apiserver because it was missing

openshift-apiserver-operator

openshift-apiserver-operator-status-controller-statussyncer_openshift-apiserver

openshift-apiserver-operator

OperatorStatusChanged

Status for clusteroperator/openshift-apiserver changed: Progressing changed from False to True ("OperatorConfigProgressing: openshiftapiserveroperatorconfigs/instance: observed generation is 0, desired generation is 2.")

openshift-kube-controller-manager-operator

kube-controller-manager-operator-status-controller-statussyncer_kube-controller-manager

kube-controller-manager-operator

OperatorStatusChanged

Status for clusteroperator/kube-controller-manager changed: Degraded message changed from "NodeControllerDegraded: All master nodes are ready\nRevisionControllerDegraded: configmap \"kube-controller-manager-pod\" not found\nInstallerControllerDegraded: missing required resources: [configmaps: cluster-policy-controller-config-0,config-0,controller-manager-kubeconfig-0,kube-controller-cert-syncer-kubeconfig-0,kube-controller-manager-pod-0,recycler-config-0,service-ca-0,serviceaccount-ca-0, secrets: localhost-recovery-client-token-0,service-account-private-key-0]\nResourceSyncControllerDegraded: configmaps \"csr-controller-ca\" already exists" to "NodeControllerDegraded: All master nodes are ready\nRevisionControllerDegraded: configmap \"kube-controller-manager-pod\" not found\nInstallerControllerDegraded: missing required resources: [configmaps: cluster-policy-controller-config-0,config-0,controller-manager-kubeconfig-0,kube-controller-cert-syncer-kubeconfig-0,kube-controller-manager-pod-0,recycler-config-0,service-ca-0,serviceaccount-ca-0, secrets: localhost-recovery-client-token-0,service-account-private-key-0]"
(x2)

openshift-kube-apiserver-operator

kube-apiserver-operator-config-observer-configobserver

kube-apiserver-operator

ObservedConfigWriteError

Failed to write observed config: Operation cannot be fulfilled on kubeapiservers.operator.openshift.io "cluster": the object has been modified; please apply your changes to the latest version and try again

openshift-apiserver

default-scheduler

apiserver-c765cd67b-xddmb

Scheduled

Successfully assigned openshift-apiserver/apiserver-c765cd67b-xddmb to master-0

openshift-authentication-operator

cluster-authentication-operator-openshiftauthenticationstaticresources-openshiftauthenticationstaticresources-staticresources

authentication-operator

SecretCreated

Created Secret/v4-0-config-system-ocp-branding-template -n openshift-authentication because it was missing

openshift-kube-controller-manager-operator

kube-controller-manager-operator-kubecontrollermanagerstaticresources-kubecontrollermanagerstaticresources-staticresources

kube-controller-manager-operator

ClusterRoleBindingCreated

Created ClusterRoleBinding.rbac.authorization.k8s.io/system:openshift:controller:cluster-csr-approver-controller because it was missing

openshift-kube-controller-manager-operator

kube-controller-manager-operator-status-controller-statussyncer_kube-controller-manager

kube-controller-manager-operator

OperatorStatusChanged

Status for clusteroperator/kube-controller-manager changed: Degraded message changed from "NodeControllerDegraded: All master nodes are ready\nRevisionControllerDegraded: configmap \"kube-controller-manager-pod\" not found\nInstallerControllerDegraded: missing required resources: [configmaps: cluster-policy-controller-config-0,config-0,controller-manager-kubeconfig-0,kube-controller-cert-syncer-kubeconfig-0,kube-controller-manager-pod-0,recycler-config-0,service-ca-0,serviceaccount-ca-0, secrets: localhost-recovery-client-token-0,service-account-private-key-0]" to "NodeControllerDegraded: All master nodes are ready\nRevisionControllerDegraded: configmap \"kube-controller-manager-pod\" not found\nInstallerControllerDegraded: missing required resources: [configmaps: cluster-policy-controller-config-0,config-0,controller-manager-kubeconfig-0,kube-controller-cert-syncer-kubeconfig-0,kube-controller-manager-pod-0,recycler-config-0,service-ca-0,serviceaccount-ca-0, secrets: localhost-recovery-client-token-0,service-account-private-key-0]\nResourceSyncControllerDegraded: configmaps \"csr-controller-ca\" already exists"

openshift-kube-controller-manager-operator

kube-controller-manager-operator-kubecontrollermanagerstaticresources-kubecontrollermanagerstaticresources-staticresources

kube-controller-manager-operator

ClusterRoleCreated

Created ClusterRole.rbac.authorization.k8s.io/system:openshift:controller:cluster-csr-approver-controller because it was missing

openshift-kube-controller-manager-operator

kube-controller-manager-operator-targetconfigcontroller

kube-controller-manager-operator

SecretCreated

Created Secret/csr-signer -n openshift-kube-controller-manager because it was missing

openshift-kube-controller-manager-operator

kube-controller-manager-operator-resource-sync-controller-resourcesynccontroller

kube-controller-manager-operator

ConfigMapCreateFailed

Failed to create ConfigMap/csr-controller-ca -n openshift-config-managed: configmaps "csr-controller-ca" already exists

openshift-kube-controller-manager-operator

kube-controller-manager-operator-kubecontrollermanagerstaticresources-kubecontrollermanagerstaticresources-staticresources

kube-controller-manager-operator

SecretCreated

Created Secret/localhost-recovery-client-token -n openshift-kube-controller-manager because it was missing
(x3)

openshift-kube-apiserver-operator

kube-apiserver-operator-config-observer-configobserver

kube-apiserver-operator

ObserveTLSSecurityProfile

minTLSVersion changed to VersionTLS12

openshift-cluster-olm-operator

CatalogdStaticResources-catalogdstaticresources-catalogdstaticresources-staticresources

cluster-olm-operator

ClusterRoleBindingCreated

Created ClusterRoleBinding.rbac.authorization.k8s.io/catalogd-proxy-rolebinding because it was missing

openshift-cluster-olm-operator

CatalogdStaticResources-catalogdstaticresources-catalogdstaticresources-staticresources

cluster-olm-operator

MutatingWebhookConfigurationCreated

Created MutatingWebhookConfiguration.admissionregistration.k8s.io/catalogd-mutating-webhook-configuration because it was missing

openshift-cluster-olm-operator

CatalogdStaticResources-catalogdstaticresources-catalogdstaticresources-staticresources

cluster-olm-operator

ServiceCreated

Created Service/catalogd-service -n openshift-catalogd because it was missing

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Degraded message changed from "OAuthServiceDegraded: Unable to get oauth server service: service \"oauth-openshift\" not found\nRevisionControllerDegraded: configmap \"audit\" not found\nWebhookAuthenticatorControllerDegraded: failed to read service-ca crt bundle: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory\nAPIServerDeploymentDegraded: waiting for .status.latestAvailableRevision to be available\nAPIServerDeploymentDegraded: \nAPIServerWorkloadDegraded: waiting for .status.latestAvailableRevision to be available\nAPIServerWorkloadDegraded: \nIngressStateEndpointsDegraded: No endpoints found for oauth-server\nRouterCertsDegraded: neither the custom secret/v4-0-config-system-router-certs -n openshift-authentication or default secret/v4-0-config-system-custom-router-certs -n openshift-authentication could be retrieved: secret \"v4-0-config-system-router-certs\" not found\nOAuthServerConfigObservationDegraded: secret \"v4-0-config-system-router-certs\" not found" to "OAuthServiceDegraded: Unable to get oauth server service: service \"oauth-openshift\" not found\nWebhookAuthenticatorControllerDegraded: failed to read service-ca crt bundle: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory\nAPIServerDeploymentDegraded: waiting for .status.latestAvailableRevision to be available\nAPIServerDeploymentDegraded: \nAPIServerWorkloadDegraded: waiting for .status.latestAvailableRevision to be available\nAPIServerWorkloadDegraded: \nIngressStateEndpointsDegraded: No endpoints found for oauth-server\nRouterCertsDegraded: neither the custom secret/v4-0-config-system-router-certs -n openshift-authentication or default secret/v4-0-config-system-custom-router-certs -n openshift-authentication could be retrieved: secret \"v4-0-config-system-router-certs\" not found\nOAuthServerConfigObservationDegraded: secret \"v4-0-config-system-router-certs\" not found"

openshift-authentication-operator

oauth-apiserver-revisioncontroller

authentication-operator

RevisionTriggered

new revision 1 triggered by "configmap \"audit-0\" not found"

openshift-authentication-operator

oauth-apiserver-apiserverstaticresources-apiserverstaticresources-staticresources

authentication-operator

ClusterRoleCreated

Created ClusterRole.rbac.authorization.k8s.io/system:openshift:useroauthaccesstoken-manager because it was missing

openshift-authentication-operator

oauth-apiserver-apiserverstaticresources-apiserverstaticresources-staticresources

authentication-operator

ClusterRoleBindingCreated

Created ClusterRoleBinding.rbac.authorization.k8s.io/system:openshift:useroauthaccesstoken-manager because it was missing

openshift-authentication-operator

oauth-apiserver-apiserverstaticresources-apiserverstaticresources-staticresources

authentication-operator

ServiceAccountCreated

Created ServiceAccount/oauth-apiserver-sa -n openshift-oauth-apiserver because it was missing

openshift-cluster-olm-operator

CatalogdStaticResources-catalogdstaticresources-catalogdstaticresources-staticresources

cluster-olm-operator

ConfigMapCreated

Created ConfigMap/catalogd-trusted-ca-bundle -n openshift-catalogd because it was missing

openshift-kube-controller-manager-operator

kube-controller-manager-operator-targetconfigcontroller

kube-controller-manager-operator

ConfigMapCreated

Created ConfigMap/serviceaccount-ca -n openshift-kube-controller-manager because it was missing

openshift-etcd-operator

openshift-cluster-etcd-operator-etcdstaticresources-etcdstaticresources-staticresources

etcd-operator

ServiceMonitorCreated

Created ServiceMonitor.monitoring.coreos.com/etcd -n openshift-etcd-operator because it was missing

openshift-cluster-olm-operator

OperatorControllerStaticResources-operatorcontrollerstaticresources-operatorcontrollerstaticresources-staticresources

cluster-olm-operator

ClusterRoleCreated

Created ClusterRole.rbac.authorization.k8s.io/operator-controller-proxy-role because it was missing

openshift-etcd-operator

openshift-cluster-etcd-operator-etcdstaticresources-etcdstaticresources-staticresources

etcd-operator

ServiceMonitorCreated

Created ServiceMonitor.monitoring.coreos.com/etcd-minimal -n openshift-etcd-operator because it was missing

openshift-kube-scheduler-operator

kubelet

openshift-kube-scheduler-operator-dddff6458-vx5d7

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ef199844317b7b012879ed8d29f9b6bc37fad8a6fdb336103cbd5cabc74c4302" in 5.426s (5.426s including waiting). Image size: 506395599 bytes.

openshift-cluster-olm-operator

OperatorControllerStaticResources-operatorcontrollerstaticresources-operatorcontrollerstaticresources-staticresources

cluster-olm-operator

RoleBindingCreated

Created RoleBinding.rbac.authorization.k8s.io/operator-controller-manager-rolebinding -n openshift-config because it was missing
(x5)

openshift-etcd-operator

openshift-cluster-etcd-operator-installer-controller

etcd-operator

RequiredInstallerResourcesMissing

configmaps: etcd-all-bundles-0,etcd-endpoints-0,etcd-pod-0, secrets: etcd-all-certs-0
(x3)

openshift-kube-apiserver-operator

kube-apiserver-operator-config-observer-configobserver

kube-apiserver-operator

ObserveTLSSecurityProfile

cipherSuites changed to ["TLS_AES_128_GCM_SHA256" "TLS_AES_256_GCM_SHA384" "TLS_CHACHA20_POLY1305_SHA256" "TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256" "TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256" "TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384" "TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384" "TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256" "TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256"]
(x2)

openshift-etcd-operator

openshift-cluster-etcd-operator-env-var-controller

etcd-operator

EnvVarControllerUpdatingStatus

Operation cannot be fulfilled on etcds.operator.openshift.io "cluster": the object has been modified; please apply your changes to the latest version and try again

openshift-etcd-operator

openshift-cluster-etcd-operator-resource-sync-controller-resourcesynccontroller

etcd-operator

ConfigMapCreated

Created ConfigMap/etcd-metric-serving-ca -n openshift-etcd-operator because it was missing

openshift-etcd-operator

openshift-cluster-etcd-operator-etcdstaticresources-etcdstaticresources-staticresources

etcd-operator

ServiceUpdated

Updated Service/etcd -n openshift-etcd because it changed
(x3)

openshift-apiserver

kubelet

apiserver-c765cd67b-xddmb

FailedMount

MountVolume.SetUp failed for volume "etcd-client" : secret "etcd-client" not found

openshift-etcd-operator

openshift-cluster-etcd-operator-config-observer-configobserver

etcd-operator

ObservedConfigWriteError

Failed to write observed config: Operation cannot be fulfilled on etcds.operator.openshift.io "cluster": the object has been modified; please apply your changes to the latest version and try again
(x3)

openshift-kube-apiserver-operator

kube-apiserver-operator-config-observer-configobserver

kube-apiserver-operator

ObserveStorageUpdated

Updated storage urls to https://192.168.32.10:2379,https://localhost:2379

openshift-cluster-olm-operator

OperatorControllerStaticResources-operatorcontrollerstaticresources-operatorcontrollerstaticresources-staticresources

cluster-olm-operator

RoleBindingCreated

Created RoleBinding.rbac.authorization.k8s.io/operator-controller-leader-election-rolebinding -n openshift-operator-controller because it was missing
(x3)

openshift-apiserver

kubelet

apiserver-c765cd67b-xddmb

FailedMount

MountVolume.SetUp failed for volume "serving-cert" : secret "serving-cert" not found

openshift-authentication-operator

cluster-authentication-operator-resource-sync-controller-resourcesynccontroller

authentication-operator

ConfigMapCreated

Created ConfigMap/etcd-serving-ca -n openshift-oauth-apiserver because it was missing
(x2)

openshift-etcd-operator

openshift-cluster-etcd-operator-config-observer-configobserver

etcd-operator

ObserveTLSSecurityProfile

minTLSVersion changed to VersionTLS12

openshift-apiserver-operator

openshift-apiserver-operator-resource-sync-controller-resourcesynccontroller

openshift-apiserver-operator

SecretCreated

Created Secret/etcd-client -n openshift-apiserver because it was missing
(x3)

openshift-kube-apiserver-operator

kube-apiserver-operator-config-observer-configobserver

kube-apiserver-operator

ObservedConfigChanged

Writing updated observed config: map[string]any{ + "admission": map[string]any{ + "pluginConfig": map[string]any{ + "PodSecurity": map[string]any{"configuration": map[string]any{...}}, + "network.openshift.io/ExternalIPRanger": map[string]any{"configuration": map[string]any{...}}, + "network.openshift.io/RestrictedEndpointsAdmission": map[string]any{"configuration": map[string]any{...}}, + }, + }, + "apiServerArguments": map[string]any{ + "api-audiences": []any{string("https://kubernetes.default.svc")}, + "etcd-servers": []any{string("https://192.168.32.10:2379"), string("https://localhost:2379")}, + "feature-gates": []any{ + string("AWSEFSDriverVolumeMetrics=true"), string("AdminNetworkPolicy=true"), + string("AlibabaPlatform=true"), string("AzureWorkloadIdentity=true"), + string("BareMetalLoadBalancer=true"), string("BuildCSIVolumes=true"), + string("ChunkSizeMiB=true"), string("CloudDualStackNodeIPs=true"), ..., + }, + "goaway-chance": []any{string("0")}, + "runtime-config": []any{string("admissionregistration.k8s.io/v1beta1=true")}, + "send-retry-after-while-not-ready-once": []any{string("true")}, + "service-account-issuer": []any{string("https://kubernetes.default.svc")}, + "service-account-jwks-uri": []any{string("https://api.sno.openstack.lab:6443/openid/v1/jwks")}, + "shutdown-delay-duration": []any{string("0s")}, + }, + "corsAllowedOrigins": []any{string(`//127\.0\.0\.1(:|$)`), string("//localhost(:|$)")}, + "gracefulTerminationDuration": string("15"), + "servicesSubnet": string("172.30.0.0/16"), + "servingInfo": map[string]any{ + "bindAddress": string("0.0.0.0:6443"), + "bindNetwork": string("tcp4"), + "cipherSuites": []any{ + string("TLS_AES_128_GCM_SHA256"), string("TLS_AES_256_GCM_SHA384"), + string("TLS_CHACHA20_POLY1305_SHA256"), + string("TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256"), + string("TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256"), + string("TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384"), + string("TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384"), + string("TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256"), ..., + }, + "minTLSVersion": string("VersionTLS12"), + "namedCertificates": []any{ + map[string]any{ + "certFile": string("/etc/kubernetes/static-pod-certs"...), + "keyFile": string("/etc/kubernetes/static-pod-certs"...), + }, + map[string]any{ + "certFile": string("/etc/kubernetes/static-pod-certs"...), + "keyFile": string("/etc/kubernetes/static-pod-certs"...), + }, + map[string]any{ + "certFile": string("/etc/kubernetes/static-pod-certs"...), + "keyFile": string("/etc/kubernetes/static-pod-certs"...), + }, + map[string]any{ + "certFile": string("/etc/kubernetes/static-pod-certs"...), + "keyFile": string("/etc/kubernetes/static-pod-certs"...), + }, + map[string]any{ + "certFile": string("/etc/kubernetes/static-pod-resou"...), + "keyFile": string("/etc/kubernetes/static-pod-resou"...), + }, + }, + }, }

openshift-kube-apiserver-operator

kube-apiserver-operator-resource-sync-controller-resourcesynccontroller

kube-apiserver-operator

SecretCreated

Created Secret/etcd-client -n openshift-kube-apiserver because it was missing
(x102)

openshift-kube-apiserver-operator

kube-apiserver-operator-target-config-controller-targetconfigcontroller

kube-apiserver-operator

ConfigMissing

no observedConfig

openshift-kube-apiserver-operator

kube-apiserver-operator-status-controller-statussyncer_kube-apiserver

kube-apiserver-operator

OperatorStatusChanged

Status for clusteroperator/kube-apiserver changed: Degraded message changed from "NodeControllerDegraded: All master nodes are ready\nNodeKubeconfigControllerDegraded: \"secret/node-kubeconfigs\": configmap \"kube-apiserver-server-ca\" not found\nRevisionControllerDegraded: configmap \"kube-apiserver-pod\" not found\nInstallerControllerDegraded: missing required resources: [configmaps: bound-sa-token-signing-certs-0,config-0,etcd-serving-ca-0,kube-apiserver-audit-policies-0,kube-apiserver-cert-syncer-kubeconfig-0,kube-apiserver-pod-0,kubelet-serving-ca-0,sa-token-signing-certs-0, secrets: etcd-client-0,localhost-recovery-client-token-0,localhost-recovery-serving-certkey-0]\nCertRotation_CheckEndpointsClient_Degraded: configmaps \"kube-control-plane-signer-ca\" already exists\nConfigObservationDegraded: error writing updated observed config: Operation cannot be fulfilled on kubeapiservers.operator.openshift.io \"cluster\": the object has been modified; please apply your changes to the latest version and try again" to "NodeControllerDegraded: All master nodes are ready\nNodeKubeconfigControllerDegraded: \"secret/node-kubeconfigs\": configmap \"kube-apiserver-server-ca\" not found\nRevisionControllerDegraded: configmap \"kube-apiserver-pod\" not found\nInstallerControllerDegraded: missing required resources: [configmaps: bound-sa-token-signing-certs-0,config-0,etcd-serving-ca-0,kube-apiserver-audit-policies-0,kube-apiserver-cert-syncer-kubeconfig-0,kube-apiserver-pod-0,kubelet-serving-ca-0,sa-token-signing-certs-0, secrets: etcd-client-0,localhost-recovery-client-token-0,localhost-recovery-serving-certkey-0]\nConfigObservationDegraded: error writing updated observed config: Operation cannot be fulfilled on kubeapiservers.operator.openshift.io \"cluster\": the object has been modified; please apply your changes to the latest version and try again"

openshift-cluster-olm-operator

OperatorControllerStaticResources-operatorcontrollerstaticresources-operatorcontrollerstaticresources-staticresources

cluster-olm-operator

RoleBindingCreated

Created RoleBinding.rbac.authorization.k8s.io/operator-controller-manager-rolebinding -n openshift-operator-controller because it was missing
(x3)

openshift-kube-apiserver-operator

kube-apiserver-operator-config-observer-configobserver

kube-apiserver-operator

ObserveFeatureFlagsUpdated

Updated apiServerArguments.feature-gates to AWSEFSDriverVolumeMetrics=true,AdminNetworkPolicy=true,AlibabaPlatform=true,AzureWorkloadIdentity=true,BareMetalLoadBalancer=true,BuildCSIVolumes=true,ChunkSizeMiB=true,CloudDualStackNodeIPs=true,DisableKubeletCloudCredentialProviders=true,GCPLabelsTags=true,HardwareSpeed=true,IngressControllerLBSubnetsAWS=true,KMSv1=true,ManagedBootImages=true,ManagedBootImagesAWS=true,MultiArchInstallAWS=true,MultiArchInstallGCP=true,NetworkDiagnosticsConfig=true,NetworkLiveMigration=true,NetworkSegmentation=true,NewOLM=true,NodeDisruptionPolicy=true,OnClusterBuild=true,PersistentIPsForVirtualization=true,PrivateHostedZoneAWS=true,SetEIPForNLBIngressController=true,StreamingCollectionEncodingToJSON=true,StreamingCollectionEncodingToProtobuf=true,VSphereControlPlaneMachineSet=true,VSphereDriverConfiguration=true,VSphereMultiVCenters=true,VSphereStaticIPs=true,ValidatingAdmissionPolicy=true,AWSClusterHostedDNS=false,AdditionalRoutingCapabilities=false,AutomatedEtcdBackup=false,BootcNodeManagement=false,CSIDriverSharedResource=false,ClusterAPIInstall=false,ClusterAPIInstallIBMCloud=false,ClusterMonitoringConfig=false,ConsolePluginContentSecurityPolicy=false,DNSNameResolver=false,DynamicResourceAllocation=false,EtcdBackendQuota=false,EventedPLEG=false,Example=false,ExternalOIDC=false,ExternalOIDCWithUIDAndExtraClaimMappings=false,GCPClusterHostedDNS=false,GatewayAPI=false,ImageStreamImportMode=false,IngressControllerDynamicConfigurationManager=false,InsightsConfig=false,InsightsConfigAPI=false,InsightsOnDemandDataGather=false,InsightsRuntimeExtractor=false,MachineAPIMigration=false,MachineAPIOperatorDisableMachineHealthCheckController=false,MachineAPIProviderOpenStack=false,MachineConfigNodes=false,MaxUnavailableStatefulSet=false,MetricsCollectionProfiles=false,MinimumKubeletVersion=false,MixedCPUsAllocation=false,MultiArchInstallAzure=false,NodeSwap=false,NutanixMultiSubnets=false,OVNObservability=false,OpenShiftPodSecurityAdmission=false,PinnedImages=false,PlatformOperators=false,ProcMountType=false,RouteAdvertisements=false,RouteExternalCertificate=false,ServiceAccountTokenNodeBinding=false,SignatureStores=false,SigstoreImageVerification=false,TranslateStreamCloseWebsocketRequests=false,UpgradeStatus=false,UserNamespacesPodSecurityStandards=false,UserNamespacesSupport=false,VSphereMultiNetworks=false,VolumeAttributesClass=false,VolumeGroupSnapshot=false

openshift-apiserver

deployment-controller

apiserver

ScalingReplicaSet

Scaled up replica set apiserver-779f85678d to 1 from 0

openshift-apiserver-operator

openshift-apiserver-operator-openshiftapiserver-workloadworkloadcontroller

openshift-apiserver-operator

DeploymentUpdated

Updated Deployment.apps/apiserver -n openshift-apiserver because it changed

openshift-authentication-operator

cluster-authentication-operator-openshiftauthenticationstaticresources-openshiftauthenticationstaticresources-staticresources

authentication-operator

ServiceAccountCreated

Created ServiceAccount/oauth-openshift -n openshift-authentication because it was missing

openshift-controller-manager-operator

openshift-controller-manager-operator

openshift-controller-manager-operator

ConfigMapUpdated

Updated ConfigMap/config -n openshift-controller-manager: cause by changes in data.openshift-controller-manager.serving-cert.secret

openshift-controller-manager-operator

openshift-controller-manager-operator

openshift-controller-manager-operator

ConfigMapUpdated

Updated ConfigMap/config -n openshift-route-controller-manager: cause by changes in data.openshift-route-controller-manager.serving-cert.secret

openshift-apiserver

replicaset-controller

apiserver-779f85678d

SuccessfulCreate

Created pod: apiserver-779f85678d-lrzfz

openshift-cluster-olm-operator

OperatorControllerStaticResources-operatorcontrollerstaticresources-operatorcontrollerstaticresources-staticresources

cluster-olm-operator

ClusterRoleBindingCreated

Created ClusterRoleBinding.rbac.authorization.k8s.io/operator-controller-manager-rolebinding because it was missing

openshift-cluster-olm-operator

CatalogdStaticResources-catalogdstaticresources-catalogdstaticresources-staticresources

cluster-olm-operator

MutatingWebhookConfigurationUpdated

Updated MutatingWebhookConfiguration.admissionregistration.k8s.io/catalogd-mutating-webhook-configuration because it changed
(x4)

openshift-apiserver

kubelet

apiserver-c765cd67b-xddmb

FailedMount

MountVolume.SetUp failed for volume "audit" : configmap "audit-0" not found

openshift-apiserver-operator

openshift-apiserver-operator-status-controller-statussyncer_openshift-apiserver

openshift-apiserver-operator

OperatorStatusChanged

Status for clusteroperator/openshift-apiserver changed: Progressing message changed from "APIServerDeploymentProgressing: deployment/apiserver.openshift-apiserver: observed generation is 0, desired generation is 1." to "APIServerDeploymentProgressing: deployment/apiserver.openshift-apiserver: observed generation is 1, desired generation is 2."

openshift-apiserver

replicaset-controller

apiserver-c765cd67b

SuccessfulDelete

Deleted pod: apiserver-c765cd67b-xddmb

openshift-apiserver

deployment-controller

apiserver

ScalingReplicaSet

Scaled down replica set apiserver-c765cd67b to 0 from 1
(x6)

openshift-controller-manager

kubelet

controller-manager-f9d74d9cf-ks2l5

FailedMount

MountVolume.SetUp failed for volume "client-ca" : configmap "client-ca" not found

openshift-apiserver-operator

openshift-apiserver-operator-status-controller-statussyncer_openshift-apiserver

openshift-apiserver-operator

OperatorStatusChanged

Status for clusteroperator/openshift-apiserver changed: Progressing message changed from "APIServerDeploymentProgressing: deployment/apiserver.openshift-apiserver: observed generation is 0, desired generation is 1.\nOperatorConfigProgressing: openshiftapiserveroperatorconfigs/instance: observed generation is 0, desired generation is 2." to "APIServerDeploymentProgressing: deployment/apiserver.openshift-apiserver: observed generation is 0, desired generation is 1."

openshift-kube-controller-manager-operator

kube-controller-manager-operator-targetconfigcontroller

kube-controller-manager-operator

ConfigMapCreated

Created ConfigMap/controller-manager-kubeconfig -n openshift-kube-controller-manager because it was missing

openshift-cluster-node-tuning-operator

kubelet

cluster-node-tuning-operator-598fbc5f8f-vxzvg

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:aa5e782406f71c048b1ac3a4bf5d1227ff4be81111114083ad4c7a209c6bfb5a" in 9.219s (9.219s including waiting). Image size: 677942383 bytes.

openshift-controller-manager

deployment-controller

controller-manager

ScalingReplicaSet

Scaled down replica set controller-manager-f9d74d9cf to 0 from 1

openshift-cluster-olm-operator

olm-status-controller-statussyncer_olm

cluster-olm-operator

OperatorStatusChanged

Status for clusteroperator/olm changed: Degraded message changed from "CatalogdClusterCatalogOpenshiftCommunityOperatorsDegraded: Internal error occurred: failed calling webhook \"inject-metadata-name.olm.operatorframework.io\": failed to call webhook: Post \"https://catalogd-service.openshift-catalogd.svc:9443/mutate-olm-operatorframework-io-v1-clustercatalog?timeout=10s\": no endpoints available for service \"catalogd-service\"\nCatalogdClusterCatalogOpenshiftRedhatMarketplaceDegraded: Internal error occurred: failed calling webhook \"inject-metadata-name.olm.operatorframework.io\": failed to call webhook: Post \"https://catalogd-service.openshift-catalogd.svc:9443/mutate-olm-operatorframework-io-v1-clustercatalog?timeout=10s\": no endpoints available for service \"catalogd-service\"\nCatalogdClusterCatalogOpenshiftCertifiedOperatorsDegraded: Internal error occurred: failed calling webhook \"inject-metadata-name.olm.operatorframework.io\": failed to call webhook: Post \"https://catalogd-service.openshift-catalogd.svc:9443/mutate-olm-operatorframework-io-v1-clustercatalog?timeout=10s\": no endpoints available for service \"catalogd-service\"" to "CatalogdClusterCatalogOpenshiftCommunityOperatorsDegraded: Internal error occurred: failed calling webhook \"inject-metadata-name.olm.operatorframework.io\": failed to call webhook: Post \"https://catalogd-service.openshift-catalogd.svc:9443/mutate-olm-operatorframework-io-v1-clustercatalog?timeout=10s\": no endpoints available for service \"catalogd-service\"\nCatalogdClusterCatalogOpenshiftRedhatMarketplaceDegraded: Internal error occurred: failed calling webhook \"inject-metadata-name.olm.operatorframework.io\": failed to call webhook: Post \"https://catalogd-service.openshift-catalogd.svc:9443/mutate-olm-operatorframework-io-v1-clustercatalog?timeout=10s\": no endpoints available for service \"catalogd-service\"\nCatalogdClusterCatalogOpenshiftCertifiedOperatorsDegraded: Internal error occurred: failed calling webhook \"inject-metadata-name.olm.operatorframework.io\": failed to call webhook: Post \"https://catalogd-service.openshift-catalogd.svc:9443/mutate-olm-operatorframework-io-v1-clustercatalog?timeout=10s\": no endpoints available for service \"catalogd-service\"\nCatalogdClusterCatalogOpenshiftRedhatOperatorsDegraded: Internal error occurred: failed calling webhook \"inject-metadata-name.olm.operatorframework.io\": failed to call webhook: Post \"https://catalogd-service.openshift-catalogd.svc:9443/mutate-olm-operatorframework-io-v1-clustercatalog?timeout=10s\": no endpoints available for service \"catalogd-service\""

openshift-cluster-olm-operator

olm-status-controller-statussyncer_olm

cluster-olm-operator

OperatorStatusChanged

Status for clusteroperator/olm changed: Progressing changed from Unknown to True ("OperatorcontrollerDeploymentOperatorControllerControllerManagerProgressing: Waiting for Deployment to act on changes"),Available changed from Unknown to False ("OperatorcontrollerDeploymentOperatorControllerControllerManagerAvailable: Waiting for Deployment")

openshift-image-registry

image-registry-operator

openshift-master-controllers

LeaderElection

cluster-image-registry-operator-5549dc66cb-fmhbq_0dc0faf4-6d2c-4b92-af57-19106e14d25f became leader

openshift-dns-operator

cluster-dns-operator

dns-operator

FeatureGatesInitialized

FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{"AWSEFSDriverVolumeMetrics", "AdminNetworkPolicy", "AlibabaPlatform", "AzureWorkloadIdentity", "BareMetalLoadBalancer", "BuildCSIVolumes", "ChunkSizeMiB", "CloudDualStackNodeIPs", "DisableKubeletCloudCredentialProviders", "GCPLabelsTags", "HardwareSpeed", "IngressControllerLBSubnetsAWS", "KMSv1", "ManagedBootImages", "ManagedBootImagesAWS", "MultiArchInstallAWS", "MultiArchInstallGCP", "NetworkDiagnosticsConfig", "NetworkLiveMigration", "NetworkSegmentation", "NewOLM", "NodeDisruptionPolicy", "OnClusterBuild", "PersistentIPsForVirtualization", "PrivateHostedZoneAWS", "SetEIPForNLBIngressController", "StreamingCollectionEncodingToJSON", "StreamingCollectionEncodingToProtobuf", "VSphereControlPlaneMachineSet", "VSphereDriverConfiguration", "VSphereMultiVCenters", "VSphereStaticIPs", "ValidatingAdmissionPolicy"}, Disabled:[]v1.FeatureGateName{"AWSClusterHostedDNS", "AdditionalRoutingCapabilities", "AutomatedEtcdBackup", "BootcNodeManagement", "CSIDriverSharedResource", "ClusterAPIInstall", "ClusterAPIInstallIBMCloud", "ClusterMonitoringConfig", "ConsolePluginContentSecurityPolicy", "DNSNameResolver", "DynamicResourceAllocation", "EtcdBackendQuota", "EventedPLEG", "Example", "ExternalOIDC", "ExternalOIDCWithUIDAndExtraClaimMappings", "GCPClusterHostedDNS", "GatewayAPI", "ImageStreamImportMode", "IngressControllerDynamicConfigurationManager", "InsightsConfig", "InsightsConfigAPI", "InsightsOnDemandDataGather", "InsightsRuntimeExtractor", "MachineAPIMigration", "MachineAPIOperatorDisableMachineHealthCheckController", "MachineAPIProviderOpenStack", "MachineConfigNodes", "MaxUnavailableStatefulSet", "MetricsCollectionProfiles", "MinimumKubeletVersion", "MixedCPUsAllocation", "MultiArchInstallAzure", "NodeSwap", "NutanixMultiSubnets", "OVNObservability", "OpenShiftPodSecurityAdmission", "PinnedImages", "PlatformOperators", "ProcMountType", "RouteAdvertisements", "RouteExternalCertificate", "ServiceAccountTokenNodeBinding", "SignatureStores", "SigstoreImageVerification", "TranslateStreamCloseWebsocketRequests", "UpgradeStatus", "UserNamespacesPodSecurityStandards", "UserNamespacesSupport", "VSphereMultiNetworks", "VolumeAttributesClass", "VolumeGroupSnapshot"}}

openshift-image-registry

kubelet

cluster-image-registry-operator-5549dc66cb-fmhbq

Started

Started container cluster-image-registry-operator

openshift-cluster-olm-operator

CatalogdDeploymentCatalogdControllerManager-catalogddeploymentcatalogdcontrollermanager-deployment-controller--catalogddeploymentcatalogdcontrollermanager

cluster-olm-operator

DeploymentCreated

Created Deployment.apps/catalogd-controller-manager -n openshift-catalogd because it was missing

openshift-cluster-olm-operator

olm-status-controller-statussyncer_olm

cluster-olm-operator

OperatorStatusChanged

Status for clusteroperator/olm changed: Progressing message changed from "OperatorcontrollerDeploymentOperatorControllerControllerManagerProgressing: Waiting for Deployment to act on changes" to "CatalogdDeploymentCatalogdControllerManagerProgressing: Waiting for Deployment to act on changes\nOperatorcontrollerDeploymentOperatorControllerControllerManagerProgressing: Waiting for Deployment to act on changes",Available message changed from "OperatorcontrollerDeploymentOperatorControllerControllerManagerAvailable: Waiting for Deployment" to "CatalogdDeploymentCatalogdControllerManagerAvailable: Waiting for Deployment\nOperatorcontrollerDeploymentOperatorControllerControllerManagerAvailable: Waiting for Deployment"

openshift-cluster-node-tuning-operator

daemonset-controller

tuned

SuccessfulCreate

Created pod: tuned-hb77b

openshift-dns-operator

kubelet

dns-operator-9c5679d8f-r6dm8

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-dns-operator

kubelet

dns-operator-9c5679d8f-r6dm8

Started

Started container dns-operator

openshift-dns-operator

kubelet

dns-operator-9c5679d8f-r6dm8

Created

Created container: dns-operator

openshift-cluster-node-tuning-operator

default-scheduler

tuned-hb77b

Scheduled

Successfully assigned openshift-cluster-node-tuning-operator/tuned-hb77b to master-0

openshift-cluster-node-tuning-operator

cluster-node-tuning-operator-598fbc5f8f-vxzvg_a082b3b9-e01a-4329-ba50-feca925640bb

node-tuning-operator-lock

LeaderElection

cluster-node-tuning-operator-598fbc5f8f-vxzvg_a082b3b9-e01a-4329-ba50-feca925640bb became leader

openshift-cluster-olm-operator

OperatorcontrollerDeploymentOperatorControllerControllerManager-operatorcontrollerdeploymentoperatorcontrollercontrollermanager-deployment-controller--operatorcontrollerdeploymentoperatorcontrollercontrollermanager

cluster-olm-operator

DeploymentCreated

Created Deployment.apps/operator-controller-controller-manager -n openshift-operator-controller because it was missing

openshift-kube-scheduler-operator

kubelet

openshift-kube-scheduler-operator-dddff6458-vx5d7

Started

Started container kube-scheduler-operator-container

openshift-kube-scheduler-operator

kubelet

openshift-kube-scheduler-operator-dddff6458-vx5d7

Created

Created container: kube-scheduler-operator-container

openshift-cluster-node-tuning-operator

performance-profile-controller

cluster-node-tuning-operator

FeatureGatesInitialized

FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{"AWSEFSDriverVolumeMetrics", "AdminNetworkPolicy", "AlibabaPlatform", "AzureWorkloadIdentity", "BareMetalLoadBalancer", "BuildCSIVolumes", "ChunkSizeMiB", "CloudDualStackNodeIPs", "DisableKubeletCloudCredentialProviders", "GCPLabelsTags", "HardwareSpeed", "IngressControllerLBSubnetsAWS", "KMSv1", "ManagedBootImages", "ManagedBootImagesAWS", "MultiArchInstallAWS", "MultiArchInstallGCP", "NetworkDiagnosticsConfig", "NetworkLiveMigration", "NetworkSegmentation", "NewOLM", "NodeDisruptionPolicy", "OnClusterBuild", "PersistentIPsForVirtualization", "PrivateHostedZoneAWS", "SetEIPForNLBIngressController", "StreamingCollectionEncodingToJSON", "StreamingCollectionEncodingToProtobuf", "VSphereControlPlaneMachineSet", "VSphereDriverConfiguration", "VSphereMultiVCenters", "VSphereStaticIPs", "ValidatingAdmissionPolicy"}, Disabled:[]v1.FeatureGateName{"AWSClusterHostedDNS", "AdditionalRoutingCapabilities", "AutomatedEtcdBackup", "BootcNodeManagement", "CSIDriverSharedResource", "ClusterAPIInstall", "ClusterAPIInstallIBMCloud", "ClusterMonitoringConfig", "ConsolePluginContentSecurityPolicy", "DNSNameResolver", "DynamicResourceAllocation", "EtcdBackendQuota", "EventedPLEG", "Example", "ExternalOIDC", "ExternalOIDCWithUIDAndExtraClaimMappings", "GCPClusterHostedDNS", "GatewayAPI", "ImageStreamImportMode", "IngressControllerDynamicConfigurationManager", "InsightsConfig", "InsightsConfigAPI", "InsightsOnDemandDataGather", "InsightsRuntimeExtractor", "MachineAPIMigration", "MachineAPIOperatorDisableMachineHealthCheckController", "MachineAPIProviderOpenStack", "MachineConfigNodes", "MaxUnavailableStatefulSet", "MetricsCollectionProfiles", "MinimumKubeletVersion", "MixedCPUsAllocation", "MultiArchInstallAzure", "NodeSwap", "NutanixMultiSubnets", "OVNObservability", "OpenShiftPodSecurityAdmission", "PinnedImages", "PlatformOperators", "ProcMountType", "RouteAdvertisements", "RouteExternalCertificate", "ServiceAccountTokenNodeBinding", "SignatureStores", "SigstoreImageVerification", "TranslateStreamCloseWebsocketRequests", "UpgradeStatus", "UserNamespacesPodSecurityStandards", "UserNamespacesSupport", "VSphereMultiNetworks", "VolumeAttributesClass", "VolumeGroupSnapshot"}}

openshift-cluster-olm-operator

olm-status-controller-statussyncer_olm

cluster-olm-operator

OperatorStatusChanged

Status for clusteroperator/olm changed: Degraded message changed from "All is well" to "CatalogdClusterCatalogOpenshiftCommunityOperatorsDegraded: Internal error occurred: failed calling webhook \"inject-metadata-name.olm.operatorframework.io\": failed to call webhook: Post \"https://catalogd-service.openshift-catalogd.svc:9443/mutate-olm-operatorframework-io-v1-clustercatalog?timeout=10s\": no endpoints available for service \"catalogd-service\""

openshift-cluster-version

openshift-cluster-version

version

LeaderElection

master-0_abd8961b-702c-4d22-aa3d-5cec899ab639 became leader

openshift-dns-operator

kubelet

dns-operator-9c5679d8f-r6dm8

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ea5c8a93f30e0a4932da5697d22c0da7eda9a7035c0555eb006b6755e62bb2fc" in 9.148s (9.148s including waiting). Image size: 468265024 bytes.

openshift-cluster-version

kubelet

cluster-version-operator-56d8475767-9gfkg

Started

Started container cluster-version-operator

openshift-route-controller-manager

deployment-controller

route-controller-manager

ScalingReplicaSet

Scaled up replica set route-controller-manager-74cf48bcc6 to 1 from 0

openshift-cluster-version

kubelet

cluster-version-operator-56d8475767-9gfkg

Created

Created container: cluster-version-operator

openshift-kube-apiserver-operator

kube-apiserver-operator-status-controller-statussyncer_kube-apiserver

kube-apiserver-operator

OperatorStatusChanged

Status for clusteroperator/kube-apiserver changed: Degraded message changed from "NodeControllerDegraded: All master nodes are ready\nNodeKubeconfigControllerDegraded: \"secret/node-kubeconfigs\": configmap \"kube-apiserver-server-ca\" not found\nRevisionControllerDegraded: configmap \"kube-apiserver-pod\" not found\nInstallerControllerDegraded: missing required resources: [configmaps: bound-sa-token-signing-certs-0,config-0,etcd-serving-ca-0,kube-apiserver-audit-policies-0,kube-apiserver-cert-syncer-kubeconfig-0,kube-apiserver-pod-0,kubelet-serving-ca-0,sa-token-signing-certs-0, secrets: etcd-client-0,localhost-recovery-client-token-0,localhost-recovery-serving-certkey-0]\nConfigObservationDegraded: error writing updated observed config: Operation cannot be fulfilled on kubeapiservers.operator.openshift.io \"cluster\": the object has been modified; please apply your changes to the latest version and try again" to "NodeControllerDegraded: All master nodes are ready\nNodeKubeconfigControllerDegraded: \"secret/node-kubeconfigs\": configmap \"kube-apiserver-server-ca\" not found\nRevisionControllerDegraded: configmap \"kube-apiserver-pod\" not found\nInstallerControllerDegraded: missing required resources: [configmaps: bound-sa-token-signing-certs-0,config-0,etcd-serving-ca-0,kube-apiserver-audit-policies-0,kube-apiserver-cert-syncer-kubeconfig-0,kube-apiserver-pod-0,kubelet-serving-ca-0,sa-token-signing-certs-0, secrets: etcd-client-0,localhost-recovery-client-token-0,localhost-recovery-serving-certkey-0]"

openshift-cluster-version

kubelet

cluster-version-operator-56d8475767-9gfkg

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-release@sha256:59727c4b3fef19e5149675cf3350735bbfe2c6588a57654b2e4552dd719f58b1" in 9.515s (9.515s including waiting). Image size: 517999161 bytes.

openshift-controller-manager

deployment-controller

controller-manager

ScalingReplicaSet

Scaled up replica set controller-manager-5dcd9ffc84 to 1 from 0

openshift-image-registry

kubelet

cluster-image-registry-operator-5549dc66cb-fmhbq

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7af9f5c5af9d529840233ef4b519120cc0e3f14c4fe28cc43b0823f2c11d8f89" in 9.228s (9.228s including waiting). Image size: 548752816 bytes.

openshift-controller-manager-operator

openshift-controller-manager-operator-status-controller-statussyncer_openshift-controller-manager

openshift-controller-manager-operator

OperatorStatusChanged

Status for clusteroperator/openshift-controller-manager changed: Progressing message changed from "Progressing: deployment/controller-manager: available replicas is 0, desired available replicas > 1\nProgressing: deployment/route-controller-manager: available replicas is 0, desired available replicas > 1" to "Progressing: deployment/controller-manager: observed generation is 2, desired generation is 3.\nProgressing: deployment/controller-manager: available replicas is 0, desired available replicas > 1\nProgressing: deployment/route-controller-manager: observed generation is 2, desired generation is 3.\nProgressing: deployment/route-controller-manager: available replicas is 0, desired available replicas > 1"

openshift-ingress-operator

kubelet

ingress-operator-66b84d69b-gzg9m

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:77fff570657d2fa0bfb709b2c8b6665bae0bf90a2be981d8dbca56c674715098" in 9.109s (9.109s including waiting). Image size: 511227324 bytes.

openshift-catalogd

default-scheduler

catalogd-controller-manager-6864dc98f7-74mgr

Scheduled

Successfully assigned openshift-catalogd/catalogd-controller-manager-6864dc98f7-74mgr to master-0

openshift-cluster-node-tuning-operator

kubelet

cluster-node-tuning-operator-598fbc5f8f-vxzvg

Started

Started container cluster-node-tuning-operator

openshift-ingress-operator

kubelet

ingress-operator-66b84d69b-gzg9m

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-ingress-operator

kubelet

ingress-operator-66b84d69b-gzg9m

Created

Created container: kube-rbac-proxy

openshift-ingress-operator

kubelet

ingress-operator-66b84d69b-gzg9m

Started

Started container kube-rbac-proxy

openshift-cluster-node-tuning-operator

kubelet

cluster-node-tuning-operator-598fbc5f8f-vxzvg

Created

Created container: cluster-node-tuning-operator

openshift-cluster-node-tuning-operator

kubelet

cluster-node-tuning-operator-598fbc5f8f-vxzvg

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:aa5e782406f71c048b1ac3a4bf5d1227ff4be81111114083ad4c7a209c6bfb5a" in 9.219s (9.219s including waiting). Image size: 677942383 bytes.

openshift-cluster-node-tuning-operator

default-scheduler

tuned-hb77b

Scheduled

Successfully assigned openshift-cluster-node-tuning-operator/tuned-hb77b to master-0

openshift-cluster-node-tuning-operator

cluster-node-tuning-operator-598fbc5f8f-vxzvg_a082b3b9-e01a-4329-ba50-feca925640bb

node-tuning-operator-lock

LeaderElection

cluster-node-tuning-operator-598fbc5f8f-vxzvg_a082b3b9-e01a-4329-ba50-feca925640bb became leader

openshift-cluster-node-tuning-operator

performance-profile-controller

cluster-node-tuning-operator

FeatureGatesInitialized

FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{"AWSEFSDriverVolumeMetrics", "AdminNetworkPolicy", "AlibabaPlatform", "AzureWorkloadIdentity", "BareMetalLoadBalancer", "BuildCSIVolumes", "ChunkSizeMiB", "CloudDualStackNodeIPs", "DisableKubeletCloudCredentialProviders", "GCPLabelsTags", "HardwareSpeed", "IngressControllerLBSubnetsAWS", "KMSv1", "ManagedBootImages", "ManagedBootImagesAWS", "MultiArchInstallAWS", "MultiArchInstallGCP", "NetworkDiagnosticsConfig", "NetworkLiveMigration", "NetworkSegmentation", "NewOLM", "NodeDisruptionPolicy", "OnClusterBuild", "PersistentIPsForVirtualization", "PrivateHostedZoneAWS", "SetEIPForNLBIngressController", "StreamingCollectionEncodingToJSON", "StreamingCollectionEncodingToProtobuf", "VSphereControlPlaneMachineSet", "VSphereDriverConfiguration", "VSphereMultiVCenters", "VSphereStaticIPs", "ValidatingAdmissionPolicy"}, Disabled:[]v1.FeatureGateName{"AWSClusterHostedDNS", "AdditionalRoutingCapabilities", "AutomatedEtcdBackup", "BootcNodeManagement", "CSIDriverSharedResource", "ClusterAPIInstall", "ClusterAPIInstallIBMCloud", "ClusterMonitoringConfig", "ConsolePluginContentSecurityPolicy", "DNSNameResolver", "DynamicResourceAllocation", "EtcdBackendQuota", "EventedPLEG", "Example", "ExternalOIDC", "ExternalOIDCWithUIDAndExtraClaimMappings", "GCPClusterHostedDNS", "GatewayAPI", "ImageStreamImportMode", "IngressControllerDynamicConfigurationManager", "InsightsConfig", "InsightsConfigAPI", "InsightsOnDemandDataGather", "InsightsRuntimeExtractor", "MachineAPIMigration", "MachineAPIOperatorDisableMachineHealthCheckController", "MachineAPIProviderOpenStack", "MachineConfigNodes", "MaxUnavailableStatefulSet", "MetricsCollectionProfiles", "MinimumKubeletVersion", "MixedCPUsAllocation", "MultiArchInstallAzure", "NodeSwap", "NutanixMultiSubnets", "OVNObservability", "OpenShiftPodSecurityAdmission", "PinnedImages", "PlatformOperators", "ProcMountType", "RouteAdvertisements", "RouteExternalCertificate", "ServiceAccountTokenNodeBinding", "SignatureStores", "SigstoreImageVerification", "TranslateStreamCloseWebsocketRequests", "UpgradeStatus", "UserNamespacesPodSecurityStandards", "UserNamespacesSupport", "VSphereMultiNetworks", "VolumeAttributesClass", "VolumeGroupSnapshot"}}

openshift-cluster-node-tuning-operator

kubelet

cluster-node-tuning-operator-598fbc5f8f-vxzvg

Started

Started container cluster-node-tuning-operator

openshift-operator-controller

deployment-controller

operator-controller-controller-manager

ScalingReplicaSet

Scaled up replica set operator-controller-controller-manager-57777556ff to 1

openshift-operator-controller

replicaset-controller

operator-controller-controller-manager-57777556ff

SuccessfulCreate

Created pod: operator-controller-controller-manager-57777556ff-nk2rf

openshift-ingress-operator

cluster-ingress-operator

ingress-operator

FeatureGatesInitialized

FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{"AWSEFSDriverVolumeMetrics", "AdminNetworkPolicy", "AlibabaPlatform", "AzureWorkloadIdentity", "BareMetalLoadBalancer", "BuildCSIVolumes", "ChunkSizeMiB", "CloudDualStackNodeIPs", "DisableKubeletCloudCredentialProviders", "GCPLabelsTags", "HardwareSpeed", "IngressControllerLBSubnetsAWS", "KMSv1", "ManagedBootImages", "ManagedBootImagesAWS", "MultiArchInstallAWS", "MultiArchInstallGCP", "NetworkDiagnosticsConfig", "NetworkLiveMigration", "NetworkSegmentation", "NewOLM", "NodeDisruptionPolicy", "OnClusterBuild", "PersistentIPsForVirtualization", "PrivateHostedZoneAWS", "SetEIPForNLBIngressController", "StreamingCollectionEncodingToJSON", "StreamingCollectionEncodingToProtobuf", "VSphereControlPlaneMachineSet", "VSphereDriverConfiguration", "VSphereMultiVCenters", "VSphereStaticIPs", "ValidatingAdmissionPolicy"}, Disabled:[]v1.FeatureGateName{"AWSClusterHostedDNS", "AdditionalRoutingCapabilities", "AutomatedEtcdBackup", "BootcNodeManagement", "CSIDriverSharedResource", "ClusterAPIInstall", "ClusterAPIInstallIBMCloud", "ClusterMonitoringConfig", "ConsolePluginContentSecurityPolicy", "DNSNameResolver", "DynamicResourceAllocation", "EtcdBackendQuota", "EventedPLEG", "Example", "ExternalOIDC", "ExternalOIDCWithUIDAndExtraClaimMappings", "GCPClusterHostedDNS", "GatewayAPI", "ImageStreamImportMode", "IngressControllerDynamicConfigurationManager", "InsightsConfig", "InsightsConfigAPI", "InsightsOnDemandDataGather", "InsightsRuntimeExtractor", "MachineAPIMigration", "MachineAPIOperatorDisableMachineHealthCheckController", "MachineAPIProviderOpenStack", "MachineConfigNodes", "MaxUnavailableStatefulSet", "MetricsCollectionProfiles", "MinimumKubeletVersion", "MixedCPUsAllocation", "MultiArchInstallAzure", "NodeSwap", "NutanixMultiSubnets", "OVNObservability", "OpenShiftPodSecurityAdmission", "PinnedImages", "PlatformOperators", "ProcMountType", "RouteAdvertisements", "RouteExternalCertificate", "ServiceAccountTokenNodeBinding", "SignatureStores", "SigstoreImageVerification", "TranslateStreamCloseWebsocketRequests", "UpgradeStatus", "UserNamespacesPodSecurityStandards", "UserNamespacesSupport", "VSphereMultiNetworks", "VolumeAttributesClass", "VolumeGroupSnapshot"}}

openshift-cluster-olm-operator

OperatorControllerStaticResources-operatorcontrollerstaticresources-operatorcontrollerstaticresources-staticresources

cluster-olm-operator

ServiceCreated

Created Service/operator-controller-controller-manager-metrics-service -n openshift-operator-controller because it was missing

openshift-cluster-node-tuning-operator

kubelet

cluster-node-tuning-operator-598fbc5f8f-vxzvg

Created

Created container: cluster-node-tuning-operator

openshift-image-registry

kubelet

cluster-image-registry-operator-5549dc66cb-fmhbq

Created

Created container: cluster-image-registry-operator

openshift-controller-manager

default-scheduler

controller-manager-5dcd9ffc84-qh6pq

FailedScheduling

0/1 nodes are available: 1 node(s) didn't match pod anti-affinity rules. preemption: 0/1 nodes are available: 1 node(s) didn't match pod anti-affinity rules.

openshift-controller-manager

replicaset-controller

controller-manager-f9d74d9cf

SuccessfulDelete

Deleted pod: controller-manager-f9d74d9cf-ks2l5

openshift-catalogd

deployment-controller

catalogd-controller-manager

ScalingReplicaSet

Scaled up replica set catalogd-controller-manager-6864dc98f7 to 1

openshift-catalogd

deployment-controller

catalogd-controller-manager

ScalingReplicaSet

Scaled up replica set catalogd-controller-manager-6864dc98f7 to 1

openshift-catalogd

replicaset-controller

catalogd-controller-manager-6864dc98f7

SuccessfulCreate

Created pod: catalogd-controller-manager-6864dc98f7-74mgr

openshift-catalogd

default-scheduler

catalogd-controller-manager-6864dc98f7-74mgr

Scheduled

Successfully assigned openshift-catalogd/catalogd-controller-manager-6864dc98f7-74mgr to master-0

openshift-controller-manager

replicaset-controller

controller-manager-5dcd9ffc84

SuccessfulCreate

Created pod: controller-manager-5dcd9ffc84-qh6pq

openshift-cluster-olm-operator

OperatorControllerStaticResources-operatorcontrollerstaticresources-operatorcontrollerstaticresources-staticresources

cluster-olm-operator

ClusterRoleBindingCreated

Created ClusterRoleBinding.rbac.authorization.k8s.io/operator-controller-proxy-rolebinding because it was missing

openshift-route-controller-manager

replicaset-controller

route-controller-manager-5ff6bdb8cf

SuccessfulDelete

Deleted pod: route-controller-manager-5ff6bdb8cf-b6499

openshift-route-controller-manager

replicaset-controller

route-controller-manager-74cf48bcc6

SuccessfulCreate

Created pod: route-controller-manager-74cf48bcc6-kwngm

openshift-catalogd

replicaset-controller

catalogd-controller-manager-6864dc98f7

SuccessfulCreate

Created pod: catalogd-controller-manager-6864dc98f7-74mgr

openshift-cluster-olm-operator

olm-status-controller-statussyncer_olm

cluster-olm-operator

OperatorStatusChanged

Status for clusteroperator/olm changed: Degraded message changed from "CatalogdClusterCatalogOpenshiftCommunityOperatorsDegraded: Internal error occurred: failed calling webhook \"inject-metadata-name.olm.operatorframework.io\": failed to call webhook: Post \"https://catalogd-service.openshift-catalogd.svc:9443/mutate-olm-operatorframework-io-v1-clustercatalog?timeout=10s\": no endpoints available for service \"catalogd-service\"" to "CatalogdClusterCatalogOpenshiftCommunityOperatorsDegraded: Internal error occurred: failed calling webhook \"inject-metadata-name.olm.operatorframework.io\": failed to call webhook: Post \"https://catalogd-service.openshift-catalogd.svc:9443/mutate-olm-operatorframework-io-v1-clustercatalog?timeout=10s\": no endpoints available for service \"catalogd-service\"\nCatalogdClusterCatalogOpenshiftRedhatMarketplaceDegraded: Internal error occurred: failed calling webhook \"inject-metadata-name.olm.operatorframework.io\": failed to call webhook: Post \"https://catalogd-service.openshift-catalogd.svc:9443/mutate-olm-operatorframework-io-v1-clustercatalog?timeout=10s\": no endpoints available for service \"catalogd-service\"\nCatalogdClusterCatalogOpenshiftCertifiedOperatorsDegraded: Internal error occurred: failed calling webhook \"inject-metadata-name.olm.operatorframework.io\": failed to call webhook: Post \"https://catalogd-service.openshift-catalogd.svc:9443/mutate-olm-operatorframework-io-v1-clustercatalog?timeout=10s\": no endpoints available for service \"catalogd-service\""

openshift-authentication-operator

cluster-authentication-operator-resource-sync-controller-resourcesynccontroller

authentication-operator

SecretCreated

Created Secret/etcd-client -n openshift-oauth-apiserver because it was missing

openshift-operator-controller

default-scheduler

operator-controller-controller-manager-57777556ff-nk2rf

Scheduled

Successfully assigned openshift-operator-controller/operator-controller-controller-manager-57777556ff-nk2rf to master-0

openshift-route-controller-manager

deployment-controller

route-controller-manager

ScalingReplicaSet

Scaled down replica set route-controller-manager-5ff6bdb8cf to 0 from 1

openshift-cluster-node-tuning-operator

daemonset-controller

tuned

SuccessfulCreate

Created pod: tuned-hb77b

openshift-cluster-olm-operator

OperatorControllerStaticResources-operatorcontrollerstaticresources-operatorcontrollerstaticresources-staticresources

cluster-olm-operator

ConfigMapCreated

Created ConfigMap/operator-controller-trusted-ca-bundle -n openshift-operator-controller because it was missing

openshift-etcd-operator

openshift-cluster-etcd-operator-etcd-endpoints-controller-etcdendpointscontroller

etcd-operator

EtcdEndpointsErrorUpdatingStatus

Operation cannot be fulfilled on etcds.operator.openshift.io "cluster": the object has been modified; please apply your changes to the latest version and try again

openshift-image-registry

image-registry-operator

cluster-image-registry-operator

FeatureGatesInitialized

FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{"AWSEFSDriverVolumeMetrics", "AdminNetworkPolicy", "AlibabaPlatform", "AzureWorkloadIdentity", "BareMetalLoadBalancer", "BuildCSIVolumes", "ChunkSizeMiB", "CloudDualStackNodeIPs", "DisableKubeletCloudCredentialProviders", "GCPLabelsTags", "HardwareSpeed", "IngressControllerLBSubnetsAWS", "KMSv1", "ManagedBootImages", "ManagedBootImagesAWS", "MultiArchInstallAWS", "MultiArchInstallGCP", "NetworkDiagnosticsConfig", "NetworkLiveMigration", "NetworkSegmentation", "NewOLM", "NodeDisruptionPolicy", "OnClusterBuild", "PersistentIPsForVirtualization", "PrivateHostedZoneAWS", "SetEIPForNLBIngressController", "StreamingCollectionEncodingToJSON", "StreamingCollectionEncodingToProtobuf", "VSphereControlPlaneMachineSet", "VSphereDriverConfiguration", "VSphereMultiVCenters", "VSphereStaticIPs", "ValidatingAdmissionPolicy"}, Disabled:[]v1.FeatureGateName{"AWSClusterHostedDNS", "AdditionalRoutingCapabilities", "AutomatedEtcdBackup", "BootcNodeManagement", "CSIDriverSharedResource", "ClusterAPIInstall", "ClusterAPIInstallIBMCloud", "ClusterMonitoringConfig", "ConsolePluginContentSecurityPolicy", "DNSNameResolver", "DynamicResourceAllocation", "EtcdBackendQuota", "EventedPLEG", "Example", "ExternalOIDC", "ExternalOIDCWithUIDAndExtraClaimMappings", "GCPClusterHostedDNS", "GatewayAPI", "ImageStreamImportMode", "IngressControllerDynamicConfigurationManager", "InsightsConfig", "InsightsConfigAPI", "InsightsOnDemandDataGather", "InsightsRuntimeExtractor", "MachineAPIMigration", "MachineAPIOperatorDisableMachineHealthCheckController", "MachineAPIProviderOpenStack", "MachineConfigNodes", "MaxUnavailableStatefulSet", "MetricsCollectionProfiles", "MinimumKubeletVersion", "MixedCPUsAllocation", "MultiArchInstallAzure", "NodeSwap", "NutanixMultiSubnets", "OVNObservability", "OpenShiftPodSecurityAdmission", "PinnedImages", "PlatformOperators", "ProcMountType", "RouteAdvertisements", "RouteExternalCertificate", "ServiceAccountTokenNodeBinding", "SignatureStores", "SigstoreImageVerification", "TranslateStreamCloseWebsocketRequests", "UpgradeStatus", "UserNamespacesPodSecurityStandards", "UserNamespacesSupport", "VSphereMultiNetworks", "VolumeAttributesClass", "VolumeGroupSnapshot"}}

openshift-catalogd

kubelet

catalogd-controller-manager-6864dc98f7-74mgr

Created

Created container: kube-rbac-proxy
(x2)

openshift-apiserver

default-scheduler

apiserver-779f85678d-lrzfz

FailedScheduling

0/1 nodes are available: 1 node(s) didn't match pod anti-affinity rules. preemption: 0/1 nodes are available: 1 node(s) didn't match pod anti-affinity rules.

openshift-cluster-node-tuning-operator

kubelet

tuned-hb77b

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:aa5e782406f71c048b1ac3a4bf5d1227ff4be81111114083ad4c7a209c6bfb5a" already present on machine

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-status-controller-statussyncer_kube-scheduler

openshift-kube-scheduler-operator

OperatorStatusChanged

Status for clusteroperator/kube-scheduler changed: Degraded changed from Unknown to False ("RevisionControllerDegraded: configmap \"kube-scheduler-pod\" not found")

openshift-operator-controller

kubelet

operator-controller-controller-manager-57777556ff-nk2rf

Started

Started container manager

openshift-dns-operator

kubelet

dns-operator-9c5679d8f-r6dm8

Started

Started container kube-rbac-proxy

openshift-operator-controller

kubelet

operator-controller-controller-manager-57777556ff-nk2rf

Created

Created container: manager

openshift-route-controller-manager

kubelet

route-controller-manager-5ff6bdb8cf-b6499

FailedMount

MountVolume.SetUp failed for volume "client-ca" : object "openshift-route-controller-manager"/"client-ca" not registered

openshift-controller-manager

default-scheduler

controller-manager-5dcd9ffc84-qh6pq

Scheduled

Successfully assigned openshift-controller-manager/controller-manager-5dcd9ffc84-qh6pq to master-0

openshift-catalogd

kubelet

catalogd-controller-manager-6864dc98f7-74mgr

Started

Started container kube-rbac-proxy

openshift-catalogd

multus

catalogd-controller-manager-6864dc98f7-74mgr

AddedInterface

Add eth0 [10.128.0.39/23] from ovn-kubernetes

openshift-catalogd

kubelet

catalogd-controller-manager-6864dc98f7-74mgr

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-catalogd

kubelet

catalogd-controller-manager-6864dc98f7-74mgr

Created

Created container: kube-rbac-proxy

openshift-catalogd

kubelet

catalogd-controller-manager-6864dc98f7-74mgr

Started

Started container kube-rbac-proxy

openshift-catalogd

kubelet

catalogd-controller-manager-6864dc98f7-74mgr

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3062f6485aec4770e60852b535c69a42527b305161fe856499c8658ead6d1e85" already present on machine

openshift-catalogd

kubelet

catalogd-controller-manager-6864dc98f7-74mgr

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3062f6485aec4770e60852b535c69a42527b305161fe856499c8658ead6d1e85" already present on machine

openshift-operator-controller

multus

operator-controller-controller-manager-57777556ff-nk2rf

AddedInterface

Add eth0 [10.128.0.38/23] from ovn-kubernetes

openshift-operator-controller

kubelet

operator-controller-controller-manager-57777556ff-nk2rf

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-operator-controller

kubelet

operator-controller-controller-manager-57777556ff-nk2rf

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:4b5ea1ef4e09b673a0c68c8848ca162ab11d9ac373a377daa52dea702ffa3023" already present on machine

openshift-ingress-operator

certificate_controller

router-ca

CreatedWildcardCACert

Created a default wildcard CA certificate

openshift-dns

default-scheduler

dns-default-v5h69

Scheduled

Successfully assigned openshift-dns/dns-default-v5h69 to master-0

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-kube-scheduler-node

openshift-kube-scheduler-operator

MasterNodesReadyChanged

All master nodes are ready

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-kube-scheduler-node

openshift-kube-scheduler-operator

MasterNodeObserved

Observed new master node master-0

openshift-ingress

default-scheduler

router-default-7dcf5569b5-xmvwz

FailedScheduling

0/1 nodes are available: 1 node(s) had untolerated taint {node-role.kubernetes.io/master: }. preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.

openshift-route-controller-manager

kubelet

route-controller-manager-5ff6bdb8cf-b6499

FailedMount

MountVolume.SetUp failed for volume "serving-cert" : object "openshift-route-controller-manager"/"serving-cert" not registered
(x2)

openshift-route-controller-manager

default-scheduler

route-controller-manager-74cf48bcc6-kwngm

FailedScheduling

0/1 nodes are available: 1 node(s) didn't match pod anti-affinity rules. preemption: 0/1 nodes are available: 1 node(s) didn't match pod anti-affinity rules.

openshift-operator-controller

operator-controller-controller-manager-57777556ff-nk2rf_647cc0bb-b249-45f7-9cc7-eec48e330833

9c4404e7.operatorframework.io

LeaderElection

operator-controller-controller-manager-57777556ff-nk2rf_647cc0bb-b249-45f7-9cc7-eec48e330833 became leader

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-status-controller-statussyncer_kube-scheduler

openshift-kube-scheduler-operator

OperatorStatusChanged

Status for clusteroperator/kube-scheduler changed: Upgradeable changed from Unknown to True ("All is well")

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-status-controller-statussyncer_kube-scheduler

openshift-kube-scheduler-operator

OperatorStatusChanged

Status for clusteroperator/kube-scheduler changed: Degraded set to Unknown (""),Progressing set to Unknown (""),Available set to Unknown (""),Upgradeable set to Unknown (""),EvaluationConditionsDetected set to Unknown (""),status.relatedObjects changed from [{"operator.openshift.io" "kubeschedulers" "" "cluster"} {"config.openshift.io" "schedulers" "" ""} {"" "namespaces" "" "openshift-config"} {"" "namespaces" "" "openshift-config-managed"} {"" "namespaces" "" "openshift-kube-scheduler-operator"} {"" "namespaces" "" "openshift-kube-scheduler"} {"controlplane.operator.openshift.io" "podnetworkconnectivitychecks" "openshift-kube-scheduler" ""}] to [{"operator.openshift.io" "kubeschedulers" "" "cluster"} {"config.openshift.io" "schedulers" "" ""} {"" "namespaces" "" "openshift-config"} {"" "namespaces" "" "openshift-config-managed"} {"" "namespaces" "" "openshift-kube-scheduler"} {"" "namespaces" "" "openshift-kube-scheduler-operator"} {"controlplane.operator.openshift.io" "podnetworkconnectivitychecks" "openshift-kube-apiserver" ""}],status.versions changed from [] to [{"raw-internal" "4.18.35"}]
(x2)

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-status-controller-statussyncer_kube-scheduler

openshift-kube-scheduler-operator

OperatorVersionChanged

clusteroperator/kube-scheduler version "raw-internal" changed from "" to "4.18.35"

kube-system

cluster-policy-controller-namespace-security-allocation-controller

bootstrap-kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openshift-ingress namespace

kube-system

cluster-policy-controller-namespace-security-allocation-controller

bootstrap-kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openshift-dns namespace

openshift-cluster-node-tuning-operator

kubelet

tuned-hb77b

Started

Started container tuned

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator

openshift-kube-scheduler-operator

FeatureGatesInitialized

FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{"AWSEFSDriverVolumeMetrics", "AdminNetworkPolicy", "AlibabaPlatform", "AzureWorkloadIdentity", "BareMetalLoadBalancer", "BuildCSIVolumes", "ChunkSizeMiB", "CloudDualStackNodeIPs", "DisableKubeletCloudCredentialProviders", "GCPLabelsTags", "HardwareSpeed", "IngressControllerLBSubnetsAWS", "KMSv1", "ManagedBootImages", "ManagedBootImagesAWS", "MultiArchInstallAWS", "MultiArchInstallGCP", "NetworkDiagnosticsConfig", "NetworkLiveMigration", "NetworkSegmentation", "NewOLM", "NodeDisruptionPolicy", "OnClusterBuild", "PersistentIPsForVirtualization", "PrivateHostedZoneAWS", "SetEIPForNLBIngressController", "StreamingCollectionEncodingToJSON", "StreamingCollectionEncodingToProtobuf", "VSphereControlPlaneMachineSet", "VSphereDriverConfiguration", "VSphereMultiVCenters", "VSphereStaticIPs", "ValidatingAdmissionPolicy"}, Disabled:[]v1.FeatureGateName{"AWSClusterHostedDNS", "AdditionalRoutingCapabilities", "AutomatedEtcdBackup", "BootcNodeManagement", "CSIDriverSharedResource", "ClusterAPIInstall", "ClusterAPIInstallIBMCloud", "ClusterMonitoringConfig", "ConsolePluginContentSecurityPolicy", "DNSNameResolver", "DynamicResourceAllocation", "EtcdBackendQuota", "EventedPLEG", "Example", "ExternalOIDC", "ExternalOIDCWithUIDAndExtraClaimMappings", "GCPClusterHostedDNS", "GatewayAPI", "ImageStreamImportMode", "IngressControllerDynamicConfigurationManager", "InsightsConfig", "InsightsConfigAPI", "InsightsOnDemandDataGather", "InsightsRuntimeExtractor", "MachineAPIMigration", "MachineAPIOperatorDisableMachineHealthCheckController", "MachineAPIProviderOpenStack", "MachineConfigNodes", "MaxUnavailableStatefulSet", "MetricsCollectionProfiles", "MinimumKubeletVersion", "MixedCPUsAllocation", "MultiArchInstallAzure", "NodeSwap", "NutanixMultiSubnets", "OVNObservability", "OpenShiftPodSecurityAdmission", "PinnedImages", "PlatformOperators", "ProcMountType", "RouteAdvertisements", "RouteExternalCertificate", "ServiceAccountTokenNodeBinding", "SignatureStores", "SigstoreImageVerification", "TranslateStreamCloseWebsocketRequests", "UpgradeStatus", "UserNamespacesPodSecurityStandards", "UserNamespacesSupport", "VSphereMultiNetworks", "VolumeAttributesClass", "VolumeGroupSnapshot"}}

openshift-dns-operator

kubelet

dns-operator-9c5679d8f-r6dm8

Created

Created container: kube-rbac-proxy

openshift-dns

daemonset-controller

dns-default

SuccessfulCreate

Created pod: dns-default-v5h69

openshift-dns

default-scheduler

node-resolver-qnp9w

Scheduled

Successfully assigned openshift-dns/node-resolver-qnp9w to master-0

openshift-catalogd

kubelet

catalogd-controller-manager-6864dc98f7-74mgr

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-kube-apiserver-operator

kube-apiserver-operator-target-config-controller-targetconfigcontroller

kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/config -n openshift-kube-apiserver because it was missing

openshift-dns

daemonset-controller

node-resolver

SuccessfulCreate

Created pod: node-resolver-qnp9w

openshift-ingress-operator

ingress_controller

default

Admitted

ingresscontroller passed validation

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-lock

LeaderElection

openshift-kube-scheduler-operator-dddff6458-vx5d7_42dfaabe-4d62-4065-a7d5-2d4427fae2d0 became leader

openshift-ingress

deployment-controller

router-default

ScalingReplicaSet

Scaled up replica set router-default-7dcf5569b5 to 1

openshift-ingress

replicaset-controller

router-default-7dcf5569b5

SuccessfulCreate

Created pod: router-default-7dcf5569b5-xmvwz

openshift-cluster-node-tuning-operator

kubelet

tuned-hb77b

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:aa5e782406f71c048b1ac3a4bf5d1227ff4be81111114083ad4c7a209c6bfb5a" already present on machine

openshift-cluster-node-tuning-operator

kubelet

tuned-hb77b

Created

Created container: tuned

openshift-cluster-node-tuning-operator

kubelet

tuned-hb77b

Started

Started container tuned

openshift-cluster-olm-operator

olm-status-controller-statussyncer_olm

cluster-olm-operator

OperatorStatusChanged

Status for clusteroperator/olm changed: Progressing message changed from "CatalogdDeploymentCatalogdControllerManagerProgressing: Waiting for Deployment to act on changes\nOperatorcontrollerDeploymentOperatorControllerControllerManagerProgressing: Waiting for Deployment to act on changes" to "CatalogdDeploymentCatalogdControllerManagerProgressing: Waiting for Deployment to act on changes\nOperatorcontrollerDeploymentOperatorControllerControllerManagerProgressing: Waiting for Deployment to deploy pods"

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-status-controller-statussyncer_kube-scheduler

openshift-kube-scheduler-operator

OperatorStatusChanged

Status for clusteroperator/kube-scheduler changed: Degraded message changed from "RevisionControllerDegraded: configmap \"kube-scheduler-pod\" not found" to "RevisionControllerDegraded: configmap \"kube-scheduler-pod\" not found\nInstallerControllerDegraded: missing required resources: [configmaps: config-0,kube-scheduler-cert-syncer-kubeconfig-0,kube-scheduler-pod-0,scheduler-kubeconfig-0,serviceaccount-ca-0, secrets: localhost-recovery-client-token-0]",Progressing changed from Unknown to False ("All is well"),Available changed from Unknown to False ("StaticPodsAvailable: 0 nodes are active; ")

openshift-cluster-node-tuning-operator

kubelet

tuned-hb77b

Created

Created container: tuned

openshift-catalogd

multus

catalogd-controller-manager-6864dc98f7-74mgr

AddedInterface

Add eth0 [10.128.0.39/23] from ovn-kubernetes

openshift-dns

kubelet

dns-default-v5h69

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:2c4d5a681595e428ff4b5083648c13615eed80be9084a3d3fc68a0295079cb12"

openshift-dns

kubelet

node-resolver-qnp9w

Created

Created container: dns-node-resolver

openshift-operator-controller

kubelet

operator-controller-controller-manager-57777556ff-nk2rf

Started

Started container kube-rbac-proxy

openshift-catalogd

catalogd-controller-manager-6864dc98f7-74mgr_5c696908-6a3a-4ac2-895b-80c7fdddaa72

catalogd-operator-lock

LeaderElection

catalogd-controller-manager-6864dc98f7-74mgr_5c696908-6a3a-4ac2-895b-80c7fdddaa72 became leader

openshift-catalogd

kubelet

catalogd-controller-manager-6864dc98f7-74mgr

Started

Started container manager

openshift-catalogd

kubelet

catalogd-controller-manager-6864dc98f7-74mgr

Created

Created container: manager

openshift-catalogd

catalogd-controller-manager-6864dc98f7-74mgr_5c696908-6a3a-4ac2-895b-80c7fdddaa72

catalogd-operator-lock

LeaderElection

catalogd-controller-manager-6864dc98f7-74mgr_5c696908-6a3a-4ac2-895b-80c7fdddaa72 became leader

openshift-dns

multus

dns-default-v5h69

AddedInterface

Add eth0 [10.128.0.40/23] from ovn-kubernetes

openshift-ingress-operator

certificate_controller

default

CreatedDefaultCertificate

Created default wildcard certificate "router-certs-default"

openshift-cluster-olm-operator

olm-status-controller-statussyncer_olm

cluster-olm-operator

OperatorStatusChanged

Status for clusteroperator/olm changed: Progressing message changed from "CatalogdDeploymentCatalogdControllerManagerProgressing: Waiting for Deployment to act on changes\nOperatorcontrollerDeploymentOperatorControllerControllerManagerProgressing: Waiting for Deployment to deploy pods" to "CatalogdDeploymentCatalogdControllerManagerProgressing: Waiting for Deployment to deploy pods\nOperatorcontrollerDeploymentOperatorControllerControllerManagerProgressing: Waiting for Deployment to deploy pods"

openshift-dns

kubelet

node-resolver-qnp9w

Started

Started container dns-node-resolver

openshift-catalogd

kubelet

catalogd-controller-manager-6864dc98f7-74mgr

Created

Created container: manager

openshift-config-managed

certificate_publisher_controller

router-certs

PublishedRouterCertificates

Published router certificates

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-status-controller-statussyncer_kube-scheduler

openshift-kube-scheduler-operator

OperatorStatusChanged

Status for clusteroperator/kube-scheduler changed: Progressing message changed from "All is well" to "NodeInstallerProgressing: 1 node is at revision 0",Available message changed from "StaticPodsAvailable: 0 nodes are active; " to "StaticPodsAvailable: 0 nodes are active; 1 node is at revision 0"

openshift-operator-controller

kubelet

operator-controller-controller-manager-57777556ff-nk2rf

Created

Created container: kube-rbac-proxy

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-status-controller-statussyncer_kube-scheduler

openshift-kube-scheduler-operator

OperatorStatusChanged

Status for clusteroperator/kube-scheduler changed: Degraded message changed from "RevisionControllerDegraded: configmap \"kube-scheduler-pod\" not found\nInstallerControllerDegraded: missing required resources: [configmaps: config-0,kube-scheduler-cert-syncer-kubeconfig-0,kube-scheduler-pod-0,scheduler-kubeconfig-0,serviceaccount-ca-0, secrets: localhost-recovery-client-token-0]" to "RevisionControllerDegraded: configmap \"kube-scheduler-pod\" not found\nInstallerControllerDegraded: missing required resources: [configmaps: config-0,kube-scheduler-cert-syncer-kubeconfig-0,kube-scheduler-pod-0,scheduler-kubeconfig-0,serviceaccount-ca-0, secrets: localhost-recovery-client-token-0]\nNodeControllerDegraded: All master nodes are ready"

openshift-dns

kubelet

node-resolver-qnp9w

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c6a4383333a1fd6d05c3f60ec793913f7937ee3d77f002d85e6c61e20507bf55" already present on machine

openshift-cluster-version

openshift-cluster-version

version

RetrievePayload

Retrieving and verifying payload version="4.18.35" image="quay.io/openshift-release-dev/ocp-release@sha256:59727c4b3fef19e5149675cf3350735bbfe2c6588a57654b2e4552dd719f58b1"

openshift-config-managed

certificate_publisher_controller

default-ingress-cert

PublishedRouterCA

Published "default-ingress-cert" in "openshift-config-managed"

openshift-cluster-version

openshift-cluster-version

version

LoadPayload

Loading payload version="4.18.35" image="quay.io/openshift-release-dev/ocp-release@sha256:59727c4b3fef19e5149675cf3350735bbfe2c6588a57654b2e4552dd719f58b1"

openshift-catalogd

kubelet

catalogd-controller-manager-6864dc98f7-74mgr

Started

Started container manager

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-backingresourcecontroller-backingresourcecontroller-staticresources

openshift-kube-scheduler-operator

ClusterRoleBindingCreated

Created ClusterRoleBinding.rbac.authorization.k8s.io/system:openshift:operator:openshift-kube-scheduler-installer because it was missing

openshift-kube-controller-manager-operator

kube-controller-manager-operator-revisioncontroller

kube-controller-manager-operator

StartingNewRevision

new revision 1 triggered by "configmap \"kube-controller-manager-pod-0\" not found"

openshift-kube-controller-manager-operator

kube-controller-manager-operator-targetconfigcontroller

kube-controller-manager-operator

ConfigMapCreated

Created ConfigMap/kube-controller-manager-pod -n openshift-kube-controller-manager because it was missing

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-resource-sync-controller-resourcesynccontroller

openshift-kube-scheduler-operator

SecretCreated

Created Secret/kube-scheduler-client-cert-key -n openshift-kube-scheduler because it was missing

openshift-apiserver-operator

openshift-apiserver-operator-status-controller-statussyncer_openshift-apiserver

openshift-apiserver-operator

OperatorStatusChanged

Status for clusteroperator/openshift-apiserver changed: Progressing message changed from "APIServerDeploymentProgressing: deployment/apiserver.openshift-apiserver: observed generation is 1, desired generation is 2." to "APIServerDeploymentProgressing: deployment/apiserver.openshift-apiserver: 1/1 pods have been updated to the latest generation and 0/1 pods are available"

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-kubecontrollermanagerstaticresources-kubecontrollermanagerstaticresources-staticresources

openshift-kube-scheduler-operator

NamespaceUpdated

Updated Namespace/openshift-kube-scheduler because it changed

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-backingresourcecontroller-backingresourcecontroller-staticresources

openshift-kube-scheduler-operator

ServiceAccountCreated

Created ServiceAccount/installer-sa -n openshift-kube-scheduler because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

kube-apiserver-operator

StartingNewRevision

new revision 1 triggered by "configmap \"kube-apiserver-pod-0\" not found"

openshift-authentication-operator

oauth-apiserver-oauthapiservercontroller-workloadworkloadcontroller

authentication-operator

DeploymentCreated

Created Deployment.apps/apiserver -n openshift-oauth-apiserver because it was missing

openshift-oauth-apiserver

default-scheduler

apiserver-6fccff6869-f98nk

Scheduled

Successfully assigned openshift-oauth-apiserver/apiserver-6fccff6869-f98nk to master-0

openshift-oauth-apiserver

replicaset-controller

apiserver-6fccff6869

SuccessfulCreate

Created pod: apiserver-6fccff6869-f98nk

openshift-oauth-apiserver

deployment-controller

apiserver

ScalingReplicaSet

Scaled up replica set apiserver-6fccff6869 to 1

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Degraded message changed from "OAuthServiceDegraded: Unable to get oauth server service: service \"oauth-openshift\" not found\nWebhookAuthenticatorControllerDegraded: failed to read service-ca crt bundle: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory\nAPIServerDeploymentDegraded: waiting for .status.latestAvailableRevision to be available\nAPIServerDeploymentDegraded: \nAPIServerWorkloadDegraded: waiting for .status.latestAvailableRevision to be available\nAPIServerWorkloadDegraded: \nIngressStateEndpointsDegraded: No endpoints found for oauth-server\nRouterCertsDegraded: neither the custom secret/v4-0-config-system-router-certs -n openshift-authentication or default secret/v4-0-config-system-custom-router-certs -n openshift-authentication could be retrieved: secret \"v4-0-config-system-router-certs\" not found\nOAuthServerConfigObservationDegraded: secret \"v4-0-config-system-router-certs\" not found" to "OAuthServiceDegraded: Unable to get oauth server service: service \"oauth-openshift\" not found\nWebhookAuthenticatorControllerDegraded: failed to read service-ca crt bundle: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory\nIngressStateEndpointsDegraded: No endpoints found for oauth-server\nRouterCertsDegraded: neither the custom secret/v4-0-config-system-router-certs -n openshift-authentication or default secret/v4-0-config-system-custom-router-certs -n openshift-authentication could be retrieved: secret \"v4-0-config-system-router-certs\" not found\nOAuthServerConfigObservationDegraded: secret \"v4-0-config-system-router-certs\" not found",Progressing changed from False to True ("APIServerDeploymentProgressing: deployment/apiserver.openshift-oauth-apiserver: observed generation is 0, desired generation is 1."),Available message changed from "APIServicesAvailable: PreconditionNotReady\nOAuthServerServiceEndpointAccessibleControllerAvailable: service \"oauth-openshift\" not found\nOAuthServerServiceEndpointsEndpointAccessibleControllerAvailable: endpoints \"oauth-openshift\" not found\nReadyIngressNodesAvailable: Authentication requires functional ingress which requires at least one schedulable and ready node. Got 0 worker nodes, 1 master nodes, 0 custom target nodes (none are schedulable or ready for ingress pods)." to "APIServerDeploymentAvailable: no apiserver.openshift-oauth-apiserver pods available on any node.\nAPIServicesAvailable: PreconditionNotReady\nOAuthServerServiceEndpointAccessibleControllerAvailable: service \"oauth-openshift\" not found\nOAuthServerServiceEndpointsEndpointAccessibleControllerAvailable: endpoints \"oauth-openshift\" not found\nReadyIngressNodesAvailable: Authentication requires functional ingress which requires at least one schedulable and ready node. Got 0 worker nodes, 1 master nodes, 0 custom target nodes (none are schedulable or ready for ingress pods)."

openshift-kube-apiserver-operator

kube-apiserver-operator-target-config-controller-targetconfigcontroller

kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/kube-apiserver-pod -n openshift-kube-apiserver because it was missing
(x2)

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-config-observer-configobserver

openshift-kube-scheduler-operator

ObserveTLSSecurityProfile

cipherSuites changed to ["TLS_AES_128_GCM_SHA256" "TLS_AES_256_GCM_SHA384" "TLS_CHACHA20_POLY1305_SHA256" "TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256" "TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256" "TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384" "TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384" "TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256" "TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256"]

openshift-etcd-operator

openshift-cluster-etcd-operator-script-controller-scriptcontroller

etcd-operator

ConfigMapCreated

Created ConfigMap/etcd-scripts -n openshift-etcd because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-target-config-controller-targetconfigcontroller

kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/client-ca -n openshift-kube-apiserver because it was missing

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Degraded message changed from "OAuthServerServiceEndpointAccessibleControllerDegraded: configmap \"v4-0-config-system-service-ca\" not found\nOAuthServiceDegraded: Unable to get oauth server service: service \"oauth-openshift\" not found\nWebhookAuthenticatorControllerDegraded: failed to read service-ca crt bundle: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory\nIngressStateEndpointsDegraded: No endpoints found for oauth-server\nRouterCertsDegraded: neither the custom secret/v4-0-config-system-router-certs -n openshift-authentication or default secret/v4-0-config-system-custom-router-certs -n openshift-authentication could be retrieved: secret \"v4-0-config-system-router-certs\" not found\nOAuthServerConfigObservationDegraded: secret \"v4-0-config-system-router-certs\" not found\nOpenshiftAuthenticationStaticResourcesDegraded: \"oauth-openshift/audit-policy.yaml\" (string): namespaces \"openshift-authentication\" not found\nOpenshiftAuthenticationStaticResourcesDegraded: " to "OAuthServerServiceEndpointAccessibleControllerDegraded: configmap \"v4-0-config-system-service-ca\" not found\nOAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\nOAuthServiceDegraded: Unable to get oauth server service: service \"oauth-openshift\" not found\nWebhookAuthenticatorControllerDegraded: failed to read service-ca crt bundle: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory\nIngressStateEndpointsDegraded: No endpoints found for oauth-server\nRouterCertsDegraded: neither the custom secret/v4-0-config-system-router-certs -n openshift-authentication or default secret/v4-0-config-system-custom-router-certs -n openshift-authentication could be retrieved: secret \"v4-0-config-system-router-certs\" not found\nOAuthServerConfigObservationDegraded: secret \"v4-0-config-system-router-certs\" not found\nOpenshiftAuthenticationStaticResourcesDegraded: \"oauth-openshift/audit-policy.yaml\" (string): namespaces \"openshift-authentication\" not found\nOpenshiftAuthenticationStaticResourcesDegraded: "

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Degraded message changed from "OAuthServerServiceEndpointAccessibleControllerDegraded: configmap \"v4-0-config-system-service-ca\" not found\nOAuthServiceDegraded: Unable to get oauth server service: service \"oauth-openshift\" not found\nWebhookAuthenticatorControllerDegraded: failed to read service-ca crt bundle: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory\nIngressStateEndpointsDegraded: No endpoints found for oauth-server\nRouterCertsDegraded: neither the custom secret/v4-0-config-system-router-certs -n openshift-authentication or default secret/v4-0-config-system-custom-router-certs -n openshift-authentication could be retrieved: secret \"v4-0-config-system-router-certs\" not found\nOAuthServerConfigObservationDegraded: secret \"v4-0-config-system-router-certs\" not found" to "OAuthServerServiceEndpointAccessibleControllerDegraded: configmap \"v4-0-config-system-service-ca\" not found\nOAuthServiceDegraded: Unable to get oauth server service: service \"oauth-openshift\" not found\nWebhookAuthenticatorControllerDegraded: failed to read service-ca crt bundle: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory\nIngressStateEndpointsDegraded: No endpoints found for oauth-server\nRouterCertsDegraded: neither the custom secret/v4-0-config-system-router-certs -n openshift-authentication or default secret/v4-0-config-system-custom-router-certs -n openshift-authentication could be retrieved: secret \"v4-0-config-system-router-certs\" not found\nOAuthServerConfigObservationDegraded: secret \"v4-0-config-system-router-certs\" not found\nOpenshiftAuthenticationStaticResourcesDegraded: \"oauth-openshift/audit-policy.yaml\" (string): namespaces \"openshift-authentication\" not found\nOpenshiftAuthenticationStaticResourcesDegraded: "

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Degraded message changed from "OAuthServiceDegraded: Unable to get oauth server service: service \"oauth-openshift\" not found\nWebhookAuthenticatorControllerDegraded: failed to read service-ca crt bundle: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory\nIngressStateEndpointsDegraded: No endpoints found for oauth-server\nRouterCertsDegraded: neither the custom secret/v4-0-config-system-router-certs -n openshift-authentication or default secret/v4-0-config-system-custom-router-certs -n openshift-authentication could be retrieved: secret \"v4-0-config-system-router-certs\" not found\nOAuthServerConfigObservationDegraded: secret \"v4-0-config-system-router-certs\" not found" to "OAuthServerServiceEndpointAccessibleControllerDegraded: configmap \"v4-0-config-system-service-ca\" not found\nOAuthServiceDegraded: Unable to get oauth server service: service \"oauth-openshift\" not found\nWebhookAuthenticatorControllerDegraded: failed to read service-ca crt bundle: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory\nIngressStateEndpointsDegraded: No endpoints found for oauth-server\nRouterCertsDegraded: neither the custom secret/v4-0-config-system-router-certs -n openshift-authentication or default secret/v4-0-config-system-custom-router-certs -n openshift-authentication could be retrieved: secret \"v4-0-config-system-router-certs\" not found\nOAuthServerConfigObservationDegraded: secret \"v4-0-config-system-router-certs\" not found"

openshift-authentication-operator

cluster-authentication-operator-openshiftauthenticationstaticresources-openshiftauthenticationstaticresources-staticresources

authentication-operator

RoleBindingCreated

Created RoleBinding.rbac.authorization.k8s.io/system:openshift:oauth-servercert-trust -n openshift-config-managed because it was missing

openshift-authentication-operator

cluster-authentication-operator-openshiftauthenticationstaticresources-openshiftauthenticationstaticresources-staticresources

authentication-operator

RoleCreated

Created Role.rbac.authorization.k8s.io/system:openshift:oauth-servercert-trust -n openshift-config-managed because it was missing

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-status-controller-statussyncer_kube-scheduler

openshift-kube-scheduler-operator

OperatorStatusChanged

Status for clusteroperator/kube-scheduler changed: Degraded message changed from "RevisionControllerDegraded: configmap \"kube-scheduler-pod\" not found\nInstallerControllerDegraded: missing required resources: [configmaps: config-0,kube-scheduler-cert-syncer-kubeconfig-0,kube-scheduler-pod-0,scheduler-kubeconfig-0,serviceaccount-ca-0, secrets: localhost-recovery-client-token-0]\nNodeControllerDegraded: All master nodes are ready" to "RevisionControllerDegraded: configmap \"kube-scheduler-pod\" not found\nInstallerControllerDegraded: missing required resources: [configmaps: config-0,kube-scheduler-cert-syncer-kubeconfig-0,kube-scheduler-pod-0,scheduler-kubeconfig-0,serviceaccount-ca-0, secrets: localhost-recovery-client-token-0]\nNodeControllerDegraded: All master nodes are ready\nConfigObservationDegraded: error writing updated observed config: Operation cannot be fulfilled on kubeschedulers.operator.openshift.io \"cluster\": the object has been modified; please apply your changes to the latest version and try again"

openshift-authentication-operator

cluster-authentication-operator-openshiftauthenticationstaticresources-openshiftauthenticationstaticresources-staticresources

authentication-operator

ServiceCreated

Created Service/oauth-openshift -n openshift-authentication because it was missing

openshift-etcd-operator

openshift-cluster-etcd-operator-target-config-controller-targetconfigcontroller

etcd-operator

ConfigMapCreated

Created ConfigMap/etcd-pod -n openshift-etcd because it was missing

openshift-etcd-operator

openshift-cluster-etcd-operator-revisioncontroller

etcd-operator

StartingNewRevision

new revision 1 triggered by "configmap \"etcd-pod-0\" not found"

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-kubecontrollermanagerstaticresources-kubecontrollermanagerstaticresources-staticresources

openshift-kube-scheduler-operator

RoleBindingCreated

Created RoleBinding.rbac.authorization.k8s.io/system:openshift:sa-listing-configmaps -n openshift-kube-scheduler because it was missing

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-kubecontrollermanagerstaticresources-kubecontrollermanagerstaticresources-staticresources

openshift-kube-scheduler-operator

RoleCreated

Created Role.rbac.authorization.k8s.io/system:openshift:sa-listing-configmaps -n openshift-kube-scheduler because it was missing

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-kubecontrollermanagerstaticresources-kubecontrollermanagerstaticresources-staticresources

openshift-kube-scheduler-operator

ClusterRoleBindingCreated

Created ClusterRoleBinding.rbac.authorization.k8s.io/system:openshift:operator:kube-scheduler:public-2 because it was missing

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-kubecontrollermanagerstaticresources-kubecontrollermanagerstaticresources-staticresources

openshift-kube-scheduler-operator

RoleBindingCreated

Created RoleBinding.rbac.authorization.k8s.io/system:openshift:leader-locking-kube-scheduler -n kube-system because it was missing

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-kubecontrollermanagerstaticresources-kubecontrollermanagerstaticresources-staticresources

openshift-kube-scheduler-operator

ConfigMapCreated

Created ConfigMap/kube-scheduler-cert-syncer-kubeconfig -n openshift-kube-scheduler because it was missing

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-targetconfigcontroller

openshift-kube-scheduler-operator

ConfigMapCreated

Created ConfigMap/config -n openshift-kube-scheduler because it was missing

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-config-observer-configobserver

openshift-kube-scheduler-operator

ObservedConfigWriteError

Failed to write observed config: Operation cannot be fulfilled on kubeschedulers.operator.openshift.io "cluster": the object has been modified; please apply your changes to the latest version and try again

openshift-route-controller-manager

default-scheduler

route-controller-manager-74cf48bcc6-kwngm

Scheduled

Successfully assigned openshift-route-controller-manager/route-controller-manager-74cf48bcc6-kwngm to master-0

openshift-apiserver

default-scheduler

apiserver-779f85678d-lrzfz

Scheduled

Successfully assigned openshift-apiserver/apiserver-779f85678d-lrzfz to master-0
(x2)

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-config-observer-configobserver

openshift-kube-scheduler-operator

ObservedConfigChanged

Writing updated observed config:   map[string]any{ +  "servingInfo": map[string]any{ +  "cipherSuites": []any{ +  string("TLS_AES_128_GCM_SHA256"), string("TLS_AES_256_GCM_SHA384"), +  string("TLS_CHACHA20_POLY1305_SHA256"), +  string("TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256"), +  string("TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256"), +  string("TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384"), +  string("TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384"), +  string("TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256"), ..., +  }, +  "minTLSVersion": string("VersionTLS12"), +  },   }
(x2)

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-config-observer-configobserver

openshift-kube-scheduler-operator

ObserveTLSSecurityProfile

minTLSVersion changed to VersionTLS12

openshift-dns

kubelet

dns-default-v5h69

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:2c4d5a681595e428ff4b5083648c13615eed80be9084a3d3fc68a0295079cb12" in 2.674s (2.674s including waiting). Image size: 484187929 bytes.

openshift-cluster-version

openshift-cluster-version

version

PayloadLoaded

Payload loaded version="4.18.35" image="quay.io/openshift-release-dev/ocp-release@sha256:59727c4b3fef19e5149675cf3350735bbfe2c6588a57654b2e4552dd719f58b1" architecture="amd64"

openshift-kube-controller-manager-operator

kube-controller-manager-operator-revisioncontroller

kube-controller-manager-operator

ConfigMapCreated

Created ConfigMap/kube-controller-manager-pod-1 -n openshift-kube-controller-manager because it was missing

openshift-dns

kubelet

dns-default-v5h69

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-status-controller-statussyncer_kube-scheduler

openshift-kube-scheduler-operator

OperatorStatusChanged

Status for clusteroperator/kube-scheduler changed: Degraded message changed from "RevisionControllerDegraded: configmap \"kube-scheduler-pod\" not found\nInstallerControllerDegraded: missing required resources: [configmaps: config-0,kube-scheduler-cert-syncer-kubeconfig-0,kube-scheduler-pod-0,scheduler-kubeconfig-0,serviceaccount-ca-0, secrets: localhost-recovery-client-token-0]\nNodeControllerDegraded: All master nodes are ready\nConfigObservationDegraded: error writing updated observed config: Operation cannot be fulfilled on kubeschedulers.operator.openshift.io \"cluster\": the object has been modified; please apply your changes to the latest version and try again" to "RevisionControllerDegraded: configmap \"kube-scheduler-pod\" not found\nInstallerControllerDegraded: missing required resources: [configmaps: config-0,kube-scheduler-cert-syncer-kubeconfig-0,kube-scheduler-pod-0,scheduler-kubeconfig-0,serviceaccount-ca-0, secrets: localhost-recovery-client-token-0]\nNodeControllerDegraded: All master nodes are ready"

openshift-apiserver

multus

apiserver-779f85678d-lrzfz

AddedInterface

Add eth0 [10.128.0.43/23] from ovn-kubernetes

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/kube-apiserver-pod-1 -n openshift-kube-apiserver because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-resource-sync-controller-resourcesynccontroller

kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/kube-apiserver-client-ca -n openshift-config-managed because it was missing

openshift-apiserver

kubelet

apiserver-779f85678d-lrzfz

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae50e496bd6ae2d27298d997470b7cb0a426eeb8b7e2e9c7187a34cb03993998"

openshift-kube-controller-manager-operator

kube-controller-manager-operator-revisioncontroller

kube-controller-manager-operator

ConfigMapCreated

Created ConfigMap/config-1 -n openshift-kube-controller-manager because it was missing

openshift-oauth-apiserver

multus

apiserver-6fccff6869-f98nk

AddedInterface

Add eth0 [10.128.0.42/23] from ovn-kubernetes

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-targetconfigcontroller

openshift-kube-scheduler-operator

ConfigMapCreated

Created ConfigMap/serviceaccount-ca -n openshift-kube-scheduler because it was missing

openshift-dns

kubelet

dns-default-v5h69

Started

Started container kube-rbac-proxy

openshift-dns

kubelet

dns-default-v5h69

Created

Created container: kube-rbac-proxy

openshift-dns

kubelet

dns-default-v5h69

Created

Created container: dns

openshift-dns

kubelet

dns-default-v5h69

Started

Started container dns

openshift-oauth-apiserver

kubelet

apiserver-6fccff6869-f98nk

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7d4a034950346bcd4e36e9e2f1343e0cf7a10cf544963f33d09c7eb2a1bfc634"

openshift-authentication-operator

cluster-authentication-operator-openshiftauthenticationstaticresources-openshiftauthenticationstaticresources-staticresources

authentication-operator

ConfigMapCreated

Created ConfigMap/audit -n openshift-authentication because it was missing

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-kubecontrollermanagerstaticresources-kubecontrollermanagerstaticresources-staticresources

openshift-kube-scheduler-operator

ServiceCreated

Created Service/scheduler -n openshift-kube-scheduler because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/config-1 -n openshift-kube-apiserver because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-target-config-controller-targetconfigcontroller

kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/kube-apiserver-server-ca -n openshift-kube-apiserver because it was missing

openshift-etcd-operator

openshift-cluster-etcd-operator-target-config-controller-targetconfigcontroller

etcd-operator

ConfigMapCreated

Created ConfigMap/restore-etcd-pod -n openshift-etcd because it was missing

openshift-kube-controller-manager-operator

kube-controller-manager-operator-revisioncontroller

kube-controller-manager-operator

ConfigMapCreated

Created ConfigMap/cluster-policy-controller-config-1 -n openshift-kube-controller-manager because it was missing

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Degraded message changed from "OAuthServerServiceEndpointAccessibleControllerDegraded: configmap \"v4-0-config-system-service-ca\" not found\nOAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\nOAuthServiceDegraded: Unable to get oauth server service: service \"oauth-openshift\" not found\nWebhookAuthenticatorControllerDegraded: failed to read service-ca crt bundle: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory\nIngressStateEndpointsDegraded: No endpoints found for oauth-server\nRouterCertsDegraded: neither the custom secret/v4-0-config-system-router-certs -n openshift-authentication or default secret/v4-0-config-system-custom-router-certs -n openshift-authentication could be retrieved: secret \"v4-0-config-system-router-certs\" not found\nOAuthServerConfigObservationDegraded: secret \"v4-0-config-system-router-certs\" not found\nOpenshiftAuthenticationStaticResourcesDegraded: \"oauth-openshift/audit-policy.yaml\" (string): namespaces \"openshift-authentication\" not found\nOpenshiftAuthenticationStaticResourcesDegraded: " to "OAuthServerServiceEndpointAccessibleControllerDegraded: configmap \"v4-0-config-system-service-ca\" not found\nOAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\nOAuthServiceDegraded: Unable to get oauth server service: service \"oauth-openshift\" not found\nWebhookAuthenticatorControllerDegraded: failed to read service-ca crt bundle: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory\nIngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\nRouterCertsDegraded: neither the custom secret/v4-0-config-system-router-certs -n openshift-authentication or default secret/v4-0-config-system-custom-router-certs -n openshift-authentication could be retrieved: secret \"v4-0-config-system-router-certs\" not found\nOAuthServerConfigObservationDegraded: secret \"v4-0-config-system-router-certs\" not found\nOpenshiftAuthenticationStaticResourcesDegraded: \"oauth-openshift/audit-policy.yaml\" (string): namespaces \"openshift-authentication\" not found\nOpenshiftAuthenticationStaticResourcesDegraded: "

openshift-etcd-operator

openshift-cluster-etcd-operator-revisioncontroller

etcd-operator

ConfigMapCreated

Created ConfigMap/etcd-pod-1 -n openshift-etcd because it was missing

openshift-operator-lifecycle-manager

multus

package-server-manager-7b95f86987-2pg77

AddedInterface

Add eth0 [10.128.0.11/23] from ovn-kubernetes

openshift-operator-lifecycle-manager

kubelet

package-server-manager-7b95f86987-2pg77

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-machine-config-operator

multus

machine-config-operator-84d549f6d5-gm4qr

AddedInterface

Add eth0 [10.128.0.19/23] from ovn-kubernetes

openshift-kube-controller-manager-operator

kube-controller-manager-operator-resource-sync-controller-resourcesynccontroller

kube-controller-manager-operator

ConfigMapCreated

Created ConfigMap/client-ca -n openshift-kube-controller-manager because it was missing

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-targetconfigcontroller

openshift-kube-scheduler-operator

ConfigMapCreated

Created ConfigMap/scheduler-kubeconfig -n openshift-kube-scheduler because it was missing

openshift-marketplace

multus

marketplace-operator-89ccd998f-mvn4t

AddedInterface

Add eth0 [10.128.0.20/23] from ovn-kubernetes

openshift-machine-api

multus

cluster-baremetal-operator-6f69995874-dv6cd

AddedInterface

Add eth0 [10.128.0.16/23] from ovn-kubernetes

openshift-multus

multus

multus-admission-controller-5dbbb8b86f-5rrrh

AddedInterface

Add eth0 [10.128.0.29/23] from ovn-kubernetes

openshift-multus

multus

network-metrics-daemon-srdjm

AddedInterface

Add eth0 [10.128.0.4/23] from ovn-kubernetes

openshift-operator-lifecycle-manager

multus

catalog-operator-68f85b4d6c-fzm28

AddedInterface

Add eth0 [10.128.0.22/23] from ovn-kubernetes

openshift-multus

multus

multus-admission-controller-5dbbb8b86f-5rrrh

AddedInterface

Add eth0 [10.128.0.29/23] from ovn-kubernetes

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Available message changed from "APIServerDeploymentAvailable: no apiserver.openshift-oauth-apiserver pods available on any node.\nAPIServicesAvailable: PreconditionNotReady\nOAuthServerServiceEndpointAccessibleControllerAvailable: service \"oauth-openshift\" not found\nOAuthServerServiceEndpointsEndpointAccessibleControllerAvailable: endpoints \"oauth-openshift\" not found\nReadyIngressNodesAvailable: Authentication requires functional ingress which requires at least one schedulable and ready node. Got 0 worker nodes, 1 master nodes, 0 custom target nodes (none are schedulable or ready for ingress pods)." to "APIServerDeploymentAvailable: no apiserver.openshift-oauth-apiserver pods available on any node.\nAPIServicesAvailable: PreconditionNotReady\nOAuthServerServiceEndpointAccessibleControllerAvailable: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerAvailable: endpoints \"oauth-openshift\" not found\nReadyIngressNodesAvailable: Authentication requires functional ingress which requires at least one schedulable and ready node. Got 0 worker nodes, 1 master nodes, 0 custom target nodes (none are schedulable or ready for ingress pods)."

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Degraded message changed from "OAuthServerServiceEndpointAccessibleControllerDegraded: \"service-ca.crt\" key of the \"openshift-authentication/v4-0-config-system-service-ca\" CM is empty\nOAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\nSystemServiceCAConfigDegraded: Config \"\" has no service CA data\nWebhookAuthenticatorControllerDegraded: failed to read service-ca crt bundle: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory\nIngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\nRouterCertsDegraded: neither the custom secret/v4-0-config-system-router-certs -n openshift-authentication or default secret/v4-0-config-system-custom-router-certs -n openshift-authentication could be retrieved: secret \"v4-0-config-system-router-certs\" not found\nOAuthServerConfigObservationDegraded: secret \"v4-0-config-system-router-certs\" not found\nOpenshiftAuthenticationStaticResourcesDegraded: \"oauth-openshift/audit-policy.yaml\" (string): namespaces \"openshift-authentication\" not found\nOpenshiftAuthenticationStaticResourcesDegraded: " to "OAuthServerServiceEndpointAccessibleControllerDegraded: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\nSystemServiceCAConfigDegraded: Config \"\" has no service CA data\nWebhookAuthenticatorControllerDegraded: failed to read service-ca crt bundle: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory\nIngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\nRouterCertsDegraded: neither the custom secret/v4-0-config-system-router-certs -n openshift-authentication or default secret/v4-0-config-system-custom-router-certs -n openshift-authentication could be retrieved: secret \"v4-0-config-system-router-certs\" not found\nOAuthServerConfigObservationDegraded: secret \"v4-0-config-system-router-certs\" not found\nOpenshiftAuthenticationStaticResourcesDegraded: \"oauth-openshift/audit-policy.yaml\" (string): namespaces \"openshift-authentication\" not found\nOpenshiftAuthenticationStaticResourcesDegraded: "

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Degraded message changed from "OAuthServerServiceEndpointAccessibleControllerDegraded: \"service-ca.crt\" key of the \"openshift-authentication/v4-0-config-system-service-ca\" CM is empty\nOAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\nOAuthServiceDegraded: Unable to get oauth server service: service \"oauth-openshift\" not found\nWebhookAuthenticatorControllerDegraded: failed to read service-ca crt bundle: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory\nIngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\nRouterCertsDegraded: neither the custom secret/v4-0-config-system-router-certs -n openshift-authentication or default secret/v4-0-config-system-custom-router-certs -n openshift-authentication could be retrieved: secret \"v4-0-config-system-router-certs\" not found\nOAuthServerConfigObservationDegraded: secret \"v4-0-config-system-router-certs\" not found\nOpenshiftAuthenticationStaticResourcesDegraded: \"oauth-openshift/audit-policy.yaml\" (string): namespaces \"openshift-authentication\" not found\nOpenshiftAuthenticationStaticResourcesDegraded: " to "OAuthServerServiceEndpointAccessibleControllerDegraded: \"service-ca.crt\" key of the \"openshift-authentication/v4-0-config-system-service-ca\" CM is empty\nOAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\nSystemServiceCAConfigDegraded: Config \"\" has no service CA data\nWebhookAuthenticatorControllerDegraded: failed to read service-ca crt bundle: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory\nIngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\nRouterCertsDegraded: neither the custom secret/v4-0-config-system-router-certs -n openshift-authentication or default secret/v4-0-config-system-custom-router-certs -n openshift-authentication could be retrieved: secret \"v4-0-config-system-router-certs\" not found\nOAuthServerConfigObservationDegraded: secret \"v4-0-config-system-router-certs\" not found\nOpenshiftAuthenticationStaticResourcesDegraded: \"oauth-openshift/audit-policy.yaml\" (string): namespaces \"openshift-authentication\" not found\nOpenshiftAuthenticationStaticResourcesDegraded: "

openshift-monitoring

multus

cluster-monitoring-operator-58845fbb57-nljsr

AddedInterface

Add eth0 [10.128.0.24/23] from ovn-kubernetes

openshift-kube-apiserver-operator

kube-apiserver-operator-resource-sync-controller-resourcesynccontroller

kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/kube-apiserver-server-ca -n openshift-config-managed because it was missing

openshift-kube-controller-manager-operator

kube-controller-manager-operator-targetconfigcontroller

kube-controller-manager-operator

ConfigMapUpdated

Updated ConfigMap/config -n openshift-kube-controller-manager: cause by changes in data.config.yaml

openshift-monitoring

multus

cluster-monitoring-operator-58845fbb57-nljsr

AddedInterface

Add eth0 [10.128.0.24/23] from ovn-kubernetes

openshift-multus

multus

network-metrics-daemon-srdjm

AddedInterface

Add eth0 [10.128.0.4/23] from ovn-kubernetes

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-kubecontrollermanagerstaticresources-kubecontrollermanagerstaticresources-staticresources

openshift-kube-scheduler-operator

ServiceAccountCreated

Created ServiceAccount/openshift-kube-scheduler-sa -n openshift-kube-scheduler because it was missing

openshift-operator-lifecycle-manager

multus

olm-operator-5c9796789-tjm9l

AddedInterface

Add eth0 [10.128.0.7/23] from ovn-kubernetes

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Degraded message changed from "OAuthServerServiceEndpointAccessibleControllerDegraded: configmap \"v4-0-config-system-service-ca\" not found\nOAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\nOAuthServiceDegraded: Unable to get oauth server service: service \"oauth-openshift\" not found\nWebhookAuthenticatorControllerDegraded: failed to read service-ca crt bundle: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory\nIngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\nRouterCertsDegraded: neither the custom secret/v4-0-config-system-router-certs -n openshift-authentication or default secret/v4-0-config-system-custom-router-certs -n openshift-authentication could be retrieved: secret \"v4-0-config-system-router-certs\" not found\nOAuthServerConfigObservationDegraded: secret \"v4-0-config-system-router-certs\" not found\nOpenshiftAuthenticationStaticResourcesDegraded: \"oauth-openshift/audit-policy.yaml\" (string): namespaces \"openshift-authentication\" not found\nOpenshiftAuthenticationStaticResourcesDegraded: " to "OAuthServerServiceEndpointAccessibleControllerDegraded: \"service-ca.crt\" key of the \"openshift-authentication/v4-0-config-system-service-ca\" CM is empty\nOAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\nOAuthServiceDegraded: Unable to get oauth server service: service \"oauth-openshift\" not found\nWebhookAuthenticatorControllerDegraded: failed to read service-ca crt bundle: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory\nIngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\nRouterCertsDegraded: neither the custom secret/v4-0-config-system-router-certs -n openshift-authentication or default secret/v4-0-config-system-custom-router-certs -n openshift-authentication could be retrieved: secret \"v4-0-config-system-router-certs\" not found\nOAuthServerConfigObservationDegraded: secret \"v4-0-config-system-router-certs\" not found\nOpenshiftAuthenticationStaticResourcesDegraded: \"oauth-openshift/audit-policy.yaml\" (string): namespaces \"openshift-authentication\" not found\nOpenshiftAuthenticationStaticResourcesDegraded: "

openshift-machine-api

multus

cluster-baremetal-operator-6f69995874-dv6cd

AddedInterface

Add eth0 [10.128.0.16/23] from ovn-kubernetes

openshift-operator-lifecycle-manager

kubelet

package-server-manager-7b95f86987-2pg77

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6bba3f73c0066e42b24839e0d29f5dce2f36436f0a11f9f5e1029bccc5ed6578"

openshift-operator-lifecycle-manager

kubelet

package-server-manager-7b95f86987-2pg77

Started

Started container kube-rbac-proxy

openshift-operator-lifecycle-manager

kubelet

package-server-manager-7b95f86987-2pg77

Created

Created container: kube-rbac-proxy

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/kube-apiserver-cert-syncer-kubeconfig-1 -n openshift-kube-apiserver because it was missing

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-kubecontrollermanagerstaticresources-kubecontrollermanagerstaticresources-staticresources

openshift-kube-scheduler-operator

ClusterRoleBindingCreated

Created ClusterRoleBinding.rbac.authorization.k8s.io/system:openshift:operator:kube-scheduler-recovery because it was missing

openshift-authentication-operator

cluster-authentication-operator-config-observer-configobserver

authentication-operator

ObservedConfigChanged

Writing updated section ("oauthServer") of observed config: " map[string]any{\n \t\"corsAllowedOrigins\": []any{string(`//127\\.0\\.0\\.1(:|$)`), string(\"//localhost(:|$)\")},\n \t\"oauthConfig\": map[string]any{\"assetPublicURL\": string(\"\"), \"loginURL\": string(\"https://api.sno.openstack.lab:6443\"), \"templates\": map[string]any{\"error\": string(\"/var/config/system/secrets/v4-0-config-system-ocp-branding-templ\"...), \"login\": string(\"/var/config/system/secrets/v4-0-config-system-ocp-branding-templ\"...), \"providerSelection\": string(\"/var/config/system/secrets/v4-0-config-system-ocp-branding-templ\"...)}, \"tokenConfig\": map[string]any{\"accessTokenMaxAgeSeconds\": float64(86400), \"authorizeTokenMaxAgeSeconds\": float64(300)}},\n \t\"serverArguments\": map[string]any{\"audit-log-format\": []any{string(\"json\")}, \"audit-log-maxbackup\": []any{string(\"10\")}, \"audit-log-maxsize\": []any{string(\"100\")}, \"audit-log-path\": []any{string(\"/var/log/oauth-server/audit.log\")}, ...},\n \t\"servingInfo\": map[string]any{\n \t\t\"cipherSuites\": []any{string(\"TLS_AES_128_GCM_SHA256\"), string(\"TLS_AES_256_GCM_SHA384\"), string(\"TLS_CHACHA20_POLY1305_SHA256\"), string(\"TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256\"), ...},\n \t\t\"minTLSVersion\": string(\"VersionTLS12\"),\n+ \t\t\"namedCertificates\": []any{\n+ \t\t\tmap[string]any{\n+ \t\t\t\t\"certFile\": string(\"/var/config/system/secrets/v4-0-config-system-router-certs/apps.\"...),\n+ \t\t\t\t\"keyFile\": string(\"/var/config/system/secrets/v4-0-config-system-router-certs/apps.\"...),\n+ \t\t\t\t\"names\": []any{string(\"*.apps.sno.openstack.lab\")},\n+ \t\t\t},\n+ \t\t},\n \t},\n \t\"volumesToMount\": map[string]any{\"identityProviders\": string(\"{}\")},\n }\n"

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-kubecontrollermanagerstaticresources-kubecontrollermanagerstaticresources-staticresources

openshift-kube-scheduler-operator

ServiceAccountCreated

Created ServiceAccount/localhost-recovery-client -n openshift-kube-scheduler because it was missing

openshift-monitoring

kubelet

cluster-monitoring-operator-58845fbb57-nljsr

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a746a87b784ea1caa278fd0e012554f9df520b6fff665ea0bc4c83f487fed113"

openshift-monitoring

kubelet

cluster-monitoring-operator-58845fbb57-nljsr

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a746a87b784ea1caa278fd0e012554f9df520b6fff665ea0bc4c83f487fed113"

openshift-multus

kubelet

network-metrics-daemon-srdjm

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:759fb1d5353dbbadd443f38631d977ca3aed9787b873be05cc9660532a252739"

openshift-machine-api

kubelet

cluster-baremetal-operator-6f69995874-dv6cd

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f933312f49083e8746fc41ab5e46a9a757b448374f14971e256ebcb36f11dd97"

openshift-authentication-operator

cluster-authentication-operator-routercertsdomainvalidationcontroller

authentication-operator

SecretCreated

Created Secret/v4-0-config-system-router-certs -n openshift-authentication because it was missing

openshift-authentication-operator

cluster-authentication-operator-config-observer-configobserver

authentication-operator

ObserveRouterSecret

namedCertificates changed to []interface {}{map[string]interface {}{"certFile":"/var/config/system/secrets/v4-0-config-system-router-certs/apps.sno.openstack.lab", "keyFile":"/var/config/system/secrets/v4-0-config-system-router-certs/apps.sno.openstack.lab", "names":[]interface {}{"*.apps.sno.openstack.lab"}}}

openshift-operator-lifecycle-manager

kubelet

olm-operator-5c9796789-tjm9l

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6bba3f73c0066e42b24839e0d29f5dce2f36436f0a11f9f5e1029bccc5ed6578"

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Degraded message changed from "OAuthServerServiceEndpointAccessibleControllerDegraded: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\nSystemServiceCAConfigDegraded: Config \"\" has no service CA data\nWebhookAuthenticatorControllerDegraded: failed to read service-ca crt bundle: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory\nIngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\nRouterCertsDegraded: neither the custom secret/v4-0-config-system-router-certs -n openshift-authentication or default secret/v4-0-config-system-custom-router-certs -n openshift-authentication could be retrieved: secret \"v4-0-config-system-router-certs\" not found\nOAuthServerConfigObservationDegraded: secret \"v4-0-config-system-router-certs\" not found\nOpenshiftAuthenticationStaticResourcesDegraded: \"oauth-openshift/audit-policy.yaml\" (string): namespaces \"openshift-authentication\" not found\nOpenshiftAuthenticationStaticResourcesDegraded: " to "OAuthServerServiceEndpointAccessibleControllerDegraded: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\nSystemServiceCAConfigDegraded: Config \"\" has no service CA data\nWebhookAuthenticatorControllerDegraded: failed to read service-ca crt bundle: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory\nIngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\nRouterCertsDegraded: neither the custom secret/v4-0-config-system-router-certs -n openshift-authentication or default secret/v4-0-config-system-custom-router-certs -n openshift-authentication could be retrieved: secret \"v4-0-config-system-router-certs\" not found\nOpenshiftAuthenticationStaticResourcesDegraded: \"oauth-openshift/audit-policy.yaml\" (string): namespaces \"openshift-authentication\" not found\nOpenshiftAuthenticationStaticResourcesDegraded: "

openshift-marketplace

kubelet

marketplace-operator-89ccd998f-mvn4t

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:632e80bba5077068ecca05fddb95aedebad4493af6f36152c01c6ae490975b62"

openshift-machine-api

kubelet

cluster-baremetal-operator-6f69995874-dv6cd

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f933312f49083e8746fc41ab5e46a9a757b448374f14971e256ebcb36f11dd97"

openshift-kube-controller-manager-operator

kube-controller-manager-operator-revisioncontroller

kube-controller-manager-operator

ConfigMapCreated

Created ConfigMap/controller-manager-kubeconfig-1 -n openshift-kube-controller-manager because it was missing

openshift-multus

kubelet

multus-admission-controller-5dbbb8b86f-5rrrh

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bcb08821551e9a5b9f82aa794bcea673279cefb93cb47492e19ccac5e2cf18fe"

openshift-multus

kubelet

network-metrics-daemon-srdjm

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:759fb1d5353dbbadd443f38631d977ca3aed9787b873be05cc9660532a252739"

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/bound-sa-token-signing-certs-1 -n openshift-kube-apiserver because it was missing

openshift-operator-lifecycle-manager

kubelet

catalog-operator-68f85b4d6c-fzm28

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6bba3f73c0066e42b24839e0d29f5dce2f36436f0a11f9f5e1029bccc5ed6578"

openshift-multus

kubelet

multus-admission-controller-5dbbb8b86f-5rrrh

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bcb08821551e9a5b9f82aa794bcea673279cefb93cb47492e19ccac5e2cf18fe"

openshift-etcd-operator

openshift-cluster-etcd-operator-revisioncontroller

etcd-operator

ConfigMapCreated

Created ConfigMap/etcd-endpoints-1 -n openshift-etcd because it was missing

openshift-oauth-apiserver

replicaset-controller

apiserver-6fccff6869

SuccessfulDelete

Deleted pod: apiserver-6fccff6869-f98nk

openshift-kube-apiserver-operator

kube-apiserver-operator-node-kubeconfig-controller-nodekubeconfigcontroller

kube-apiserver-operator

SecretCreated

Created Secret/node-kubeconfigs -n openshift-kube-apiserver because it was missing

openshift-oauth-apiserver

default-scheduler

apiserver-bc9b556d6-vdnq2

FailedScheduling

0/1 nodes are available: 1 node(s) didn't match pod anti-affinity rules. preemption: 0/1 nodes are available: 1 node(s) didn't match pod anti-affinity rules.

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Progressing message changed from "APIServerDeploymentProgressing: deployment/apiserver.openshift-oauth-apiserver: observed generation is 0, desired generation is 1." to "APIServerDeploymentProgressing: deployment/apiserver.openshift-oauth-apiserver: observed generation is 1, desired generation is 2."

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-kubecontrollermanagerstaticresources-kubecontrollermanagerstaticresources-staticresources

openshift-kube-scheduler-operator

SecretCreated

Created Secret/localhost-recovery-client-token -n openshift-kube-scheduler because it was missing

openshift-kube-controller-manager-operator

kube-controller-manager-operator-revisioncontroller

kube-controller-manager-operator

ConfigMapCreated

Created ConfigMap/kube-controller-cert-syncer-kubeconfig-1 -n openshift-kube-controller-manager because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-status-controller-statussyncer_kube-apiserver

kube-apiserver-operator

OperatorStatusChanged

Status for clusteroperator/kube-apiserver changed: Degraded message changed from "NodeControllerDegraded: All master nodes are ready\nNodeKubeconfigControllerDegraded: \"secret/node-kubeconfigs\": configmap \"kube-apiserver-server-ca\" not found\nRevisionControllerDegraded: configmap \"kube-apiserver-pod\" not found\nInstallerControllerDegraded: missing required resources: [configmaps: bound-sa-token-signing-certs-0,config-0,etcd-serving-ca-0,kube-apiserver-audit-policies-0,kube-apiserver-cert-syncer-kubeconfig-0,kube-apiserver-pod-0,kubelet-serving-ca-0,sa-token-signing-certs-0, secrets: etcd-client-0,localhost-recovery-client-token-0,localhost-recovery-serving-certkey-0]" to "NodeControllerDegraded: All master nodes are ready\nRevisionControllerDegraded: configmap \"kube-apiserver-pod\" not found\nInstallerControllerDegraded: missing required resources: [configmaps: bound-sa-token-signing-certs-0,config-0,etcd-serving-ca-0,kube-apiserver-audit-policies-0,kube-apiserver-cert-syncer-kubeconfig-0,kube-apiserver-pod-0,kubelet-serving-ca-0,sa-token-signing-certs-0, secrets: etcd-client-0,localhost-recovery-client-token-0,localhost-recovery-serving-certkey-0]"

openshift-oauth-apiserver

replicaset-controller

apiserver-bc9b556d6

SuccessfulCreate

Created pod: apiserver-bc9b556d6-vdnq2

openshift-oauth-apiserver

deployment-controller

apiserver

ScalingReplicaSet

Scaled down replica set apiserver-6fccff6869 to 0 from 1

openshift-oauth-apiserver

deployment-controller

apiserver

ScalingReplicaSet

Scaled up replica set apiserver-bc9b556d6 to 1 from 0

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/etcd-serving-ca-1 -n openshift-kube-apiserver because it was missing

openshift-machine-config-operator

kubelet

machine-config-operator-84d549f6d5-gm4qr

Started

Started container machine-config-operator

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-status-controller-statussyncer_kube-scheduler

openshift-kube-scheduler-operator

OperatorStatusChanged

Status for clusteroperator/kube-scheduler changed: Degraded message changed from "RevisionControllerDegraded: configmap \"kube-scheduler-pod\" not found\nInstallerControllerDegraded: missing required resources: [configmaps: config-0,kube-scheduler-cert-syncer-kubeconfig-0,kube-scheduler-pod-0,scheduler-kubeconfig-0,serviceaccount-ca-0, secrets: localhost-recovery-client-token-0]\nNodeControllerDegraded: All master nodes are ready" to "RevisionControllerDegraded: configmap \"kube-scheduler-pod\" not found\nInstallerControllerDegraded: missing required resources: [configmaps: config-0,kube-scheduler-cert-syncer-kubeconfig-0,kube-scheduler-pod-0,scheduler-kubeconfig-0,serviceaccount-ca-0, secrets: localhost-recovery-client-token-0]\nNodeControllerDegraded: All master nodes are ready\nTargetConfigControllerDegraded: \"serviceaccount/localhost-recovery-client\": serviceaccounts \"localhost-recovery-client\" not found"

openshift-etcd-operator

openshift-cluster-etcd-operator-revisioncontroller

etcd-operator

ConfigMapCreated

Created ConfigMap/etcd-all-bundles-1 -n openshift-etcd because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/kube-apiserver-server-ca-1 -n openshift-kube-apiserver because it was missing

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-targetconfigcontroller

openshift-kube-scheduler-operator

ConfigMapCreated

Created ConfigMap/kube-scheduler-pod -n openshift-kube-scheduler because it was missing

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-revisioncontroller

openshift-kube-scheduler-operator

StartingNewRevision

new revision 1 triggered by "configmap \"kube-scheduler-pod-0\" not found"

openshift-apiserver

kubelet

apiserver-779f85678d-lrzfz

Started

Started container fix-audit-permissions

openshift-apiserver

kubelet

apiserver-779f85678d-lrzfz

Created

Created container: fix-audit-permissions

openshift-oauth-apiserver

kubelet

apiserver-6fccff6869-f98nk

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7d4a034950346bcd4e36e9e2f1343e0cf7a10cf544963f33d09c7eb2a1bfc634" in 5.003s (5.003s including waiting). Image size: 505345991 bytes.

openshift-oauth-apiserver

kubelet

apiserver-6fccff6869-f98nk

Created

Created container: fix-audit-permissions

openshift-oauth-apiserver

kubelet

apiserver-6fccff6869-f98nk

Started

Started container fix-audit-permissions

openshift-apiserver

kubelet

apiserver-779f85678d-lrzfz

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae50e496bd6ae2d27298d997470b7cb0a426eeb8b7e2e9c7187a34cb03993998" in 4.947s (4.947s including waiting). Image size: 589386806 bytes.

openshift-machine-config-operator

kubelet

machine-config-operator-84d549f6d5-gm4qr

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:af0fe0ca926422a6471d5bf22fc0e682c36c24fba05496a3bdfac0b7d3733015" already present on machine

openshift-machine-config-operator

kubelet

machine-config-operator-84d549f6d5-gm4qr

Created

Created container: machine-config-operator

openshift-machine-config-operator

machine-config-operator

master-0

FeatureGatesInitialized

FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{"AWSEFSDriverVolumeMetrics", "AdminNetworkPolicy", "AlibabaPlatform", "AzureWorkloadIdentity", "BareMetalLoadBalancer", "BuildCSIVolumes", "ChunkSizeMiB", "CloudDualStackNodeIPs", "DisableKubeletCloudCredentialProviders", "GCPLabelsTags", "HardwareSpeed", "IngressControllerLBSubnetsAWS", "KMSv1", "ManagedBootImages", "ManagedBootImagesAWS", "MultiArchInstallAWS", "MultiArchInstallGCP", "NetworkDiagnosticsConfig", "NetworkLiveMigration", "NetworkSegmentation", "NewOLM", "NodeDisruptionPolicy", "OnClusterBuild", "PersistentIPsForVirtualization", "PrivateHostedZoneAWS", "SetEIPForNLBIngressController", "StreamingCollectionEncodingToJSON", "StreamingCollectionEncodingToProtobuf", "VSphereControlPlaneMachineSet", "VSphereDriverConfiguration", "VSphereMultiVCenters", "VSphereStaticIPs", "ValidatingAdmissionPolicy"}, Disabled:[]v1.FeatureGateName{"AWSClusterHostedDNS", "AdditionalRoutingCapabilities", "AutomatedEtcdBackup", "BootcNodeManagement", "CSIDriverSharedResource", "ClusterAPIInstall", "ClusterAPIInstallIBMCloud", "ClusterMonitoringConfig", "ConsolePluginContentSecurityPolicy", "DNSNameResolver", "DynamicResourceAllocation", "EtcdBackendQuota", "EventedPLEG", "Example", "ExternalOIDC", "ExternalOIDCWithUIDAndExtraClaimMappings", "GCPClusterHostedDNS", "GatewayAPI", "ImageStreamImportMode", "IngressControllerDynamicConfigurationManager", "InsightsConfig", "InsightsConfigAPI", "InsightsOnDemandDataGather", "InsightsRuntimeExtractor", "MachineAPIMigration", "MachineAPIOperatorDisableMachineHealthCheckController", "MachineAPIProviderOpenStack", "MachineConfigNodes", "MaxUnavailableStatefulSet", "MetricsCollectionProfiles", "MinimumKubeletVersion", "MixedCPUsAllocation", "MultiArchInstallAzure", "NodeSwap", "NutanixMultiSubnets", "OVNObservability", "OpenShiftPodSecurityAdmission", "PinnedImages", "PlatformOperators", "ProcMountType", "RouteAdvertisements", "RouteExternalCertificate", "ServiceAccountTokenNodeBinding", "SignatureStores", "SigstoreImageVerification", "TranslateStreamCloseWebsocketRequests", "UpgradeStatus", "UserNamespacesPodSecurityStandards", "UserNamespacesSupport", "VSphereMultiNetworks", "VolumeAttributesClass", "VolumeGroupSnapshot"}}

openshift-machine-config-operator

kubelet

machine-config-operator-84d549f6d5-gm4qr

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-machine-config-operator

kubelet

machine-config-operator-84d549f6d5-gm4qr

Created

Created container: kube-rbac-proxy

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Degraded message changed from "OAuthServerServiceEndpointAccessibleControllerDegraded: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\nSystemServiceCAConfigDegraded: Config \"\" has no service CA data\nWebhookAuthenticatorControllerDegraded: failed to read service-ca crt bundle: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory\nIngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\nRouterCertsDegraded: neither the custom secret/v4-0-config-system-router-certs -n openshift-authentication or default secret/v4-0-config-system-custom-router-certs -n openshift-authentication could be retrieved: secret \"v4-0-config-system-router-certs\" not found\nOpenshiftAuthenticationStaticResourcesDegraded: \"oauth-openshift/audit-policy.yaml\" (string): namespaces \"openshift-authentication\" not found\nOpenshiftAuthenticationStaticResourcesDegraded: " to "OAuthServerServiceEndpointAccessibleControllerDegraded: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\nWebhookAuthenticatorControllerDegraded: failed to read service-ca crt bundle: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory\nIngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\nRouterCertsDegraded: neither the custom secret/v4-0-config-system-router-certs -n openshift-authentication or default secret/v4-0-config-system-custom-router-certs -n openshift-authentication could be retrieved: secret \"v4-0-config-system-router-certs\" not found\nOpenshiftAuthenticationStaticResourcesDegraded: \"oauth-openshift/audit-policy.yaml\" (string): namespaces \"openshift-authentication\" not found\nOpenshiftAuthenticationStaticResourcesDegraded: "

openshift-machine-config-operator

kubelet

machine-config-operator-84d549f6d5-gm4qr

Started

Started container kube-rbac-proxy

openshift-kube-controller-manager-operator

kube-controller-manager-operator-revisioncontroller

kube-controller-manager-operator

ConfigMapCreated

Created ConfigMap/serviceaccount-ca-1 -n openshift-kube-controller-manager because it was missing

openshift-kube-controller-manager-operator

kube-controller-manager-operator-targetconfigcontroller

kube-controller-manager-operator

ConfigMapUpdated

Updated ConfigMap/cluster-policy-controller-config -n openshift-kube-controller-manager: cause by changes in data.config.yaml

openshift-machine-config-operator

machine-config-operator

machine-config-operator

SecretCreated

Created Secret/master-user-data-managed -n openshift-machine-api because it was missing

openshift-kube-controller-manager-operator

kube-controller-manager-operator-revisioncontroller

kube-controller-manager-operator

ConfigMapCreated

Created ConfigMap/service-ca-1 -n openshift-kube-controller-manager because it was missing

openshift-machine-config-operator

machine-config-operator

machine-config-operator

ClusterRoleCreated

Created ClusterRole.rbac.authorization.k8s.io/machine-config-daemon because it was missing

openshift-machine-config-operator

machine-config-operator

machine-config-operator

ClusterRoleCreated

Created ClusterRole.rbac.authorization.k8s.io/machine-config-daemon-events because it was missing

openshift-etcd-operator

openshift-cluster-etcd-operator-revisioncontroller

etcd-operator

SecretCreated

Created Secret/etcd-all-certs-1 -n openshift-etcd because it was missing

openshift-machine-config-operator

machine-config-operator

machine-config-operator

RoleCreated

Created Role.rbac.authorization.k8s.io/mcd-prometheus-k8s -n openshift-machine-config-operator because it was missing

openshift-machine-config-operator

machineconfigoperator

machine-config

OperatorVersionChanged

clusteroperator/machine-config started a version change from [] to [{operator 4.18.35} {operator-image quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:af0fe0ca926422a6471d5bf22fc0e682c36c24fba05496a3bdfac0b7d3733015}]

openshift-apiserver

kubelet

apiserver-779f85678d-lrzfz

Created

Created container: openshift-apiserver

openshift-machine-config-operator

machine-config-operator

machine-config-operator

RoleBindingCreated

Created RoleBinding.rbac.authorization.k8s.io/machine-config-daemon-events -n default because it was missing

openshift-machine-config-operator

machine-config-operator

machine-config-operator

RoleBindingCreated

Created RoleBinding.rbac.authorization.k8s.io/mcd-prometheus-k8s -n openshift-machine-config-operator because it was missing

openshift-apiserver

kubelet

apiserver-779f85678d-lrzfz

Started

Started container openshift-apiserver

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-revisioncontroller

openshift-kube-scheduler-operator

ConfigMapCreated

Created ConfigMap/config-1 -n openshift-kube-scheduler because it was missing

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-revisioncontroller

openshift-kube-scheduler-operator

ConfigMapCreated

Created ConfigMap/kube-scheduler-pod-1 -n openshift-kube-scheduler because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/kubelet-serving-ca-1 -n openshift-kube-apiserver because it was missing

openshift-machine-config-operator

machine-config-operator

machine-config-operator

RoleBindingCreated

Created RoleBinding.rbac.authorization.k8s.io/machine-config-daemon-events -n openshift-machine-config-operator because it was missing

openshift-machine-config-operator

machine-config-operator

machine-config-operator

RoleCreated

Created Role.rbac.authorization.k8s.io/machine-config-daemon -n openshift-machine-config-operator because it was missing

openshift-kube-controller-manager-operator

kube-controller-manager-operator-revisioncontroller

kube-controller-manager-operator

ConfigMapCreated

Created ConfigMap/recycler-config-1 -n openshift-kube-controller-manager because it was missing

openshift-apiserver

kubelet

apiserver-779f85678d-lrzfz

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae50e496bd6ae2d27298d997470b7cb0a426eeb8b7e2e9c7187a34cb03993998" already present on machine

openshift-cluster-olm-operator

olm-status-controller-statussyncer_olm

cluster-olm-operator

OperatorStatusChanged

Status for clusteroperator/olm changed: Degraded message changed from "CatalogdClusterCatalogOpenshiftCommunityOperatorsDegraded: Internal error occurred: failed calling webhook \"inject-metadata-name.olm.operatorframework.io\": failed to call webhook: Post \"https://catalogd-service.openshift-catalogd.svc:9443/mutate-olm-operatorframework-io-v1-clustercatalog?timeout=10s\": no endpoints available for service \"catalogd-service\"\nCatalogdClusterCatalogOpenshiftRedhatMarketplaceDegraded: Internal error occurred: failed calling webhook \"inject-metadata-name.olm.operatorframework.io\": failed to call webhook: Post \"https://catalogd-service.openshift-catalogd.svc:9443/mutate-olm-operatorframework-io-v1-clustercatalog?timeout=10s\": no endpoints available for service \"catalogd-service\"\nCatalogdClusterCatalogOpenshiftCertifiedOperatorsDegraded: Internal error occurred: failed calling webhook \"inject-metadata-name.olm.operatorframework.io\": failed to call webhook: Post \"https://catalogd-service.openshift-catalogd.svc:9443/mutate-olm-operatorframework-io-v1-clustercatalog?timeout=10s\": no endpoints available for service \"catalogd-service\"" to "CatalogdClusterCatalogOpenshiftRedhatMarketplaceDegraded: Internal error occurred: failed calling webhook \"inject-metadata-name.olm.operatorframework.io\": failed to call webhook: Post \"https://catalogd-service.openshift-catalogd.svc:9443/mutate-olm-operatorframework-io-v1-clustercatalog?timeout=10s\": no endpoints available for service \"catalogd-service\"\nCatalogdClusterCatalogOpenshiftCertifiedOperatorsDegraded: Internal error occurred: failed calling webhook \"inject-metadata-name.olm.operatorframework.io\": failed to call webhook: Post \"https://catalogd-service.openshift-catalogd.svc:9443/mutate-olm-operatorframework-io-v1-clustercatalog?timeout=10s\": no endpoints available for service \"catalogd-service\""

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-targetconfigcontroller

openshift-kube-scheduler-operator

ConfigMapUpdated

Updated ConfigMap/serviceaccount-ca -n openshift-kube-scheduler: cause by changes in data.ca-bundle.crt

openshift-machine-config-operator

machine-config-operator

machine-config-operator

RoleBindingCreated

Created RoleBinding.rbac.authorization.k8s.io/machine-config-daemon -n openshift-machine-config-operator because it was missing

openshift-kube-controller-manager-operator

kube-controller-manager-operator-revisioncontroller

kube-controller-manager-operator

SecretCreated

Created Secret/serving-cert-1 -n openshift-kube-controller-manager because it was missing

openshift-kube-controller-manager-operator

kube-controller-manager-operator-revisioncontroller

kube-controller-manager-operator

SecretCreated

Created Secret/service-account-private-key-1 -n openshift-kube-controller-manager because it was missing

openshift-cluster-olm-operator

olm-status-controller-statussyncer_olm

cluster-olm-operator

OperatorStatusChanged

Status for clusteroperator/olm changed: Degraded message changed from "CatalogdClusterCatalogOpenshiftRedhatMarketplaceDegraded: Internal error occurred: failed calling webhook \"inject-metadata-name.olm.operatorframework.io\": failed to call webhook: Post \"https://catalogd-service.openshift-catalogd.svc:9443/mutate-olm-operatorframework-io-v1-clustercatalog?timeout=10s\": no endpoints available for service \"catalogd-service\"\nCatalogdClusterCatalogOpenshiftCertifiedOperatorsDegraded: Internal error occurred: failed calling webhook \"inject-metadata-name.olm.operatorframework.io\": failed to call webhook: Post \"https://catalogd-service.openshift-catalogd.svc:9443/mutate-olm-operatorframework-io-v1-clustercatalog?timeout=10s\": no endpoints available for service \"catalogd-service\"" to "CatalogdClusterCatalogOpenshiftRedhatMarketplaceDegraded: Internal error occurred: failed calling webhook \"inject-metadata-name.olm.operatorframework.io\": failed to call webhook: Post \"https://catalogd-service.openshift-catalogd.svc:9443/mutate-olm-operatorframework-io-v1-clustercatalog?timeout=10s\": no endpoints available for service \"catalogd-service\""

openshift-machine-config-operator

machine-config-operator

machine-config-operator

ClusterRoleBindingCreated

Created ClusterRoleBinding.rbac.authorization.k8s.io/machine-config-daemon because it was missing

openshift-cluster-olm-operator

olm-status-controller-statussyncer_olm

cluster-olm-operator

OperatorStatusChanged

Status for clusteroperator/olm changed: Degraded message changed from "CatalogdClusterCatalogOpenshiftCommunityOperatorsDegraded: Internal error occurred: failed calling webhook \"inject-metadata-name.olm.operatorframework.io\": failed to call webhook: Post \"https://catalogd-service.openshift-catalogd.svc:9443/mutate-olm-operatorframework-io-v1-clustercatalog?timeout=10s\": no endpoints available for service \"catalogd-service\"\nCatalogdClusterCatalogOpenshiftRedhatMarketplaceDegraded: Internal error occurred: failed calling webhook \"inject-metadata-name.olm.operatorframework.io\": failed to call webhook: Post \"https://catalogd-service.openshift-catalogd.svc:9443/mutate-olm-operatorframework-io-v1-clustercatalog?timeout=10s\": no endpoints available for service \"catalogd-service\"\nCatalogdClusterCatalogOpenshiftCertifiedOperatorsDegraded: Internal error occurred: failed calling webhook \"inject-metadata-name.olm.operatorframework.io\": failed to call webhook: Post \"https://catalogd-service.openshift-catalogd.svc:9443/mutate-olm-operatorframework-io-v1-clustercatalog?timeout=10s\": no endpoints available for service \"catalogd-service\"\nCatalogdClusterCatalogOpenshiftRedhatOperatorsDegraded: Internal error occurred: failed calling webhook \"inject-metadata-name.olm.operatorframework.io\": failed to call webhook: Post \"https://catalogd-service.openshift-catalogd.svc:9443/mutate-olm-operatorframework-io-v1-clustercatalog?timeout=10s\": no endpoints available for service \"catalogd-service\"" to "CatalogdClusterCatalogOpenshiftCommunityOperatorsDegraded: Internal error occurred: failed calling webhook \"inject-metadata-name.olm.operatorframework.io\": failed to call webhook: Post \"https://catalogd-service.openshift-catalogd.svc:9443/mutate-olm-operatorframework-io-v1-clustercatalog?timeout=10s\": no endpoints available for service \"catalogd-service\"\nCatalogdClusterCatalogOpenshiftRedhatMarketplaceDegraded: Internal error occurred: failed calling webhook \"inject-metadata-name.olm.operatorframework.io\": failed to call webhook: Post \"https://catalogd-service.openshift-catalogd.svc:9443/mutate-olm-operatorframework-io-v1-clustercatalog?timeout=10s\": no endpoints available for service \"catalogd-service\"\nCatalogdClusterCatalogOpenshiftCertifiedOperatorsDegraded: Internal error occurred: failed calling webhook \"inject-metadata-name.olm.operatorframework.io\": failed to call webhook: Post \"https://catalogd-service.openshift-catalogd.svc:9443/mutate-olm-operatorframework-io-v1-clustercatalog?timeout=10s\": no endpoints available for service \"catalogd-service\""

openshift-machine-config-operator

machine-config-operator

machine-config-operator

ValidatingAdmissionPolicyCreated

Created ValidatingAdmissionPolicy.admissionregistration.k8s.io/mcn-guards because it was missing

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Degraded message changed from "OAuthServerServiceEndpointAccessibleControllerDegraded: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\nWebhookAuthenticatorControllerDegraded: failed to read service-ca crt bundle: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory\nIngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\nRouterCertsDegraded: neither the custom secret/v4-0-config-system-router-certs -n openshift-authentication or default secret/v4-0-config-system-custom-router-certs -n openshift-authentication could be retrieved: secret \"v4-0-config-system-router-certs\" not found\nOpenshiftAuthenticationStaticResourcesDegraded: \"oauth-openshift/audit-policy.yaml\" (string): namespaces \"openshift-authentication\" not found\nOpenshiftAuthenticationStaticResourcesDegraded: " to "OAuthServerServiceEndpointAccessibleControllerDegraded: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\nWebhookAuthenticatorControllerDegraded: failed to read service-ca crt bundle: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory\nIngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\nRouterCertsDegraded: neither the custom secret/v4-0-config-system-router-certs -n openshift-authentication or default secret/v4-0-config-system-custom-router-certs -n openshift-authentication could be retrieved: secret \"v4-0-config-system-router-certs\" not found"

openshift-authentication-operator

cluster-authentication-operator-trust-distribution-trustdistributioncontroller

authentication-operator

ConfigMapCreated

Created ConfigMap/oauth-serving-cert -n openshift-config-managed because it was missing

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Degraded message changed from "OAuthServerServiceEndpointAccessibleControllerDegraded: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\nWebhookAuthenticatorControllerDegraded: failed to read service-ca crt bundle: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory\nIngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\nRouterCertsDegraded: neither the custom secret/v4-0-config-system-router-certs -n openshift-authentication or default secret/v4-0-config-system-custom-router-certs -n openshift-authentication could be retrieved: secret \"v4-0-config-system-router-certs\" not found" to "OAuthServerServiceEndpointAccessibleControllerDegraded: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\nWebhookAuthenticatorControllerDegraded: failed to read service-ca crt bundle: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory\nIngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server"

openshift-machine-config-operator

machine-config-operator

machine-config-operator

ServiceAccountCreated

Created ServiceAccount/machine-config-daemon -n openshift-machine-config-operator because it was missing
(x5)

openshift-route-controller-manager

kubelet

route-controller-manager-74cf48bcc6-kwngm

FailedMount

MountVolume.SetUp failed for volume "client-ca" : configmap "client-ca" not found

openshift-apiserver

kubelet

apiserver-779f85678d-lrzfz

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c5ce3d1134d6500e2b8528516c1889d7bbc6259aba4981c6983395b0e9eeff65" already present on machine

openshift-oauth-apiserver

default-scheduler

apiserver-bc9b556d6-vdnq2

Scheduled

Successfully assigned openshift-oauth-apiserver/apiserver-bc9b556d6-vdnq2 to master-0

openshift-machine-config-operator

machine-config-operator

machine-config-operator

ConfigMapUpdated

Updated ConfigMap/kube-rbac-proxy -n openshift-machine-config-operator: cause by changes in data.config-file.yaml

openshift-cluster-olm-operator

olm-status-controller-statussyncer_olm

cluster-olm-operator

OperatorStatusChanged

Status for clusteroperator/olm changed: Degraded message changed from "CatalogdClusterCatalogOpenshiftRedhatMarketplaceDegraded: Internal error occurred: failed calling webhook \"inject-metadata-name.olm.operatorframework.io\": failed to call webhook: Post \"https://catalogd-service.openshift-catalogd.svc:9443/mutate-olm-operatorframework-io-v1-clustercatalog?timeout=10s\": no endpoints available for service \"catalogd-service\"" to "All is well"

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-revisioncontroller

openshift-kube-scheduler-operator

ConfigMapCreated

Created ConfigMap/serviceaccount-ca-1 -n openshift-kube-scheduler because it was missing

openshift-machine-config-operator

default-scheduler

machine-config-daemon-9t8x6

Scheduled

Successfully assigned openshift-machine-config-operator/machine-config-daemon-9t8x6 to master-0

openshift-machine-config-operator

daemonset-controller

machine-config-daemon

SuccessfulCreate

Created pod: machine-config-daemon-9t8x6

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/sa-token-signing-certs-1 -n openshift-kube-apiserver because it was missing

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-revisioncontroller

openshift-kube-scheduler-operator

ConfigMapCreated

Created ConfigMap/scheduler-kubeconfig-1 -n openshift-kube-scheduler because it was missing

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-revisioncontroller

openshift-kube-scheduler-operator

ConfigMapCreated

Created ConfigMap/kube-scheduler-cert-syncer-kubeconfig-1 -n openshift-kube-scheduler because it was missing

openshift-machine-config-operator

machine-config-operator

machine-config-operator

ValidatingAdmissionPolicyBindingCreated

Created ValidatingAdmissionPolicyBinding.admissionregistration.k8s.io/mcn-guards-binding because it was missing

openshift-kube-controller-manager-operator

kube-controller-manager-operator-revisioncontroller

kube-controller-manager-operator

SecretCreated

Created Secret/localhost-recovery-client-token-1 -n openshift-kube-controller-manager because it was missing

openshift-kube-controller-manager-operator

kube-controller-manager-operator-status-controller-statussyncer_kube-controller-manager

kube-controller-manager-operator

OperatorStatusChanged

Status for clusteroperator/kube-controller-manager changed: Degraded message changed from "NodeControllerDegraded: All master nodes are ready\nRevisionControllerDegraded: configmap \"kube-controller-manager-pod\" not found\nInstallerControllerDegraded: missing required resources: [configmaps: cluster-policy-controller-config-0,config-0,controller-manager-kubeconfig-0,kube-controller-cert-syncer-kubeconfig-0,kube-controller-manager-pod-0,recycler-config-0,service-ca-0,serviceaccount-ca-0, secrets: localhost-recovery-client-token-0,service-account-private-key-0]" to "NodeControllerDegraded: All master nodes are ready\nInstallerControllerDegraded: missing required resources: [configmaps: cluster-policy-controller-config-0,config-0,controller-manager-kubeconfig-0,kube-controller-cert-syncer-kubeconfig-0,kube-controller-manager-pod-0,recycler-config-0,service-ca-0,serviceaccount-ca-0, secrets: localhost-recovery-client-token-0,service-account-private-key-0]"

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-revisioncontroller

openshift-kube-scheduler-operator

SecretCreated

Created Secret/serving-cert-1 -n openshift-kube-scheduler because it was missing

openshift-kube-controller-manager-operator

kube-controller-manager-operator-revisioncontroller

kube-controller-manager-operator

RevisionTriggered

new revision 1 triggered by "configmap \"kube-controller-manager-pod-0\" not found"

openshift-kube-controller-manager-operator

kube-controller-manager-operator-revisioncontroller

kube-controller-manager-operator

StartingNewRevision

new revision 2 triggered by "required configmap/config has changed,required configmap/cluster-policy-controller-config has changed"

kube-system

cluster-policy-controller-namespace-security-allocation-controller

bootstrap-kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openshift-node namespace

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/kube-apiserver-audit-policies-1 -n openshift-kube-apiserver because it was missing
(x80)

openshift-kube-controller-manager-operator

kube-controller-manager-operator-installer-controller

kube-controller-manager-operator

RequiredInstallerResourcesMissing

configmaps: cluster-policy-controller-config-0,config-0,controller-manager-kubeconfig-0,kube-controller-cert-syncer-kubeconfig-0,kube-controller-manager-pod-0,recycler-config-0,service-ca-0,serviceaccount-ca-0, secrets: localhost-recovery-client-token-0,service-account-private-key-0

kube-system

cluster-policy-controller-namespace-security-allocation-controller

bootstrap-kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openshift namespace

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-revisioncontroller

openshift-kube-scheduler-operator

SecretCreated

Created Secret/localhost-recovery-client-token-1 -n openshift-kube-scheduler because it was missing

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Progressing message changed from "APIServerDeploymentProgressing: deployment/apiserver.openshift-oauth-apiserver: observed generation is 1, desired generation is 2." to "APIServerDeploymentProgressing: deployment/apiserver.openshift-oauth-apiserver: 1/1 pods have been updated to the latest generation and 0/1 pods are available"

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

kube-apiserver-operator

SecretCreated

Created Secret/etcd-client-1 -n openshift-kube-apiserver because it was missing
(x2)

openshift-authentication-operator

oauth-apiserver-oauthapiservercontroller-workloadworkloadcontroller

authentication-operator

DeploymentUpdated

Updated Deployment.apps/apiserver -n openshift-oauth-apiserver because it changed

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-revisioncontroller

openshift-kube-scheduler-operator

StartingNewRevision

new revision 2 triggered by "required configmap/serviceaccount-ca has changed"

openshift-cluster-version

kubelet

cluster-version-operator-56d8475767-9gfkg

Killing

Stopping container cluster-version-operator

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-status-controller-statussyncer_kube-scheduler

openshift-kube-scheduler-operator

OperatorStatusChanged

Status for clusteroperator/kube-scheduler changed: Degraded message changed from "RevisionControllerDegraded: configmap \"kube-scheduler-pod\" not found\nInstallerControllerDegraded: missing required resources: [configmaps: config-0,kube-scheduler-cert-syncer-kubeconfig-0,kube-scheduler-pod-0,scheduler-kubeconfig-0,serviceaccount-ca-0, secrets: localhost-recovery-client-token-0]\nNodeControllerDegraded: All master nodes are ready\nTargetConfigControllerDegraded: \"serviceaccount/localhost-recovery-client\": serviceaccounts \"localhost-recovery-client\" not found" to "InstallerControllerDegraded: missing required resources: [configmaps: config-0,kube-scheduler-cert-syncer-kubeconfig-0,kube-scheduler-pod-0,scheduler-kubeconfig-0,serviceaccount-ca-0, secrets: localhost-recovery-client-token-0]\nNodeControllerDegraded: All master nodes are ready\nTargetConfigControllerDegraded: \"serviceaccount/localhost-recovery-client\": serviceaccounts \"localhost-recovery-client\" not found"

openshift-kube-controller-manager-operator

kube-controller-manager-operator-installer-controller

kube-controller-manager-operator

NodeTargetRevisionChanged

Updating node "master-0" from revision 0 to 1 because node master-0 static pod not found

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-revisioncontroller

openshift-kube-scheduler-operator

RevisionTriggered

new revision 1 triggered by "configmap \"kube-scheduler-pod-0\" not found"

openshift-cluster-version

deployment-controller

cluster-version-operator

ScalingReplicaSet

Scaled down replica set cluster-version-operator-56d8475767 to 0 from 1

openshift-kube-controller-manager-operator

kube-controller-manager-operator-status-controller-statussyncer_kube-controller-manager

kube-controller-manager-operator

OperatorStatusChanged

Status for clusteroperator/kube-controller-manager changed: Degraded message changed from "NodeControllerDegraded: All master nodes are ready\nInstallerControllerDegraded: missing required resources: [configmaps: cluster-policy-controller-config-0,config-0,controller-manager-kubeconfig-0,kube-controller-cert-syncer-kubeconfig-0,kube-controller-manager-pod-0,recycler-config-0,service-ca-0,serviceaccount-ca-0, secrets: localhost-recovery-client-token-0,service-account-private-key-0]" to "NodeControllerDegraded: All master nodes are ready",Progressing changed from False to True ("NodeInstallerProgressing: 1 node is at revision 0; 0 nodes have achieved new revision 1"),Available message changed from "StaticPodsAvailable: 0 nodes are active; 1 node is at revision 0" to "StaticPodsAvailable: 0 nodes are active; 1 node is at revision 0; 0 nodes have achieved new revision 1"

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-status-controller-statussyncer_kube-scheduler

openshift-kube-scheduler-operator

OperatorStatusChanged

Status for clusteroperator/kube-scheduler changed: Degraded message changed from "InstallerControllerDegraded: missing required resources: [configmaps: config-0,kube-scheduler-cert-syncer-kubeconfig-0,kube-scheduler-pod-0,scheduler-kubeconfig-0,serviceaccount-ca-0, secrets: localhost-recovery-client-token-0]\nNodeControllerDegraded: All master nodes are ready\nTargetConfigControllerDegraded: \"serviceaccount/localhost-recovery-client\": serviceaccounts \"localhost-recovery-client\" not found" to "InstallerControllerDegraded: missing required resources: [configmaps: config-0,kube-scheduler-cert-syncer-kubeconfig-0,kube-scheduler-pod-0,scheduler-kubeconfig-0,serviceaccount-ca-0, secrets: localhost-recovery-client-token-0]\nNodeControllerDegraded: All master nodes are ready"

openshift-cluster-version

replicaset-controller

cluster-version-operator-56d8475767

SuccessfulDelete

Deleted pod: cluster-version-operator-56d8475767-9gfkg
(x46)

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-installer-controller

openshift-kube-scheduler-operator

RequiredInstallerResourcesMissing

configmaps: config-0,kube-scheduler-cert-syncer-kubeconfig-0,kube-scheduler-pod-0,scheduler-kubeconfig-0,serviceaccount-ca-0, secrets: localhost-recovery-client-token-0

openshift-kube-controller-manager-operator

kube-controller-manager-operator-targetconfigcontroller

kube-controller-manager-operator

ConfigMapUpdated

Updated ConfigMap/serviceaccount-ca -n openshift-kube-controller-manager: cause by changes in data.ca-bundle.crt

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-targetconfigcontroller

openshift-kube-scheduler-operator

ConfigMapUpdated

Updated ConfigMap/kube-scheduler-pod -n openshift-kube-scheduler: cause by changes in data.pod.yaml

openshift-cluster-version

openshift-cluster-version

version

LeaderElection

master-0_abd8961b-702c-4d22-aa3d-5cec899ab639 stopped leading
(x6)

openshift-controller-manager

kubelet

controller-manager-5dcd9ffc84-qh6pq

FailedMount

MountVolume.SetUp failed for volume "client-ca" : configmap "client-ca" not found

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

kube-apiserver-operator

SecretCreated

Created Secret/localhost-recovery-serving-certkey-1 -n openshift-kube-apiserver because it was missing

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-status-controller-statussyncer_kube-scheduler

openshift-kube-scheduler-operator

OperatorStatusChanged

Status for clusteroperator/kube-scheduler changed: Degraded message changed from "InstallerControllerDegraded: missing required resources: [configmaps: config-0,kube-scheduler-cert-syncer-kubeconfig-0,kube-scheduler-pod-0,scheduler-kubeconfig-0,serviceaccount-ca-0, secrets: localhost-recovery-client-token-0]\nNodeControllerDegraded: All master nodes are ready" to "NodeControllerDegraded: All master nodes are ready",Progressing changed from False to True ("NodeInstallerProgressing: 1 node is at revision 0; 0 nodes have achieved new revision 1"),Available message changed from "StaticPodsAvailable: 0 nodes are active; 1 node is at revision 0" to "StaticPodsAvailable: 0 nodes are active; 1 node is at revision 0; 0 nodes have achieved new revision 1"

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-installer-controller

openshift-kube-scheduler-operator

NodeTargetRevisionChanged

Updating node "master-0" from revision 0 to 1 because node master-0 static pod not found

openshift-kube-controller-manager-operator

kube-controller-manager-operator-revisioncontroller

kube-controller-manager-operator

ConfigMapCreated

Created ConfigMap/kube-controller-manager-pod-2 -n openshift-kube-controller-manager because it was missing

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-revisioncontroller

openshift-kube-scheduler-operator

ConfigMapCreated

Created ConfigMap/kube-scheduler-pod-2 -n openshift-kube-scheduler because it was missing

openshift-kube-controller-manager-operator

kube-controller-manager-operator-installer-controller

kube-controller-manager-operator

PodCreated

Created Pod/installer-1-master-0 -n openshift-kube-controller-manager because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

kube-apiserver-operator

SecretCreated

Created Secret/localhost-recovery-client-token-1 -n openshift-kube-apiserver because it was missing

openshift-controller-manager

replicaset-controller

controller-manager-5dcd9ffc84

SuccessfulDelete

Deleted pod: controller-manager-5dcd9ffc84-qh6pq

openshift-route-controller-manager

deployment-controller

route-controller-manager

ScalingReplicaSet

Scaled up replica set route-controller-manager-7ffc895647 to 1 from 0

openshift-route-controller-manager

deployment-controller

route-controller-manager

ScalingReplicaSet

Scaled down replica set route-controller-manager-74cf48bcc6 to 0 from 1

openshift-route-controller-manager

replicaset-controller

route-controller-manager-7ffc895647

SuccessfulCreate

Created pod: route-controller-manager-7ffc895647-6j97v

openshift-controller-manager

deployment-controller

controller-manager

ScalingReplicaSet

Scaled up replica set controller-manager-6486d766f9 to 1 from 0

openshift-controller-manager-operator

openshift-controller-manager-operator

openshift-controller-manager-operator

ConfigMapCreated

Created ConfigMap/client-ca -n openshift-controller-manager because it was missing

openshift-controller-manager

replicaset-controller

controller-manager-6486d766f9

SuccessfulCreate

Created pod: controller-manager-6486d766f9-5b77h

openshift-route-controller-manager

replicaset-controller

route-controller-manager-74cf48bcc6

SuccessfulDelete

Deleted pod: route-controller-manager-74cf48bcc6-kwngm

openshift-kube-controller-manager-operator

kube-controller-manager-operator-revisioncontroller

kube-controller-manager-operator

ConfigMapCreated

Created ConfigMap/config-2 -n openshift-kube-controller-manager because it was missing

openshift-controller-manager-operator

openshift-controller-manager-operator

openshift-controller-manager-operator

ConfigMapCreated

Created ConfigMap/client-ca -n openshift-route-controller-manager because it was missing

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-installer-controller

openshift-kube-scheduler-operator

PodCreated

Created Pod/installer-1-master-0 -n openshift-kube-scheduler because it was missing

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-revisioncontroller

openshift-kube-scheduler-operator

ConfigMapCreated

Created ConfigMap/config-2 -n openshift-kube-scheduler because it was missing

openshift-controller-manager

deployment-controller

controller-manager

ScalingReplicaSet

Scaled down replica set controller-manager-5dcd9ffc84 to 0 from 1

openshift-controller-manager-operator

openshift-controller-manager-operator-status-controller-statussyncer_openshift-controller-manager

openshift-controller-manager-operator

OperatorStatusChanged

Status for clusteroperator/openshift-controller-manager changed: Progressing message changed from "Progressing: deployment/controller-manager: observed generation is 2, desired generation is 3.\nProgressing: deployment/controller-manager: available replicas is 0, desired available replicas > 1\nProgressing: deployment/route-controller-manager: observed generation is 2, desired generation is 3.\nProgressing: deployment/route-controller-manager: available replicas is 0, desired available replicas > 1" to "Progressing: deployment/controller-manager: observed generation is 3, desired generation is 4.\nProgressing: deployment/controller-manager: available replicas is 0, desired available replicas > 1\nProgressing: deployment/route-controller-manager: observed generation is 3, desired generation is 4.\nProgressing: deployment/route-controller-manager: available replicas is 0, desired available replicas > 1"

openshift-kube-controller-manager-operator

kube-controller-manager-operator-revisioncontroller

kube-controller-manager-operator

ConfigMapCreated

Created ConfigMap/cluster-policy-controller-config-2 -n openshift-kube-controller-manager because it was missing

openshift-multus

kubelet

multus-admission-controller-5dbbb8b86f-5rrrh

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bcb08821551e9a5b9f82aa794bcea673279cefb93cb47492e19ccac5e2cf18fe" in 12.792s (12.793s including waiting). Image size: 456576198 bytes.

openshift-multus

kubelet

multus-admission-controller-5dbbb8b86f-5rrrh

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bcb08821551e9a5b9f82aa794bcea673279cefb93cb47492e19ccac5e2cf18fe" in 12.792s (12.793s including waiting). Image size: 456576198 bytes.

openshift-machine-api

kubelet

cluster-baremetal-operator-6f69995874-dv6cd

Started

Started container cluster-baremetal-operator

openshift-kube-controller-manager

multus

installer-1-master-0

AddedInterface

Add eth0 [10.128.0.47/23] from ovn-kubernetes

openshift-multus

kubelet

network-metrics-daemon-srdjm

Created

Created container: network-metrics-daemon

openshift-machine-config-operator

kubelet

machine-config-daemon-9t8x6

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-machine-config-operator

kubelet

machine-config-daemon-9t8x6

Started

Started container machine-config-daemon

openshift-machine-config-operator

kubelet

machine-config-daemon-9t8x6

Created

Created container: machine-config-daemon

openshift-machine-config-operator

kubelet

machine-config-daemon-9t8x6

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:af0fe0ca926422a6471d5bf22fc0e682c36c24fba05496a3bdfac0b7d3733015" already present on machine

openshift-multus

kubelet

network-metrics-daemon-srdjm

Started

Started container network-metrics-daemon

openshift-multus

kubelet

network-metrics-daemon-srdjm

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:759fb1d5353dbbadd443f38631d977ca3aed9787b873be05cc9660532a252739" in 12.956s (12.956s including waiting). Image size: 448828620 bytes.

openshift-multus

kubelet

network-metrics-daemon-srdjm

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-etcd

multus

installer-1-master-0

AddedInterface

Add eth0 [10.128.0.46/23] from ovn-kubernetes

openshift-apiserver

kubelet

apiserver-779f85678d-lrzfz

Created

Created container: openshift-apiserver-check-endpoints

openshift-apiserver

kubelet

apiserver-779f85678d-lrzfz

Started

Started container openshift-apiserver-check-endpoints

openshift-multus

kubelet

multus-admission-controller-5dbbb8b86f-5rrrh

Created

Created container: multus-admission-controller

openshift-multus

kubelet

multus-admission-controller-5dbbb8b86f-5rrrh

Started

Started container multus-admission-controller

openshift-multus

kubelet

multus-admission-controller-5dbbb8b86f-5rrrh

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-multus

kubelet

multus-admission-controller-5dbbb8b86f-5rrrh

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-machine-api

kubelet

cluster-baremetal-operator-6f69995874-dv6cd

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-machine-api

kubelet

cluster-baremetal-operator-6f69995874-dv6cd

Started

Started container cluster-baremetal-operator

openshift-machine-api

kubelet

cluster-baremetal-operator-6f69995874-dv6cd

Created

Created container: cluster-baremetal-operator

openshift-machine-api

kubelet

cluster-baremetal-operator-6f69995874-dv6cd

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f933312f49083e8746fc41ab5e46a9a757b448374f14971e256ebcb36f11dd97" in 12.968s (12.968s including waiting). Image size: 470826739 bytes.

openshift-multus

kubelet

multus-admission-controller-5dbbb8b86f-5rrrh

Started

Started container multus-admission-controller

openshift-multus

kubelet

multus-admission-controller-5dbbb8b86f-5rrrh

Created

Created container: multus-admission-controller
(x2)

openshift-route-controller-manager

default-scheduler

route-controller-manager-7ffc895647-6j97v

FailedScheduling

0/1 nodes are available: 1 node(s) didn't match pod anti-affinity rules. preemption: 0/1 nodes are available: 1 node(s) didn't match pod anti-affinity rules.

openshift-multus

kubelet

network-metrics-daemon-srdjm

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:759fb1d5353dbbadd443f38631d977ca3aed9787b873be05cc9660532a252739" in 12.956s (12.956s including waiting). Image size: 448828620 bytes.

openshift-multus

kubelet

network-metrics-daemon-srdjm

Created

Created container: network-metrics-daemon

openshift-multus

kubelet

network-metrics-daemon-srdjm

Started

Started container network-metrics-daemon

openshift-kube-apiserver-operator

kube-apiserver-operator-status-controller-statussyncer_kube-apiserver

kube-apiserver-operator

OperatorStatusChanged

Status for clusteroperator/kube-apiserver changed: Degraded message changed from "NodeControllerDegraded: All master nodes are ready\nRevisionControllerDegraded: configmap \"kube-apiserver-pod\" not found\nInstallerControllerDegraded: missing required resources: [configmaps: bound-sa-token-signing-certs-0,config-0,etcd-serving-ca-0,kube-apiserver-audit-policies-0,kube-apiserver-cert-syncer-kubeconfig-0,kube-apiserver-pod-0,kubelet-serving-ca-0,sa-token-signing-certs-0, secrets: etcd-client-0,localhost-recovery-client-token-0,localhost-recovery-serving-certkey-0]" to "NodeControllerDegraded: All master nodes are ready\nInstallerControllerDegraded: missing required resources: [configmaps: bound-sa-token-signing-certs-0,config-0,etcd-serving-ca-0,kube-apiserver-audit-policies-0,kube-apiserver-cert-syncer-kubeconfig-0,kube-apiserver-pod-0,kubelet-serving-ca-0,sa-token-signing-certs-0, secrets: etcd-client-0,localhost-recovery-client-token-0,localhost-recovery-serving-certkey-0]"

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

kube-apiserver-operator

RevisionTriggered

new revision 1 triggered by "configmap \"kube-apiserver-pod-0\" not found"

openshift-kube-controller-manager-operator

kube-controller-manager-operator-revisioncontroller

kube-controller-manager-operator

ConfigMapCreated

Created ConfigMap/controller-manager-kubeconfig-2 -n openshift-kube-controller-manager because it was missing

openshift-kube-scheduler

multus

installer-1-master-0

AddedInterface

Add eth0 [10.128.0.48/23] from ovn-kubernetes

openshift-kube-scheduler

kubelet

installer-1-master-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ef199844317b7b012879ed8d29f9b6bc37fad8a6fdb336103cbd5cabc74c4302" already present on machine

openshift-multus

kubelet

network-metrics-daemon-srdjm

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-cluster-version

deployment-controller

cluster-version-operator

ScalingReplicaSet

Scaled up replica set cluster-version-operator-7d58488df to 1

openshift-cluster-version

replicaset-controller

cluster-version-operator-7d58488df

SuccessfulCreate

Created pod: cluster-version-operator-7d58488df-qmm8h

openshift-cluster-version

default-scheduler

cluster-version-operator-7d58488df-qmm8h

Scheduled

Successfully assigned openshift-cluster-version/cluster-version-operator-7d58488df-qmm8h to master-0
(x60)

openshift-kube-apiserver-operator

kube-apiserver-operator-installer-controller

kube-apiserver-operator

RequiredInstallerResourcesMissing

configmaps: bound-sa-token-signing-certs-0,config-0,etcd-serving-ca-0,kube-apiserver-audit-policies-0,kube-apiserver-cert-syncer-kubeconfig-0,kube-apiserver-pod-0,kubelet-serving-ca-0,sa-token-signing-certs-0, secrets: etcd-client-0,localhost-recovery-client-token-0,localhost-recovery-serving-certkey-0

openshift-monitoring

kubelet

cluster-monitoring-operator-58845fbb57-nljsr

Started

Started container cluster-monitoring-operator

openshift-monitoring

kubelet

cluster-monitoring-operator-58845fbb57-nljsr

Created

Created container: cluster-monitoring-operator
(x2)

openshift-controller-manager

default-scheduler

controller-manager-6486d766f9-5b77h

FailedScheduling

0/1 nodes are available: 1 node(s) didn't match pod anti-affinity rules. preemption: 0/1 nodes are available: 1 node(s) didn't match pod anti-affinity rules.

openshift-monitoring

kubelet

cluster-monitoring-operator-58845fbb57-nljsr

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a746a87b784ea1caa278fd0e012554f9df520b6fff665ea0bc4c83f487fed113" in 12.96s (12.96s including waiting). Image size: 484450894 bytes.

openshift-marketplace

kubelet

marketplace-operator-89ccd998f-mvn4t

Started

Started container marketplace-operator

openshift-marketplace

kubelet

marketplace-operator-89ccd998f-mvn4t

Created

Created container: marketplace-operator

openshift-marketplace

kubelet

marketplace-operator-89ccd998f-mvn4t

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:632e80bba5077068ecca05fddb95aedebad4493af6f36152c01c6ae490975b62" in 12.957s (12.957s including waiting). Image size: 458126937 bytes.

openshift-machine-api

kubelet

cluster-baremetal-operator-6f69995874-dv6cd

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f933312f49083e8746fc41ab5e46a9a757b448374f14971e256ebcb36f11dd97" in 12.968s (12.968s including waiting). Image size: 470826739 bytes.

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-revisioncontroller

openshift-kube-scheduler-operator

ConfigMapCreated

Created ConfigMap/serviceaccount-ca-2 -n openshift-kube-scheduler because it was missing

openshift-machine-api

kubelet

cluster-baremetal-operator-6f69995874-dv6cd

Created

Created container: cluster-baremetal-operator

openshift-machine-api

kubelet

cluster-baremetal-operator-6f69995874-dv6cd

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-oauth-apiserver

multus

apiserver-bc9b556d6-vdnq2

AddedInterface

Add eth0 [10.128.0.45/23] from ovn-kubernetes

openshift-oauth-apiserver

kubelet

apiserver-bc9b556d6-vdnq2

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7d4a034950346bcd4e36e9e2f1343e0cf7a10cf544963f33d09c7eb2a1bfc634" already present on machine

openshift-monitoring

kubelet

cluster-monitoring-operator-58845fbb57-nljsr

Started

Started container cluster-monitoring-operator

openshift-monitoring

kubelet

cluster-monitoring-operator-58845fbb57-nljsr

Created

Created container: cluster-monitoring-operator

openshift-kube-controller-manager

kubelet

installer-1-master-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c032f87ae61d6f757ff3ce52620a70a43516591987731f25da77aba152f17458" already present on machine

openshift-monitoring

kubelet

cluster-monitoring-operator-58845fbb57-nljsr

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a746a87b784ea1caa278fd0e012554f9df520b6fff665ea0bc4c83f487fed113" in 12.96s (12.96s including waiting). Image size: 484450894 bytes.

openshift-kube-controller-manager-operator

kube-controller-manager-operator-status-controller-statussyncer_kube-controller-manager

kube-controller-manager-operator

OperatorStatusChanged

Status for clusteroperator/kube-controller-manager changed: Degraded message changed from "NodeControllerDegraded: All master nodes are ready" to "NodeControllerDegraded: All master nodes are ready\nGarbageCollectorDegraded: error fetching rules: Get \"https://thanos-querier.openshift-monitoring.svc:9091/api/v1/rules\": dial tcp: lookup thanos-querier.openshift-monitoring.svc on 172.30.0.10:53: no such host"

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ServiceCreated

Created Service/prometheus-operator-admission-webhook -n openshift-monitoring because it was missing

openshift-kube-controller-manager

kubelet

installer-1-master-0

Created

Created container: installer

openshift-kube-controller-manager

kubelet

installer-1-master-0

Started

Started container installer

openshift-monitoring

default-scheduler

prometheus-operator-admission-webhook-69c6b55594-tkc2j

FailedScheduling

0/1 nodes are available: 1 node(s) had untolerated taint {node-role.kubernetes.io/master: }. preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.

openshift-machine-config-operator

kubelet

machine-config-daemon-9t8x6

Created

Created container: kube-rbac-proxy

openshift-oauth-apiserver

kubelet

apiserver-bc9b556d6-vdnq2

Started

Started container fix-audit-permissions

openshift-oauth-apiserver

kubelet

apiserver-bc9b556d6-vdnq2

Created

Created container: fix-audit-permissions

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

FeatureGatesInitialized

FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{"AWSEFSDriverVolumeMetrics", "AdminNetworkPolicy", "AlibabaPlatform", "AzureWorkloadIdentity", "BareMetalLoadBalancer", "BuildCSIVolumes", "ChunkSizeMiB", "CloudDualStackNodeIPs", "DisableKubeletCloudCredentialProviders", "GCPLabelsTags", "HardwareSpeed", "IngressControllerLBSubnetsAWS", "KMSv1", "ManagedBootImages", "ManagedBootImagesAWS", "MultiArchInstallAWS", "MultiArchInstallGCP", "NetworkDiagnosticsConfig", "NetworkLiveMigration", "NetworkSegmentation", "NewOLM", "NodeDisruptionPolicy", "OnClusterBuild", "PersistentIPsForVirtualization", "PrivateHostedZoneAWS", "SetEIPForNLBIngressController", "StreamingCollectionEncodingToJSON", "StreamingCollectionEncodingToProtobuf", "VSphereControlPlaneMachineSet", "VSphereDriverConfiguration", "VSphereMultiVCenters", "VSphereStaticIPs", "ValidatingAdmissionPolicy"}, Disabled:[]v1.FeatureGateName{"AWSClusterHostedDNS", "AdditionalRoutingCapabilities", "AutomatedEtcdBackup", "BootcNodeManagement", "CSIDriverSharedResource", "ClusterAPIInstall", "ClusterAPIInstallIBMCloud", "ClusterMonitoringConfig", "ConsolePluginContentSecurityPolicy", "DNSNameResolver", "DynamicResourceAllocation", "EtcdBackendQuota", "EventedPLEG", "Example", "ExternalOIDC", "ExternalOIDCWithUIDAndExtraClaimMappings", "GCPClusterHostedDNS", "GatewayAPI", "ImageStreamImportMode", "IngressControllerDynamicConfigurationManager", "InsightsConfig", "InsightsConfigAPI", "InsightsOnDemandDataGather", "InsightsRuntimeExtractor", "MachineAPIMigration", "MachineAPIOperatorDisableMachineHealthCheckController", "MachineAPIProviderOpenStack", "MachineConfigNodes", "MaxUnavailableStatefulSet", "MetricsCollectionProfiles", "MinimumKubeletVersion", "MixedCPUsAllocation", "MultiArchInstallAzure", "NodeSwap", "NutanixMultiSubnets", "OVNObservability", "OpenShiftPodSecurityAdmission", "PinnedImages", "PlatformOperators", "ProcMountType", "RouteAdvertisements", "RouteExternalCertificate", "ServiceAccountTokenNodeBinding", "SignatureStores", "SigstoreImageVerification", "TranslateStreamCloseWebsocketRequests", "UpgradeStatus", "UserNamespacesPodSecurityStandards", "UserNamespacesSupport", "VSphereMultiNetworks", "VolumeAttributesClass", "VolumeGroupSnapshot"}}

openshift-monitoring

cluster-monitoring-operator-openshiftmonitoringtelemeterclientcertrequester

cluster-monitoring-operator

NoValidCertificateFound

No valid client certificate for OpenShiftMonitoringTelemeterClientCertRequester is found: unable to parse certificate: data does not contain any valid RSA or ECDSA certificates

openshift-machine-api

cluster-baremetal-operator-6f69995874-dv6cd_d3f73e8f-2c31-494a-9a70-fc6a896e6c7d

cluster-baremetal-operator

LeaderElection

cluster-baremetal-operator-6f69995874-dv6cd_d3f73e8f-2c31-494a-9a70-fc6a896e6c7d became leader

openshift-machine-api

kubelet

cluster-baremetal-operator-6f69995874-dv6cd

Started

Started container baremetal-kube-rbac-proxy

openshift-machine-api

kubelet

cluster-baremetal-operator-6f69995874-dv6cd

Created

Created container: baremetal-kube-rbac-proxy

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ClusterRoleCreated

Created ClusterRole.rbac.authorization.k8s.io/prometheus-operator because it was missing

openshift-machine-config-operator

machine-config-operator

machine-config-operator

ClusterRoleCreated

Created ClusterRole.rbac.authorization.k8s.io/machine-config-controller because it was missing

openshift-operator-lifecycle-manager

kubelet

catalog-operator-68f85b4d6c-fzm28

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6bba3f73c0066e42b24839e0d29f5dce2f36436f0a11f9f5e1029bccc5ed6578" in 14.137s (14.137s including waiting). Image size: 862657321 bytes.

openshift-monitoring

replicaset-controller

prometheus-operator-admission-webhook-69c6b55594

SuccessfulCreate

Created pod: prometheus-operator-admission-webhook-69c6b55594-tkc2j

openshift-operator-lifecycle-manager

kubelet

catalog-operator-68f85b4d6c-fzm28

Created

Created container: catalog-operator

openshift-monitoring

cluster-monitoring-operator-openshiftmonitoringtelemeterclientcertrequester

cluster-monitoring-operator

CSRCreated

A csr "system:openshift:openshift-monitoring-5vbxd" is created for OpenShiftMonitoringTelemeterClientCertRequester

openshift-monitoring

cluster-monitoring-operator-openshiftmonitoringclientcertrequester

cluster-monitoring-operator

NoValidCertificateFound

No valid client certificate for OpenShiftMonitoringClientCertRequester is found: unable to parse certificate: data does not contain any valid RSA or ECDSA certificates

openshift-marketplace

kubelet

marketplace-operator-89ccd998f-mvn4t

ProbeError

Readiness probe error: Get "http://10.128.0.20:8080/healthz": dial tcp 10.128.0.20:8080: connect: connection refused body:

openshift-marketplace

kubelet

marketplace-operator-89ccd998f-mvn4t

Unhealthy

Readiness probe failed: Get "http://10.128.0.20:8080/healthz": dial tcp 10.128.0.20:8080: connect: connection refused

openshift-operator-lifecycle-manager

kubelet

olm-operator-5c9796789-tjm9l

Started

Started container olm-operator

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-revisioncontroller

openshift-kube-scheduler-operator

ConfigMapCreated

Created ConfigMap/scheduler-kubeconfig-2 -n openshift-kube-scheduler because it was missing

kube-system

cluster-policy-controller-webhook-authenticator-cert-approver-csr-approver-controller-webhookauthenticatorcertapprover_csr-approver-controller

bootstrap-kube-controller-manager-master-0

CSRApproval

The CSR "system:openshift:openshift-monitoring-mk9bw" has been approved

kube-system

cluster-policy-controller-webhook-authenticator-cert-approver-csr-approver-controller-webhookauthenticatorcertapprover_csr-approver-controller

bootstrap-kube-controller-manager-master-0

CSRApproval

The CSR "system:openshift:openshift-monitoring-5vbxd" has been approved

openshift-multus

kubelet

network-metrics-daemon-srdjm

Created

Created container: kube-rbac-proxy

openshift-multus

kubelet

network-metrics-daemon-srdjm

Started

Started container kube-rbac-proxy

openshift-machine-config-operator

machine-config-operator

machine-config-operator

ClusterRoleCreated

Created ClusterRole.rbac.authorization.k8s.io/machine-config-controller-events because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

FeatureGatesInitialized

FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{"AWSEFSDriverVolumeMetrics", "AdminNetworkPolicy", "AlibabaPlatform", "AzureWorkloadIdentity", "BareMetalLoadBalancer", "BuildCSIVolumes", "ChunkSizeMiB", "CloudDualStackNodeIPs", "DisableKubeletCloudCredentialProviders", "GCPLabelsTags", "HardwareSpeed", "IngressControllerLBSubnetsAWS", "KMSv1", "ManagedBootImages", "ManagedBootImagesAWS", "MultiArchInstallAWS", "MultiArchInstallGCP", "NetworkDiagnosticsConfig", "NetworkLiveMigration", "NetworkSegmentation", "NewOLM", "NodeDisruptionPolicy", "OnClusterBuild", "PersistentIPsForVirtualization", "PrivateHostedZoneAWS", "SetEIPForNLBIngressController", "StreamingCollectionEncodingToJSON", "StreamingCollectionEncodingToProtobuf", "VSphereControlPlaneMachineSet", "VSphereDriverConfiguration", "VSphereMultiVCenters", "VSphereStaticIPs", "ValidatingAdmissionPolicy"}, Disabled:[]v1.FeatureGateName{"AWSClusterHostedDNS", "AdditionalRoutingCapabilities", "AutomatedEtcdBackup", "BootcNodeManagement", "CSIDriverSharedResource", "ClusterAPIInstall", "ClusterAPIInstallIBMCloud", "ClusterMonitoringConfig", "ConsolePluginContentSecurityPolicy", "DNSNameResolver", "DynamicResourceAllocation", "EtcdBackendQuota", "EventedPLEG", "Example", "ExternalOIDC", "ExternalOIDCWithUIDAndExtraClaimMappings", "GCPClusterHostedDNS", "GatewayAPI", "ImageStreamImportMode", "IngressControllerDynamicConfigurationManager", "InsightsConfig", "InsightsConfigAPI", "InsightsOnDemandDataGather", "InsightsRuntimeExtractor", "MachineAPIMigration", "MachineAPIOperatorDisableMachineHealthCheckController", "MachineAPIProviderOpenStack", "MachineConfigNodes", "MaxUnavailableStatefulSet", "MetricsCollectionProfiles", "MinimumKubeletVersion", "MixedCPUsAllocation", "MultiArchInstallAzure", "NodeSwap", "NutanixMultiSubnets", "OVNObservability", "OpenShiftPodSecurityAdmission", "PinnedImages", "PlatformOperators", "ProcMountType", "RouteAdvertisements", "RouteExternalCertificate", "ServiceAccountTokenNodeBinding", "SignatureStores", "SigstoreImageVerification", "TranslateStreamCloseWebsocketRequests", "UpgradeStatus", "UserNamespacesPodSecurityStandards", "UserNamespacesSupport", "VSphereMultiNetworks", "VolumeAttributesClass", "VolumeGroupSnapshot"}}

openshift-monitoring

cluster-monitoring-operator-openshiftmonitoringtelemeterclientcertrequester

cluster-monitoring-operator

NoValidCertificateFound

No valid client certificate for OpenShiftMonitoringTelemeterClientCertRequester is found: unable to parse certificate: data does not contain any valid RSA or ECDSA certificates

openshift-monitoring

cluster-monitoring-operator-openshiftmonitoringtelemeterclientcertrequester

cluster-monitoring-operator

CSRCreated

A csr "system:openshift:openshift-monitoring-5vbxd" is created for OpenShiftMonitoringTelemeterClientCertRequester

openshift-monitoring

cluster-monitoring-operator-openshiftmonitoringclientcertrequester

cluster-monitoring-operator

NoValidCertificateFound

No valid client certificate for OpenShiftMonitoringClientCertRequester is found: unable to parse certificate: data does not contain any valid RSA or ECDSA certificates

openshift-monitoring

cluster-monitoring-operator-openshiftmonitoringclientcertrequester

cluster-monitoring-operator

CSRCreated

A csr "system:openshift:openshift-monitoring-mk9bw" is created for OpenShiftMonitoringClientCertRequester

openshift-monitoring

cluster-monitoring-operator-openshiftmonitoringtelemeterclientcertrequester

cluster-monitoring-operator

ClientCertificateCreated

A new client certificate for OpenShiftMonitoringTelemeterClientCertRequester is available

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

SecretCreated

Created Secret/alert-relabel-configs -n openshift-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator-openshiftmonitoringclientcertrequester

cluster-monitoring-operator

ClientCertificateCreated

A new client certificate for OpenShiftMonitoringClientCertRequester is available

openshift-monitoring

cluster-monitoring-operator-openshiftmonitoringclientcertrequester

cluster-monitoring-operator

CSRCreated

A csr "system:openshift:openshift-monitoring-mk9bw" is created for OpenShiftMonitoringClientCertRequester

openshift-monitoring

deployment-controller

prometheus-operator-admission-webhook

ScalingReplicaSet

Scaled up replica set prometheus-operator-admission-webhook-69c6b55594 to 1

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ServiceAccountCreated

Created ServiceAccount/prometheus-operator -n openshift-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ConfigMapCreated

Created ConfigMap/metrics-client-ca -n openshift-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ClusterRoleCreated

Created ClusterRole.rbac.authorization.k8s.io/prometheus-operator because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ClusterRoleBindingCreated

Created ClusterRoleBinding.rbac.authorization.k8s.io/prometheus-operator because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ServiceAccountCreated

Created ServiceAccount/prometheus-operator-admission-webhook -n openshift-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ServiceCreated

Created Service/prometheus-operator-admission-webhook -n openshift-monitoring because it was missing

openshift-operator-lifecycle-manager

kubelet

package-server-manager-7b95f86987-2pg77

Started

Started container package-server-manager

openshift-etcd

kubelet

installer-1-master-0

Started

Started container installer

openshift-etcd

kubelet

installer-1-master-0

Created

Created container: installer

openshift-operator-lifecycle-manager

kubelet

package-server-manager-7b95f86987-2pg77

Created

Created container: package-server-manager

openshift-machine-config-operator

machine-config-operator

master-0

FeatureGatesInitialized

FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{"AWSEFSDriverVolumeMetrics", "AdminNetworkPolicy", "AlibabaPlatform", "AzureWorkloadIdentity", "BareMetalLoadBalancer", "BuildCSIVolumes", "ChunkSizeMiB", "CloudDualStackNodeIPs", "DisableKubeletCloudCredentialProviders", "GCPLabelsTags", "HardwareSpeed", "IngressControllerLBSubnetsAWS", "KMSv1", "ManagedBootImages", "ManagedBootImagesAWS", "MultiArchInstallAWS", "MultiArchInstallGCP", "NetworkDiagnosticsConfig", "NetworkLiveMigration", "NetworkSegmentation", "NewOLM", "NodeDisruptionPolicy", "OnClusterBuild", "PersistentIPsForVirtualization", "PrivateHostedZoneAWS", "SetEIPForNLBIngressController", "StreamingCollectionEncodingToJSON", "StreamingCollectionEncodingToProtobuf", "VSphereControlPlaneMachineSet", "VSphereDriverConfiguration", "VSphereMultiVCenters", "VSphereStaticIPs", "ValidatingAdmissionPolicy"}, Disabled:[]v1.FeatureGateName{"AWSClusterHostedDNS", "AdditionalRoutingCapabilities", "AutomatedEtcdBackup", "BootcNodeManagement", "CSIDriverSharedResource", "ClusterAPIInstall", "ClusterAPIInstallIBMCloud", "ClusterMonitoringConfig", "ConsolePluginContentSecurityPolicy", "DNSNameResolver", "DynamicResourceAllocation", "EtcdBackendQuota", "EventedPLEG", "Example", "ExternalOIDC", "ExternalOIDCWithUIDAndExtraClaimMappings", "GCPClusterHostedDNS", "GatewayAPI", "ImageStreamImportMode", "IngressControllerDynamicConfigurationManager", "InsightsConfig", "InsightsConfigAPI", "InsightsOnDemandDataGather", "InsightsRuntimeExtractor", "MachineAPIMigration", "MachineAPIOperatorDisableMachineHealthCheckController", "MachineAPIProviderOpenStack", "MachineConfigNodes", "MaxUnavailableStatefulSet", "MetricsCollectionProfiles", "MinimumKubeletVersion", "MixedCPUsAllocation", "MultiArchInstallAzure", "NodeSwap", "NutanixMultiSubnets", "OVNObservability", "OpenShiftPodSecurityAdmission", "PinnedImages", "PlatformOperators", "ProcMountType", "RouteAdvertisements", "RouteExternalCertificate", "ServiceAccountTokenNodeBinding", "SignatureStores", "SigstoreImageVerification", "TranslateStreamCloseWebsocketRequests", "UpgradeStatus", "UserNamespacesPodSecurityStandards", "UserNamespacesSupport", "VSphereMultiNetworks", "VolumeAttributesClass", "VolumeGroupSnapshot"}}

openshift-multus

kubelet

network-metrics-daemon-srdjm

Started

Started container kube-rbac-proxy

openshift-multus

kubelet

network-metrics-daemon-srdjm

Created

Created container: kube-rbac-proxy

openshift-etcd

kubelet

installer-1-master-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d5a971d5889f167cfe61a64c366424b87c17a6dc141ffcc43406cdcbb50cae2a" already present on machine

openshift-operator-lifecycle-manager

kubelet

package-server-manager-7b95f86987-2pg77

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6bba3f73c0066e42b24839e0d29f5dce2f36436f0a11f9f5e1029bccc5ed6578" in 14.887s (14.887s including waiting). Image size: 862657321 bytes.

openshift-kube-scheduler

kubelet

installer-1-master-0

Started

Started container installer

openshift-kube-scheduler

kubelet

installer-1-master-0

Created

Created container: installer

openshift-monitoring

cluster-monitoring-operator-openshiftmonitoringtelemeterclientcertrequester

cluster-monitoring-operator

ClientCertificateCreated

A new client certificate for OpenShiftMonitoringTelemeterClientCertRequester is available

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

SecretCreated

Created Secret/alert-relabel-configs -n openshift-monitoring because it was missing

openshift-operator-lifecycle-manager

kubelet

catalog-operator-68f85b4d6c-fzm28

Started

Started container catalog-operator

openshift-kube-controller-manager-operator

kube-controller-manager-operator-status-controller-statussyncer_kube-controller-manager

kube-controller-manager-operator

OperatorStatusChanged

Status for clusteroperator/kube-controller-manager changed: Degraded message changed from "NodeControllerDegraded: All master nodes are ready\nGarbageCollectorDegraded: error fetching rules: Get \"https://thanos-querier.openshift-monitoring.svc:9091/api/v1/rules\": dial tcp: lookup thanos-querier.openshift-monitoring.svc on 172.30.0.10:53: no such host" to "NodeControllerDegraded: All master nodes are ready"

openshift-machine-config-operator

kubelet

machine-config-daemon-9t8x6

Started

Started container kube-rbac-proxy

openshift-monitoring

cluster-monitoring-operator-openshiftmonitoringclientcertrequester

cluster-monitoring-operator

ClientCertificateCreated

A new client certificate for OpenShiftMonitoringClientCertRequester is available

openshift-machine-config-operator

machine-config-operator

machine-config-operator

RoleCreated

Created Role.rbac.authorization.k8s.io/mcc-prometheus-k8s -n openshift-machine-config-operator because it was missing

openshift-machine-api

cluster-baremetal-operator-6f69995874-dv6cd_d3f73e8f-2c31-494a-9a70-fc6a896e6c7d

cluster-baremetal-operator

LeaderElection

cluster-baremetal-operator-6f69995874-dv6cd_d3f73e8f-2c31-494a-9a70-fc6a896e6c7d became leader

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ConfigMapCreated

Created ConfigMap/metrics-client-ca -n openshift-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ServiceAccountCreated

Created ServiceAccount/prometheus-operator-admission-webhook -n openshift-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ClusterRoleBindingCreated

Created ClusterRoleBinding.rbac.authorization.k8s.io/prometheus-operator because it was missing

openshift-machine-config-operator

machine-config-operator

machine-config-operator

RoleBindingCreated

Created RoleBinding.rbac.authorization.k8s.io/machine-config-controller-events -n default because it was missing

openshift-operator-lifecycle-manager

kubelet

olm-operator-5c9796789-tjm9l

Created

Created container: olm-operator

openshift-machine-config-operator

machine-config-operator

machine-config-operator

RoleBindingCreated

Created RoleBinding.rbac.authorization.k8s.io/mcc-prometheus-k8s -n openshift-machine-config-operator because it was missing

openshift-machine-config-operator

machine-config-operator

machine-config-operator

RoleBindingCreated

Created RoleBinding.rbac.authorization.k8s.io/machine-config-controller-events -n openshift-machine-config-operator because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ServiceAccountCreated

Created ServiceAccount/prometheus-operator -n openshift-monitoring because it was missing

openshift-machine-api

kubelet

cluster-baremetal-operator-6f69995874-dv6cd

Created

Created container: baremetal-kube-rbac-proxy

openshift-multus

kubelet

multus-admission-controller-5dbbb8b86f-5rrrh

Created

Created container: kube-rbac-proxy

openshift-multus

kubelet

multus-admission-controller-5dbbb8b86f-5rrrh

Started

Started container kube-rbac-proxy

openshift-monitoring

default-scheduler

prometheus-operator-admission-webhook-69c6b55594-tkc2j

FailedScheduling

0/1 nodes are available: 1 node(s) had untolerated taint {node-role.kubernetes.io/master: }. preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.

openshift-monitoring

replicaset-controller

prometheus-operator-admission-webhook-69c6b55594

SuccessfulCreate

Created pod: prometheus-operator-admission-webhook-69c6b55594-tkc2j

openshift-monitoring

deployment-controller

prometheus-operator-admission-webhook

ScalingReplicaSet

Scaled up replica set prometheus-operator-admission-webhook-69c6b55594 to 1

openshift-multus

kubelet

multus-admission-controller-5dbbb8b86f-5rrrh

Started

Started container kube-rbac-proxy

openshift-multus

kubelet

multus-admission-controller-5dbbb8b86f-5rrrh

Created

Created container: kube-rbac-proxy

openshift-machine-api

kubelet

cluster-baremetal-operator-6f69995874-dv6cd

Started

Started container baremetal-kube-rbac-proxy

openshift-operator-lifecycle-manager

kubelet

olm-operator-5c9796789-tjm9l

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6bba3f73c0066e42b24839e0d29f5dce2f36436f0a11f9f5e1029bccc5ed6578" in 14.102s (14.102s including waiting). Image size: 862657321 bytes.

openshift-cluster-version

kubelet

cluster-version-operator-7d58488df-qmm8h

Pulled

Container image "quay.io/openshift-release-dev/ocp-release@sha256:59727c4b3fef19e5149675cf3350735bbfe2c6588a57654b2e4552dd719f58b1" already present on machine

openshift-machine-config-operator

machine-config-operator

machine-config-operator

ClusterRoleBindingCreated

Created ClusterRoleBinding.rbac.authorization.k8s.io/machine-config-controller because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-status-controller-statussyncer_kube-apiserver

kube-apiserver-operator

OperatorStatusChanged

Status for clusteroperator/kube-apiserver changed: Degraded message changed from "NodeControllerDegraded: All master nodes are ready\nInstallerControllerDegraded: missing required resources: [configmaps: bound-sa-token-signing-certs-0,config-0,etcd-serving-ca-0,kube-apiserver-audit-policies-0,kube-apiserver-cert-syncer-kubeconfig-0,kube-apiserver-pod-0,kubelet-serving-ca-0,sa-token-signing-certs-0, secrets: etcd-client-0,localhost-recovery-client-token-0,localhost-recovery-serving-certkey-0]" to "NodeControllerDegraded: All master nodes are ready",Progressing changed from False to True ("NodeInstallerProgressing: 1 node is at revision 0; 0 nodes have achieved new revision 1"),Available message changed from "StaticPodsAvailable: 0 nodes are active; 1 node is at revision 0" to "StaticPodsAvailable: 0 nodes are active; 1 node is at revision 0; 0 nodes have achieved new revision 1"

openshift-marketplace

default-scheduler

redhat-marketplace-srjqw

Scheduled

Successfully assigned openshift-marketplace/redhat-marketplace-srjqw to master-0

openshift-cluster-version

kubelet

cluster-version-operator-7d58488df-qmm8h

FailedMount

MountVolume.SetUp failed for volume "kube-api-access" : failed to sync configmap cache: timed out waiting for the condition

openshift-machine-config-operator

machine-config-operator

machine-config-operator

ServiceAccountCreated

Created ServiceAccount/machine-config-controller -n openshift-machine-config-operator because it was missing

openshift-kube-controller-manager-operator

kube-controller-manager-operator-revisioncontroller

kube-controller-manager-operator

ConfigMapCreated

Created ConfigMap/kube-controller-cert-syncer-kubeconfig-2 -n openshift-kube-controller-manager because it was missing

openshift-machine-config-operator

machine-config-operator

machine-config-operator

ValidatingAdmissionPolicyBindingCreated

Created ValidatingAdmissionPolicyBinding.admissionregistration.k8s.io/machine-configuration-guards-binding because it was missing

openshift-machine-config-operator

machine-config-operator

machine-config-operator

RoleBindingCreated

Created RoleBinding.rbac.authorization.k8s.io/machine-os-puller-binding -n openshift-machine-config-operator because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-installer-controller

kube-apiserver-operator

NodeTargetRevisionChanged

Updating node "master-0" from revision 0 to 1 because node master-0 static pod not found

openshift-operator-lifecycle-manager

package-server-manager-7b95f86987-2pg77_48be34a1-8287-4af5-a0e6-3e5bb9d5e838

packageserver-controller-lock

LeaderElection

package-server-manager-7b95f86987-2pg77_48be34a1-8287-4af5-a0e6-3e5bb9d5e838 became leader

openshift-machine-config-operator

machine-config-operator

machine-config-operator

ValidatingAdmissionPolicyCreated

Created ValidatingAdmissionPolicy.admissionregistration.k8s.io/custom-machine-config-pool-selector because it was missing

openshift-machine-config-operator

machine-config-operator

machine-config-operator

ValidatingAdmissionPolicyCreated

Created ValidatingAdmissionPolicy.admissionregistration.k8s.io/managed-bootimages-platform-check because it was missing

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-revisioncontroller

openshift-kube-scheduler-operator

ConfigMapCreated

Created ConfigMap/kube-scheduler-cert-syncer-kubeconfig-2 -n openshift-kube-scheduler because it was missing

openshift-machine-config-operator

machine-config-operator

machine-config-operator

ValidatingAdmissionPolicyCreated

Created ValidatingAdmissionPolicy.admissionregistration.k8s.io/machine-configuration-guards because it was missing

openshift-machine-config-operator

machine-config-operator

machine-config-operator

ServiceAccountCreated

Created ServiceAccount/machine-os-puller -n openshift-machine-config-operator because it was missing

openshift-oauth-apiserver

kubelet

apiserver-bc9b556d6-vdnq2

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7d4a034950346bcd4e36e9e2f1343e0cf7a10cf544963f33d09c7eb2a1bfc634" already present on machine

openshift-marketplace

kubelet

redhat-marketplace-srjqw

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6bba3f73c0066e42b24839e0d29f5dce2f36436f0a11f9f5e1029bccc5ed6578" already present on machine

openshift-machine-config-operator

default-scheduler

machine-config-controller-b4f87c5b9-pj7rj

Scheduled

Successfully assigned openshift-machine-config-operator/machine-config-controller-b4f87c5b9-pj7rj to master-0

openshift-marketplace

multus

redhat-marketplace-srjqw

AddedInterface

Add eth0 [10.128.0.49/23] from ovn-kubernetes

openshift-cluster-version

kubelet

cluster-version-operator-7d58488df-qmm8h

Created

Created container: cluster-version-operator

openshift-cluster-version

kubelet

cluster-version-operator-7d58488df-qmm8h

Started

Started container cluster-version-operator

openshift-cluster-version

openshift-cluster-version

version

LeaderElection

master-0_d00985cf-7208-44bf-bd7d-3c5284c2a8bf became leader

openshift-marketplace

default-scheduler

redhat-operators-mgcb9

Scheduled

Successfully assigned openshift-marketplace/redhat-operators-mgcb9 to master-0
(x2)

openshift-operator-lifecycle-manager

operator-lifecycle-manager

packageserver

RequirementsUnknown

requirements not yet checked

openshift-controller-manager

default-scheduler

controller-manager-6486d766f9-5b77h

Scheduled

Successfully assigned openshift-controller-manager/controller-manager-6486d766f9-5b77h to master-0

openshift-machine-config-operator

deployment-controller

machine-config-controller

ScalingReplicaSet

Scaled up replica set machine-config-controller-b4f87c5b9 to 1

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-revisioncontroller

openshift-kube-scheduler-operator

SecretCreated

Created Secret/serving-cert-2 -n openshift-kube-scheduler because it was missing

openshift-route-controller-manager

default-scheduler

route-controller-manager-7ffc895647-6j97v

Scheduled

Successfully assigned openshift-route-controller-manager/route-controller-manager-7ffc895647-6j97v to master-0

openshift-machine-config-operator

machine-config-operator

machine-config-operator

ValidatingAdmissionPolicyBindingCreated

Created ValidatingAdmissionPolicyBinding.admissionregistration.k8s.io/custom-machine-config-pool-selector-binding because it was missing

openshift-machine-config-operator

machine-config-operator

machine-config-operator

ValidatingAdmissionPolicyBindingCreated

Created ValidatingAdmissionPolicyBinding.admissionregistration.k8s.io/managed-bootimages-platform-check-binding because it was missing

openshift-oauth-apiserver

kubelet

apiserver-bc9b556d6-vdnq2

Created

Created container: oauth-apiserver

openshift-oauth-apiserver

kubelet

apiserver-bc9b556d6-vdnq2

Started

Started container oauth-apiserver

openshift-kube-controller-manager-operator

kube-controller-manager-operator-revisioncontroller

kube-controller-manager-operator

ConfigMapCreated

Created ConfigMap/serviceaccount-ca-2 -n openshift-kube-controller-manager because it was missing

openshift-machine-config-operator

replicaset-controller

machine-config-controller-b4f87c5b9

SuccessfulCreate

Created pod: machine-config-controller-b4f87c5b9-pj7rj

openshift-marketplace

kubelet

redhat-marketplace-srjqw

Created

Created container: extract-utilities

openshift-route-controller-manager

kubelet

route-controller-manager-7ffc895647-6j97v

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:446bedea4916d3c1ee52be94137e484659e9561bd1de95c8189eee279aae984b"

openshift-route-controller-manager

multus

route-controller-manager-7ffc895647-6j97v

AddedInterface

Add eth0 [10.128.0.51/23] from ovn-kubernetes

openshift-machine-config-operator

multus

machine-config-controller-b4f87c5b9-pj7rj

AddedInterface

Add eth0 [10.128.0.53/23] from ovn-kubernetes

openshift-marketplace

kubelet

redhat-marketplace-srjqw

Started

Started container extract-utilities

openshift-marketplace

kubelet

redhat-marketplace-srjqw

Pulling

Pulling image "registry.redhat.io/redhat/redhat-marketplace-index:v4.18"
(x2)

openshift-operator-lifecycle-manager

operator-lifecycle-manager

packageserver

AllRequirementsMet

all requirements found, attempting install

openshift-marketplace

multus

redhat-operators-mgcb9

AddedInterface

Add eth0 [10.128.0.50/23] from ovn-kubernetes

openshift-marketplace

kubelet

redhat-operators-mgcb9

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6bba3f73c0066e42b24839e0d29f5dce2f36436f0a11f9f5e1029bccc5ed6578" already present on machine

openshift-marketplace

kubelet

redhat-operators-mgcb9

Created

Created container: extract-utilities

openshift-marketplace

kubelet

redhat-operators-mgcb9

Started

Started container extract-utilities

openshift-controller-manager

kubelet

controller-manager-6486d766f9-5b77h

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c2dd7a03348212e49876f5359f233d893a541ed9b934df390201a05133a06982"

openshift-controller-manager

multus

controller-manager-6486d766f9-5b77h

AddedInterface

Add eth0 [10.128.0.52/23] from ovn-kubernetes

openshift-machine-config-operator

kubelet

machine-config-controller-b4f87c5b9-pj7rj

Created

Created container: kube-rbac-proxy

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Available message changed from "APIServerDeploymentAvailable: no apiserver.openshift-oauth-apiserver pods available on any node.\nAPIServicesAvailable: PreconditionNotReady\nOAuthServerServiceEndpointAccessibleControllerAvailable: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerAvailable: endpoints \"oauth-openshift\" not found\nReadyIngressNodesAvailable: Authentication requires functional ingress which requires at least one schedulable and ready node. Got 0 worker nodes, 1 master nodes, 0 custom target nodes (none are schedulable or ready for ingress pods)." to "APIServerDeploymentAvailable: no apiserver.openshift-oauth-apiserver pods available on any node.\nAPIServicesAvailable: PreconditionNotReady\nOAuthServerServiceEndpointAccessibleControllerAvailable: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerAvailable: endpoints \"oauth-openshift\" not found"

openshift-kube-controller-manager-operator

kube-controller-manager-operator-revisioncontroller

kube-controller-manager-operator

ConfigMapCreated

Created ConfigMap/service-ca-2 -n openshift-kube-controller-manager because it was missing

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-revisioncontroller

openshift-kube-scheduler-operator

RevisionTriggered

new revision 2 triggered by "required configmap/serviceaccount-ca has changed"

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-revisioncontroller

openshift-kube-scheduler-operator

SecretCreated

Created Secret/localhost-recovery-client-token-2 -n openshift-kube-scheduler because it was missing

openshift-operator-lifecycle-manager

default-scheduler

packageserver-6c85f64bb9-fmpsg

Scheduled

Successfully assigned openshift-operator-lifecycle-manager/packageserver-6c85f64bb9-fmpsg to master-0

openshift-machine-config-operator

machine-config-operator

master-0

FeatureGatesInitialized

FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{"AWSEFSDriverVolumeMetrics", "AdminNetworkPolicy", "AlibabaPlatform", "AzureWorkloadIdentity", "BareMetalLoadBalancer", "BuildCSIVolumes", "ChunkSizeMiB", "CloudDualStackNodeIPs", "DisableKubeletCloudCredentialProviders", "GCPLabelsTags", "HardwareSpeed", "IngressControllerLBSubnetsAWS", "KMSv1", "ManagedBootImages", "ManagedBootImagesAWS", "MultiArchInstallAWS", "MultiArchInstallGCP", "NetworkDiagnosticsConfig", "NetworkLiveMigration", "NetworkSegmentation", "NewOLM", "NodeDisruptionPolicy", "OnClusterBuild", "PersistentIPsForVirtualization", "PrivateHostedZoneAWS", "SetEIPForNLBIngressController", "StreamingCollectionEncodingToJSON", "StreamingCollectionEncodingToProtobuf", "VSphereControlPlaneMachineSet", "VSphereDriverConfiguration", "VSphereMultiVCenters", "VSphereStaticIPs", "ValidatingAdmissionPolicy"}, Disabled:[]v1.FeatureGateName{"AWSClusterHostedDNS", "AdditionalRoutingCapabilities", "AutomatedEtcdBackup", "BootcNodeManagement", "CSIDriverSharedResource", "ClusterAPIInstall", "ClusterAPIInstallIBMCloud", "ClusterMonitoringConfig", "ConsolePluginContentSecurityPolicy", "DNSNameResolver", "DynamicResourceAllocation", "EtcdBackendQuota", "EventedPLEG", "Example", "ExternalOIDC", "ExternalOIDCWithUIDAndExtraClaimMappings", "GCPClusterHostedDNS", "GatewayAPI", "ImageStreamImportMode", "IngressControllerDynamicConfigurationManager", "InsightsConfig", "InsightsConfigAPI", "InsightsOnDemandDataGather", "InsightsRuntimeExtractor", "MachineAPIMigration", "MachineAPIOperatorDisableMachineHealthCheckController", "MachineAPIProviderOpenStack", "MachineConfigNodes", "MaxUnavailableStatefulSet", "MetricsCollectionProfiles", "MinimumKubeletVersion", "MixedCPUsAllocation", "MultiArchInstallAzure", "NodeSwap", "NutanixMultiSubnets", "OVNObservability", "OpenShiftPodSecurityAdmission", "PinnedImages", "PlatformOperators", "ProcMountType", "RouteAdvertisements", "RouteExternalCertificate", "ServiceAccountTokenNodeBinding", "SignatureStores", "SigstoreImageVerification", "TranslateStreamCloseWebsocketRequests", "UpgradeStatus", "UserNamespacesPodSecurityStandards", "UserNamespacesSupport", "VSphereMultiNetworks", "VolumeAttributesClass", "VolumeGroupSnapshot"}}

openshift-machine-config-operator

kubelet

machine-config-controller-b4f87c5b9-pj7rj

Started

Started container kube-rbac-proxy

openshift-marketplace

kubelet

redhat-operators-mgcb9

Pulling

Pulling image "registry.redhat.io/redhat/redhat-operator-index:v4.18"

openshift-apiserver-operator

openshift-apiserver-operator-apiservice-openshift-apiserver-controller-openshift-apiserver-apiservice

openshift-apiserver-operator

Created

Created <unknown>/v1.build.openshift.io because it was missing

openshift-machine-config-operator

kubelet

machine-config-controller-b4f87c5b9-pj7rj

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-monitoring

default-scheduler

prometheus-operator-admission-webhook-69c6b55594-tkc2j

Scheduled

Successfully assigned openshift-monitoring/prometheus-operator-admission-webhook-69c6b55594-tkc2j to master-0

openshift-machine-config-operator

kubelet

machine-config-controller-b4f87c5b9-pj7rj

Started

Started container machine-config-controller

openshift-machine-config-operator

kubelet

machine-config-controller-b4f87c5b9-pj7rj

Created

Created container: machine-config-controller

openshift-marketplace

default-scheduler

community-operators-zspn5

Scheduled

Successfully assigned openshift-marketplace/community-operators-zspn5 to master-0

openshift-monitoring

default-scheduler

prometheus-operator-admission-webhook-69c6b55594-tkc2j

Scheduled

Successfully assigned openshift-monitoring/prometheus-operator-admission-webhook-69c6b55594-tkc2j to master-0

openshift-apiserver-operator

openshift-apiserver-operator-apiservice-openshift-apiserver-controller-openshift-apiserver-apiservice

openshift-apiserver-operator

Created

Created <unknown>/v1.apps.openshift.io because it was missing

openshift-apiserver-operator

openshift-apiserver-operator-apiservice-openshift-apiserver-controller-openshift-apiserver-apiservice

openshift-apiserver-operator

Created

Created <unknown>/v1.authorization.openshift.io because it was missing

openshift-machine-config-operator

kubelet

machine-config-controller-b4f87c5b9-pj7rj

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:af0fe0ca926422a6471d5bf22fc0e682c36c24fba05496a3bdfac0b7d3733015" already present on machine

openshift-ingress

default-scheduler

router-default-7dcf5569b5-xmvwz

Scheduled

Successfully assigned openshift-ingress/router-default-7dcf5569b5-xmvwz to master-0

openshift-network-diagnostics

default-scheduler

network-check-source-b4bf74f6-fhvg6

Scheduled

Successfully assigned openshift-network-diagnostics/network-check-source-b4bf74f6-fhvg6 to master-0

openshift-apiserver-operator

openshift-apiserver-operator-apiservice-openshift-apiserver-controller-openshift-apiserver-apiservice

openshift-apiserver-operator

Created

Created <unknown>/v1.quota.openshift.io because it was missing

openshift-operator-lifecycle-manager

replicaset-controller

packageserver-6c85f64bb9

SuccessfulCreate

Created pod: packageserver-6c85f64bb9-fmpsg

openshift-marketplace

default-scheduler

certified-operators-86g9n

Scheduled

Successfully assigned openshift-marketplace/certified-operators-86g9n to master-0

openshift-apiserver-operator

openshift-apiserver-operator-apiservice-openshift-apiserver-controller-openshift-apiserver-apiservice

openshift-apiserver-operator

Created

Created <unknown>/v1.project.openshift.io because it was missing

openshift-apiserver-operator

openshift-apiserver-operator-apiservice-openshift-apiserver-controller-openshift-apiserver-apiservice

openshift-apiserver-operator

Created

Created <unknown>/v1.image.openshift.io because it was missing
(x2)

openshift-operator-lifecycle-manager

operator-lifecycle-manager

packageserver

InstallWaiting

apiServices not installed
(x2)

openshift-operator-lifecycle-manager

operator-lifecycle-manager

packageserver

InstallSucceeded

waiting for install components to report healthy

openshift-cluster-version

openshift-cluster-version

version

LoadPayload

Loading payload version="4.18.35" image="quay.io/openshift-release-dev/ocp-release@sha256:59727c4b3fef19e5149675cf3350735bbfe2c6588a57654b2e4552dd719f58b1"

openshift-cluster-version

openshift-cluster-version

version

RetrievePayload

Retrieving and verifying payload version="4.18.35" image="quay.io/openshift-release-dev/ocp-release@sha256:59727c4b3fef19e5149675cf3350735bbfe2c6588a57654b2e4552dd719f58b1"

openshift-operator-lifecycle-manager

deployment-controller

packageserver

ScalingReplicaSet

Scaled up replica set packageserver-6c85f64bb9 to 1

openshift-kube-controller-manager-operator

kube-controller-manager-operator-revisioncontroller

kube-controller-manager-operator

ConfigMapCreated

Created ConfigMap/recycler-config-2 -n openshift-kube-controller-manager because it was missing

openshift-operator-lifecycle-manager

multus

packageserver-6c85f64bb9-fmpsg

AddedInterface

Add eth0 [10.128.0.55/23] from ovn-kubernetes

openshift-marketplace

kubelet

certified-operators-86g9n

Created

Created container: extract-utilities

openshift-kube-apiserver-operator

kube-apiserver-operator-installer-controller

kube-apiserver-operator

PodCreated

Created Pod/installer-1-master-0 -n openshift-kube-apiserver because it was missing

openshift-monitoring

multus

prometheus-operator-admission-webhook-69c6b55594-tkc2j

AddedInterface

Add eth0 [10.128.0.56/23] from ovn-kubernetes

openshift-marketplace

multus

community-operators-zspn5

AddedInterface

Add eth0 [10.128.0.58/23] from ovn-kubernetes

openshift-operator-lifecycle-manager

kubelet

packageserver-6c85f64bb9-fmpsg

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6bba3f73c0066e42b24839e0d29f5dce2f36436f0a11f9f5e1029bccc5ed6578" already present on machine

openshift-network-diagnostics

multus

network-check-source-b4bf74f6-fhvg6

AddedInterface

Add eth0 [10.128.0.57/23] from ovn-kubernetes

openshift-marketplace

kubelet

certified-operators-86g9n

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6bba3f73c0066e42b24839e0d29f5dce2f36436f0a11f9f5e1029bccc5ed6578" already present on machine

openshift-marketplace

multus

certified-operators-86g9n

AddedInterface

Add eth0 [10.128.0.54/23] from ovn-kubernetes

openshift-monitoring

multus

prometheus-operator-admission-webhook-69c6b55594-tkc2j

AddedInterface

Add eth0 [10.128.0.56/23] from ovn-kubernetes

openshift-apiserver-operator

openshift-apiserver-operator-apiservice-openshift-apiserver-controller-openshift-apiserver-apiservice

openshift-apiserver-operator

Created

Created <unknown>/v1.route.openshift.io because it was missing

openshift-apiserver-operator

openshift-apiserver-operator-apiservice-openshift-apiserver-controller-openshift-apiserver-apiservice

openshift-apiserver-operator

Created

Created <unknown>/v1.security.openshift.io because it was missing

openshift-ingress

kubelet

router-default-7dcf5569b5-xmvwz

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:002dfb86e17ad8f5cc232a7d2dce183b23335c8ecb7e7d31dcf3e4446b390777"

openshift-marketplace

kubelet

community-operators-zspn5

Created

Created container: extract-utilities

openshift-apiserver-operator

openshift-apiserver-operator-apiservice-openshift-apiserver-controller-openshift-apiserver-apiservice

openshift-apiserver-operator

OpenShiftAPICheckFailed

"authorization.openshift.io.v1" failed with an attempt failed with statusCode = 503, err = the server is currently unable to handle the request

openshift-kube-apiserver

multus

installer-1-master-0

AddedInterface

Add eth0 [10.128.0.59/23] from ovn-kubernetes

openshift-network-diagnostics

kubelet

network-check-source-b4bf74f6-fhvg6

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ec8fd46dfb35ed10e8f98933166f69ce579c2f35b8db03d21e4c34fc544553e4" already present on machine

openshift-apiserver-operator

openshift-apiserver-operator-apiservice-openshift-apiserver-controller-openshift-apiserver-apiservice

openshift-apiserver-operator

Created

Created <unknown>/v1.template.openshift.io because it was missing

openshift-network-diagnostics

kubelet

network-check-source-b4bf74f6-fhvg6

Created

Created container: check-endpoints

openshift-marketplace

kubelet

community-operators-zspn5

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6bba3f73c0066e42b24839e0d29f5dce2f36436f0a11f9f5e1029bccc5ed6578" already present on machine

openshift-operator-lifecycle-manager

kubelet

packageserver-6c85f64bb9-fmpsg

Started

Started container packageserver

openshift-apiserver-operator

openshift-apiserver-operator-status-controller-statussyncer_openshift-apiserver

openshift-apiserver-operator

OperatorStatusChanged

Status for clusteroperator/openshift-apiserver changed: status.versions changed from [{"operator" "4.18.35"}] to [{"operator" "4.18.35"} {"openshift-apiserver" "4.18.35"}]

openshift-apiserver-operator

openshift-apiserver-operator-apiservice-openshift-apiserver-controller-openshift-apiserver-apiservice

openshift-apiserver-operator

OpenShiftAPICheckFailed

"apps.openshift.io.v1" failed with an attempt failed with statusCode = 503, err = the server is currently unable to handle the request

openshift-monitoring

kubelet

prometheus-operator-admission-webhook-69c6b55594-tkc2j

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:97c7a635130c574a2c501091bb44f17cd92e05e29b5102e59578b5885d9bfec0"

openshift-network-diagnostics

kubelet

network-check-source-b4bf74f6-fhvg6

Started

Started container check-endpoints

openshift-monitoring

kubelet

prometheus-operator-admission-webhook-69c6b55594-tkc2j

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:97c7a635130c574a2c501091bb44f17cd92e05e29b5102e59578b5885d9bfec0"

openshift-cluster-version

openshift-cluster-version

version

PayloadLoaded

Payload loaded version="4.18.35" image="quay.io/openshift-release-dev/ocp-release@sha256:59727c4b3fef19e5149675cf3350735bbfe2c6588a57654b2e4552dd719f58b1" architecture="amd64"

openshift-apiserver-operator

openshift-apiserver-operator-apiservice-openshift-apiserver-controller-openshift-apiserver-apiservice

openshift-apiserver-operator

OpenShiftAPICheckFailed

"build.openshift.io.v1" failed with an attempt failed with statusCode = 503, err = the server is currently unable to handle the request

openshift-apiserver-operator

openshift-apiserver-operator-apiservice-openshift-apiserver-controller-openshift-apiserver-apiservice

openshift-apiserver-operator

OpenShiftAPICheckFailed

"image.openshift.io.v1" failed with an attempt failed with statusCode = 503, err = the server is currently unable to handle the request

openshift-apiserver-operator

openshift-apiserver-operator-apiservice-openshift-apiserver-controller-openshift-apiserver-apiservice

openshift-apiserver-operator

OpenShiftAPICheckFailed

"project.openshift.io.v1" failed with an attempt failed with statusCode = 503, err = the server is currently unable to handle the request

openshift-apiserver-operator

openshift-apiserver-operator-apiservice-openshift-apiserver-controller-openshift-apiserver-apiservice

openshift-apiserver-operator

OpenShiftAPICheckFailed

"quota.openshift.io.v1" failed with an attempt failed with statusCode = 503, err = the server is currently unable to handle the request

openshift-operator-lifecycle-manager

kubelet

packageserver-6c85f64bb9-fmpsg

Created

Created container: packageserver

openshift-kube-controller-manager-operator

kube-controller-manager-operator-revisioncontroller

kube-controller-manager-operator

SecretCreated

Created Secret/service-account-private-key-2 -n openshift-kube-controller-manager because it was missing

openshift-marketplace

kubelet

certified-operators-86g9n

Started

Started container extract-utilities

openshift-apiserver-operator

openshift-apiserver-operator-status-controller-statussyncer_openshift-apiserver

openshift-apiserver-operator

OperatorVersionChanged

clusteroperator/openshift-apiserver version "openshift-apiserver" changed from "" to "4.18.35"

openshift-apiserver-operator

openshift-apiserver-operator-apiservice-openshift-apiserver-controller-openshift-apiserver-apiservice

openshift-apiserver-operator

OpenShiftAPICheckFailed

"route.openshift.io.v1" failed with an attempt failed with statusCode = 503, err = the server is currently unable to handle the request

openshift-apiserver-operator

openshift-apiserver-operator-apiservice-openshift-apiserver-controller-openshift-apiserver-apiservice

openshift-apiserver-operator

OpenShiftAPICheckFailed

"security.openshift.io.v1" failed with an attempt failed with statusCode = 503, err = the server is currently unable to handle the request

openshift-apiserver-operator

openshift-apiserver-operator-status-controller-statussyncer_openshift-apiserver

openshift-apiserver-operator

OperatorStatusChanged

Status for clusteroperator/openshift-apiserver changed: Progressing changed from True to False ("All is well"),Available message changed from "APIServerDeploymentAvailable: no apiserver.openshift-apiserver pods available on any node.\nAPIServicesAvailable: PreconditionNotReady" to "APIServicesAvailable: PreconditionNotReady"

openshift-marketplace

kubelet

certified-operators-86g9n

Pulling

Pulling image "registry.redhat.io/redhat/certified-operator-index:v4.18"

openshift-apiserver-operator

openshift-apiserver-operator-apiservice-openshift-apiserver-controller-openshift-apiserver-apiservice

openshift-apiserver-operator

OpenShiftAPICheckFailed

"template.openshift.io.v1" failed with an attempt failed with statusCode = 503, err = the server is currently unable to handle the request

openshift-kube-controller-manager-operator

kube-controller-manager-operator-revisioncontroller

kube-controller-manager-operator

SecretCreated

Created Secret/serving-cert-2 -n openshift-kube-controller-manager because it was missing

openshift-operator-lifecycle-manager

kubelet

packageserver-6c85f64bb9-fmpsg

ProbeError

Readiness probe error: Get "https://10.128.0.55:5443/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) body:

openshift-operator-lifecycle-manager

kubelet

packageserver-6c85f64bb9-fmpsg

Unhealthy

Readiness probe failed: Get "https://10.128.0.55:5443/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)

openshift-apiserver-operator

openshift-apiserver-operator-status-controller-statussyncer_openshift-apiserver

openshift-apiserver-operator

OperatorStatusChanged

Status for clusteroperator/openshift-apiserver changed: Available message changed from "APIServicesAvailable: PreconditionNotReady" to "APIServicesAvailable: \"apps.openshift.io.v1\" is not ready: an attempt failed with statusCode = 503, err = the server is currently unable to handle the request\nAPIServicesAvailable: \"authorization.openshift.io.v1\" is not ready: an attempt failed with statusCode = 503, err = the server is currently unable to handle the request\nAPIServicesAvailable: \"build.openshift.io.v1\" is not ready: an attempt failed with statusCode = 503, err = the server is currently unable to handle the request\nAPIServicesAvailable: \"image.openshift.io.v1\" is not ready: an attempt failed with statusCode = 503, err = the server is currently unable to handle the request\nAPIServicesAvailable: \"project.openshift.io.v1\" is not ready: an attempt failed with statusCode = 503, err = the server is currently unable to handle the request\nAPIServicesAvailable: \"quota.openshift.io.v1\" is not ready: an attempt failed with statusCode = 503, err = the server is currently unable to handle the request\nAPIServicesAvailable: \"route.openshift.io.v1\" is not ready: an attempt failed with statusCode = 503, err = the server is currently unable to handle the request\nAPIServicesAvailable: \"security.openshift.io.v1\" is not ready: an attempt failed with statusCode = 503, err = the server is currently unable to handle the request\nAPIServicesAvailable: \"template.openshift.io.v1\" is not ready: an attempt failed with statusCode = 503, err = the server is currently unable to handle the request"

openshift-marketplace

kubelet

community-operators-zspn5

Started

Started container extract-utilities

openshift-marketplace

kubelet

community-operators-zspn5

Pulling

Pulling image "registry.redhat.io/redhat/community-operator-index:v4.18"

openshift-kube-controller-manager-operator

kube-controller-manager-operator-revisioncontroller

kube-controller-manager-operator

RevisionTriggered

new revision 2 triggered by "required configmap/config has changed,required configmap/cluster-policy-controller-config has changed"

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-status-controller-statussyncer_kube-scheduler

openshift-kube-scheduler-operator

OperatorStatusChanged

Status for clusteroperator/kube-scheduler changed: Progressing message changed from "NodeInstallerProgressing: 1 node is at revision 0; 0 nodes have achieved new revision 1" to "NodeInstallerProgressing: 1 node is at revision 0; 0 nodes have achieved new revision 2",Available message changed from "StaticPodsAvailable: 0 nodes are active; 1 node is at revision 0; 0 nodes have achieved new revision 1" to "StaticPodsAvailable: 0 nodes are active; 1 node is at revision 0; 0 nodes have achieved new revision 2"

openshift-operator-lifecycle-manager

kubelet

packageserver-6c85f64bb9-fmpsg

Unhealthy

Readiness probe failed: HTTP probe failed with statuscode: 500

openshift-authentication-operator

oauth-apiserver-apiservice-openshift-apiserver-controller-openshift-apiserver-apiservice

authentication-operator

OpenShiftAPICheckFailed

"oauth.openshift.io.v1" failed with an attempt failed with statusCode = 503, err = the server is currently unable to handle the request

openshift-apiserver-operator

openshift-apiserver-operator-status-controller-statussyncer_openshift-apiserver

openshift-apiserver-operator

OperatorStatusChanged

Status for clusteroperator/openshift-apiserver changed: Available message changed from "APIServicesAvailable: \"apps.openshift.io.v1\" is not ready: an attempt failed with statusCode = 503, err = the server is currently unable to handle the request\nAPIServicesAvailable: \"authorization.openshift.io.v1\" is not ready: an attempt failed with statusCode = 503, err = the server is currently unable to handle the request\nAPIServicesAvailable: \"build.openshift.io.v1\" is not ready: an attempt failed with statusCode = 503, err = the server is currently unable to handle the request\nAPIServicesAvailable: \"image.openshift.io.v1\" is not ready: an attempt failed with statusCode = 503, err = the server is currently unable to handle the request\nAPIServicesAvailable: \"project.openshift.io.v1\" is not ready: an attempt failed with statusCode = 503, err = the server is currently unable to handle the request\nAPIServicesAvailable: \"quota.openshift.io.v1\" is not ready: an attempt failed with statusCode = 503, err = the server is currently unable to handle the request\nAPIServicesAvailable: \"route.openshift.io.v1\" is not ready: an attempt failed with statusCode = 503, err = the server is currently unable to handle the request\nAPIServicesAvailable: \"security.openshift.io.v1\" is not ready: an attempt failed with statusCode = 503, err = the server is currently unable to handle the request\nAPIServicesAvailable: \"template.openshift.io.v1\" is not ready: an attempt failed with statusCode = 503, err = the server is currently unable to handle the request" to "APIServicesAvailable: apiservices.apiregistration.k8s.io/v1.apps.openshift.io: not available: failing or missing response from https://10.128.0.43:8443/apis/apps.openshift.io/v1: bad status from https://10.128.0.43:8443/apis/apps.openshift.io/v1: 401\nAPIServicesAvailable: apiservices.apiregistration.k8s.io/v1.authorization.openshift.io: not available: failing or missing response from https://10.128.0.43:8443/apis/authorization.openshift.io/v1: bad status from https://10.128.0.43:8443/apis/authorization.openshift.io/v1: 401\nAPIServicesAvailable: apiservices.apiregistration.k8s.io/v1.build.openshift.io: not available: failing or missing response from https://10.128.0.43:8443/apis/build.openshift.io/v1: bad status from https://10.128.0.43:8443/apis/build.openshift.io/v1: 401\nAPIServicesAvailable: apiservices.apiregistration.k8s.io/v1.image.openshift.io: not available: failing or missing response from https://10.128.0.43:8443/apis/image.openshift.io/v1: bad status from https://10.128.0.43:8443/apis/image.openshift.io/v1: 401\nAPIServicesAvailable: apiservices.apiregistration.k8s.io/v1.project.openshift.io: not available: failing or missing response from https://10.128.0.43:8443/apis/project.openshift.io/v1: bad status from https://10.128.0.43:8443/apis/project.openshift.io/v1: 401\nAPIServicesAvailable: apiservices.apiregistration.k8s.io/v1.quota.openshift.io: not available: failing or missing response from https://10.128.0.43:8443/apis/quota.openshift.io/v1: bad status from https://10.128.0.43:8443/apis/quota.openshift.io/v1: 401\nAPIServicesAvailable: apiservices.apiregistration.k8s.io/v1.route.openshift.io: not available: failing or missing response from https://10.128.0.43:8443/apis/route.openshift.io/v1: bad status from https://10.128.0.43:8443/apis/route.openshift.io/v1: 401\nAPIServicesAvailable: apiservices.apiregistration.k8s.io/v1.security.openshift.io: not available: failing or missing response from https://10.128.0.43:8443/apis/security.openshift.io/v1: bad status from https://10.128.0.43:8443/apis/security.openshift.io/v1: 401\nAPIServicesAvailable: apiservices.apiregistration.k8s.io/v1.template.openshift.io: not available: failing or missing response from https://10.128.0.43:8443/apis/template.openshift.io/v1: bad status from https://10.128.0.43:8443/apis/template.openshift.io/v1: 401"

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorVersionChanged

clusteroperator/authentication version "oauth-apiserver" changed from "" to "4.18.35"

openshift-kube-scheduler

kubelet

installer-1-master-0

Killing

Stopping container installer

openshift-kube-controller-manager-operator

kube-controller-manager-operator-revisioncontroller

kube-controller-manager-operator

SecretCreated

Created Secret/localhost-recovery-client-token-2 -n openshift-kube-controller-manager because it was missing

openshift-operator-lifecycle-manager

kubelet

packageserver-6c85f64bb9-fmpsg

ProbeError

Readiness probe error: HTTP probe failed with statuscode: 500 body: [+]ping ok [+]log ok [-]poststarthook/generic-apiserver-start-informers failed: reason withheld [-]poststarthook/max-in-flight-filter failed: reason withheld [-]poststarthook/storage-object-count-tracker-hook failed: reason withheld healthz check failed

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Progressing changed from True to False ("AuthenticatorCertKeyProgressing: All is well"),Available message changed from "APIServerDeploymentAvailable: no apiserver.openshift-oauth-apiserver pods available on any node.\nAPIServicesAvailable: apiservices.apiregistration.k8s.io/v1.oauth.openshift.io: not available: failing or missing response from https://10.128.0.45:8443/apis/oauth.openshift.io/v1: bad status from https://10.128.0.45:8443/apis/oauth.openshift.io/v1: 401\nAPIServicesAvailable: apiservices.apiregistration.k8s.io/v1.user.openshift.io: not available: failing or missing response from https://10.128.0.45:8443/apis/user.openshift.io/v1: bad status from https://10.128.0.45:8443/apis/user.openshift.io/v1: 401\nOAuthServerServiceEndpointAccessibleControllerAvailable: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerAvailable: endpoints \"oauth-openshift\" not found" to "APIServicesAvailable: apiservices.apiregistration.k8s.io/v1.oauth.openshift.io: not available: failing or missing response from https://10.128.0.45:8443/apis/oauth.openshift.io/v1: bad status from https://10.128.0.45:8443/apis/oauth.openshift.io/v1: 401\nAPIServicesAvailable: apiservices.apiregistration.k8s.io/v1.user.openshift.io: not available: failing or missing response from https://10.128.0.45:8443/apis/user.openshift.io/v1: bad status from https://10.128.0.45:8443/apis/user.openshift.io/v1: 401\nOAuthServerServiceEndpointAccessibleControllerAvailable: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerAvailable: endpoints \"oauth-openshift\" not found"

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Available message changed from "APIServerDeploymentAvailable: no apiserver.openshift-oauth-apiserver pods available on any node.\nAPIServicesAvailable: \"oauth.openshift.io.v1\" is not ready: an attempt failed with statusCode = 503, err = the server is currently unable to handle the request\nAPIServicesAvailable: \"user.openshift.io.v1\" is not ready: an attempt failed with statusCode = 503, err = the server is currently unable to handle the request\nOAuthServerServiceEndpointAccessibleControllerAvailable: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerAvailable: endpoints \"oauth-openshift\" not found" to "APIServerDeploymentAvailable: no apiserver.openshift-oauth-apiserver pods available on any node.\nAPIServicesAvailable: apiservices.apiregistration.k8s.io/v1.oauth.openshift.io: not available: failing or missing response from https://10.128.0.45:8443/apis/oauth.openshift.io/v1: bad status from https://10.128.0.45:8443/apis/oauth.openshift.io/v1: 401\nAPIServicesAvailable: apiservices.apiregistration.k8s.io/v1.user.openshift.io: not available: failing or missing response from https://10.128.0.45:8443/apis/user.openshift.io/v1: bad status from https://10.128.0.45:8443/apis/user.openshift.io/v1: 401\nOAuthServerServiceEndpointAccessibleControllerAvailable: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerAvailable: endpoints \"oauth-openshift\" not found"

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Available message changed from "APIServerDeploymentAvailable: no apiserver.openshift-oauth-apiserver pods available on any node.\nAPIServicesAvailable: PreconditionNotReady\nOAuthServerServiceEndpointAccessibleControllerAvailable: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerAvailable: endpoints \"oauth-openshift\" not found" to "APIServerDeploymentAvailable: no apiserver.openshift-oauth-apiserver pods available on any node.\nAPIServicesAvailable: \"oauth.openshift.io.v1\" is not ready: an attempt failed with statusCode = 503, err = the server is currently unable to handle the request\nAPIServicesAvailable: \"user.openshift.io.v1\" is not ready: an attempt failed with statusCode = 503, err = the server is currently unable to handle the request\nOAuthServerServiceEndpointAccessibleControllerAvailable: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerAvailable: endpoints \"oauth-openshift\" not found"

openshift-authentication-operator

oauth-apiserver-apiservice-openshift-apiserver-controller-openshift-apiserver-apiservice

authentication-operator

OpenShiftAPICheckFailed

"user.openshift.io.v1" failed with an attempt failed with statusCode = 503, err = the server is currently unable to handle the request

openshift-authentication-operator

oauth-apiserver-apiservice-openshift-apiserver-controller-openshift-apiserver-apiservice

authentication-operator

Created

Created <unknown>/v1.user.openshift.io because it was missing

openshift-authentication-operator

oauth-apiserver-apiservice-openshift-apiserver-controller-openshift-apiserver-apiservice

authentication-operator

Created

Created <unknown>/v1.oauth.openshift.io because it was missing

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: status.versions changed from [{"operator" "4.18.35"}] to [{"operator" "4.18.35"} {"oauth-apiserver" "4.18.35"}]

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-installer-controller

openshift-kube-scheduler-operator

PodCreated

Created Pod/installer-2-master-0 -n openshift-kube-scheduler because it was missing

openshift-machine-config-operator

machineconfigcontroller-rendercontroller

worker

RenderedConfigGenerated

rendered-worker-6c41cf327075224fbef5e7f434bc5d69 successfully generated (release version: 4.18.35, controller version: 393b8dc2c216dbbbf68cd1ccde5cbc2b551b2fe8)

openshift-machine-config-operator

machineconfigcontroller-rendercontroller

master

RenderedConfigGenerated

rendered-master-47246abc0cee5fbe972e47e23f330f37 successfully generated (release version: 4.18.35, controller version: 393b8dc2c216dbbbf68cd1ccde5cbc2b551b2fe8)

openshift-kube-controller-manager-operator

kube-controller-manager-operator-status-controller-statussyncer_kube-controller-manager

kube-controller-manager-operator

OperatorStatusChanged

Status for clusteroperator/kube-controller-manager changed: Progressing message changed from "NodeInstallerProgressing: 1 node is at revision 0; 0 nodes have achieved new revision 1" to "NodeInstallerProgressing: 1 node is at revision 0; 0 nodes have achieved new revision 2",Available message changed from "StaticPodsAvailable: 0 nodes are active; 1 node is at revision 0; 0 nodes have achieved new revision 1" to "StaticPodsAvailable: 0 nodes are active; 1 node is at revision 0; 0 nodes have achieved new revision 2"

openshift-machine-config-operator

machine-config-operator

machine-config-operator

ClusterRoleBindingCreated

Created ClusterRoleBinding.rbac.authorization.k8s.io/system-bootstrap-node-renewal because it was missing

openshift-machine-config-operator

machine-config-operator

machine-config-operator

ClusterRoleBindingCreated

Created ClusterRoleBinding.rbac.authorization.k8s.io/machine-config-server because it was missing

openshift-machine-config-operator

machine-config-operator

machine-config-operator

ServiceAccountCreated

Created ServiceAccount/machine-config-server -n openshift-machine-config-operator because it was missing

openshift-machine-config-operator

machine-config-operator

machine-config-operator

ClusterRoleCreated

Created ClusterRole.rbac.authorization.k8s.io/machine-config-server because it was missing

openshift-machine-config-operator

machine-config-operator

machine-config-operator

ServiceAccountCreated

Created ServiceAccount/node-bootstrapper -n openshift-machine-config-operator because it was missing

openshift-machine-config-operator

machine-config-operator

machine-config-operator

SecretCreated

Created Secret/node-bootstrapper-token -n openshift-machine-config-operator because it was missing

openshift-machine-config-operator

daemonset-controller

machine-config-server

SuccessfulCreate

Created pod: machine-config-server-gj4pm

openshift-kube-apiserver

kubelet

installer-1-master-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c5ce3d1134d6500e2b8528516c1889d7bbc6259aba4981c6983395b0e9eeff65" already present on machine

openshift-machine-config-operator

default-scheduler

machine-config-server-gj4pm

Scheduled

Successfully assigned openshift-machine-config-operator/machine-config-server-gj4pm to master-0

openshift-kube-controller-manager

kubelet

installer-1-master-0

Killing

Stopping container installer

openshift-machine-config-operator

machineconfigcontroller-nodecontroller

master

AnnotationChange

Node master-0 now has machineconfiguration.openshift.io/currentConfig=rendered-master-47246abc0cee5fbe972e47e23f330f37

openshift-machine-config-operator

machineconfigcontroller-nodecontroller

master

AnnotationChange

Node master-0 now has machineconfiguration.openshift.io/desiredConfig=rendered-master-47246abc0cee5fbe972e47e23f330f37

openshift-machine-config-operator

machineconfigcontroller-nodecontroller

master

AnnotationChange

Node master-0 now has machineconfiguration.openshift.io/state=Done

openshift-kube-controller-manager-operator

kube-controller-manager-operator-installer-controller

kube-controller-manager-operator

PodCreated

Created Pod/installer-2-master-0 -n openshift-kube-controller-manager because it was missing

openshift-kube-scheduler

multus

installer-2-master-0

AddedInterface

Add eth0 [10.128.0.60/23] from ovn-kubernetes

openshift-route-controller-manager

kubelet

route-controller-manager-7ffc895647-6j97v

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:446bedea4916d3c1ee52be94137e484659e9561bd1de95c8189eee279aae984b" in 13.127s (13.127s including waiting). Image size: 487096305 bytes.

openshift-machine-config-operator

kubelet

machine-config-server-gj4pm

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:af0fe0ca926422a6471d5bf22fc0e682c36c24fba05496a3bdfac0b7d3733015" already present on machine

openshift-route-controller-manager

replicaset-controller

route-controller-manager-56f686584b

SuccessfulCreate

Created pod: route-controller-manager-56f686584b-fdcx5
(x4)

openshift-controller-manager-operator

openshift-controller-manager-operator

openshift-controller-manager-operator

DeploymentUpdated

Updated Deployment.apps/route-controller-manager -n openshift-route-controller-manager because it changed

openshift-route-controller-manager

kubelet

route-controller-manager-7ffc895647-6j97v

Started

Started container route-controller-manager

openshift-machine-api

replicaset-controller

control-plane-machine-set-operator-6f97756bc8

SuccessfulCreate

Created pod: control-plane-machine-set-operator-6f97756bc8-7t5qv

openshift-machine-config-operator

kubelet

machine-config-server-gj4pm

Created

Created container: machine-config-server

openshift-machine-api

default-scheduler

control-plane-machine-set-operator-6f97756bc8-7t5qv

Scheduled

Successfully assigned openshift-machine-api/control-plane-machine-set-operator-6f97756bc8-7t5qv to master-0

openshift-machine-config-operator

kubelet

machine-config-server-gj4pm

Started

Started container machine-config-server

openshift-kube-apiserver

kubelet

installer-1-master-0

Started

Started container installer

openshift-route-controller-manager

deployment-controller

route-controller-manager

ScalingReplicaSet

Scaled down replica set route-controller-manager-7ffc895647 to 0 from 1

openshift-controller-manager

default-scheduler

controller-manager-fc56bb77c-qd4sn

FailedScheduling

0/1 nodes are available: 1 node(s) didn't match pod anti-affinity rules. preemption: 0/1 nodes are available: 1 node(s) didn't match pod anti-affinity rules.

openshift-controller-manager

deployment-controller

controller-manager

ScalingReplicaSet

Scaled up replica set controller-manager-fc56bb77c to 1 from 0

openshift-machine-api

replicaset-controller

control-plane-machine-set-operator-6f97756bc8

SuccessfulCreate

Created pod: control-plane-machine-set-operator-6f97756bc8-7t5qv

openshift-controller-manager

deployment-controller

controller-manager

ScalingReplicaSet

Scaled down replica set controller-manager-6486d766f9 to 0 from 1

openshift-route-controller-manager

deployment-controller

route-controller-manager

ScalingReplicaSet

Scaled up replica set route-controller-manager-56f686584b to 1 from 0

openshift-kube-apiserver

kubelet

installer-1-master-0

Created

Created container: installer

openshift-machine-api

default-scheduler

control-plane-machine-set-operator-6f97756bc8-7t5qv

Scheduled

Successfully assigned openshift-machine-api/control-plane-machine-set-operator-6f97756bc8-7t5qv to master-0

openshift-route-controller-manager

replicaset-controller

route-controller-manager-7ffc895647

SuccessfulDelete

Deleted pod: route-controller-manager-7ffc895647-6j97v

openshift-controller-manager

replicaset-controller

controller-manager-6486d766f9

SuccessfulDelete

Deleted pod: controller-manager-6486d766f9-5b77h

openshift-route-controller-manager

default-scheduler

route-controller-manager-56f686584b-fdcx5

FailedScheduling

0/1 nodes are available: 1 node(s) didn't match pod anti-affinity rules. preemption: 0/1 nodes are available: 1 node(s) didn't match pod anti-affinity rules.
(x4)

openshift-controller-manager-operator

openshift-controller-manager-operator

openshift-controller-manager-operator

DeploymentUpdated

Updated Deployment.apps/controller-manager -n openshift-controller-manager because it changed

openshift-controller-manager

replicaset-controller

controller-manager-fc56bb77c

SuccessfulCreate

Created pod: controller-manager-fc56bb77c-qd4sn

openshift-route-controller-manager

kubelet

route-controller-manager-7ffc895647-6j97v

Created

Created container: route-controller-manager

openshift-route-controller-manager

route-controller-manager

openshift-route-controllers

LeaderElection

route-controller-manager-7ffc895647-6j97v_0ca6b232-7717-4cf9-bac0-64ff04f0ff1a became leader

openshift-controller-manager-operator

openshift-controller-manager-operator-status-controller-statussyncer_openshift-controller-manager

openshift-controller-manager-operator

OperatorStatusChanged

Status for clusteroperator/openshift-controller-manager changed: Progressing message changed from "Progressing: deployment/controller-manager: observed generation is 3, desired generation is 4.\nProgressing: deployment/controller-manager: available replicas is 0, desired available replicas > 1\nProgressing: deployment/route-controller-manager: observed generation is 3, desired generation is 4.\nProgressing: deployment/route-controller-manager: available replicas is 0, desired available replicas > 1" to "Progressing: deployment/controller-manager: observed generation is 4, desired generation is 5.\nProgressing: deployment/controller-manager: available replicas is 0, desired available replicas > 1\nProgressing: deployment/route-controller-manager: observed generation is 4, desired generation is 5.\nProgressing: deployment/route-controller-manager: available replicas is 0, desired available replicas > 1"

openshift-controller-manager-operator

openshift-controller-manager-operator

openshift-controller-manager-operator

ConfigMapUpdated

Updated ConfigMap/config -n openshift-route-controller-manager: cause by changes in data.openshift-route-controller-manager.client-ca.configmap

openshift-controller-manager-operator

openshift-controller-manager-operator

openshift-controller-manager-operator

ConfigMapUpdated

Updated ConfigMap/config -n openshift-controller-manager: cause by changes in data.openshift-controller-manager.client-ca.configmap

openshift-machine-api

deployment-controller

control-plane-machine-set-operator

ScalingReplicaSet

Scaled up replica set control-plane-machine-set-operator-6f97756bc8 to 1

openshift-machine-api

deployment-controller

control-plane-machine-set-operator

ScalingReplicaSet

Scaled up replica set control-plane-machine-set-operator-6f97756bc8 to 1

openshift-machine-config-operator

machine-config-operator

machine-config-operator

ServiceAccountCreated

Created ServiceAccount/machine-os-builder -n openshift-machine-config-operator because it was missing

openshift-machine-config-operator

machine-config-operator

machine-config-operator

ClusterRoleCreated

Created ClusterRole.rbac.authorization.k8s.io/machine-os-builder-events because it was missing

openshift-machine-config-operator

machine-config-operator

machine-config-operator

RoleBindingCreated

Created RoleBinding.rbac.authorization.k8s.io/machine-os-builder-events -n default because it was missing

openshift-machine-config-operator

machine-config-operator

machine-config-operator

RoleBindingCreated

Created RoleBinding.rbac.authorization.k8s.io/machine-os-builder-events -n openshift-machine-config-operator because it was missing

openshift-machine-config-operator

machine-config-operator

machine-config-operator

ClusterRoleBindingCreated

Created ClusterRoleBinding.rbac.authorization.k8s.io/machine-os-builder because it was missing

openshift-machine-config-operator

machine-config-operator

machine-config-operator

ClusterRoleBindingCreated

Created ClusterRoleBinding.rbac.authorization.k8s.io/machine-os-builder-anyuid because it was missing

openshift-machine-config-operator

machine-config-operator

machine-config-operator

ClusterRoleCreated

Created ClusterRole.rbac.authorization.k8s.io/machine-os-builder because it was missing

openshift-cluster-machine-approver

default-scheduler

machine-approver-6cb57bb5db-mt758

Scheduled

Successfully assigned openshift-cluster-machine-approver/machine-approver-6cb57bb5db-mt758 to master-0

openshift-cluster-machine-approver

replicaset-controller

machine-approver-6cb57bb5db

SuccessfulCreate

Created pod: machine-approver-6cb57bb5db-mt758

openshift-cluster-machine-approver

deployment-controller

machine-approver

ScalingReplicaSet

Scaled up replica set machine-approver-6cb57bb5db to 1

openshift-route-controller-manager

kubelet

route-controller-manager-7ffc895647-6j97v

Killing

Stopping container route-controller-manager

openshift-machine-config-operator

machine-config-operator

machine-config-operator

SecretCreated

Created Secret/worker-user-data-managed -n openshift-machine-api because it was missing

openshift-machine-config-operator

machineconfigoperator

machine-config

OperatorVersionChanged

clusteroperator/machine-config version changed from [] to [{operator 4.18.35} {operator-image quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:af0fe0ca926422a6471d5bf22fc0e682c36c24fba05496a3bdfac0b7d3733015}]

openshift-route-controller-manager

kubelet

route-controller-manager-7ffc895647-6j97v

ProbeError

Readiness probe error: Get "https://10.128.0.51:8443/healthz": dial tcp 10.128.0.51:8443: connect: connection refused body:

openshift-cluster-samples-operator

default-scheduler

cluster-samples-operator-85f7577d78-mwfgx

Scheduled

Successfully assigned openshift-cluster-samples-operator/cluster-samples-operator-85f7577d78-mwfgx to master-0

openshift-cloud-credential-operator

replicaset-controller

cloud-credential-operator-744f9dbf77

SuccessfulCreate

Created pod: cloud-credential-operator-744f9dbf77-r4qvh

openshift-cloud-credential-operator

deployment-controller

cloud-credential-operator

ScalingReplicaSet

Scaled up replica set cloud-credential-operator-744f9dbf77 to 1

openshift-kube-scheduler

kubelet

installer-2-master-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ef199844317b7b012879ed8d29f9b6bc37fad8a6fdb336103cbd5cabc74c4302" already present on machine

openshift-cluster-samples-operator

replicaset-controller

cluster-samples-operator-85f7577d78

SuccessfulCreate

Created pod: cluster-samples-operator-85f7577d78-mwfgx

openshift-cloud-credential-operator

default-scheduler

cloud-credential-operator-744f9dbf77-r4qvh

Scheduled

Successfully assigned openshift-cloud-credential-operator/cloud-credential-operator-744f9dbf77-r4qvh to master-0

openshift-cluster-samples-operator

deployment-controller

cluster-samples-operator

ScalingReplicaSet

Scaled up replica set cluster-samples-operator-85f7577d78 to 1

openshift-route-controller-manager

kubelet

route-controller-manager-7ffc895647-6j97v

Unhealthy

Readiness probe failed: Get "https://10.128.0.51:8443/healthz": dial tcp 10.128.0.51:8443: connect: connection refused
(x31)

openshift-kube-controller-manager-operator

kube-controller-manager-operator-satokensignercontroller

kube-controller-manager-operator

SATokenSignerControllerStuck

unexpected addresses: 192.168.32.10

openshift-machine-api

default-scheduler

cluster-autoscaler-operator-866dc4744-xwxg7

Scheduled

Successfully assigned openshift-machine-api/cluster-autoscaler-operator-866dc4744-xwxg7 to master-0

openshift-machine-api

replicaset-controller

cluster-autoscaler-operator-866dc4744

SuccessfulCreate

Created pod: cluster-autoscaler-operator-866dc4744-xwxg7

openshift-machine-api

deployment-controller

cluster-autoscaler-operator

ScalingReplicaSet

Scaled up replica set cluster-autoscaler-operator-866dc4744 to 1

openshift-machine-api

deployment-controller

cluster-autoscaler-operator

ScalingReplicaSet

Scaled up replica set cluster-autoscaler-operator-866dc4744 to 1

openshift-machine-api

default-scheduler

cluster-autoscaler-operator-866dc4744-xwxg7

Scheduled

Successfully assigned openshift-machine-api/cluster-autoscaler-operator-866dc4744-xwxg7 to master-0

openshift-machine-api

replicaset-controller

cluster-autoscaler-operator-866dc4744

SuccessfulCreate

Created pod: cluster-autoscaler-operator-866dc4744-xwxg7

openshift-insights

replicaset-controller

insights-operator-68bf6ff9d6

SuccessfulCreate

Created pod: insights-operator-68bf6ff9d6-mvfn5

openshift-insights

default-scheduler

insights-operator-68bf6ff9d6-mvfn5

Scheduled

Successfully assigned openshift-insights/insights-operator-68bf6ff9d6-mvfn5 to master-0

openshift-insights

deployment-controller

insights-operator

ScalingReplicaSet

Scaled up replica set insights-operator-68bf6ff9d6 to 1

openshift-controller-manager

kubelet

controller-manager-6486d766f9-5b77h

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c2dd7a03348212e49876f5359f233d893a541ed9b934df390201a05133a06982" in 22.556s (22.556s including waiting). Image size: 558211175 bytes.

openshift-cloud-controller-manager-operator

default-scheduler

cluster-cloud-controller-manager-operator-7559f7c68c-xhq29

Scheduled

Successfully assigned openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-7559f7c68c-xhq29 to master-0

openshift-cloud-controller-manager-operator

replicaset-controller

cluster-cloud-controller-manager-operator-7559f7c68c

SuccessfulCreate

Created pod: cluster-cloud-controller-manager-operator-7559f7c68c-xhq29

openshift-cloud-controller-manager-operator

deployment-controller

cluster-cloud-controller-manager-operator

ScalingReplicaSet

Scaled up replica set cluster-cloud-controller-manager-operator-7559f7c68c to 1

openshift-machine-api

default-scheduler

machine-api-operator-6fbb6cf6f9-n8tnn

Scheduled

Successfully assigned openshift-machine-api/machine-api-operator-6fbb6cf6f9-n8tnn to master-0

openshift-machine-api

deployment-controller

machine-api-operator

ScalingReplicaSet

Scaled up replica set machine-api-operator-6fbb6cf6f9 to 1

openshift-machine-api

replicaset-controller

machine-api-operator-6fbb6cf6f9

SuccessfulCreate

Created pod: machine-api-operator-6fbb6cf6f9-n8tnn

openshift-machine-api

default-scheduler

machine-api-operator-6fbb6cf6f9-n8tnn

Scheduled

Successfully assigned openshift-machine-api/machine-api-operator-6fbb6cf6f9-n8tnn to master-0

openshift-machine-api

replicaset-controller

machine-api-operator-6fbb6cf6f9

SuccessfulCreate

Created pod: machine-api-operator-6fbb6cf6f9-n8tnn

openshift-machine-api

deployment-controller

machine-api-operator

ScalingReplicaSet

Scaled up replica set machine-api-operator-6fbb6cf6f9 to 1

openshift-machine-config-operator

machine-config-operator

machine-config-operator

ConfigMapUpdated

Updated ConfigMap/kube-rbac-proxy -n openshift-machine-config-operator: cause by changes in data.config-file.yaml

openshift-etcd

kubelet

etcd-master-0-master-0

Killing

Stopping container etcdctl

openshift-route-controller-manager

kubelet

route-controller-manager-7ffc895647-6j97v

ProbeError

Readiness probe error: Get "https://10.128.0.51:8443/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) body:

openshift-marketplace

kubelet

certified-operators-86g9n

Pulled

Successfully pulled image "registry.redhat.io/redhat/certified-operator-index:v4.18" in 26.654s (26.654s including waiting). Image size: 1252654287 bytes.

openshift-route-controller-manager

kubelet

route-controller-manager-7ffc895647-6j97v

Unhealthy

Readiness probe failed: Get "https://10.128.0.51:8443/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)

openshift-kube-apiserver-operator

kubelet

kube-apiserver-operator-8b68b9d9b-c6vkz

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c5ce3d1134d6500e2b8528516c1889d7bbc6259aba4981c6983395b0e9eeff65" already present on machine

openshift-marketplace

kubelet

redhat-marketplace-srjqw

Pulled

Successfully pulled image "registry.redhat.io/redhat/redhat-marketplace-index:v4.18" in 31.508s (31.508s including waiting). Image size: 1231028434 bytes.

openshift-ingress

kubelet

router-default-7dcf5569b5-xmvwz

Started

Started container router

openshift-ingress

kubelet

router-default-7dcf5569b5-xmvwz

Created

Created container: router

openshift-controller-manager

kubelet

controller-manager-6486d766f9-5b77h

Created

Created container: controller-manager

openshift-cloud-controller-manager-operator

kubelet

cluster-cloud-controller-manager-operator-7559f7c68c-xhq29

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:112a03f2411f871cdaca5f20daef71024dac710113d5f30897117a5a02f6b6f5"

openshift-monitoring

kubelet

prometheus-operator-admission-webhook-69c6b55594-tkc2j

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:97c7a635130c574a2c501091bb44f17cd92e05e29b5102e59578b5885d9bfec0" in 28.994s (28.994s including waiting). Image size: 444573129 bytes.
(x2)

openshift-service-ca-operator

kubelet

service-ca-operator-b865698dc-qzb2h

Created

Created container: service-ca-operator

openshift-marketplace

kubelet

redhat-operators-mgcb9

Pulled

Successfully pulled image "registry.redhat.io/redhat/redhat-operator-index:v4.18" in 30.365s (30.365s including waiting). Image size: 1746912226 bytes.
(x2)

openshift-service-ca-operator

kubelet

service-ca-operator-b865698dc-qzb2h

Started

Started container service-ca-operator

openshift-monitoring

kubelet

prometheus-operator-admission-webhook-69c6b55594-tkc2j

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:97c7a635130c574a2c501091bb44f17cd92e05e29b5102e59578b5885d9bfec0" in 28.994s (28.994s including waiting). Image size: 444573129 bytes.

openshift-controller-manager

kubelet

controller-manager-6486d766f9-5b77h

Started

Started container controller-manager

openshift-cluster-machine-approver

kubelet

machine-approver-6cb57bb5db-mt758

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-controller-manager

kubelet

controller-manager-6486d766f9-5b77h

Killing

Stopping container controller-manager

openshift-marketplace

kubelet

community-operators-zspn5

Pulled

Successfully pulled image "registry.redhat.io/redhat/community-operator-index:v4.18" in 26.353s (26.353s including waiting). Image size: 1223740630 bytes.

openshift-service-ca-operator

kubelet

service-ca-operator-b865698dc-qzb2h

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:812819a9d712b9e345ef5f1404b242c281e2518ad724baebc393ec0fd3b3d263" already present on machine

openshift-marketplace

kubelet

certified-operators-86g9n

Started

Started container extract-content

openshift-marketplace

kubelet

redhat-operators-mgcb9

Started

Started container extract-content

openshift-marketplace

kubelet

certified-operators-86g9n

Created

Created container: extract-content

openshift-marketplace

kubelet

redhat-marketplace-srjqw

Created

Created container: extract-content

openshift-marketplace

kubelet

redhat-marketplace-srjqw

Started

Started container extract-content

openshift-marketplace

kubelet

certified-operators-86g9n

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b00c42562d477ef44d51f35950253a26d7debc7de86e53270831aafef5795c1"

openshift-kube-apiserver-operator

kubelet

kube-apiserver-operator-8b68b9d9b-c6vkz

Started

Started container kube-apiserver-operator

openshift-marketplace

kubelet

redhat-marketplace-srjqw

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b00c42562d477ef44d51f35950253a26d7debc7de86e53270831aafef5795c1"

openshift-kube-apiserver-operator

kubelet

kube-apiserver-operator-8b68b9d9b-c6vkz

Created

Created container: kube-apiserver-operator

openshift-monitoring

kubelet

prometheus-operator-admission-webhook-69c6b55594-tkc2j

Started

Started container prometheus-operator-admission-webhook

openshift-marketplace

kubelet

community-operators-zspn5

Created

Created container: extract-content

openshift-marketplace

kubelet

community-operators-zspn5

Started

Started container extract-content

openshift-marketplace

kubelet

community-operators-zspn5

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b00c42562d477ef44d51f35950253a26d7debc7de86e53270831aafef5795c1"

openshift-monitoring

kubelet

prometheus-operator-admission-webhook-69c6b55594-tkc2j

Created

Created container: prometheus-operator-admission-webhook

openshift-monitoring

kubelet

prometheus-operator-admission-webhook-69c6b55594-tkc2j

Started

Started container prometheus-operator-admission-webhook

openshift-cluster-machine-approver

kubelet

machine-approver-6cb57bb5db-mt758

Created

Created container: kube-rbac-proxy

openshift-marketplace

kubelet

redhat-operators-mgcb9

Created

Created container: extract-content

openshift-cluster-machine-approver

kubelet

machine-approver-6cb57bb5db-mt758

Started

Started container kube-rbac-proxy

openshift-cluster-machine-approver

kubelet

machine-approver-6cb57bb5db-mt758

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:cdd28dfe7132e19af9f013f72cf120d970bc31b6b74693af262f8d2e82a096e1"

openshift-monitoring

kubelet

prometheus-operator-admission-webhook-69c6b55594-tkc2j

Created

Created container: prometheus-operator-admission-webhook

openshift-kube-scheduler

kubelet

installer-2-master-0

Started

Started container installer

openshift-kube-scheduler

kubelet

installer-2-master-0

Created

Created container: installer

openshift-controller-manager-operator

kubelet

openshift-controller-manager-operator-8c94f4649-p7pt6

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:313d1d8ca85e65236a59f058a3316c49436dde691b3a3930d5bc5e3b4b8c8a71" already present on machine
(x2)

openshift-controller-manager-operator

kubelet

openshift-controller-manager-operator-8c94f4649-p7pt6

Created

Created container: openshift-controller-manager-operator

openshift-marketplace

kubelet

redhat-operators-mgcb9

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b00c42562d477ef44d51f35950253a26d7debc7de86e53270831aafef5795c1"
(x2)

openshift-controller-manager-operator

kubelet

openshift-controller-manager-operator-8c94f4649-p7pt6

Started

Started container openshift-controller-manager-operator

openshift-cloud-controller-manager-operator

kubelet

cluster-cloud-controller-manager-operator-7559f7c68c-xhq29

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:112a03f2411f871cdaca5f20daef71024dac710113d5f30897117a5a02f6b6f5" in 6.552s (6.552s including waiting). Image size: 557428271 bytes.

openshift-cluster-machine-approver

kubelet

machine-approver-6cb57bb5db-mt758

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:cdd28dfe7132e19af9f013f72cf120d970bc31b6b74693af262f8d2e82a096e1" in 5.763s (5.763s including waiting). Image size: 467235741 bytes.

openshift-marketplace

kubelet

certified-operators-86g9n

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b00c42562d477ef44d51f35950253a26d7debc7de86e53270831aafef5795c1" in 5.164s (5.164s including waiting). Image size: 918289953 bytes.

openshift-marketplace

kubelet

redhat-marketplace-srjqw

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b00c42562d477ef44d51f35950253a26d7debc7de86e53270831aafef5795c1" in 5.082s (5.082s including waiting). Image size: 918289953 bytes.

openshift-marketplace

kubelet

community-operators-zspn5

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b00c42562d477ef44d51f35950253a26d7debc7de86e53270831aafef5795c1" in 4.795s (4.795s including waiting). Image size: 918289953 bytes.

openshift-marketplace

kubelet

certified-operators-86g9n

Created

Created container: registry-server

openshift-cluster-machine-approver

kubelet

machine-approver-6cb57bb5db-mt758

Created

Created container: machine-approver-controller

openshift-cloud-controller-manager-operator

kubelet

cluster-cloud-controller-manager-operator-7559f7c68c-xhq29

Created

Created container: cluster-cloud-controller-manager

openshift-marketplace

kubelet

redhat-marketplace-srjqw

Started

Started container registry-server

openshift-marketplace

kubelet

community-operators-zspn5

Created

Created container: registry-server

openshift-marketplace

kubelet

redhat-marketplace-srjqw

Created

Created container: registry-server

openshift-marketplace

kubelet

community-operators-zspn5

Started

Started container registry-server

openshift-marketplace

kubelet

redhat-operators-mgcb9

Created

Created container: registry-server

openshift-marketplace

kubelet

certified-operators-86g9n

Started

Started container registry-server

openshift-cloud-controller-manager-operator

kubelet

cluster-cloud-controller-manager-operator-7559f7c68c-xhq29

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:112a03f2411f871cdaca5f20daef71024dac710113d5f30897117a5a02f6b6f5" already present on machine

openshift-marketplace

kubelet

redhat-operators-mgcb9

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b00c42562d477ef44d51f35950253a26d7debc7de86e53270831aafef5795c1" in 5.774s (5.774s including waiting). Image size: 918289953 bytes.

openshift-cluster-machine-approver

kubelet

machine-approver-6cb57bb5db-mt758

Started

Started container machine-approver-controller

openshift-cloud-controller-manager-operator

kubelet

cluster-cloud-controller-manager-operator-7559f7c68c-xhq29

Started

Started container cluster-cloud-controller-manager

openshift-cloud-controller-manager-operator

kubelet

cluster-cloud-controller-manager-operator-7559f7c68c-xhq29

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-marketplace

kubelet

redhat-operators-mgcb9

Started

Started container registry-server

openshift-cloud-controller-manager-operator

kubelet

cluster-cloud-controller-manager-operator-7559f7c68c-xhq29

Created

Created container: kube-rbac-proxy

openshift-cloud-controller-manager-operator

kubelet

cluster-cloud-controller-manager-operator-7559f7c68c-xhq29

Started

Started container config-sync-controllers

openshift-cloud-controller-manager-operator

kubelet

cluster-cloud-controller-manager-operator-7559f7c68c-xhq29

Started

Started container kube-rbac-proxy

openshift-etcd

kubelet

etcd-master-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e29dc9f042f2d0471171a0611070886cb2f7c57338ab7f112613417bcd33b278" already present on machine

openshift-cloud-controller-manager-operator

kubelet

cluster-cloud-controller-manager-operator-7559f7c68c-xhq29

Created

Created container: config-sync-controllers

openshift-marketplace

kubelet

certified-operators-86g9n

Unhealthy

Startup probe failed: timeout: failed to connect service ":50051" within 1s

openshift-etcd

kubelet

etcd-master-0

Created

Created container: setup
(x10)

openshift-ingress

kubelet

router-default-7dcf5569b5-xmvwz

Unhealthy

Startup probe failed: HTTP probe failed with statuscode: 500

openshift-etcd

kubelet

etcd-master-0

Started

Started container setup
(x11)

openshift-ingress

kubelet

router-default-7dcf5569b5-xmvwz

ProbeError

Startup probe error: HTTP probe failed with statuscode: 500 body: [-]backend-http failed: reason withheld [-]has-synced failed: reason withheld [+]process-running ok healthz check failed

kube-system

kubelet

bootstrap-kube-scheduler-master-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b23c544d3894e5b31f66a18c554f03b0d29f92c2000c46b57b1c96da7ec25db9" already present on machine

kube-system

kubelet

bootstrap-kube-scheduler-master-0

Created

Created container: kube-scheduler

kube-system

kubelet

bootstrap-kube-scheduler-master-0

Started

Started container kube-scheduler

openshift-etcd

kubelet

etcd-master-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e29dc9f042f2d0471171a0611070886cb2f7c57338ab7f112613417bcd33b278" already present on machine

openshift-etcd

kubelet

etcd-master-0

Created

Created container: etcd-ensure-env-vars

openshift-etcd

kubelet

etcd-master-0

Started

Started container etcd-ensure-env-vars
(x3)

kube-system

kubelet

bootstrap-kube-controller-manager-master-0

Unhealthy

Startup probe failed: Get "https://192.168.32.10:10257/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)

openshift-etcd

kubelet

etcd-master-0

Started

Started container etcd-resources-copy

openshift-etcd

kubelet

etcd-master-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e29dc9f042f2d0471171a0611070886cb2f7c57338ab7f112613417bcd33b278" already present on machine

openshift-etcd

kubelet

etcd-master-0

Created

Created container: etcd-resources-copy
(x3)

openshift-authentication-operator

kubelet

authentication-operator-5885bfd7f4-62zrx

Unhealthy

Liveness probe failed: Get "https://10.128.0.25:8443/healthz": dial tcp 10.128.0.25:8443: connect: connection refused
(x3)

openshift-authentication-operator

kubelet

authentication-operator-5885bfd7f4-62zrx

ProbeError

Liveness probe error: Get "https://10.128.0.25:8443/healthz": dial tcp 10.128.0.25:8443: connect: connection refused body:

openshift-etcd

kubelet

etcd-master-0

Started

Started container etcd

openshift-etcd

kubelet

etcd-master-0

Started

Started container etcd-metrics

openshift-etcd

kubelet

etcd-master-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d5a971d5889f167cfe61a64c366424b87c17a6dc141ffcc43406cdcbb50cae2a" already present on machine

openshift-etcd

kubelet

etcd-master-0

Created

Created container: etcdctl

openshift-etcd

kubelet

etcd-master-0

Created

Created container: etcd-metrics

openshift-etcd

kubelet

etcd-master-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e29dc9f042f2d0471171a0611070886cb2f7c57338ab7f112613417bcd33b278" already present on machine

openshift-etcd

kubelet

etcd-master-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e29dc9f042f2d0471171a0611070886cb2f7c57338ab7f112613417bcd33b278" already present on machine

openshift-etcd

kubelet

etcd-master-0

Started

Started container etcdctl

openshift-etcd

kubelet

etcd-master-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e29dc9f042f2d0471171a0611070886cb2f7c57338ab7f112613417bcd33b278" already present on machine

openshift-etcd

kubelet

etcd-master-0

Created

Created container: etcd

openshift-etcd

kubelet

etcd-master-0

Created

Created container: etcd-readyz

openshift-etcd

kubelet

etcd-master-0

Started

Started container etcd-readyz

openshift-etcd

kubelet

etcd-master-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d5a971d5889f167cfe61a64c366424b87c17a6dc141ffcc43406cdcbb50cae2a" already present on machine

openshift-etcd

kubelet

etcd-master-0

Created

Created container: etcd-rev

openshift-etcd

kubelet

etcd-master-0

Started

Started container etcd-rev

openshift-etcd-operator

kubelet

etcd-operator-8544cbcf9c-brhw4

Unhealthy

Liveness probe failed: Get "https://10.128.0.28:8443/healthz": dial tcp 10.128.0.28:8443: connect: connection refused

openshift-etcd-operator

kubelet

etcd-operator-8544cbcf9c-brhw4

ProbeError

Liveness probe error: Get "https://10.128.0.28:8443/healthz": dial tcp 10.128.0.28:8443: connect: connection refused body:
(x2)

openshift-cluster-samples-operator

multus

cluster-samples-operator-85f7577d78-mwfgx

AddedInterface

Add eth0 [10.128.0.64/23] from ovn-kubernetes

openshift-machine-api

kubelet

control-plane-machine-set-operator-6f97756bc8-7t5qv

FailedCreatePodSandBox

Failed to create pod sandbox: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_control-plane-machine-set-operator-6f97756bc8-7t5qv_openshift-machine-api_e9e912c1-a2d4-4d78-98c1-ea3b232ddd7e_0(9dbc34ec166af95a7853428b41cee350fe85c8c54be4aa6e328035aa02204cfc): error adding pod openshift-machine-api_control-plane-machine-set-operator-6f97756bc8-7t5qv to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"9dbc34ec166af95a7853428b41cee350fe85c8c54be4aa6e328035aa02204cfc" Netns:"/var/run/netns/528b4df4-56d0-4831-a96c-32b3710c2f88" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-machine-api;K8S_POD_NAME=control-plane-machine-set-operator-6f97756bc8-7t5qv;K8S_POD_INFRA_CONTAINER_ID=9dbc34ec166af95a7853428b41cee350fe85c8c54be4aa6e328035aa02204cfc;K8S_POD_UID=e9e912c1-a2d4-4d78-98c1-ea3b232ddd7e" Path:"" ERRORED: error configuring pod [openshift-machine-api/control-plane-machine-set-operator-6f97756bc8-7t5qv] networking: Multus: [openshift-machine-api/control-plane-machine-set-operator-6f97756bc8-7t5qv/e9e912c1-a2d4-4d78-98c1-ea3b232ddd7e]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod control-plane-machine-set-operator-6f97756bc8-7t5qv in out of cluster comm: SetNetworkStatus: failed to update the pod control-plane-machine-set-operator-6f97756bc8-7t5qv in out of cluster comm: status update failed for pod /: Get "https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-machine-api/pods/control-plane-machine-set-operator-6f97756bc8-7t5qv?timeout=1m0s": net/http: request canceled (Client.Timeout exceeded while awaiting headers) ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"}

openshift-cluster-samples-operator

kubelet

cluster-samples-operator-85f7577d78-mwfgx

FailedCreatePodSandBox

Failed to create pod sandbox: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_cluster-samples-operator-85f7577d78-mwfgx_openshift-cluster-samples-operator_e3bf8eaf-5f6c-41a6-aaeb-6c921d789466_0(ca850c7655dc9ec8e9a1fbbcce67531da5122d6e0d326f12b4ff2fa81952ebba): error adding pod openshift-cluster-samples-operator_cluster-samples-operator-85f7577d78-mwfgx to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"ca850c7655dc9ec8e9a1fbbcce67531da5122d6e0d326f12b4ff2fa81952ebba" Netns:"/var/run/netns/87752d08-ed34-4f8b-a26a-f4ed4ca16e8e" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-cluster-samples-operator;K8S_POD_NAME=cluster-samples-operator-85f7577d78-mwfgx;K8S_POD_INFRA_CONTAINER_ID=ca850c7655dc9ec8e9a1fbbcce67531da5122d6e0d326f12b4ff2fa81952ebba;K8S_POD_UID=e3bf8eaf-5f6c-41a6-aaeb-6c921d789466" Path:"" ERRORED: error configuring pod [openshift-cluster-samples-operator/cluster-samples-operator-85f7577d78-mwfgx] networking: Multus: [openshift-cluster-samples-operator/cluster-samples-operator-85f7577d78-mwfgx/e3bf8eaf-5f6c-41a6-aaeb-6c921d789466]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod cluster-samples-operator-85f7577d78-mwfgx in out of cluster comm: SetNetworkStatus: failed to update the pod cluster-samples-operator-85f7577d78-mwfgx in out of cluster comm: status update failed for pod /: Get "https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-cluster-samples-operator/pods/cluster-samples-operator-85f7577d78-mwfgx?timeout=1m0s": net/http: request canceled (Client.Timeout exceeded while awaiting headers) ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"}

openshift-insights

kubelet

insights-operator-68bf6ff9d6-mvfn5

FailedCreatePodSandBox

Failed to create pod sandbox: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_insights-operator-68bf6ff9d6-mvfn5_openshift-insights_f2217de0-7805-4f5f-8ea5-93b81b7e0236_0(6e679c83d9e07e8539cd4ee0b9a66261e4e8939cdf7af15cbfbb7e5cfaad82f9): error adding pod openshift-insights_insights-operator-68bf6ff9d6-mvfn5 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"6e679c83d9e07e8539cd4ee0b9a66261e4e8939cdf7af15cbfbb7e5cfaad82f9" Netns:"/var/run/netns/06f0a4c8-4a3a-400f-a230-8a6ef7f51225" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-insights;K8S_POD_NAME=insights-operator-68bf6ff9d6-mvfn5;K8S_POD_INFRA_CONTAINER_ID=6e679c83d9e07e8539cd4ee0b9a66261e4e8939cdf7af15cbfbb7e5cfaad82f9;K8S_POD_UID=f2217de0-7805-4f5f-8ea5-93b81b7e0236" Path:"" ERRORED: error configuring pod [openshift-insights/insights-operator-68bf6ff9d6-mvfn5] networking: Multus: [openshift-insights/insights-operator-68bf6ff9d6-mvfn5/f2217de0-7805-4f5f-8ea5-93b81b7e0236]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod insights-operator-68bf6ff9d6-mvfn5 in out of cluster comm: SetNetworkStatus: failed to update the pod insights-operator-68bf6ff9d6-mvfn5 in out of cluster comm: status update failed for pod /: Get "https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-insights/pods/insights-operator-68bf6ff9d6-mvfn5?timeout=1m0s": net/http: request canceled (Client.Timeout exceeded while awaiting headers) ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"}

openshift-cloud-credential-operator

kubelet

cloud-credential-operator-744f9dbf77-r4qvh

FailedCreatePodSandBox

Failed to create pod sandbox: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_cloud-credential-operator-744f9dbf77-r4qvh_openshift-cloud-credential-operator_469183dd-dc54-467d-82a1-611132ae8ec4_0(41dcb504e61c8c9db109ecb73d18ade9019dd987172fffe0bc5955fea5e2bd2a): error adding pod openshift-cloud-credential-operator_cloud-credential-operator-744f9dbf77-r4qvh to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"41dcb504e61c8c9db109ecb73d18ade9019dd987172fffe0bc5955fea5e2bd2a" Netns:"/var/run/netns/9ad75b80-3551-465a-a3ca-a85581db5fef" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-cloud-credential-operator;K8S_POD_NAME=cloud-credential-operator-744f9dbf77-r4qvh;K8S_POD_INFRA_CONTAINER_ID=41dcb504e61c8c9db109ecb73d18ade9019dd987172fffe0bc5955fea5e2bd2a;K8S_POD_UID=469183dd-dc54-467d-82a1-611132ae8ec4" Path:"" ERRORED: error configuring pod [openshift-cloud-credential-operator/cloud-credential-operator-744f9dbf77-r4qvh] networking: Multus: [openshift-cloud-credential-operator/cloud-credential-operator-744f9dbf77-r4qvh/469183dd-dc54-467d-82a1-611132ae8ec4]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod cloud-credential-operator-744f9dbf77-r4qvh in out of cluster comm: SetNetworkStatus: failed to update the pod cloud-credential-operator-744f9dbf77-r4qvh in out of cluster comm: status update failed for pod /: Get "https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-cloud-credential-operator/pods/cloud-credential-operator-744f9dbf77-r4qvh?timeout=1m0s": net/http: request canceled (Client.Timeout exceeded while awaiting headers) ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"}

openshift-machine-api

kubelet

machine-api-operator-6fbb6cf6f9-n8tnn

FailedCreatePodSandBox

Failed to create pod sandbox: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_machine-api-operator-6fbb6cf6f9-n8tnn_openshift-machine-api_c7f5e6cd-e093-409a-8758-d3db7a7eb32c_0(a48640ea906a4e679db30d8ed63617d252b5836d193a04691bdbafc25f9a7a5d): error adding pod openshift-machine-api_machine-api-operator-6fbb6cf6f9-n8tnn to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"a48640ea906a4e679db30d8ed63617d252b5836d193a04691bdbafc25f9a7a5d" Netns:"/var/run/netns/60ecd76f-554b-4846-9272-ed180fe57e60" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-machine-api;K8S_POD_NAME=machine-api-operator-6fbb6cf6f9-n8tnn;K8S_POD_INFRA_CONTAINER_ID=a48640ea906a4e679db30d8ed63617d252b5836d193a04691bdbafc25f9a7a5d;K8S_POD_UID=c7f5e6cd-e093-409a-8758-d3db7a7eb32c" Path:"" ERRORED: error configuring pod [openshift-machine-api/machine-api-operator-6fbb6cf6f9-n8tnn] networking: Multus: [openshift-machine-api/machine-api-operator-6fbb6cf6f9-n8tnn/c7f5e6cd-e093-409a-8758-d3db7a7eb32c]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod machine-api-operator-6fbb6cf6f9-n8tnn in out of cluster comm: SetNetworkStatus: failed to update the pod machine-api-operator-6fbb6cf6f9-n8tnn in out of cluster comm: status update failed for pod /: Get "https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-machine-api/pods/machine-api-operator-6fbb6cf6f9-n8tnn?timeout=1m0s": context deadline exceeded (Client.Timeout exceeded while awaiting headers) ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"}
(x2)

openshift-machine-api

multus

cluster-autoscaler-operator-866dc4744-xwxg7

AddedInterface

Add eth0 [10.128.0.65/23] from ovn-kubernetes

openshift-machine-api

kubelet

control-plane-machine-set-operator-6f97756bc8-7t5qv

FailedCreatePodSandBox

Failed to create pod sandbox: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_control-plane-machine-set-operator-6f97756bc8-7t5qv_openshift-machine-api_e9e912c1-a2d4-4d78-98c1-ea3b232ddd7e_0(9dbc34ec166af95a7853428b41cee350fe85c8c54be4aa6e328035aa02204cfc): error adding pod openshift-machine-api_control-plane-machine-set-operator-6f97756bc8-7t5qv to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"9dbc34ec166af95a7853428b41cee350fe85c8c54be4aa6e328035aa02204cfc" Netns:"/var/run/netns/528b4df4-56d0-4831-a96c-32b3710c2f88" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-machine-api;K8S_POD_NAME=control-plane-machine-set-operator-6f97756bc8-7t5qv;K8S_POD_INFRA_CONTAINER_ID=9dbc34ec166af95a7853428b41cee350fe85c8c54be4aa6e328035aa02204cfc;K8S_POD_UID=e9e912c1-a2d4-4d78-98c1-ea3b232ddd7e" Path:"" ERRORED: error configuring pod [openshift-machine-api/control-plane-machine-set-operator-6f97756bc8-7t5qv] networking: Multus: [openshift-machine-api/control-plane-machine-set-operator-6f97756bc8-7t5qv/e9e912c1-a2d4-4d78-98c1-ea3b232ddd7e]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod control-plane-machine-set-operator-6f97756bc8-7t5qv in out of cluster comm: SetNetworkStatus: failed to update the pod control-plane-machine-set-operator-6f97756bc8-7t5qv in out of cluster comm: status update failed for pod /: Get "https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-machine-api/pods/control-plane-machine-set-operator-6f97756bc8-7t5qv?timeout=1m0s": net/http: request canceled (Client.Timeout exceeded while awaiting headers) ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"}

openshift-kube-controller-manager

kubelet

installer-2-master-0

FailedCreatePodSandBox

Failed to create pod sandbox: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_installer-2-master-0_openshift-kube-controller-manager_8b1c7a56-5d00-468a-bb8d-dbaf8e854951_0(c7f364ac5ac343cc49d7eb09efc4fc8a8e9a655adeeba4e984a9c0ca3e7b9153): error adding pod openshift-kube-controller-manager_installer-2-master-0 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"c7f364ac5ac343cc49d7eb09efc4fc8a8e9a655adeeba4e984a9c0ca3e7b9153" Netns:"/var/run/netns/60c6990c-6415-4a0b-ba8b-e62820d5aea5" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-kube-controller-manager;K8S_POD_NAME=installer-2-master-0;K8S_POD_INFRA_CONTAINER_ID=c7f364ac5ac343cc49d7eb09efc4fc8a8e9a655adeeba4e984a9c0ca3e7b9153;K8S_POD_UID=8b1c7a56-5d00-468a-bb8d-dbaf8e854951" Path:"" ERRORED: error configuring pod [openshift-kube-controller-manager/installer-2-master-0] networking: Multus: [openshift-kube-controller-manager/installer-2-master-0/8b1c7a56-5d00-468a-bb8d-dbaf8e854951]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod installer-2-master-0 in out of cluster comm: SetNetworkStatus: failed to update the pod installer-2-master-0 in out of cluster comm: status update failed for pod /: Get "https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/installer-2-master-0?timeout=1m0s": context deadline exceeded ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"}

openshift-machine-api

kubelet

cluster-autoscaler-operator-866dc4744-xwxg7

FailedCreatePodSandBox

Failed to create pod sandbox: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_cluster-autoscaler-operator-866dc4744-xwxg7_openshift-machine-api_21bebade-17fa-444e-92a9-eea53d6cd673_0(eea5c777fab069276340fb8b8e8bc932a48b37c137af14866a4f948ceb435c50): error adding pod openshift-machine-api_cluster-autoscaler-operator-866dc4744-xwxg7 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"eea5c777fab069276340fb8b8e8bc932a48b37c137af14866a4f948ceb435c50" Netns:"/var/run/netns/cf3886d6-d600-4c57-baf9-25c23fb5edd0" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-machine-api;K8S_POD_NAME=cluster-autoscaler-operator-866dc4744-xwxg7;K8S_POD_INFRA_CONTAINER_ID=eea5c777fab069276340fb8b8e8bc932a48b37c137af14866a4f948ceb435c50;K8S_POD_UID=21bebade-17fa-444e-92a9-eea53d6cd673" Path:"" ERRORED: error configuring pod [openshift-machine-api/cluster-autoscaler-operator-866dc4744-xwxg7] networking: Multus: [openshift-machine-api/cluster-autoscaler-operator-866dc4744-xwxg7/21bebade-17fa-444e-92a9-eea53d6cd673]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod cluster-autoscaler-operator-866dc4744-xwxg7 in out of cluster comm: SetNetworkStatus: failed to update the pod cluster-autoscaler-operator-866dc4744-xwxg7 in out of cluster comm: status update failed for pod /: Get "https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-machine-api/pods/cluster-autoscaler-operator-866dc4744-xwxg7?timeout=1m0s": context deadline exceeded ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"}
(x2)

openshift-machine-api

multus

cluster-autoscaler-operator-866dc4744-xwxg7

AddedInterface

Add eth0 [10.128.0.65/23] from ovn-kubernetes

openshift-machine-api

kubelet

cluster-autoscaler-operator-866dc4744-xwxg7

FailedCreatePodSandBox

Failed to create pod sandbox: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_cluster-autoscaler-operator-866dc4744-xwxg7_openshift-machine-api_21bebade-17fa-444e-92a9-eea53d6cd673_0(eea5c777fab069276340fb8b8e8bc932a48b37c137af14866a4f948ceb435c50): error adding pod openshift-machine-api_cluster-autoscaler-operator-866dc4744-xwxg7 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"eea5c777fab069276340fb8b8e8bc932a48b37c137af14866a4f948ceb435c50" Netns:"/var/run/netns/cf3886d6-d600-4c57-baf9-25c23fb5edd0" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-machine-api;K8S_POD_NAME=cluster-autoscaler-operator-866dc4744-xwxg7;K8S_POD_INFRA_CONTAINER_ID=eea5c777fab069276340fb8b8e8bc932a48b37c137af14866a4f948ceb435c50;K8S_POD_UID=21bebade-17fa-444e-92a9-eea53d6cd673" Path:"" ERRORED: error configuring pod [openshift-machine-api/cluster-autoscaler-operator-866dc4744-xwxg7] networking: Multus: [openshift-machine-api/cluster-autoscaler-operator-866dc4744-xwxg7/21bebade-17fa-444e-92a9-eea53d6cd673]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod cluster-autoscaler-operator-866dc4744-xwxg7 in out of cluster comm: SetNetworkStatus: failed to update the pod cluster-autoscaler-operator-866dc4744-xwxg7 in out of cluster comm: status update failed for pod /: Get "https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-machine-api/pods/cluster-autoscaler-operator-866dc4744-xwxg7?timeout=1m0s": context deadline exceeded ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"}

openshift-machine-api

kubelet

machine-api-operator-6fbb6cf6f9-n8tnn

FailedCreatePodSandBox

Failed to create pod sandbox: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_machine-api-operator-6fbb6cf6f9-n8tnn_openshift-machine-api_c7f5e6cd-e093-409a-8758-d3db7a7eb32c_0(a48640ea906a4e679db30d8ed63617d252b5836d193a04691bdbafc25f9a7a5d): error adding pod openshift-machine-api_machine-api-operator-6fbb6cf6f9-n8tnn to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"a48640ea906a4e679db30d8ed63617d252b5836d193a04691bdbafc25f9a7a5d" Netns:"/var/run/netns/60ecd76f-554b-4846-9272-ed180fe57e60" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-machine-api;K8S_POD_NAME=machine-api-operator-6fbb6cf6f9-n8tnn;K8S_POD_INFRA_CONTAINER_ID=a48640ea906a4e679db30d8ed63617d252b5836d193a04691bdbafc25f9a7a5d;K8S_POD_UID=c7f5e6cd-e093-409a-8758-d3db7a7eb32c" Path:"" ERRORED: error configuring pod [openshift-machine-api/machine-api-operator-6fbb6cf6f9-n8tnn] networking: Multus: [openshift-machine-api/machine-api-operator-6fbb6cf6f9-n8tnn/c7f5e6cd-e093-409a-8758-d3db7a7eb32c]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod machine-api-operator-6fbb6cf6f9-n8tnn in out of cluster comm: SetNetworkStatus: failed to update the pod machine-api-operator-6fbb6cf6f9-n8tnn in out of cluster comm: status update failed for pod /: Get "https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-machine-api/pods/machine-api-operator-6fbb6cf6f9-n8tnn?timeout=1m0s": context deadline exceeded (Client.Timeout exceeded while awaiting headers) ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"}

openshift-network-node-identity

kubelet

network-node-identity-6t5vb

Started

Started container approver
(x2)

openshift-insights

multus

insights-operator-68bf6ff9d6-mvfn5

AddedInterface

Add eth0 [10.128.0.66/23] from ovn-kubernetes
(x2)

openshift-cloud-credential-operator

multus

cloud-credential-operator-744f9dbf77-r4qvh

AddedInterface

Add eth0 [10.128.0.63/23] from ovn-kubernetes
(x2)

openshift-kube-controller-manager

multus

installer-2-master-0

AddedInterface

Add eth0 [10.128.0.61/23] from ovn-kubernetes
(x2)

openshift-machine-api

multus

machine-api-operator-6fbb6cf6f9-n8tnn

AddedInterface

Add eth0 [10.128.0.67/23] from ovn-kubernetes

openshift-network-node-identity

kubelet

network-node-identity-6t5vb

Created

Created container: approver

openshift-network-node-identity

kubelet

network-node-identity-6t5vb

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:2abc1fd79e7781634ed5ed9e8f2b98b9094ea51f40ac3a773c5e5224607bf3d7" already present on machine
(x2)

openshift-machine-api

multus

machine-api-operator-6fbb6cf6f9-n8tnn

AddedInterface

Add eth0 [10.128.0.67/23] from ovn-kubernetes

openshift-cloud-controller-manager-operator

master-0_c6320a43-08b8-478c-8885-83a32042d3db

cluster-cloud-config-sync-leader

LeaderElection

master-0_c6320a43-08b8-478c-8885-83a32042d3db became leader

openshift-cluster-machine-approver

master-0_090054cf-9390-4fb3-bbc7-9ac8cd1c86e5

cluster-machine-approver-leader

LeaderElection

master-0_090054cf-9390-4fb3-bbc7-9ac8cd1c86e5 became leader

openshift-machine-api

kubelet

cluster-autoscaler-operator-866dc4744-xwxg7

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-machine-api

kubelet

cluster-autoscaler-operator-866dc4744-xwxg7

Created

Created container: kube-rbac-proxy
(x2)

openshift-apiserver-operator

kubelet

openshift-apiserver-operator-d65958b8-th2vj

Created

Created container: openshift-apiserver-operator

kube-system

kubelet

bootstrap-kube-controller-manager-master-0

Killing

Container kube-controller-manager failed startup probe, will be restarted
(x2)

openshift-apiserver-operator

kubelet

openshift-apiserver-operator-d65958b8-th2vj

Started

Started container openshift-apiserver-operator

openshift-machine-api

kubelet

cluster-autoscaler-operator-866dc4744-xwxg7

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-cluster-samples-operator

kubelet

cluster-samples-operator-85f7577d78-mwfgx

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ab745a9e15dadc862548ceb5740b8f5d02075232760c6715d82b4c3b70eddca9"

openshift-network-operator

kubelet

network-operator-7bd846bfc4-mt454

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ec8fd46dfb35ed10e8f98933166f69ce579c2f35b8db03d21e4c34fc544553e4" already present on machine

openshift-cloud-credential-operator

kubelet

cloud-credential-operator-744f9dbf77-r4qvh

Created

Created container: kube-rbac-proxy

openshift-insights

kubelet

insights-operator-68bf6ff9d6-mvfn5

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1973d56a1097a48ea0ebf2c4dbae1ed86fa67bb0116f4962f7720d48aa337d27"

openshift-kube-controller-manager

kubelet

installer-2-master-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c032f87ae61d6f757ff3ce52620a70a43516591987731f25da77aba152f17458" already present on machine

openshift-cloud-controller-manager-operator

master-0_a5a49fc9-1301-4ccc-80e0-fdcecac8bc2d

cluster-cloud-controller-manager-leader

LeaderElection

master-0_a5a49fc9-1301-4ccc-80e0-fdcecac8bc2d became leader

openshift-machine-api

kubelet

cluster-autoscaler-operator-866dc4744-xwxg7

Started

Started container kube-rbac-proxy

openshift-machine-api

kubelet

machine-api-operator-6fbb6cf6f9-n8tnn

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-cloud-credential-operator

kubelet

cloud-credential-operator-744f9dbf77-r4qvh

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine
(x2)

kube-system

kubelet

bootstrap-kube-controller-manager-master-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b23c544d3894e5b31f66a18c554f03b0d29f92c2000c46b57b1c96da7ec25db9" already present on machine

openshift-machine-api

kubelet

cluster-autoscaler-operator-866dc4744-xwxg7

Started

Started container kube-rbac-proxy

openshift-machine-api

kubelet

cluster-autoscaler-operator-866dc4744-xwxg7

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ce8e3088493b4a72dd766b3b5b4ccb83b7d72d514bbf64063a913dfe961273db"

openshift-etcd-operator

kubelet

etcd-operator-8544cbcf9c-brhw4

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d5a971d5889f167cfe61a64c366424b87c17a6dc141ffcc43406cdcbb50cae2a" already present on machine

openshift-authentication-operator

kubelet

authentication-operator-5885bfd7f4-62zrx

Killing

Container authentication-operator failed liveness probe, will be restarted

openshift-ingress-operator

kubelet

ingress-operator-66b84d69b-gzg9m

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:77fff570657d2fa0bfb709b2c8b6665bae0bf90a2be981d8dbca56c674715098" already present on machine
(x2)

openshift-ingress-operator

kubelet

ingress-operator-66b84d69b-gzg9m

Created

Created container: ingress-operator

openshift-kube-controller-manager-operator

kubelet

kube-controller-manager-operator-ff989d6cc-qcpb4

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c032f87ae61d6f757ff3ce52620a70a43516591987731f25da77aba152f17458" already present on machine

openshift-machine-api

kubelet

cluster-autoscaler-operator-866dc4744-xwxg7

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ce8e3088493b4a72dd766b3b5b4ccb83b7d72d514bbf64063a913dfe961273db"

openshift-kube-storage-version-migrator-operator

kubelet

kube-storage-version-migrator-operator-6bb5bfb6fd-64dg5

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5c983016b9ceed0fca1f51bd49c2653243c7e5af91cbf2f478b091db6e028252" already present on machine
(x2)

openshift-kube-storage-version-migrator-operator

kubelet

kube-storage-version-migrator-operator-6bb5bfb6fd-64dg5

Started

Started container kube-storage-version-migrator-operator

openshift-apiserver-operator

kubelet

openshift-apiserver-operator-d65958b8-th2vj

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1f23bac0a2a6cfd638e4af679dc787a8790d99c391f6e2ade8087dc477ff765e" already present on machine

openshift-machine-api

kubelet

cluster-autoscaler-operator-866dc4744-xwxg7

Created

Created container: kube-rbac-proxy

openshift-cloud-controller-manager

cloud-controller-manager-operator

openshift-cloud-controller-manager

FeatureGatesInitialized

FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{"AWSEFSDriverVolumeMetrics", "AdminNetworkPolicy", "AlibabaPlatform", "AzureWorkloadIdentity", "BareMetalLoadBalancer", "BuildCSIVolumes", "ChunkSizeMiB", "CloudDualStackNodeIPs", "DisableKubeletCloudCredentialProviders", "GCPLabelsTags", "HardwareSpeed", "IngressControllerLBSubnetsAWS", "KMSv1", "ManagedBootImages", "ManagedBootImagesAWS", "MultiArchInstallAWS", "MultiArchInstallGCP", "NetworkDiagnosticsConfig", "NetworkLiveMigration", "NetworkSegmentation", "NewOLM", "NodeDisruptionPolicy", "OnClusterBuild", "PersistentIPsForVirtualization", "PrivateHostedZoneAWS", "SetEIPForNLBIngressController", "StreamingCollectionEncodingToJSON", "StreamingCollectionEncodingToProtobuf", "VSphereControlPlaneMachineSet", "VSphereDriverConfiguration", "VSphereMultiVCenters", "VSphereStaticIPs", "ValidatingAdmissionPolicy"}, Disabled:[]v1.FeatureGateName{"AWSClusterHostedDNS", "AdditionalRoutingCapabilities", "AutomatedEtcdBackup", "BootcNodeManagement", "CSIDriverSharedResource", "ClusterAPIInstall", "ClusterAPIInstallIBMCloud", "ClusterMonitoringConfig", "ConsolePluginContentSecurityPolicy", "DNSNameResolver", "DynamicResourceAllocation", "EtcdBackendQuota", "EventedPLEG", "Example", "ExternalOIDC", "ExternalOIDCWithUIDAndExtraClaimMappings", "GCPClusterHostedDNS", "GatewayAPI", "ImageStreamImportMode", "IngressControllerDynamicConfigurationManager", "InsightsConfig", "InsightsConfigAPI", "InsightsOnDemandDataGather", "InsightsRuntimeExtractor", "MachineAPIMigration", "MachineAPIOperatorDisableMachineHealthCheckController", "MachineAPIProviderOpenStack", "MachineConfigNodes", "MaxUnavailableStatefulSet", "MetricsCollectionProfiles", "MinimumKubeletVersion", "MixedCPUsAllocation", "MultiArchInstallAzure", "NodeSwap", "NutanixMultiSubnets", "OVNObservability", "OpenShiftPodSecurityAdmission", "PinnedImages", "PlatformOperators", "ProcMountType", "RouteAdvertisements", "RouteExternalCertificate", "ServiceAccountTokenNodeBinding", "SignatureStores", "SigstoreImageVerification", "TranslateStreamCloseWebsocketRequests", "UpgradeStatus", "UserNamespacesPodSecurityStandards", "UserNamespacesSupport", "VSphereMultiNetworks", "VolumeAttributesClass", "VolumeGroupSnapshot"}}

openshift-machine-api

kubelet

machine-api-operator-6fbb6cf6f9-n8tnn

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine
(x2)

openshift-kube-storage-version-migrator-operator

kubelet

kube-storage-version-migrator-operator-6bb5bfb6fd-64dg5

Created

Created container: kube-storage-version-migrator-operator

openshift-ingress-operator

cluster-ingress-operator

ingress-operator

FeatureGatesInitialized

FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{"AWSEFSDriverVolumeMetrics", "AdminNetworkPolicy", "AlibabaPlatform", "AzureWorkloadIdentity", "BareMetalLoadBalancer", "BuildCSIVolumes", "ChunkSizeMiB", "CloudDualStackNodeIPs", "DisableKubeletCloudCredentialProviders", "GCPLabelsTags", "HardwareSpeed", "IngressControllerLBSubnetsAWS", "KMSv1", "ManagedBootImages", "ManagedBootImagesAWS", "MultiArchInstallAWS", "MultiArchInstallGCP", "NetworkDiagnosticsConfig", "NetworkLiveMigration", "NetworkSegmentation", "NewOLM", "NodeDisruptionPolicy", "OnClusterBuild", "PersistentIPsForVirtualization", "PrivateHostedZoneAWS", "SetEIPForNLBIngressController", "StreamingCollectionEncodingToJSON", "StreamingCollectionEncodingToProtobuf", "VSphereControlPlaneMachineSet", "VSphereDriverConfiguration", "VSphereMultiVCenters", "VSphereStaticIPs", "ValidatingAdmissionPolicy"}, Disabled:[]v1.FeatureGateName{"AWSClusterHostedDNS", "AdditionalRoutingCapabilities", "AutomatedEtcdBackup", "BootcNodeManagement", "CSIDriverSharedResource", "ClusterAPIInstall", "ClusterAPIInstallIBMCloud", "ClusterMonitoringConfig", "ConsolePluginContentSecurityPolicy", "DNSNameResolver", "DynamicResourceAllocation", "EtcdBackendQuota", "EventedPLEG", "Example", "ExternalOIDC", "ExternalOIDCWithUIDAndExtraClaimMappings", "GCPClusterHostedDNS", "GatewayAPI", "ImageStreamImportMode", "IngressControllerDynamicConfigurationManager", "InsightsConfig", "InsightsConfigAPI", "InsightsOnDemandDataGather", "InsightsRuntimeExtractor", "MachineAPIMigration", "MachineAPIOperatorDisableMachineHealthCheckController", "MachineAPIProviderOpenStack", "MachineConfigNodes", "MaxUnavailableStatefulSet", "MetricsCollectionProfiles", "MinimumKubeletVersion", "MixedCPUsAllocation", "MultiArchInstallAzure", "NodeSwap", "NutanixMultiSubnets", "OVNObservability", "OpenShiftPodSecurityAdmission", "PinnedImages", "PlatformOperators", "ProcMountType", "RouteAdvertisements", "RouteExternalCertificate", "ServiceAccountTokenNodeBinding", "SignatureStores", "SigstoreImageVerification", "TranslateStreamCloseWebsocketRequests", "UpgradeStatus", "UserNamespacesPodSecurityStandards", "UserNamespacesSupport", "VSphereMultiNetworks", "VolumeAttributesClass", "VolumeGroupSnapshot"}}

openshift-network-operator

kubelet

network-operator-7bd846bfc4-mt454

Created

Created container: network-operator

openshift-machine-api

kubelet

machine-api-operator-6fbb6cf6f9-n8tnn

Started

Started container kube-rbac-proxy
(x2)

kube-system

kubelet

bootstrap-kube-controller-manager-master-0

Created

Created container: kube-controller-manager
(x2)

openshift-ingress-operator

kubelet

ingress-operator-66b84d69b-gzg9m

Started

Started container ingress-operator

openshift-machine-api

kubelet

machine-api-operator-6fbb6cf6f9-n8tnn

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:de91abd5ad76fb491881a75a0feb4b8ca5600ceb5e15a4b0b687ada01ea0a44c"
(x2)

openshift-kube-controller-manager-operator

kubelet

kube-controller-manager-operator-ff989d6cc-qcpb4

Started

Started container kube-controller-manager-operator

openshift-machine-api

kubelet

machine-api-operator-6fbb6cf6f9-n8tnn

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:de91abd5ad76fb491881a75a0feb4b8ca5600ceb5e15a4b0b687ada01ea0a44c"

openshift-machine-api

kubelet

machine-api-operator-6fbb6cf6f9-n8tnn

Started

Started container kube-rbac-proxy

openshift-machine-api

kubelet

machine-api-operator-6fbb6cf6f9-n8tnn

Created

Created container: kube-rbac-proxy
(x2)

kube-system

kubelet

bootstrap-kube-controller-manager-master-0

Started

Started container kube-controller-manager
(x2)

openshift-kube-controller-manager-operator

kubelet

kube-controller-manager-operator-ff989d6cc-qcpb4

Created

Created container: kube-controller-manager-operator

openshift-kube-controller-manager

kubelet

installer-2-master-0

Created

Created container: installer

openshift-kube-controller-manager

kubelet

installer-2-master-0

Started

Started container installer

openshift-cloud-credential-operator

kubelet

cloud-credential-operator-744f9dbf77-r4qvh

Started

Started container kube-rbac-proxy
(x2)

openshift-etcd-operator

kubelet

etcd-operator-8544cbcf9c-brhw4

Started

Started container etcd-operator

openshift-machine-api

kubelet

machine-api-operator-6fbb6cf6f9-n8tnn

Created

Created container: kube-rbac-proxy

openshift-cloud-credential-operator

kubelet

cloud-credential-operator-744f9dbf77-r4qvh

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:82f2c59d19eb73ad5c0f93b0a63003c1885f5297959c9c45b401d1a74aea6e76"

openshift-network-operator

kubelet

network-operator-7bd846bfc4-mt454

Started

Started container network-operator
(x2)

openshift-etcd-operator

kubelet

etcd-operator-8544cbcf9c-brhw4

Created

Created container: etcd-operator

openshift-machine-config-operator

machineconfigoperator

machine-config

OperatorDegraded: MachineConfigControllerFailed

Failed to resync 4.18.35 because: Could not make kubeconfig-data CM, configmaps "kubeconfig-data" already exists

openshift-cluster-samples-operator

kubelet

cluster-samples-operator-85f7577d78-mwfgx

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ab745a9e15dadc862548ceb5740b8f5d02075232760c6715d82b4c3b70eddca9" already present on machine

openshift-machine-config-operator

machineconfigdaemon

master-0

Uncordon

Update completed for config rendered-master-47246abc0cee5fbe972e47e23f330f37 and node has been uncordoned

openshift-machine-api

kubelet

cluster-autoscaler-operator-866dc4744-xwxg7

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ce8e3088493b4a72dd766b3b5b4ccb83b7d72d514bbf64063a913dfe961273db" in 4.623s (4.623s including waiting). Image size: 456375453 bytes.

openshift-insights

kubelet

insights-operator-68bf6ff9d6-mvfn5

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1973d56a1097a48ea0ebf2c4dbae1ed86fa67bb0116f4962f7720d48aa337d27" in 4.939s (4.939s including waiting). Image size: 504662731 bytes.

openshift-machine-api

kubelet

cluster-autoscaler-operator-866dc4744-xwxg7

Created

Created container: cluster-autoscaler-operator

openshift-cluster-samples-operator

kubelet

cluster-samples-operator-85f7577d78-mwfgx

Created

Created container: cluster-samples-operator-watch
(x3)

openshift-authentication-operator

kubelet

authentication-operator-5885bfd7f4-62zrx

Started

Started container authentication-operator

openshift-machine-api

kubelet

cluster-autoscaler-operator-866dc4744-xwxg7

Started

Started container cluster-autoscaler-operator

openshift-cluster-samples-operator

kubelet

cluster-samples-operator-85f7577d78-mwfgx

Started

Started container cluster-samples-operator-watch

openshift-machine-api

kubelet

cluster-autoscaler-operator-866dc4744-xwxg7

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ce8e3088493b4a72dd766b3b5b4ccb83b7d72d514bbf64063a913dfe961273db" in 4.623s (4.623s including waiting). Image size: 456375453 bytes.

openshift-cluster-olm-operator

olm-status-controller-statussyncer_olm

cluster-olm-operator

OperatorStatusChanged

Status for clusteroperator/olm changed: Degraded message changed from "CatalogdDeploymentCatalogdControllerManagerDegraded: the server was unable to return a response in the time allotted, but may still be processing the request (get olms.operator.openshift.io cluster)" to "All is well"

openshift-machine-api

kubelet

cluster-autoscaler-operator-866dc4744-xwxg7

Started

Started container cluster-autoscaler-operator
(x3)

openshift-authentication-operator

kubelet

authentication-operator-5885bfd7f4-62zrx

Created

Created container: authentication-operator

openshift-cluster-samples-operator

kubelet

cluster-samples-operator-85f7577d78-mwfgx

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ab745a9e15dadc862548ceb5740b8f5d02075232760c6715d82b4c3b70eddca9" in 4.638s (4.638s including waiting). Image size: 455417803 bytes.

openshift-cluster-samples-operator

kubelet

cluster-samples-operator-85f7577d78-mwfgx

Created

Created container: cluster-samples-operator

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-status-controller-statussyncer_kube-scheduler

openshift-kube-scheduler-operator

OperatorStatusChanged

Status for clusteroperator/kube-scheduler changed: Degraded message changed from "NodeControllerDegraded: All master nodes are ready" to "StaticPodsDegraded: the server was unable to return a response in the time allotted, but may still be processing the request (get pods openshift-kube-scheduler-master-0)\nNodeControllerDegraded: All master nodes are ready"

openshift-cluster-olm-operator

olm-status-controller-statussyncer_olm

cluster-olm-operator

OperatorStatusChanged

Status for clusteroperator/olm changed: Degraded message changed from "All is well" to "CatalogdDeploymentCatalogdControllerManagerDegraded: the server was unable to return a response in the time allotted, but may still be processing the request (get olms.operator.openshift.io cluster)"

openshift-cluster-samples-operator

kubelet

cluster-samples-operator-85f7577d78-mwfgx

Started

Started container cluster-samples-operator

openshift-insights

kubelet

insights-operator-68bf6ff9d6-mvfn5

Created

Created container: insights-operator

openshift-machine-config-operator

machineconfigdaemon

master-0

NodeDone

Setting node master-0, currentConfig rendered-master-47246abc0cee5fbe972e47e23f330f37 to Done

openshift-machine-config-operator

machineconfigdaemon

master-0

ConfigDriftMonitorStarted

Config Drift Monitor started, watching against rendered-master-47246abc0cee5fbe972e47e23f330f37
(x2)

openshift-authentication-operator

kubelet

authentication-operator-5885bfd7f4-62zrx

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bfe394b58ec6195de8b8420e781b7630d85a412b9112d892fea903f92b783427" already present on machine

openshift-machine-config-operator

machineconfigcontroller-nodecontroller

master

AnnotationChange

Node master-0 now has machineconfiguration.openshift.io/reason=

openshift-insights

kubelet

insights-operator-68bf6ff9d6-mvfn5

Started

Started container insights-operator

openshift-machine-api

kubelet

cluster-autoscaler-operator-866dc4744-xwxg7

Created

Created container: cluster-autoscaler-operator

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-status-controller-statussyncer_kube-scheduler

openshift-kube-scheduler-operator

OperatorStatusChanged

Status for clusteroperator/kube-scheduler changed: Degraded message changed from "StaticPodsDegraded: the server was unable to return a response in the time allotted, but may still be processing the request (get pods openshift-kube-scheduler-master-0)\nNodeControllerDegraded: All master nodes are ready" to "NodeControllerDegraded: All master nodes are ready\nKubeControllerManagerStaticResourcesDegraded: \"assets/kube-scheduler/ns.yaml\" (string): the server was unable to return a response in the time allotted, but may still be processing the request (get namespaces openshift-kube-scheduler)\nKubeControllerManagerStaticResourcesDegraded: "

openshift-machine-api

kubelet

control-plane-machine-set-operator-6f97756bc8-7t5qv

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:908eaaf624959bc7645f6d585d160431d1efb070e9a1f37fefed73a3be42b0d3"

openshift-cluster-samples-operator

file-change-watchdog

cluster-samples-operator

FileChangeWatchdogStarted

Started watching files for process cluster-samples-operator[2]
(x2)

openshift-machine-api

multus

control-plane-machine-set-operator-6f97756bc8-7t5qv

AddedInterface

Add eth0 [10.128.0.62/23] from ovn-kubernetes

openshift-insights

openshift-insights-operator

insights-operator

FeatureGatesInitialized

FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{"AWSEFSDriverVolumeMetrics", "AdminNetworkPolicy", "AlibabaPlatform", "AzureWorkloadIdentity", "BareMetalLoadBalancer", "BuildCSIVolumes", "ChunkSizeMiB", "CloudDualStackNodeIPs", "DisableKubeletCloudCredentialProviders", "GCPLabelsTags", "HardwareSpeed", "IngressControllerLBSubnetsAWS", "KMSv1", "ManagedBootImages", "ManagedBootImagesAWS", "MultiArchInstallAWS", "MultiArchInstallGCP", "NetworkDiagnosticsConfig", "NetworkLiveMigration", "NetworkSegmentation", "NewOLM", "NodeDisruptionPolicy", "OnClusterBuild", "PersistentIPsForVirtualization", "PrivateHostedZoneAWS", "SetEIPForNLBIngressController", "StreamingCollectionEncodingToJSON", "StreamingCollectionEncodingToProtobuf", "VSphereControlPlaneMachineSet", "VSphereDriverConfiguration", "VSphereMultiVCenters", "VSphereStaticIPs", "ValidatingAdmissionPolicy"}, Disabled:[]v1.FeatureGateName{"AWSClusterHostedDNS", "AdditionalRoutingCapabilities", "AutomatedEtcdBackup", "BootcNodeManagement", "CSIDriverSharedResource", "ClusterAPIInstall", "ClusterAPIInstallIBMCloud", "ClusterMonitoringConfig", "ConsolePluginContentSecurityPolicy", "DNSNameResolver", "DynamicResourceAllocation", "EtcdBackendQuota", "EventedPLEG", "Example", "ExternalOIDC", "ExternalOIDCWithUIDAndExtraClaimMappings", "GCPClusterHostedDNS", "GatewayAPI", "ImageStreamImportMode", "IngressControllerDynamicConfigurationManager", "InsightsConfig", "InsightsConfigAPI", "InsightsOnDemandDataGather", "InsightsRuntimeExtractor", "MachineAPIMigration", "MachineAPIOperatorDisableMachineHealthCheckController", "MachineAPIProviderOpenStack", "MachineConfigNodes", "MaxUnavailableStatefulSet", "MetricsCollectionProfiles", "MinimumKubeletVersion", "MixedCPUsAllocation", "MultiArchInstallAzure", "NodeSwap", "NutanixMultiSubnets", "OVNObservability", "OpenShiftPodSecurityAdmission", "PinnedImages", "PlatformOperators", "ProcMountType", "RouteAdvertisements", "RouteExternalCertificate", "ServiceAccountTokenNodeBinding", "SignatureStores", "SigstoreImageVerification", "TranslateStreamCloseWebsocketRequests", "UpgradeStatus", "UserNamespacesPodSecurityStandards", "UserNamespacesSupport", "VSphereMultiNetworks", "VolumeAttributesClass", "VolumeGroupSnapshot"}}

openshift-machine-api

kubelet

control-plane-machine-set-operator-6f97756bc8-7t5qv

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:908eaaf624959bc7645f6d585d160431d1efb070e9a1f37fefed73a3be42b0d3"
(x2)

openshift-machine-api

multus

control-plane-machine-set-operator-6f97756bc8-7t5qv

AddedInterface

Add eth0 [10.128.0.62/23] from ovn-kubernetes

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-status-controller-statussyncer_kube-scheduler

openshift-kube-scheduler-operator

OperatorStatusChanged

Status for clusteroperator/kube-scheduler changed: Degraded message changed from "NodeControllerDegraded: All master nodes are ready\nKubeControllerManagerStaticResourcesDegraded: \"assets/kube-scheduler/ns.yaml\" (string): the server was unable to return a response in the time allotted, but may still be processing the request (get namespaces openshift-kube-scheduler)\nKubeControllerManagerStaticResourcesDegraded: " to "NodeControllerDegraded: All master nodes are ready"

openshift-authentication-operator

oauth-apiserver-audit-policy-controller-auditpolicycontroller

authentication-operator

FastControllerResync

Controller "auditPolicyController" resync interval is set to 10s which might lead to client request throttling

openshift-service-ca-operator

service-ca-operator

service-ca-operator-lock

LeaderElection

service-ca-operator-b865698dc-qzb2h_ead37c96-0b80-43b3-a3f5-85a031c18364 became leader

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Degraded changed from False to True ("IngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\nOAuthServerServiceEndpointAccessibleControllerDegraded: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\nWebhookAuthenticatorControllerDegraded: failed to read service-ca crt bundle: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory")

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-status-controller-statussyncer_kube-scheduler

openshift-kube-scheduler-operator

OperatorStatusChanged

Status for clusteroperator/kube-scheduler changed: Degraded message changed from "NodeControllerDegraded: All master nodes are ready" to "NodeControllerDegraded: All master nodes are ready\nTargetConfigControllerDegraded: \"configmap\": the server was unable to return a response in the time allotted, but may still be processing the request (get configmaps config)"

openshift-machine-api

cluster-autoscaler-operator-866dc4744-xwxg7_92578c1c-5f2d-4e15-87b8-bebcc7511b80

cluster-autoscaler-operator-leader

LeaderElection

cluster-autoscaler-operator-866dc4744-xwxg7_92578c1c-5f2d-4e15-87b8-bebcc7511b80 became leader

openshift-kube-apiserver-operator

kube-apiserver-operator

kube-apiserver-operator-lock

LeaderElection

kube-apiserver-operator-8b68b9d9b-c6vkz_87b6548b-7a58-46f2-9005-cfa70afd4557 became leader

openshift-machine-api

cluster-autoscaler-operator-866dc4744-xwxg7_92578c1c-5f2d-4e15-87b8-bebcc7511b80

cluster-autoscaler-operator-leader

LeaderElection

cluster-autoscaler-operator-866dc4744-xwxg7_92578c1c-5f2d-4e15-87b8-bebcc7511b80 became leader

openshift-authentication-operator

cluster-authentication-operator

cluster-authentication-operator-lock

LeaderElection

authentication-operator-5885bfd7f4-62zrx_218d68c0-baaf-4f14-94ad-45cca8749e7d became leader

openshift-kube-apiserver-operator

kube-apiserver-operator

openshift-kube-apiserver-operator

FeatureGatesInitialized

FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{"AWSEFSDriverVolumeMetrics", "AdminNetworkPolicy", "AlibabaPlatform", "AzureWorkloadIdentity", "BareMetalLoadBalancer", "BuildCSIVolumes", "ChunkSizeMiB", "CloudDualStackNodeIPs", "DisableKubeletCloudCredentialProviders", "GCPLabelsTags", "HardwareSpeed", "IngressControllerLBSubnetsAWS", "KMSv1", "ManagedBootImages", "ManagedBootImagesAWS", "MultiArchInstallAWS", "MultiArchInstallGCP", "NetworkDiagnosticsConfig", "NetworkLiveMigration", "NetworkSegmentation", "NewOLM", "NodeDisruptionPolicy", "OnClusterBuild", "PersistentIPsForVirtualization", "PrivateHostedZoneAWS", "SetEIPForNLBIngressController", "StreamingCollectionEncodingToJSON", "StreamingCollectionEncodingToProtobuf", "VSphereControlPlaneMachineSet", "VSphereDriverConfiguration", "VSphereMultiVCenters", "VSphereStaticIPs", "ValidatingAdmissionPolicy"}, Disabled:[]v1.FeatureGateName{"AWSClusterHostedDNS", "AdditionalRoutingCapabilities", "AutomatedEtcdBackup", "BootcNodeManagement", "CSIDriverSharedResource", "ClusterAPIInstall", "ClusterAPIInstallIBMCloud", "ClusterMonitoringConfig", "ConsolePluginContentSecurityPolicy", "DNSNameResolver", "DynamicResourceAllocation", "EtcdBackendQuota", "EventedPLEG", "Example", "ExternalOIDC", "ExternalOIDCWithUIDAndExtraClaimMappings", "GCPClusterHostedDNS", "GatewayAPI", "ImageStreamImportMode", "IngressControllerDynamicConfigurationManager", "InsightsConfig", "InsightsConfigAPI", "InsightsOnDemandDataGather", "InsightsRuntimeExtractor", "MachineAPIMigration", "MachineAPIOperatorDisableMachineHealthCheckController", "MachineAPIProviderOpenStack", "MachineConfigNodes", "MaxUnavailableStatefulSet", "MetricsCollectionProfiles", "MinimumKubeletVersion", "MixedCPUsAllocation", "MultiArchInstallAzure", "NodeSwap", "NutanixMultiSubnets", "OVNObservability", "OpenShiftPodSecurityAdmission", "PinnedImages", "PlatformOperators", "ProcMountType", "RouteAdvertisements", "RouteExternalCertificate", "ServiceAccountTokenNodeBinding", "SignatureStores", "SigstoreImageVerification", "TranslateStreamCloseWebsocketRequests", "UpgradeStatus", "UserNamespacesPodSecurityStandards", "UserNamespacesSupport", "VSphereMultiNetworks", "VolumeAttributesClass", "VolumeGroupSnapshot"}}

openshift-authentication-operator

oauth-apiserver-apiservice-openshift-apiserver-controller-openshift-apiserver-apiservice

authentication-operator

FastControllerResync

Controller "openshift-apiserver-APIService" resync interval is set to 10s which might lead to client request throttling

openshift-kube-apiserver-operator

kube-apiserver-operator-audit-policy-controller-auditpolicycontroller

openshift-kube-apiserver-operator

FastControllerResync

Controller "auditPolicyController" resync interval is set to 10s which might lead to client request throttling

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-status-controller-statussyncer_kube-scheduler

openshift-kube-scheduler-operator

OperatorStatusChanged

Status for clusteroperator/kube-scheduler changed: Degraded message changed from "NodeControllerDegraded: All master nodes are ready\nTargetConfigControllerDegraded: \"configmap\": the server was unable to return a response in the time allotted, but may still be processing the request (get configmaps config)" to "NodeControllerDegraded: All master nodes are ready"

openshift-machine-config-operator

machine-config-operator

machine-config-operator

ConfigMapUpdated

Updated ConfigMap/kube-rbac-proxy -n openshift-machine-config-operator: cause by changes in data.config-file.yaml

openshift-machine-api

kubelet

control-plane-machine-set-operator-6f97756bc8-7t5qv

Started

Started container control-plane-machine-set-operator

default

machineapioperator

machine-api

Status upgrade

Progressing towards operator: 4.18.35

openshift-kube-apiserver-operator

kube-apiserver-operator-installer-controller

openshift-kube-apiserver-operator

InstallerPodFailed

installer errors: installer: ving-cert", (string) (len=21) "user-serving-cert-000", (string) (len=21) "user-serving-cert-001", (string) (len=21) "user-serving-cert-002", (string) (len=21) "user-serving-cert-003", (string) (len=21) "user-serving-cert-004", (string) (len=21) "user-serving-cert-005", (string) (len=21) "user-serving-cert-006", (string) (len=21) "user-serving-cert-007", (string) (len=21) "user-serving-cert-008", (string) (len=21) "user-serving-cert-009" }, CertConfigMapNamePrefixes: ([]string) (len=4 cap=4) { (string) (len=20) "aggregator-client-ca", (string) (len=9) "client-ca", (string) (len=29) "control-plane-node-kubeconfig", (string) (len=26) "check-endpoints-kubeconfig" }, OptionalCertConfigMapNamePrefixes: ([]string) (len=1 cap=1) { (string) (len=17) "trusted-ca-bundle" }, CertDir: (string) (len=57) "/etc/kubernetes/static-pod-resources/kube-apiserver-certs", ResourceDir: (string) (len=36) "/etc/kubernetes/static-pod-resources", PodManifestDir: (string) (len=25) "/etc/kubernetes/manifests", Timeout: (time.Duration) 2m0s, StaticPodManifestsLockFile: (string) "", PodMutationFns: ([]installerpod.PodMutationFunc) <nil>, KubeletVersion: (string) "" }) I0320 08:37:20.888074 1 cmd.go:413] Getting controller reference for node master-0 I0320 08:37:20.919755 1 cmd.go:426] Waiting for installer revisions to settle for node master-0 I0320 08:37:20.919828 1 envvar.go:172] "Feature gate default state" feature="WatchListClient" enabled=false I0320 08:37:20.919840 1 envvar.go:172] "Feature gate default state" feature="InformerResourceVersion" enabled=false I0320 08:37:20.938866 1 cmd.go:518] Waiting additional period after revisions have settled for node master-0 I0320 08:37:50.939911 1 cmd.go:524] Getting installer pods for node master-0 F0320 08:38:04.943427 1 cmd.go:109] Get "https://172.30.0.1:443/api/v1/namespaces/openshift-kube-apiserver/pods?labelSelector=app%3Dinstaller": net/http: request canceled (Client.Timeout exceeded while awaiting headers)

openshift-machine-api

kubelet

machine-api-operator-6fbb6cf6f9-n8tnn

Created

Created container: machine-api-operator

openshift-machine-api

kubelet

machine-api-operator-6fbb6cf6f9-n8tnn

Started

Started container machine-api-operator

openshift-machine-api

kubelet

control-plane-machine-set-operator-6f97756bc8-7t5qv

Created

Created container: control-plane-machine-set-operator

openshift-machine-api

kubelet

control-plane-machine-set-operator-6f97756bc8-7t5qv

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:908eaaf624959bc7645f6d585d160431d1efb070e9a1f37fefed73a3be42b0d3" in 7.467s (7.467s including waiting). Image size: 470681292 bytes.

openshift-machine-api

machineapioperator

machine-api-operator

FeatureGatesInitialized

FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{"AWSEFSDriverVolumeMetrics", "AdminNetworkPolicy", "AlibabaPlatform", "AzureWorkloadIdentity", "BareMetalLoadBalancer", "BuildCSIVolumes", "ChunkSizeMiB", "CloudDualStackNodeIPs", "DisableKubeletCloudCredentialProviders", "GCPLabelsTags", "HardwareSpeed", "IngressControllerLBSubnetsAWS", "KMSv1", "ManagedBootImages", "ManagedBootImagesAWS", "MultiArchInstallAWS", "MultiArchInstallGCP", "NetworkDiagnosticsConfig", "NetworkLiveMigration", "NetworkSegmentation", "NewOLM", "NodeDisruptionPolicy", "OnClusterBuild", "PersistentIPsForVirtualization", "PrivateHostedZoneAWS", "SetEIPForNLBIngressController", "StreamingCollectionEncodingToJSON", "StreamingCollectionEncodingToProtobuf", "VSphereControlPlaneMachineSet", "VSphereDriverConfiguration", "VSphereMultiVCenters", "VSphereStaticIPs", "ValidatingAdmissionPolicy"}, Disabled:[]v1.FeatureGateName{"AWSClusterHostedDNS", "AdditionalRoutingCapabilities", "AutomatedEtcdBackup", "BootcNodeManagement", "CSIDriverSharedResource", "ClusterAPIInstall", "ClusterAPIInstallIBMCloud", "ClusterMonitoringConfig", "ConsolePluginContentSecurityPolicy", "DNSNameResolver", "DynamicResourceAllocation", "EtcdBackendQuota", "EventedPLEG", "Example", "ExternalOIDC", "ExternalOIDCWithUIDAndExtraClaimMappings", "GCPClusterHostedDNS", "GatewayAPI", "ImageStreamImportMode", "IngressControllerDynamicConfigurationManager", "InsightsConfig", "InsightsConfigAPI", "InsightsOnDemandDataGather", "InsightsRuntimeExtractor", "MachineAPIMigration", "MachineAPIOperatorDisableMachineHealthCheckController", "MachineAPIProviderOpenStack", "MachineConfigNodes", "MaxUnavailableStatefulSet", "MetricsCollectionProfiles", "MinimumKubeletVersion", "MixedCPUsAllocation", "MultiArchInstallAzure", "NodeSwap", "NutanixMultiSubnets", "OVNObservability", "OpenShiftPodSecurityAdmission", "PinnedImages", "PlatformOperators", "ProcMountType", "RouteAdvertisements", "RouteExternalCertificate", "ServiceAccountTokenNodeBinding", "SignatureStores", "SigstoreImageVerification", "TranslateStreamCloseWebsocketRequests", "UpgradeStatus", "UserNamespacesPodSecurityStandards", "UserNamespacesSupport", "VSphereMultiNetworks", "VolumeAttributesClass", "VolumeGroupSnapshot"}}

openshift-machine-api

machineapioperator

machine-api-operator

FeatureGatesInitialized

FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{"AWSEFSDriverVolumeMetrics", "AdminNetworkPolicy", "AlibabaPlatform", "AzureWorkloadIdentity", "BareMetalLoadBalancer", "BuildCSIVolumes", "ChunkSizeMiB", "CloudDualStackNodeIPs", "DisableKubeletCloudCredentialProviders", "GCPLabelsTags", "HardwareSpeed", "IngressControllerLBSubnetsAWS", "KMSv1", "ManagedBootImages", "ManagedBootImagesAWS", "MultiArchInstallAWS", "MultiArchInstallGCP", "NetworkDiagnosticsConfig", "NetworkLiveMigration", "NetworkSegmentation", "NewOLM", "NodeDisruptionPolicy", "OnClusterBuild", "PersistentIPsForVirtualization", "PrivateHostedZoneAWS", "SetEIPForNLBIngressController", "StreamingCollectionEncodingToJSON", "StreamingCollectionEncodingToProtobuf", "VSphereControlPlaneMachineSet", "VSphereDriverConfiguration", "VSphereMultiVCenters", "VSphereStaticIPs", "ValidatingAdmissionPolicy"}, Disabled:[]v1.FeatureGateName{"AWSClusterHostedDNS", "AdditionalRoutingCapabilities", "AutomatedEtcdBackup", "BootcNodeManagement", "CSIDriverSharedResource", "ClusterAPIInstall", "ClusterAPIInstallIBMCloud", "ClusterMonitoringConfig", "ConsolePluginContentSecurityPolicy", "DNSNameResolver", "DynamicResourceAllocation", "EtcdBackendQuota", "EventedPLEG", "Example", "ExternalOIDC", "ExternalOIDCWithUIDAndExtraClaimMappings", "GCPClusterHostedDNS", "GatewayAPI", "ImageStreamImportMode", "IngressControllerDynamicConfigurationManager", "InsightsConfig", "InsightsConfigAPI", "InsightsOnDemandDataGather", "InsightsRuntimeExtractor", "MachineAPIMigration", "MachineAPIOperatorDisableMachineHealthCheckController", "MachineAPIProviderOpenStack", "MachineConfigNodes", "MaxUnavailableStatefulSet", "MetricsCollectionProfiles", "MinimumKubeletVersion", "MixedCPUsAllocation", "MultiArchInstallAzure", "NodeSwap", "NutanixMultiSubnets", "OVNObservability", "OpenShiftPodSecurityAdmission", "PinnedImages", "PlatformOperators", "ProcMountType", "RouteAdvertisements", "RouteExternalCertificate", "ServiceAccountTokenNodeBinding", "SignatureStores", "SigstoreImageVerification", "TranslateStreamCloseWebsocketRequests", "UpgradeStatus", "UserNamespacesPodSecurityStandards", "UserNamespacesSupport", "VSphereMultiNetworks", "VolumeAttributesClass", "VolumeGroupSnapshot"}}

openshift-machine-api

control-plane-machine-set-operator-6f97756bc8-7t5qv_487c5efe-9830-446f-8545-7ebf924edd1a

control-plane-machine-set-leader

LeaderElection

control-plane-machine-set-operator-6f97756bc8-7t5qv_487c5efe-9830-446f-8545-7ebf924edd1a became leader

openshift-cloud-credential-operator

kubelet

cloud-credential-operator-744f9dbf77-r4qvh

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:82f2c59d19eb73ad5c0f93b0a63003c1885f5297959c9c45b401d1a74aea6e76" in 12.543s (12.543s including waiting). Image size: 880382887 bytes.

openshift-machine-api

kubelet

machine-api-operator-6fbb6cf6f9-n8tnn

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:de91abd5ad76fb491881a75a0feb4b8ca5600ceb5e15a4b0b687ada01ea0a44c" in 12.448s (12.448s including waiting). Image size: 862205633 bytes.

openshift-cloud-credential-operator

kubelet

cloud-credential-operator-744f9dbf77-r4qvh

Created

Created container: cloud-credential-operator

openshift-cloud-credential-operator

kubelet

cloud-credential-operator-744f9dbf77-r4qvh

Started

Started container cloud-credential-operator

openshift-machine-api

kubelet

machine-api-operator-6fbb6cf6f9-n8tnn

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:de91abd5ad76fb491881a75a0feb4b8ca5600ceb5e15a4b0b687ada01ea0a44c" in 12.448s (12.448s including waiting). Image size: 862205633 bytes.

openshift-machine-api

kubelet

machine-api-operator-6fbb6cf6f9-n8tnn

Created

Created container: machine-api-operator

openshift-machine-api

kubelet

control-plane-machine-set-operator-6f97756bc8-7t5qv

Started

Started container control-plane-machine-set-operator

openshift-machine-api

kubelet

control-plane-machine-set-operator-6f97756bc8-7t5qv

Created

Created container: control-plane-machine-set-operator

openshift-machine-api

kubelet

control-plane-machine-set-operator-6f97756bc8-7t5qv

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:908eaaf624959bc7645f6d585d160431d1efb070e9a1f37fefed73a3be42b0d3" in 7.467s (7.467s including waiting). Image size: 470681292 bytes.

openshift-machine-api

kubelet

machine-api-operator-6fbb6cf6f9-n8tnn

Started

Started container machine-api-operator

openshift-machine-api

control-plane-machine-set-operator-6f97756bc8-7t5qv_487c5efe-9830-446f-8545-7ebf924edd1a

control-plane-machine-set-leader

LeaderElection

control-plane-machine-set-operator-6f97756bc8-7t5qv_487c5efe-9830-446f-8545-7ebf924edd1a became leader

openshift-authentication-operator

oauth-apiserver-webhook-authenticator-controller-webhookauthenticatorcontroller

authentication-operator

SecretCreated

Created Secret/webhook-authentication-integrated-oauth -n openshift-config because it was missing

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Degraded message changed from "IngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\nOAuthServerServiceEndpointAccessibleControllerDegraded: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\nWebhookAuthenticatorControllerDegraded: failed to read service-ca crt bundle: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory" to "IngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\nOAuthServerServiceEndpointAccessibleControllerDegraded: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready"

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

openshift-kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/config-2 -n openshift-kube-apiserver because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

openshift-kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/kube-apiserver-pod-2 -n openshift-kube-apiserver because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

openshift-kube-apiserver-operator

StartingNewRevision

new revision 2 triggered by "optional secret/webhook-authenticator has been created"

openshift-kube-apiserver-operator

kube-apiserver-operator-resource-sync-controller-resourcesynccontroller

openshift-kube-apiserver-operator

SecretCreated

Created Secret/webhook-authenticator -n openshift-kube-apiserver because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

openshift-kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/kube-apiserver-cert-syncer-kubeconfig-2 -n openshift-kube-apiserver because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

openshift-kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/bound-sa-token-signing-certs-2 -n openshift-kube-apiserver because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

openshift-kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/etcd-serving-ca-2 -n openshift-kube-apiserver because it was missing

kube-system

default-scheduler

kube-scheduler

LeaderElection

master-0_c28f64e2-3f59-4922-8dca-15f8e2495eb6 became leader

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

openshift-kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/kube-apiserver-server-ca-2 -n openshift-kube-apiserver because it was missing

openshift-route-controller-manager

kubelet

route-controller-manager-56f686584b-fdcx5

Created

Created container: route-controller-manager

openshift-route-controller-manager

kubelet

route-controller-manager-56f686584b-fdcx5

Started

Started container route-controller-manager

openshift-controller-manager

kubelet

controller-manager-fc56bb77c-qd4sn

Created

Created container: controller-manager

openshift-route-controller-manager

route-controller-manager

openshift-route-controllers

LeaderElection

route-controller-manager-56f686584b-fdcx5_8713fa8f-8896-4d05-8304-d0a8ed9f240e became leader

openshift-controller-manager

multus

controller-manager-fc56bb77c-qd4sn

AddedInterface

Add eth0 [10.128.0.69/23] from ovn-kubernetes

openshift-route-controller-manager

multus

route-controller-manager-56f686584b-fdcx5

AddedInterface

Add eth0 [10.128.0.68/23] from ovn-kubernetes

openshift-controller-manager

kubelet

controller-manager-fc56bb77c-qd4sn

Started

Started container controller-manager

openshift-controller-manager

openshift-controller-manager

openshift-master-controllers

LeaderElection

controller-manager-fc56bb77c-qd4sn became leader

openshift-route-controller-manager

kubelet

route-controller-manager-56f686584b-fdcx5

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:446bedea4916d3c1ee52be94137e484659e9561bd1de95c8189eee279aae984b" already present on machine

openshift-controller-manager

kubelet

controller-manager-fc56bb77c-qd4sn

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c2dd7a03348212e49876f5359f233d893a541ed9b934df390201a05133a06982" already present on machine

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

openshift-kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/kubelet-serving-ca-2 -n openshift-kube-apiserver because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

openshift-kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/sa-token-signing-certs-2 -n openshift-kube-apiserver because it was missing

openshift-marketplace

kubelet

community-operators-zspn5

Killing

Stopping container registry-server

openshift-marketplace

kubelet

certified-operators-86g9n

Killing

Stopping container registry-server

kube-system

kube-controller-manager

kube-controller-manager

LeaderElection

master-0_95469d5a-03d8-487b-961e-4187fe645153 became leader

openshift-cluster-machine-approver

replicaset-controller

machine-approver-6cb57bb5db

SuccessfulDelete

Deleted pod: machine-approver-6cb57bb5db-mt758

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ValidatingWebhookConfigurationCreated

Created ValidatingWebhookConfiguration.admissionregistration.k8s.io/prometheusrules.openshift.io because it was missing

openshift-cluster-machine-approver

deployment-controller

machine-approver

ScalingReplicaSet

Scaled down replica set machine-approver-6cb57bb5db to 0 from 1

openshift-monitoring

deployment-controller

prometheus-operator

ScalingReplicaSet

Scaled up replica set prometheus-operator-6c8df6d4b to 1

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ServiceCreated

Created Service/prometheus-operator -n openshift-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ValidatingWebhookConfigurationCreated

Created ValidatingWebhookConfiguration.admissionregistration.k8s.io/alertmanagerconfigs.openshift.io because it was missing

default

node-controller

master-0

RegisteredNode

Node master-0 event: Registered Node master-0 in Controller

openshift-cloud-controller-manager-operator

kubelet

cluster-cloud-controller-manager-operator-7559f7c68c-xhq29

Killing

Stopping container cluster-cloud-controller-manager

openshift-cloud-controller-manager-operator

kubelet

cluster-cloud-controller-manager-operator-7559f7c68c-xhq29

Killing

Stopping container config-sync-controllers

openshift-cloud-controller-manager-operator

replicaset-controller

cluster-cloud-controller-manager-operator-7559f7c68c

SuccessfulDelete

Deleted pod: cluster-cloud-controller-manager-operator-7559f7c68c-xhq29

openshift-monitoring

replicaset-controller

prometheus-operator-6c8df6d4b

SuccessfulCreate

Created pod: prometheus-operator-6c8df6d4b-5m857

openshift-monitoring

replicaset-controller

prometheus-operator-6c8df6d4b

SuccessfulCreate

Created pod: prometheus-operator-6c8df6d4b-5m857

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ValidatingWebhookConfigurationCreated

Created ValidatingWebhookConfiguration.admissionregistration.k8s.io/prometheusrules.openshift.io because it was missing

openshift-cloud-controller-manager-operator

deployment-controller

cluster-cloud-controller-manager-operator

ScalingReplicaSet

Scaled down replica set cluster-cloud-controller-manager-operator-7559f7c68c to 0 from 1

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ValidatingWebhookConfigurationCreated

Created ValidatingWebhookConfiguration.admissionregistration.k8s.io/alertmanagerconfigs.openshift.io because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ServiceCreated

Created Service/prometheus-operator -n openshift-monitoring because it was missing

openshift-monitoring

deployment-controller

prometheus-operator

ScalingReplicaSet

Scaled up replica set prometheus-operator-6c8df6d4b to 1

openshift-cluster-machine-approver

kubelet

machine-approver-6cb57bb5db-mt758

Killing

Stopping container machine-approver-controller

openshift-cloud-controller-manager-operator

kubelet

cluster-cloud-controller-manager-operator-7559f7c68c-xhq29

Killing

Stopping container kube-rbac-proxy

openshift-cluster-machine-approver

kubelet

machine-approver-6cb57bb5db-mt758

Killing

Stopping container kube-rbac-proxy

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

openshift-kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/kube-apiserver-audit-policies-2 -n openshift-kube-apiserver because it was missing

openshift-cluster-machine-approver

deployment-controller

machine-approver

ScalingReplicaSet

Scaled up replica set machine-approver-5c6485487f to 1

openshift-monitoring

kubelet

prometheus-operator-6c8df6d4b-5m857

FailedMount

MountVolume.SetUp failed for volume "prometheus-operator-tls" : secret "prometheus-operator-tls" not found

openshift-marketplace

multus

community-operators-dtqgc

AddedInterface

Add eth0 [10.128.0.71/23] from ovn-kubernetes

openshift-monitoring

kubelet

prometheus-operator-6c8df6d4b-5m857

FailedMount

MountVolume.SetUp failed for volume "prometheus-operator-tls" : secret "prometheus-operator-tls" not found

openshift-marketplace

kubelet

certified-operators-cc955

Started

Started container extract-utilities

openshift-cluster-machine-approver

replicaset-controller

machine-approver-5c6485487f

SuccessfulCreate

Created pod: machine-approver-5c6485487f-qb94j

openshift-marketplace

kubelet

community-operators-dtqgc

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6bba3f73c0066e42b24839e0d29f5dce2f36436f0a11f9f5e1029bccc5ed6578" already present on machine

openshift-marketplace

kubelet

certified-operators-cc955

Created

Created container: extract-utilities

openshift-cloud-controller-manager-operator

deployment-controller

cluster-cloud-controller-manager-operator

ScalingReplicaSet

Scaled up replica set cluster-cloud-controller-manager-operator-7dff898856 to 1

openshift-marketplace

kubelet

certified-operators-cc955

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6bba3f73c0066e42b24839e0d29f5dce2f36436f0a11f9f5e1029bccc5ed6578" already present on machine

openshift-marketplace

kubelet

community-operators-dtqgc

Started

Started container extract-utilities

openshift-cluster-machine-approver

kubelet

machine-approver-5c6485487f-qb94j

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-cloud-controller-manager-operator

replicaset-controller

cluster-cloud-controller-manager-operator-7dff898856

SuccessfulCreate

Created pod: cluster-cloud-controller-manager-operator-7dff898856-7vxxr

openshift-marketplace

multus

certified-operators-cc955

AddedInterface

Add eth0 [10.128.0.70/23] from ovn-kubernetes

openshift-cloud-controller-manager-operator

kubelet

cluster-cloud-controller-manager-operator-7dff898856-7vxxr

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:112a03f2411f871cdaca5f20daef71024dac710113d5f30897117a5a02f6b6f5" already present on machine

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

openshift-kube-apiserver-operator

SecretCreated

Created Secret/etcd-client-2 -n openshift-kube-apiserver because it was missing

openshift-marketplace

kubelet

community-operators-dtqgc

Created

Created container: extract-utilities

openshift-monitoring

kubelet

prometheus-operator-6c8df6d4b-5m857

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9174864cd47431966d033d283bc7836e7ca579139ef85c36275db542fda80803"

openshift-cloud-controller-manager-operator

kubelet

cluster-cloud-controller-manager-operator-7dff898856-7vxxr

Created

Created container: kube-rbac-proxy

openshift-cloud-controller-manager-operator

kubelet

cluster-cloud-controller-manager-operator-7dff898856-7vxxr

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:112a03f2411f871cdaca5f20daef71024dac710113d5f30897117a5a02f6b6f5" already present on machine

openshift-cloud-controller-manager-operator

kubelet

cluster-cloud-controller-manager-operator-7dff898856-7vxxr

Created

Created container: cluster-cloud-controller-manager

openshift-cluster-machine-approver

kubelet

machine-approver-5c6485487f-qb94j

Created

Created container: machine-approver-controller

openshift-cluster-machine-approver

kubelet

machine-approver-5c6485487f-qb94j

Started

Started container machine-approver-controller

openshift-marketplace

kubelet

redhat-marketplace-srjqw

Killing

Stopping container registry-server

openshift-monitoring

kubelet

prometheus-operator-6c8df6d4b-5m857

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9174864cd47431966d033d283bc7836e7ca579139ef85c36275db542fda80803"

openshift-marketplace

kubelet

community-operators-dtqgc

Pulling

Pulling image "registry.redhat.io/redhat/community-operator-index:v4.18"

openshift-marketplace

kubelet

redhat-operators-mgcb9

Killing

Stopping container registry-server

openshift-cloud-controller-manager-operator

kubelet

cluster-cloud-controller-manager-operator-7dff898856-7vxxr

Started

Started container kube-rbac-proxy

openshift-cloud-controller-manager-operator

kubelet

cluster-cloud-controller-manager-operator-7dff898856-7vxxr

Started

Started container cluster-cloud-controller-manager

openshift-cloud-controller-manager-operator

kubelet

cluster-cloud-controller-manager-operator-7dff898856-7vxxr

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-cloud-controller-manager-operator

kubelet

cluster-cloud-controller-manager-operator-7dff898856-7vxxr

Started

Started container config-sync-controllers

openshift-cluster-machine-approver

kubelet

machine-approver-5c6485487f-qb94j

Created

Created container: kube-rbac-proxy

openshift-cloud-controller-manager

cloud-controller-manager-operator

openshift-cloud-controller-manager

FeatureGatesInitialized

FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{"AWSEFSDriverVolumeMetrics", "AdminNetworkPolicy", "AlibabaPlatform", "AzureWorkloadIdentity", "BareMetalLoadBalancer", "BuildCSIVolumes", "ChunkSizeMiB", "CloudDualStackNodeIPs", "DisableKubeletCloudCredentialProviders", "GCPLabelsTags", "HardwareSpeed", "IngressControllerLBSubnetsAWS", "KMSv1", "ManagedBootImages", "ManagedBootImagesAWS", "MultiArchInstallAWS", "MultiArchInstallGCP", "NetworkDiagnosticsConfig", "NetworkLiveMigration", "NetworkSegmentation", "NewOLM", "NodeDisruptionPolicy", "OnClusterBuild", "PersistentIPsForVirtualization", "PrivateHostedZoneAWS", "SetEIPForNLBIngressController", "StreamingCollectionEncodingToJSON", "StreamingCollectionEncodingToProtobuf", "VSphereControlPlaneMachineSet", "VSphereDriverConfiguration", "VSphereMultiVCenters", "VSphereStaticIPs", "ValidatingAdmissionPolicy"}, Disabled:[]v1.FeatureGateName{"AWSClusterHostedDNS", "AdditionalRoutingCapabilities", "AutomatedEtcdBackup", "BootcNodeManagement", "CSIDriverSharedResource", "ClusterAPIInstall", "ClusterAPIInstallIBMCloud", "ClusterMonitoringConfig", "ConsolePluginContentSecurityPolicy", "DNSNameResolver", "DynamicResourceAllocation", "EtcdBackendQuota", "EventedPLEG", "Example", "ExternalOIDC", "ExternalOIDCWithUIDAndExtraClaimMappings", "GCPClusterHostedDNS", "GatewayAPI", "ImageStreamImportMode", "IngressControllerDynamicConfigurationManager", "InsightsConfig", "InsightsConfigAPI", "InsightsOnDemandDataGather", "InsightsRuntimeExtractor", "MachineAPIMigration", "MachineAPIOperatorDisableMachineHealthCheckController", "MachineAPIProviderOpenStack", "MachineConfigNodes", "MaxUnavailableStatefulSet", "MetricsCollectionProfiles", "MinimumKubeletVersion", "MixedCPUsAllocation", "MultiArchInstallAzure", "NodeSwap", "NutanixMultiSubnets", "OVNObservability", "OpenShiftPodSecurityAdmission", "PinnedImages", "PlatformOperators", "ProcMountType", "RouteAdvertisements", "RouteExternalCertificate", "ServiceAccountTokenNodeBinding", "SignatureStores", "SigstoreImageVerification", "TranslateStreamCloseWebsocketRequests", "UpgradeStatus", "UserNamespacesPodSecurityStandards", "UserNamespacesSupport", "VSphereMultiNetworks", "VolumeAttributesClass", "VolumeGroupSnapshot"}}

openshift-cluster-machine-approver

kubelet

machine-approver-5c6485487f-qb94j

Started

Started container kube-rbac-proxy

openshift-cloud-controller-manager-operator

kubelet

cluster-cloud-controller-manager-operator-7dff898856-7vxxr

Created

Created container: config-sync-controllers

openshift-monitoring

multus

prometheus-operator-6c8df6d4b-5m857

AddedInterface

Add eth0 [10.128.0.72/23] from ovn-kubernetes

openshift-cluster-machine-approver

master-0_f90c4172-40c8-4446-b37c-db81dc3f6514

cluster-machine-approver-leader

LeaderElection

master-0_f90c4172-40c8-4446-b37c-db81dc3f6514 became leader

openshift-marketplace

kubelet

certified-operators-cc955

Pulling

Pulling image "registry.redhat.io/redhat/certified-operator-index:v4.18"

openshift-marketplace

kubelet

certified-operators-cc955

Pulled

Successfully pulled image "registry.redhat.io/redhat/certified-operator-index:v4.18" in 635ms (635ms including waiting). Image size: 1252654287 bytes.

openshift-cluster-machine-approver

kubelet

machine-approver-5c6485487f-qb94j

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:cdd28dfe7132e19af9f013f72cf120d970bc31b6b74693af262f8d2e82a096e1" already present on machine

openshift-monitoring

multus

prometheus-operator-6c8df6d4b-5m857

AddedInterface

Add eth0 [10.128.0.72/23] from ovn-kubernetes

openshift-marketplace

kubelet

community-operators-dtqgc

Created

Created container: extract-content

openshift-marketplace

kubelet

redhat-operators-jstrn

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6bba3f73c0066e42b24839e0d29f5dce2f36436f0a11f9f5e1029bccc5ed6578" already present on machine

openshift-marketplace

kubelet

certified-operators-cc955

Started

Started container extract-content

openshift-marketplace

multus

redhat-marketplace-hqqrk

AddedInterface

Add eth0 [10.128.0.73/23] from ovn-kubernetes

openshift-marketplace

kubelet

redhat-marketplace-hqqrk

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6bba3f73c0066e42b24839e0d29f5dce2f36436f0a11f9f5e1029bccc5ed6578" already present on machine

openshift-marketplace

kubelet

community-operators-dtqgc

Pulled

Successfully pulled image "registry.redhat.io/redhat/community-operator-index:v4.18" in 796ms (796ms including waiting). Image size: 1223740630 bytes.

openshift-marketplace

kubelet

certified-operators-cc955

Created

Created container: extract-content

openshift-marketplace

kubelet

community-operators-dtqgc

Started

Started container extract-content

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

openshift-kube-apiserver-operator

SecretCreated

Created Secret/localhost-recovery-serving-certkey-2 -n openshift-kube-apiserver because it was missing

openshift-marketplace

multus

redhat-operators-jstrn

AddedInterface

Add eth0 [10.128.0.74/23] from ovn-kubernetes

openshift-monitoring

kubelet

prometheus-operator-6c8df6d4b-5m857

Created

Created container: prometheus-operator

openshift-marketplace

kubelet

certified-operators-cc955

Started

Started container registry-server

openshift-marketplace

kubelet

redhat-operators-jstrn

Started

Started container extract-utilities

openshift-marketplace

kubelet

redhat-operators-jstrn

Created

Created container: extract-utilities

openshift-monitoring

kubelet

prometheus-operator-6c8df6d4b-5m857

Created

Created container: prometheus-operator

openshift-marketplace

kubelet

redhat-marketplace-hqqrk

Created

Created container: extract-utilities

openshift-monitoring

kubelet

prometheus-operator-6c8df6d4b-5m857

Started

Started container kube-rbac-proxy

openshift-marketplace

kubelet

certified-operators-cc955

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b00c42562d477ef44d51f35950253a26d7debc7de86e53270831aafef5795c1"

openshift-marketplace

kubelet

certified-operators-cc955

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b00c42562d477ef44d51f35950253a26d7debc7de86e53270831aafef5795c1" in 379ms (379ms including waiting). Image size: 918289953 bytes.

openshift-monitoring

kubelet

prometheus-operator-6c8df6d4b-5m857

Started

Started container kube-rbac-proxy

openshift-marketplace

kubelet

certified-operators-cc955

Created

Created container: registry-server

openshift-marketplace

kubelet

community-operators-dtqgc

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b00c42562d477ef44d51f35950253a26d7debc7de86e53270831aafef5795c1" in 562ms (563ms including waiting). Image size: 918289953 bytes.

openshift-monitoring

kubelet

prometheus-operator-6c8df6d4b-5m857

Created

Created container: kube-rbac-proxy

openshift-marketplace

kubelet

redhat-marketplace-hqqrk

Started

Started container extract-utilities

openshift-monitoring

kubelet

prometheus-operator-6c8df6d4b-5m857

Started

Started container prometheus-operator

openshift-monitoring

kubelet

prometheus-operator-6c8df6d4b-5m857

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-monitoring

kubelet

prometheus-operator-6c8df6d4b-5m857

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

openshift-kube-apiserver-operator

SecretCreated

Created Secret/localhost-recovery-client-token-2 -n openshift-kube-apiserver because it was missing

openshift-monitoring

kubelet

prometheus-operator-6c8df6d4b-5m857

Started

Started container prometheus-operator

openshift-monitoring

kubelet

prometheus-operator-6c8df6d4b-5m857

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9174864cd47431966d033d283bc7836e7ca579139ef85c36275db542fda80803" in 2.091s (2.091s including waiting). Image size: 461569068 bytes.

openshift-monitoring

kubelet

prometheus-operator-6c8df6d4b-5m857

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9174864cd47431966d033d283bc7836e7ca579139ef85c36275db542fda80803" in 2.091s (2.091s including waiting). Image size: 461569068 bytes.

openshift-monitoring

kubelet

prometheus-operator-6c8df6d4b-5m857

Created

Created container: kube-rbac-proxy

openshift-marketplace

kubelet

community-operators-dtqgc

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b00c42562d477ef44d51f35950253a26d7debc7de86e53270831aafef5795c1"

openshift-marketplace

kubelet

community-operators-dtqgc

Started

Started container registry-server

openshift-marketplace

kubelet

redhat-marketplace-hqqrk

Pulling

Pulling image "registry.redhat.io/redhat/redhat-marketplace-index:v4.18"

openshift-marketplace

kubelet

redhat-operators-jstrn

Pulling

Pulling image "registry.redhat.io/redhat/redhat-operator-index:v4.18"

openshift-marketplace

kubelet

community-operators-dtqgc

Created

Created container: registry-server

openshift-marketplace

kubelet

redhat-marketplace-hqqrk

Pulled

Successfully pulled image "registry.redhat.io/redhat/redhat-marketplace-index:v4.18" in 540ms (540ms including waiting). Image size: 1231028434 bytes.

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

openshift-kube-apiserver-operator

SecretCreated

Created Secret/webhook-authenticator-2 -n openshift-kube-apiserver because it was missing

openshift-marketplace

kubelet

redhat-marketplace-hqqrk

Created

Created container: extract-content

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ServiceAccountCreated

Created ServiceAccount/openshift-state-metrics -n openshift-monitoring because it was missing

openshift-marketplace

kubelet

redhat-operators-jstrn

Started

Started container extract-content

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ServiceAccountCreated

Created ServiceAccount/kube-state-metrics -n openshift-monitoring because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

openshift-kube-apiserver-operator

RevisionTriggered

new revision 2 triggered by "optional secret/webhook-authenticator has been created"

openshift-marketplace

kubelet

redhat-marketplace-hqqrk

Started

Started container extract-content

openshift-marketplace

kubelet

redhat-marketplace-hqqrk

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b00c42562d477ef44d51f35950253a26d7debc7de86e53270831aafef5795c1"

openshift-marketplace

kubelet

redhat-marketplace-hqqrk

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b00c42562d477ef44d51f35950253a26d7debc7de86e53270831aafef5795c1" in 606ms (606ms including waiting). Image size: 918289953 bytes.

openshift-marketplace

kubelet

redhat-operators-jstrn

Created

Created container: extract-content

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ServiceAccountCreated

Created ServiceAccount/openshift-state-metrics -n openshift-monitoring because it was missing

openshift-marketplace

kubelet

redhat-operators-jstrn

Pulled

Successfully pulled image "registry.redhat.io/redhat/redhat-operator-index:v4.18" in 861ms (861ms including waiting). Image size: 1746912226 bytes.

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ServiceAccountCreated

Created ServiceAccount/kube-state-metrics -n openshift-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ServiceAccountCreated

Created ServiceAccount/node-exporter -n openshift-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ClusterRoleBindingCreated

Created ClusterRoleBinding.rbac.authorization.k8s.io/metrics-server:system:auth-delegator because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ClusterRoleBindingCreated

Created ClusterRoleBinding.rbac.authorization.k8s.io/prometheus-k8s because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ServiceCreated

Created Service/openshift-state-metrics -n openshift-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ServiceAccountCreated

Created ServiceAccount/metrics-server -n openshift-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ServiceCreated

Created Service/node-exporter -n openshift-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ServiceCreated

Created Service/kube-state-metrics -n openshift-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ConfigMapCreated

Created ConfigMap/kube-state-metrics-custom-resource-state-configmap -n openshift-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ClusterRoleCreated

Created ClusterRole.rbac.authorization.k8s.io/system:aggregated-metrics-reader because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ClusterRoleCreated

Created ClusterRole.rbac.authorization.k8s.io/pod-metrics-reader because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ClusterRoleBindingCreated

Created ClusterRoleBinding.rbac.authorization.k8s.io/node-exporter because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ClusterRoleBindingCreated

Created ClusterRoleBinding.rbac.authorization.k8s.io/openshift-state-metrics because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ClusterRoleCreated

Created ClusterRole.rbac.authorization.k8s.io/telemeter-client because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ClusterRoleCreated

Created ClusterRole.rbac.authorization.k8s.io/system:metrics-server because it was missing

openshift-monitoring

kubelet

node-exporter-lb4t5

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:0bccc03fd9ffe278e15c8f4be1db030307e4cd5020b78d711fc62f104fd6a980"

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ClusterRoleCreated

Created ClusterRole.rbac.authorization.k8s.io/openshift-state-metrics because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ClusterRoleBindingCreated

Created ClusterRoleBinding.rbac.authorization.k8s.io/kube-state-metrics because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ClusterRoleCreated

Created ClusterRole.rbac.authorization.k8s.io/system:aggregated-metrics-reader because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ClusterRoleCreated

Created ClusterRole.rbac.authorization.k8s.io/telemeter-client because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ClusterRoleCreated

Created ClusterRole.rbac.authorization.k8s.io/openshift-state-metrics because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ClusterRoleCreated

Created ClusterRole.rbac.authorization.k8s.io/node-exporter because it was missing

openshift-monitoring

replicaset-controller

kube-state-metrics-7bbc969446

SuccessfulCreate

Created pod: kube-state-metrics-7bbc969446-qh6vq

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ServiceAccountCreated

Created ServiceAccount/telemeter-client -n openshift-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ClusterRoleCreated

Created ClusterRole.rbac.authorization.k8s.io/cluster-monitoring-view because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ClusterRoleBindingCreated

Created ClusterRoleBinding.rbac.authorization.k8s.io/telemeter-client-view because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ServiceCreated

Created Service/telemeter-client -n openshift-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ClusterRoleCreated

Created ClusterRole.rbac.authorization.k8s.io/monitoring-rules-edit because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

SecretCreated

Created Secret/telemeter-client -n openshift-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ClusterRoleCreated

Created ClusterRole.rbac.authorization.k8s.io/kube-state-metrics because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

SecretCreated

Created Secret/telemeter-client-kube-rbac-proxy-config -n openshift-monitoring because it was missing

openshift-monitoring

deployment-controller

openshift-state-metrics

ScalingReplicaSet

Scaled up replica set openshift-state-metrics-5dc6c74576 to 1

openshift-monitoring

replicaset-controller

openshift-state-metrics-5dc6c74576

SuccessfulCreate

Created pod: openshift-state-metrics-5dc6c74576-j8tmv

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ClusterRoleBindingCreated

Created ClusterRoleBinding.rbac.authorization.k8s.io/system:metrics-server because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ClusterRoleCreated

Created ClusterRole.rbac.authorization.k8s.io/prometheus-k8s because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ClusterRoleCreated

Created ClusterRole.rbac.authorization.k8s.io/monitoring-rules-view because it was missing

openshift-monitoring

deployment-controller

kube-state-metrics

ScalingReplicaSet

Scaled up replica set kube-state-metrics-7bbc969446 to 1

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ClusterRoleCreated

Created ClusterRole.rbac.authorization.k8s.io/node-exporter because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ClusterRoleCreated

Created ClusterRole.rbac.authorization.k8s.io/cluster-monitoring-view because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ClusterRoleCreated

Created ClusterRole.rbac.authorization.k8s.io/kube-state-metrics because it was missing

openshift-marketplace

kubelet

redhat-operators-jstrn

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b00c42562d477ef44d51f35950253a26d7debc7de86e53270831aafef5795c1"

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ClusterRoleBindingCreated

Created ClusterRoleBinding.rbac.authorization.k8s.io/telemeter-client because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ClusterRoleCreated

Created ClusterRole.rbac.authorization.k8s.io/alert-routing-edit because it was missing

openshift-monitoring

deployment-controller

openshift-state-metrics

ScalingReplicaSet

Scaled up replica set openshift-state-metrics-5dc6c74576 to 1

openshift-monitoring

replicaset-controller

openshift-state-metrics-5dc6c74576

SuccessfulCreate

Created pod: openshift-state-metrics-5dc6c74576-j8tmv

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ClusterRoleBindingCreated

Created ClusterRoleBinding.rbac.authorization.k8s.io/openshift-state-metrics because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ClusterRoleCreated

Created ClusterRole.rbac.authorization.k8s.io/prometheus-k8s because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ClusterRoleBindingCreated

Created ClusterRoleBinding.rbac.authorization.k8s.io/node-exporter because it was missing

openshift-marketplace

kubelet

redhat-operators-jstrn

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b00c42562d477ef44d51f35950253a26d7debc7de86e53270831aafef5795c1" in 491ms (491ms including waiting). Image size: 918289953 bytes.

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ClusterRoleBindingCreated

Created ClusterRoleBinding.rbac.authorization.k8s.io/telemeter-client because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ConfigMapCreated

Created ConfigMap/kubelet-serving-ca-bundle -n openshift-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ClusterRoleCreated

Created ClusterRole.rbac.authorization.k8s.io/alert-routing-edit because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ClusterRoleBindingCreated

Created ClusterRoleBinding.rbac.authorization.k8s.io/prometheus-k8s because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ConfigMapCreated

Created ConfigMap/kubelet-serving-ca-bundle -n openshift-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ServiceAccountCreated

Created ServiceAccount/node-exporter -n openshift-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ServiceAccountCreated

Created ServiceAccount/telemeter-client -n openshift-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ServiceCreated

Created Service/openshift-state-metrics -n openshift-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ServiceAccountCreated

Created ServiceAccount/metrics-server -n openshift-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ServiceCreated

Created Service/node-exporter -n openshift-monitoring because it was missing

openshift-monitoring

kubelet

node-exporter-lb4t5

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:0bccc03fd9ffe278e15c8f4be1db030307e4cd5020b78d711fc62f104fd6a980"

openshift-monitoring

daemonset-controller

node-exporter

SuccessfulCreate

Created pod: node-exporter-lb4t5

openshift-monitoring

daemonset-controller

node-exporter

SuccessfulCreate

Created pod: node-exporter-lb4t5

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ClusterRoleBindingCreated

Created ClusterRoleBinding.rbac.authorization.k8s.io/kube-state-metrics because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

RoleCreated

Created Role.rbac.authorization.k8s.io/monitoring-alertmanager-api-writer -n openshift-user-workload-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

RoleCreated

Created Role.rbac.authorization.k8s.io/monitoring-alertmanager-api-reader -n openshift-user-workload-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ClusterRoleCreated

Created ClusterRole.rbac.authorization.k8s.io/monitoring-edit because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

RoleCreated

Created Role.rbac.authorization.k8s.io/user-workload-monitoring-config-edit -n openshift-user-workload-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

RoleBindingCreated

Created RoleBinding.rbac.authorization.k8s.io/metrics-server-auth-reader -n kube-system because it was missing

openshift-monitoring

deployment-controller

kube-state-metrics

ScalingReplicaSet

Scaled up replica set kube-state-metrics-7bbc969446 to 1

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

RoleBindingCreated

Created RoleBinding.rbac.authorization.k8s.io/metrics-server-auth-reader -n kube-system because it was missing

openshift-marketplace

kubelet

redhat-marketplace-hqqrk

Started

Started container registry-server

openshift-marketplace

kubelet

redhat-marketplace-hqqrk

Created

Created container: registry-server

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

RoleCreated

Created Role.rbac.authorization.k8s.io/user-workload-monitoring-config-edit -n openshift-user-workload-monitoring because it was missing

openshift-monitoring

replicaset-controller

kube-state-metrics-7bbc969446

SuccessfulCreate

Created pod: kube-state-metrics-7bbc969446-qh6vq

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ClusterRoleCreated

Created ClusterRole.rbac.authorization.k8s.io/monitoring-edit because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ClusterRoleBindingCreated

Created ClusterRoleBinding.rbac.authorization.k8s.io/metrics-server:system:auth-delegator because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ClusterRoleCreated

Created ClusterRole.rbac.authorization.k8s.io/monitoring-rules-view because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ClusterRoleBindingCreated

Created ClusterRoleBinding.rbac.authorization.k8s.io/system:metrics-server because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

SecretCreated

Created Secret/telemeter-client-kube-rbac-proxy-config -n openshift-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

RoleCreated

Created Role.rbac.authorization.k8s.io/monitoring-alertmanager-api-reader -n openshift-user-workload-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

SecretCreated

Created Secret/telemeter-client -n openshift-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ClusterRoleCreated

Created ClusterRole.rbac.authorization.k8s.io/monitoring-rules-edit because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ServiceCreated

Created Service/telemeter-client -n openshift-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ClusterRoleBindingCreated

Created ClusterRoleBinding.rbac.authorization.k8s.io/telemeter-client-view because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ClusterRoleCreated

Created ClusterRole.rbac.authorization.k8s.io/system:metrics-server because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

RoleCreated

Created Role.rbac.authorization.k8s.io/monitoring-alertmanager-api-writer -n openshift-user-workload-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ClusterRoleCreated

Created ClusterRole.rbac.authorization.k8s.io/pod-metrics-reader because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ConfigMapCreated

Created ConfigMap/kube-state-metrics-custom-resource-state-configmap -n openshift-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ServiceCreated

Created Service/kube-state-metrics -n openshift-monitoring because it was missing

openshift-monitoring

kubelet

openshift-state-metrics-5dc6c74576-j8tmv

Created

Created container: kube-rbac-proxy-self

openshift-monitoring

kubelet

kube-state-metrics-7bbc969446-qh6vq

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f264240fe2a46d7aa95e56ee202a8403c3dad6c220cf29caff0936c82e0c086f"

openshift-monitoring

multus

openshift-state-metrics-5dc6c74576-j8tmv

AddedInterface

Add eth0 [10.128.0.75/23] from ovn-kubernetes

openshift-monitoring

kubelet

openshift-state-metrics-5dc6c74576-j8tmv

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-monitoring

multus

kube-state-metrics-7bbc969446-qh6vq

AddedInterface

Add eth0 [10.128.0.76/23] from ovn-kubernetes

openshift-marketplace

kubelet

redhat-operators-jstrn

Started

Started container registry-server

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

RoleCreated

Created Role.rbac.authorization.k8s.io/monitoring-alertmanager-edit -n openshift-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

RoleCreated

Created Role.rbac.authorization.k8s.io/monitoring-alertmanager-view -n openshift-monitoring because it was missing

openshift-monitoring

multus

openshift-state-metrics-5dc6c74576-j8tmv

AddedInterface

Add eth0 [10.128.0.75/23] from ovn-kubernetes

openshift-monitoring

kubelet

openshift-state-metrics-5dc6c74576-j8tmv

Started

Started container kube-rbac-proxy-self

openshift-monitoring

kubelet

openshift-state-metrics-5dc6c74576-j8tmv

Created

Created container: kube-rbac-proxy-main

openshift-monitoring

multus

kube-state-metrics-7bbc969446-qh6vq

AddedInterface

Add eth0 [10.128.0.76/23] from ovn-kubernetes

openshift-monitoring

kubelet

openshift-state-metrics-5dc6c74576-j8tmv

Started

Started container kube-rbac-proxy-main

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

RoleCreated

Created Role.rbac.authorization.k8s.io/monitoring-alertmanager-view -n openshift-monitoring because it was missing

openshift-monitoring

kubelet

openshift-state-metrics-5dc6c74576-j8tmv

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-monitoring

kubelet

openshift-state-metrics-5dc6c74576-j8tmv

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:96ae39329a45e017d3444b3794dc95126641ca54fe645bb8729b3d501bd47c64"

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

RoleCreated

Created Role.rbac.authorization.k8s.io/monitoring-alertmanager-edit -n openshift-monitoring because it was missing

openshift-monitoring

kubelet

kube-state-metrics-7bbc969446-qh6vq

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f264240fe2a46d7aa95e56ee202a8403c3dad6c220cf29caff0936c82e0c086f"

openshift-monitoring

kubelet

openshift-state-metrics-5dc6c74576-j8tmv

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-monitoring

kubelet

openshift-state-metrics-5dc6c74576-j8tmv

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:96ae39329a45e017d3444b3794dc95126641ca54fe645bb8729b3d501bd47c64"

openshift-monitoring

kubelet

openshift-state-metrics-5dc6c74576-j8tmv

Started

Started container kube-rbac-proxy-self

openshift-marketplace

kubelet

redhat-operators-jstrn

Created

Created container: registry-server

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ServiceCreated

Created Service/thanos-querier -n openshift-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ServiceCreated

Created Service/metrics-server -n openshift-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

RoleCreated

Created Role.rbac.authorization.k8s.io/cluster-monitoring-metrics-api -n openshift-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

SecretCreated

Created Secret/grpc-tls -n openshift-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ServiceCreated

Created Service/thanos-querier -n openshift-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ServiceCreated

Created Service/metrics-server -n openshift-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

RoleCreated

Created Role.rbac.authorization.k8s.io/cluster-monitoring-metrics-api -n openshift-monitoring because it was missing

openshift-monitoring

kubelet

openshift-state-metrics-5dc6c74576-j8tmv

Started

Started container kube-rbac-proxy-main

openshift-monitoring

kubelet

openshift-state-metrics-5dc6c74576-j8tmv

Created

Created container: kube-rbac-proxy-main

openshift-monitoring

kubelet

openshift-state-metrics-5dc6c74576-j8tmv

Created

Created container: kube-rbac-proxy-self

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

SecretCreated

Created Secret/grpc-tls -n openshift-monitoring because it was missing

openshift-monitoring

kubelet

openshift-state-metrics-5dc6c74576-j8tmv

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-monitoring

kubelet

node-exporter-lb4t5

Started

Started container init-textfile

openshift-monitoring

kubelet

node-exporter-lb4t5

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:0bccc03fd9ffe278e15c8f4be1db030307e4cd5020b78d711fc62f104fd6a980" in 1.639s (1.639s including waiting). Image size: 417688124 bytes.

openshift-monitoring

kubelet

node-exporter-lb4t5

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:0bccc03fd9ffe278e15c8f4be1db030307e4cd5020b78d711fc62f104fd6a980" in 1.639s (1.639s including waiting). Image size: 417688124 bytes.

openshift-monitoring

kubelet

node-exporter-lb4t5

Created

Created container: init-textfile

openshift-monitoring

kubelet

node-exporter-lb4t5

Started

Started container init-textfile

openshift-monitoring

kubelet

node-exporter-lb4t5

Created

Created container: init-textfile

openshift-monitoring

kubelet

kube-state-metrics-7bbc969446-qh6vq

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-monitoring

kubelet

node-exporter-lb4t5

Started

Started container node-exporter

openshift-monitoring

kubelet

kube-state-metrics-7bbc969446-qh6vq

Started

Started container kube-state-metrics

openshift-monitoring

kubelet

node-exporter-lb4t5

Started

Started container kube-rbac-proxy

openshift-monitoring

kubelet

kube-state-metrics-7bbc969446-qh6vq

Created

Created container: kube-state-metrics

openshift-monitoring

kubelet

kube-state-metrics-7bbc969446-qh6vq

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f264240fe2a46d7aa95e56ee202a8403c3dad6c220cf29caff0936c82e0c086f" in 2.291s (2.291s including waiting). Image size: 440559529 bytes.

openshift-monitoring

kubelet

node-exporter-lb4t5

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:0bccc03fd9ffe278e15c8f4be1db030307e4cd5020b78d711fc62f104fd6a980" already present on machine

openshift-monitoring

kubelet

node-exporter-lb4t5

Created

Created container: kube-rbac-proxy

openshift-monitoring

kubelet

node-exporter-lb4t5

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-monitoring

kubelet

node-exporter-lb4t5

Started

Started container node-exporter

openshift-monitoring

kubelet

node-exporter-lb4t5

Created

Created container: node-exporter

openshift-monitoring

kubelet

kube-state-metrics-7bbc969446-qh6vq

Created

Created container: kube-rbac-proxy-main

openshift-monitoring

kubelet

kube-state-metrics-7bbc969446-qh6vq

Started

Started container kube-rbac-proxy-main

openshift-monitoring

kubelet

openshift-state-metrics-5dc6c74576-j8tmv

Started

Started container openshift-state-metrics

openshift-monitoring

kubelet

node-exporter-lb4t5

Created

Created container: node-exporter

openshift-monitoring

kubelet

kube-state-metrics-7bbc969446-qh6vq

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-monitoring

kubelet

node-exporter-lb4t5

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-monitoring

kubelet

node-exporter-lb4t5

Created

Created container: kube-rbac-proxy

openshift-monitoring

kubelet

node-exporter-lb4t5

Started

Started container kube-rbac-proxy

openshift-monitoring

kubelet

kube-state-metrics-7bbc969446-qh6vq

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f264240fe2a46d7aa95e56ee202a8403c3dad6c220cf29caff0936c82e0c086f" in 2.291s (2.291s including waiting). Image size: 440559529 bytes.

openshift-monitoring

kubelet

kube-state-metrics-7bbc969446-qh6vq

Created

Created container: kube-state-metrics

openshift-monitoring

kubelet

kube-state-metrics-7bbc969446-qh6vq

Started

Started container kube-state-metrics

openshift-monitoring

kubelet

openshift-state-metrics-5dc6c74576-j8tmv

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:96ae39329a45e017d3444b3794dc95126641ca54fe645bb8729b3d501bd47c64" in 2.148s (2.148s including waiting). Image size: 431974228 bytes.

openshift-monitoring

kubelet

openshift-state-metrics-5dc6c74576-j8tmv

Created

Created container: openshift-state-metrics

openshift-monitoring

kubelet

openshift-state-metrics-5dc6c74576-j8tmv

Started

Started container openshift-state-metrics

openshift-monitoring

kubelet

node-exporter-lb4t5

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:0bccc03fd9ffe278e15c8f4be1db030307e4cd5020b78d711fc62f104fd6a980" already present on machine

openshift-monitoring

kubelet

openshift-state-metrics-5dc6c74576-j8tmv

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:96ae39329a45e017d3444b3794dc95126641ca54fe645bb8729b3d501bd47c64" in 2.148s (2.148s including waiting). Image size: 431974228 bytes.

openshift-monitoring

kubelet

kube-state-metrics-7bbc969446-qh6vq

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-monitoring

kubelet

kube-state-metrics-7bbc969446-qh6vq

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-monitoring

kubelet

openshift-state-metrics-5dc6c74576-j8tmv

Created

Created container: openshift-state-metrics

openshift-monitoring

kubelet

kube-state-metrics-7bbc969446-qh6vq

Created

Created container: kube-rbac-proxy-main

openshift-monitoring

kubelet

kube-state-metrics-7bbc969446-qh6vq

Started

Started container kube-rbac-proxy-main

openshift-kube-controller-manager

static-pod-installer

installer-2-master-0

StaticPodInstallerCompleted

Successfully installed revision 2

openshift-monitoring

kubelet

kube-state-metrics-7bbc969446-qh6vq

Started

Started container kube-rbac-proxy-self

openshift-kube-controller-manager

kubelet

kube-controller-manager-master-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1fbbcb390de2563a0177b92fba1b5a65777366e2dc80e2808b61d87c41b47a2d" already present on machine

kube-system

kubelet

bootstrap-kube-controller-manager-master-0

Killing

Stopping container kube-controller-manager

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ConfigMapCreated

Created ConfigMap/metrics-server-audit-profiles -n openshift-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ConfigMapCreated

Created ConfigMap/metrics-server-audit-profiles -n openshift-monitoring because it was missing

kube-system

kubelet

bootstrap-kube-controller-manager-master-0

Killing

Stopping container cluster-policy-controller

openshift-kube-controller-manager

kubelet

kube-controller-manager-master-0

Started

Started container kube-controller-manager

openshift-monitoring

kubelet

kube-state-metrics-7bbc969446-qh6vq

Created

Created container: kube-rbac-proxy-self

openshift-monitoring

kubelet

kube-state-metrics-7bbc969446-qh6vq

Started

Started container kube-rbac-proxy-self

openshift-kube-controller-manager

kubelet

kube-controller-manager-master-0

Created

Created container: kube-controller-manager

openshift-kube-controller-manager

kubelet

kube-controller-manager-master-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b23c544d3894e5b31f66a18c554f03b0d29f92c2000c46b57b1c96da7ec25db9" already present on machine

openshift-monitoring

kubelet

kube-state-metrics-7bbc969446-qh6vq

Created

Created container: kube-rbac-proxy-self

openshift-kube-controller-manager

cluster-policy-controller

cluster-policy-controller-lock

LeaderElection

master-0_6aec5505-4a23-4987-b1f0-6d61a746b83c became leader

openshift-kube-controller-manager

kubelet

kube-controller-manager-master-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c032f87ae61d6f757ff3ce52620a70a43516591987731f25da77aba152f17458" already present on machine

openshift-kube-controller-manager

kubelet

kube-controller-manager-master-0

Started

Started container kube-controller-manager-cert-syncer

openshift-kube-controller-manager

kubelet

kube-controller-manager-master-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c032f87ae61d6f757ff3ce52620a70a43516591987731f25da77aba152f17458" already present on machine

openshift-kube-controller-manager

cluster-policy-controller

kube-controller-manager-master-0

ControlPlaneTopology

unable to get control plane topology, using HA cluster values for leader election: infrastructures.config.openshift.io "cluster" is forbidden: User "system:kube-controller-manager" cannot get resource "infrastructures" in API group "config.openshift.io" at the cluster scope

openshift-kube-controller-manager

kubelet

kube-controller-manager-master-0

Created

Created container: kube-controller-manager-recovery-controller

openshift-kube-controller-manager

kubelet

kube-controller-manager-master-0

Started

Started container kube-controller-manager-recovery-controller

openshift-kube-controller-manager

kubelet

kube-controller-manager-master-0

Created

Created container: kube-controller-manager-cert-syncer

openshift-kube-controller-manager

kubelet

kube-controller-manager-master-0

Created

Created container: cluster-policy-controller

openshift-kube-controller-manager

kubelet

kube-controller-manager-master-0

Started

Started container cluster-policy-controller

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

SecretCreated

Created Secret/metrics-server-40c28rqu4fltf -n openshift-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

SecretCreated

Created Secret/metrics-server-40c28rqu4fltf -n openshift-monitoring because it was missing

openshift-marketplace

kubelet

redhat-operators-jstrn

Unhealthy

Startup probe failed: timeout: failed to connect service ":50051" within 1s

openshift-kube-apiserver-operator

kube-apiserver-operator-status-controller-statussyncer_kube-apiserver

openshift-kube-apiserver-operator

OperatorStatusChanged

Status for clusteroperator/kube-apiserver changed: Degraded message changed from "NodeControllerDegraded: All master nodes are ready" to "NodeControllerDegraded: All master nodes are ready\nNodeInstallerDegraded: 1 nodes are failing on revision 1:\nNodeInstallerDegraded: installer: ving-cert\",\nNodeInstallerDegraded: (string) (len=21) \"user-serving-cert-000\",\nNodeInstallerDegraded: (string) (len=21) \"user-serving-cert-001\",\nNodeInstallerDegraded: (string) (len=21) \"user-serving-cert-002\",\nNodeInstallerDegraded: (string) (len=21) \"user-serving-cert-003\",\nNodeInstallerDegraded: (string) (len=21) \"user-serving-cert-004\",\nNodeInstallerDegraded: (string) (len=21) \"user-serving-cert-005\",\nNodeInstallerDegraded: (string) (len=21) \"user-serving-cert-006\",\nNodeInstallerDegraded: (string) (len=21) \"user-serving-cert-007\",\nNodeInstallerDegraded: (string) (len=21) \"user-serving-cert-008\",\nNodeInstallerDegraded: (string) (len=21) \"user-serving-cert-009\"\nNodeInstallerDegraded: },\nNodeInstallerDegraded: CertConfigMapNamePrefixes: ([]string) (len=4 cap=4) {\nNodeInstallerDegraded: (string) (len=20) \"aggregator-client-ca\",\nNodeInstallerDegraded: (string) (len=9) \"client-ca\",\nNodeInstallerDegraded: (string) (len=29) \"control-plane-node-kubeconfig\",\nNodeInstallerDegraded: (string) (len=26) \"check-endpoints-kubeconfig\"\nNodeInstallerDegraded: OptionalCertConfigMapNamePrefixes: ([]string) (len=1 cap=1) {\nNodeInstallerDegraded: (string) (len=17) \"trusted-ca-bundle\"\nNodeInstallerDegraded: CertDir: (string) (len=57) \"/etc/kubernetes/static-pod-resources/kube-apiserver-certs\",\nNodeInstallerDegraded: ResourceDir: (string) (len=36) \"/etc/kubernetes/static-pod-resources\",\nNodeInstallerDegraded: PodManifestDir: (string) (len=25) \"/etc/kubernetes/manifests\",\nNodeInstallerDegraded: Timeout: (time.Duration) 2m0s,\nNodeInstallerDegraded: StaticPodManifestsLockFile: (string) \"\",\nNodeInstallerDegraded: PodMutationFns: ([]installerpod.PodMutationFunc) <nil>,\nNodeInstallerDegraded: KubeletVersion: (string) \"\"\nNodeInstallerDegraded: })\nNodeInstallerDegraded: I0320 08:37:20.888074 1 cmd.go:413] Getting controller reference for node master-0\nNodeInstallerDegraded: I0320 08:37:20.919755 1 cmd.go:426] Waiting for installer revisions to settle for node master-0\nNodeInstallerDegraded: I0320 08:37:20.919828 1 envvar.go:172] \"Feature gate default state\" feature=\"WatchListClient\" enabled=false\nNodeInstallerDegraded: I0320 08:37:20.919840 1 envvar.go:172] \"Feature gate default state\" feature=\"InformerResourceVersion\" enabled=false\nNodeInstallerDegraded: I0320 08:37:20.938866 1 cmd.go:518] Waiting additional period after revisions have settled for node master-0\nNodeInstallerDegraded: I0320 08:37:50.939911 1 cmd.go:524] Getting installer pods for node master-0\nNodeInstallerDegraded: F0320 08:38:04.943427 1 cmd.go:109] Get \"https://172.30.0.1:443/api/v1/namespaces/openshift-kube-apiserver/pods?labelSelector=app%3Dinstaller\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)\nNodeInstallerDegraded: "
(x2)

openshift-kube-apiserver-operator

kube-apiserver-operator-config-observer-configobserver

openshift-kube-apiserver-operator

ObservedConfigChanged

Writing updated observed config:   map[string]any{    "admission": map[string]any{"pluginConfig": map[string]any{"PodSecurity": map[string]any{"configuration": map[string]any{"defaults": map[string]any{"audit": string("restricted"), "audit-version": string("latest"), "enforce": string("privileged"), "enforce-version": string("latest"), ...}}}, "network.openshift.io/ExternalIPRanger": map[string]any{"configuration": map[string]any{"allowIngressIP": bool(false), "apiVersion": string("network.openshift.io/v1"), "kind": string("ExternalIPRangerAdmissionConfig")}}, "network.openshift.io/RestrictedEndpointsAdmission": map[string]any{"configuration": map[string]any{"apiVersion": string("network.openshift.io/v1"), "kind": string("RestrictedEndpointsAdmissionConfig"), "restrictedCIDRs": []any{string("10.128.0.0/16"), string("172.30.0.0/16")}}}}},    "apiServerArguments": map[string]any{    "api-audiences": []any{string("https://kubernetes.default.svc")}, +  "authentication-token-webhook-config-file": []any{ +  string("/etc/kubernetes/static-pod-resources/secrets/webhook-authenticator/kubeConfig"), +  }, +  "authentication-token-webhook-version": []any{string("v1")},    "etcd-servers": []any{string("https://192.168.32.10:2379"), string("https://localhost:2379")},    "feature-gates": []any{string("AWSEFSDriverVolumeMetrics=true"), string("AdminNetworkPolicy=true"), string("AlibabaPlatform=true"), string("AzureWorkloadIdentity=true"), ...},    ... // 6 identical entries    },    "corsAllowedOrigins": []any{string(`//127\.0\.0\.1(:|$)`), string("//localhost(:|$)")},    "gracefulTerminationDuration": string("15"),    ... // 2 identical entries   }
(x2)

openshift-kube-apiserver-operator

kube-apiserver-operator-config-observer-configobserver

openshift-kube-apiserver-operator

ObserveWebhookTokenAuthenticator

authentication-token webhook configuration status changed from false to true

openshift-kube-apiserver-operator

kube-apiserver-operator-status-controller-statussyncer_kube-apiserver

openshift-kube-apiserver-operator

OperatorStatusChanged

Status for clusteroperator/kube-apiserver changed: Degraded message changed from "NodeControllerDegraded: All master nodes are ready\nNodeInstallerDegraded: 1 nodes are failing on revision 1:\nNodeInstallerDegraded: installer: ving-cert\",\nNodeInstallerDegraded: (string) (len=21) \"user-serving-cert-000\",\nNodeInstallerDegraded: (string) (len=21) \"user-serving-cert-001\",\nNodeInstallerDegraded: (string) (len=21) \"user-serving-cert-002\",\nNodeInstallerDegraded: (string) (len=21) \"user-serving-cert-003\",\nNodeInstallerDegraded: (string) (len=21) \"user-serving-cert-004\",\nNodeInstallerDegraded: (string) (len=21) \"user-serving-cert-005\",\nNodeInstallerDegraded: (string) (len=21) \"user-serving-cert-006\",\nNodeInstallerDegraded: (string) (len=21) \"user-serving-cert-007\",\nNodeInstallerDegraded: (string) (len=21) \"user-serving-cert-008\",\nNodeInstallerDegraded: (string) (len=21) \"user-serving-cert-009\"\nNodeInstallerDegraded: },\nNodeInstallerDegraded: CertConfigMapNamePrefixes: ([]string) (len=4 cap=4) {\nNodeInstallerDegraded: (string) (len=20) \"aggregator-client-ca\",\nNodeInstallerDegraded: (string) (len=9) \"client-ca\",\nNodeInstallerDegraded: (string) (len=29) \"control-plane-node-kubeconfig\",\nNodeInstallerDegraded: (string) (len=26) \"check-endpoints-kubeconfig\"\nNodeInstallerDegraded: OptionalCertConfigMapNamePrefixes: ([]string) (len=1 cap=1) {\nNodeInstallerDegraded: (string) (len=17) \"trusted-ca-bundle\"\nNodeInstallerDegraded: CertDir: (string) (len=57) \"/etc/kubernetes/static-pod-resources/kube-apiserver-certs\",\nNodeInstallerDegraded: ResourceDir: (string) (len=36) \"/etc/kubernetes/static-pod-resources\",\nNodeInstallerDegraded: PodManifestDir: (string) (len=25) \"/etc/kubernetes/manifests\",\nNodeInstallerDegraded: Timeout: (time.Duration) 2m0s,\nNodeInstallerDegraded: StaticPodManifestsLockFile: (string) \"\",\nNodeInstallerDegraded: PodMutationFns: ([]installerpod.PodMutationFunc) <nil>,\nNodeInstallerDegraded: KubeletVersion: (string) \"\"\nNodeInstallerDegraded: })\nNodeInstallerDegraded: I0320 08:37:20.888074 1 cmd.go:413] Getting controller reference for node master-0\nNodeInstallerDegraded: I0320 08:37:20.919755 1 cmd.go:426] Waiting for installer revisions to settle for node master-0\nNodeInstallerDegraded: I0320 08:37:20.919828 1 envvar.go:172] \"Feature gate default state\" feature=\"WatchListClient\" enabled=false\nNodeInstallerDegraded: I0320 08:37:20.919840 1 envvar.go:172] \"Feature gate default state\" feature=\"InformerResourceVersion\" enabled=false\nNodeInstallerDegraded: I0320 08:37:20.938866 1 cmd.go:518] Waiting additional period after revisions have settled for node master-0\nNodeInstallerDegraded: I0320 08:37:50.939911 1 cmd.go:524] Getting installer pods for node master-0\nNodeInstallerDegraded: F0320 08:38:04.943427 1 cmd.go:109] Get \"https://172.30.0.1:443/api/v1/namespaces/openshift-kube-apiserver/pods?labelSelector=app%3Dinstaller\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)\nNodeInstallerDegraded: " to "NodeControllerDegraded: All master nodes are ready",Progressing message changed from "NodeInstallerProgressing: 1 node is at revision 0; 0 nodes have achieved new revision 1" to "NodeInstallerProgressing: 1 node is at revision 0; 0 nodes have achieved new revision 2",Available message changed from "StaticPodsAvailable: 0 nodes are active; 1 node is at revision 0; 0 nodes have achieved new revision 1" to "StaticPodsAvailable: 0 nodes are active; 1 node is at revision 0; 0 nodes have achieved new revision 2"

openshift-kube-scheduler

static-pod-installer

openshift-kube-scheduler

StaticPodInstallerCompleted

Successfully installed revision 2

kube-system

kubelet

bootstrap-kube-scheduler-master-0

Killing

Stopping container kube-scheduler

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-status-controller-statussyncer_kube-scheduler

openshift-kube-scheduler-operator

OperatorVersionChanged

clusteroperator/kube-scheduler version "kube-scheduler" changed from "" to "1.31.14"

openshift-kube-scheduler

kubelet

openshift-kube-scheduler-master-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b23c544d3894e5b31f66a18c554f03b0d29f92c2000c46b57b1c96da7ec25db9" already present on machine

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-status-controller-statussyncer_kube-scheduler

openshift-kube-scheduler-operator

OperatorStatusChanged

Status for clusteroperator/kube-scheduler changed: status.versions changed from [{"raw-internal" "4.18.35"}] to [{"raw-internal" "4.18.35"} {"kube-scheduler" "1.31.14"} {"operator" "4.18.35"}]

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-status-controller-statussyncer_kube-scheduler

openshift-kube-scheduler-operator

OperatorVersionChanged

clusteroperator/kube-scheduler version "operator" changed from "" to "4.18.35"

openshift-kube-scheduler

kubelet

openshift-kube-scheduler-master-0

Started

Started container wait-for-host-port

openshift-kube-scheduler

kubelet

openshift-kube-scheduler-master-0

Created

Created container: wait-for-host-port

openshift-kube-scheduler

kubelet

openshift-kube-scheduler-master-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ef199844317b7b012879ed8d29f9b6bc37fad8a6fdb336103cbd5cabc74c4302" already present on machine

openshift-kube-scheduler

kubelet

openshift-kube-scheduler-master-0

Created

Created container: kube-scheduler

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

openshift-kube-apiserver-operator

StartingNewRevision

new revision 3 triggered by "required configmap/config has changed"

openshift-kube-apiserver-operator

kube-apiserver-operator-installer-controller

openshift-kube-apiserver-operator

PodCreated

Created Pod/installer-2-master-0 -n openshift-kube-apiserver because it was missing

openshift-kube-scheduler

kubelet

openshift-kube-scheduler-master-0

Started

Started container kube-scheduler

openshift-kube-scheduler

kubelet

openshift-kube-scheduler-master-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b23c544d3894e5b31f66a18c554f03b0d29f92c2000c46b57b1c96da7ec25db9" already present on machine

openshift-kube-scheduler

kubelet

openshift-kube-scheduler-master-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ef199844317b7b012879ed8d29f9b6bc37fad8a6fdb336103cbd5cabc74c4302" already present on machine

openshift-kube-apiserver

kubelet

installer-2-master-0

Started

Started container installer

openshift-kube-apiserver

kubelet

installer-2-master-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c5ce3d1134d6500e2b8528516c1889d7bbc6259aba4981c6983395b0e9eeff65" already present on machine

openshift-kube-scheduler

kubelet

openshift-kube-scheduler-master-0

Created

Created container: kube-scheduler-cert-syncer

openshift-kube-scheduler

kubelet

openshift-kube-scheduler-master-0

Started

Started container kube-scheduler-recovery-controller

openshift-kube-scheduler

kubelet

openshift-kube-scheduler-master-0

Created

Created container: kube-scheduler-recovery-controller

openshift-kube-scheduler

default-scheduler

kube-scheduler

LeaderElection

master-0_56d86407-0345-4d04-82d0-eeab60a13e88 became leader

openshift-kube-scheduler

kubelet

openshift-kube-scheduler-master-0

Started

Started container kube-scheduler-cert-syncer

openshift-kube-apiserver

multus

installer-2-master-0

AddedInterface

Add eth0 [10.128.0.77/23] from ovn-kubernetes

openshift-kube-apiserver

kubelet

installer-2-master-0

Created

Created container: installer

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

openshift-kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/kube-apiserver-pod-3 -n openshift-kube-apiserver because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

openshift-kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/config-3 -n openshift-kube-apiserver because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

openshift-kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/kube-apiserver-cert-syncer-kubeconfig-3 -n openshift-kube-apiserver because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

openshift-kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/bound-sa-token-signing-certs-3 -n openshift-kube-apiserver because it was missing

openshift-monitoring

deployment-controller

metrics-server

ScalingReplicaSet

Scaled up replica set metrics-server-64c67d44c4 to 1

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-revisioncontroller

openshift-kube-scheduler-operator

ConfigMapCreated

Created ConfigMap/kube-scheduler-pod-3 -n openshift-kube-scheduler because it was missing

kube-system

kube-controller-manager

kube-controller-manager

LeaderElection

master-0_b4a63c7e-12d1-45f7-92d4-9f2524fc0aed became leader

openshift-monitoring

deployment-controller

metrics-server

ScalingReplicaSet

Scaled up replica set metrics-server-64c67d44c4 to 1

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-revisioncontroller

openshift-kube-scheduler-operator

ConfigMapCreated

Created ConfigMap/serviceaccount-ca-3 -n openshift-kube-scheduler because it was missing

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-revisioncontroller

openshift-kube-scheduler-operator

ConfigMapCreated

Created ConfigMap/config-3 -n openshift-kube-scheduler because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

openshift-kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/etcd-serving-ca-3 -n openshift-kube-apiserver because it was missing

openshift-monitoring

replicaset-controller

metrics-server-64c67d44c4

SuccessfulCreate

Created pod: metrics-server-64c67d44c4-s7vfs

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-revisioncontroller

openshift-kube-scheduler-operator

StartingNewRevision

new revision 3 triggered by "required secret/localhost-recovery-client-token has changed"

default

node-controller

master-0

RegisteredNode

Node master-0 event: Registered Node master-0 in Controller

openshift-monitoring

replicaset-controller

metrics-server-64c67d44c4

SuccessfulCreate

Created pod: metrics-server-64c67d44c4-s7vfs

openshift-monitoring

multus

metrics-server-64c67d44c4-s7vfs

AddedInterface

Add eth0 [10.128.0.78/23] from ovn-kubernetes

openshift-monitoring

kubelet

metrics-server-64c67d44c4-s7vfs

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:39c122c726d1bf520dd481350fee5ad940762d5d4c9f8c012db6bf56b0ca8757"

openshift-monitoring

kubelet

metrics-server-64c67d44c4-s7vfs

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:39c122c726d1bf520dd481350fee5ad940762d5d4c9f8c012db6bf56b0ca8757"

openshift-monitoring

multus

metrics-server-64c67d44c4-s7vfs

AddedInterface

Add eth0 [10.128.0.78/23] from ovn-kubernetes

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

openshift-kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/kube-apiserver-server-ca-3 -n openshift-kube-apiserver because it was missing

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-revisioncontroller

openshift-kube-scheduler-operator

ConfigMapCreated

Created ConfigMap/scheduler-kubeconfig-3 -n openshift-kube-scheduler because it was missing

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-revisioncontroller

openshift-kube-scheduler-operator

ConfigMapCreated

Created ConfigMap/kube-scheduler-cert-syncer-kubeconfig-3 -n openshift-kube-scheduler because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

openshift-kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/kubelet-serving-ca-3 -n openshift-kube-apiserver because it was missing

openshift-monitoring

kubelet

metrics-server-64c67d44c4-s7vfs

Created

Created container: metrics-server

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-revisioncontroller

openshift-kube-scheduler-operator

SecretCreated

Created Secret/serving-cert-3 -n openshift-kube-scheduler because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

openshift-kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/sa-token-signing-certs-3 -n openshift-kube-apiserver because it was missing

openshift-monitoring

kubelet

metrics-server-64c67d44c4-s7vfs

Started

Started container metrics-server

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-revisioncontroller

openshift-kube-scheduler-operator

SecretCreated

Created Secret/localhost-recovery-client-token-3 -n openshift-kube-scheduler because it was missing

openshift-monitoring

kubelet

metrics-server-64c67d44c4-s7vfs

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:39c122c726d1bf520dd481350fee5ad940762d5d4c9f8c012db6bf56b0ca8757" in 1.444s (1.444s including waiting). Image size: 471431303 bytes.

openshift-monitoring

kubelet

metrics-server-64c67d44c4-s7vfs

Created

Created container: metrics-server

openshift-monitoring

kubelet

metrics-server-64c67d44c4-s7vfs

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:39c122c726d1bf520dd481350fee5ad940762d5d4c9f8c012db6bf56b0ca8757" in 1.444s (1.444s including waiting). Image size: 471431303 bytes.

openshift-monitoring

kubelet

metrics-server-64c67d44c4-s7vfs

Started

Started container metrics-server

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-revisioncontroller

openshift-kube-scheduler-operator

RevisionTriggered

new revision 3 triggered by "required secret/localhost-recovery-client-token has changed"

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

openshift-kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/kube-apiserver-audit-policies-3 -n openshift-kube-apiserver because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

openshift-kube-apiserver-operator

SecretCreated

Created Secret/etcd-client-3 -n openshift-kube-apiserver because it was missing

openshift-machine-config-operator

machine-config-operator

machine-config-operator

ConfigMapUpdated

Updated ConfigMap/kube-rbac-proxy -n openshift-machine-config-operator: cause by changes in data.config-file.yaml

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-status-controller-statussyncer_kube-scheduler

openshift-kube-scheduler-operator

OperatorStatusChanged

Status for clusteroperator/kube-scheduler changed: Progressing message changed from "NodeInstallerProgressing: 1 node is at revision 0; 0 nodes have achieved new revision 2" to "NodeInstallerProgressing: 1 node is at revision 0; 0 nodes have achieved new revision 3",Available message changed from "StaticPodsAvailable: 0 nodes are active; 1 node is at revision 0; 0 nodes have achieved new revision 2" to "StaticPodsAvailable: 0 nodes are active; 1 node is at revision 0; 0 nodes have achieved new revision 3"

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

openshift-kube-apiserver-operator

SecretCreated

Created Secret/localhost-recovery-serving-certkey-3 -n openshift-kube-apiserver because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

openshift-kube-apiserver-operator

SecretCreated

Created Secret/localhost-recovery-client-token-3 -n openshift-kube-apiserver because it was missing

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-installer-controller

openshift-kube-scheduler-operator

PodCreated

Created Pod/installer-3-master-0 -n openshift-kube-scheduler because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

openshift-kube-apiserver-operator

SecretCreated

Created Secret/webhook-authenticator-3 -n openshift-kube-apiserver because it was missing

openshift-kube-scheduler

kubelet

installer-3-master-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ef199844317b7b012879ed8d29f9b6bc37fad8a6fdb336103cbd5cabc74c4302" already present on machine

openshift-kube-scheduler

kubelet

installer-3-master-0

Created

Created container: installer

openshift-kube-scheduler

kubelet

installer-3-master-0

Started

Started container installer

openshift-kube-scheduler

multus

installer-3-master-0

AddedInterface

Add eth0 [10.128.0.79/23] from ovn-kubernetes

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

openshift-kube-apiserver-operator

RevisionTriggered

new revision 3 triggered by "required configmap/config has changed"

openshift-kube-apiserver-operator

kube-apiserver-operator-status-controller-statussyncer_kube-apiserver

openshift-kube-apiserver-operator

OperatorStatusChanged

Status for clusteroperator/kube-apiserver changed: Progressing message changed from "NodeInstallerProgressing: 1 node is at revision 0; 0 nodes have achieved new revision 2" to "NodeInstallerProgressing: 1 node is at revision 0; 0 nodes have achieved new revision 3",Available message changed from "StaticPodsAvailable: 0 nodes are active; 1 node is at revision 0; 0 nodes have achieved new revision 2" to "StaticPodsAvailable: 0 nodes are active; 1 node is at revision 0; 0 nodes have achieved new revision 3"

openshift-kube-apiserver

kubelet

installer-2-master-0

Killing

Stopping container installer

openshift-kube-apiserver-operator

kube-apiserver-operator-installer-controller

openshift-kube-apiserver-operator

PodCreated

Created Pod/installer-3-master-0 -n openshift-kube-apiserver because it was missing

openshift-kube-apiserver

kubelet

installer-3-master-0

Started

Started container installer

openshift-kube-apiserver

kubelet

installer-3-master-0

Created

Created container: installer

openshift-kube-apiserver

multus

installer-3-master-0

AddedInterface

Add eth0 [10.128.0.80/23] from ovn-kubernetes

openshift-kube-apiserver

kubelet

installer-3-master-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c5ce3d1134d6500e2b8528516c1889d7bbc6259aba4981c6983395b0e9eeff65" already present on machine

openshift-machine-config-operator

machine-config-operator

machine-config-operator

ConfigMapUpdated

Updated ConfigMap/kube-rbac-proxy -n openshift-machine-config-operator: cause by changes in data.config-file.yaml

openshift-network-node-identity

master-0_47f28bbd-5bfa-4d79-be20-4069134b6551

ovnkube-identity

LeaderElection

master-0_47f28bbd-5bfa-4d79-be20-4069134b6551 became leader

openshift-kube-scheduler

kubelet

openshift-kube-scheduler-master-0

Killing

Stopping container kube-scheduler-cert-syncer

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

APIServiceCreated

Created APIService.apiregistration.k8s.io/v1beta1.metrics.k8s.io because it was missing

openshift-kube-scheduler

static-pod-installer

installer-3-master-0

StaticPodInstallerCompleted

Successfully installed revision 3

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

APIServiceCreated

Created APIService.apiregistration.k8s.io/v1beta1.metrics.k8s.io because it was missing

openshift-kube-scheduler

kubelet

openshift-kube-scheduler-master-0

Killing

Stopping container kube-scheduler

openshift-kube-scheduler

kubelet

openshift-kube-scheduler-master-0

Killing

Stopping container kube-scheduler-recovery-controller

openshift-kube-scheduler

kubelet

openshift-kube-scheduler-master-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b23c544d3894e5b31f66a18c554f03b0d29f92c2000c46b57b1c96da7ec25db9" already present on machine

openshift-kube-scheduler

kubelet

openshift-kube-scheduler-master-0

Created

Created container: wait-for-host-port

openshift-kube-scheduler

kubelet

openshift-kube-scheduler-master-0

Started

Started container wait-for-host-port

openshift-kube-scheduler

kubelet

openshift-kube-scheduler-master-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ef199844317b7b012879ed8d29f9b6bc37fad8a6fdb336103cbd5cabc74c4302" already present on machine

openshift-kube-scheduler

kubelet

openshift-kube-scheduler-master-0

Started

Started container kube-scheduler

openshift-kube-scheduler

kubelet

openshift-kube-scheduler-master-0

Created

Created container: kube-scheduler

openshift-kube-scheduler

kubelet

openshift-kube-scheduler-master-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b23c544d3894e5b31f66a18c554f03b0d29f92c2000c46b57b1c96da7ec25db9" already present on machine

openshift-kube-scheduler

kubelet

openshift-kube-scheduler-master-0

Created

Created container: kube-scheduler-cert-syncer

openshift-kube-scheduler

kubelet

openshift-kube-scheduler-master-0

Created

Created container: kube-scheduler-recovery-controller

openshift-kube-scheduler

default-scheduler

kube-scheduler

LeaderElection

master-0_340730bb-6556-4f21-bc5b-1a8749acd927 became leader

openshift-kube-scheduler

cert-recovery-controller

openshift-kube-scheduler

ControlPlaneTopology

unable to get control plane topology, using HA cluster values for leader election: Get "https://localhost:6443/apis/config.openshift.io/v1/infrastructures/cluster": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery

openshift-kube-scheduler

kubelet

openshift-kube-scheduler-master-0

Started

Started container kube-scheduler-recovery-controller

openshift-kube-scheduler

kubelet

openshift-kube-scheduler-master-0

Started

Started container kube-scheduler-cert-syncer

openshift-kube-scheduler

kubelet

openshift-kube-scheduler-master-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ef199844317b7b012879ed8d29f9b6bc37fad8a6fdb336103cbd5cabc74c4302" already present on machine

openshift-kube-apiserver

kubelet

bootstrap-kube-apiserver-master-0

ProbeError

Readiness probe error: HTTP probe failed with statuscode: 500 body: [+]ping ok [+]log ok [+]api-openshift-apiserver-available ok [+]api-openshift-oauth-apiserver-available ok [+]informer-sync ok [+]poststarthook/openshift.io-oauth-apiserver-reachable ok [+]poststarthook/start-apiserver-admission-initializer ok [+]poststarthook/quota.openshift.io-clusterquotamapping ok [+]poststarthook/openshift.io-api-request-count-filter ok [+]poststarthook/openshift.io-startkubeinformers ok [+]poststarthook/openshift.io-openshift-apiserver-reachable ok [+]poststarthook/generic-apiserver-start-informers ok [+]poststarthook/priority-and-fairness-config-consumer ok [+]poststarthook/priority-and-fairness-filter ok [+]poststarthook/storage-object-count-tracker-hook ok [+]poststarthook/start-apiextensions-informers ok [+]poststarthook/start-apiextensions-controllers ok [+]poststarthook/crd-informer-synced ok [+]poststarthook/start-system-namespaces-controller ok [+]poststarthook/start-cluster-authentication-info-controller ok [+]poststarthook/start-kube-apiserver-identity-lease-controller ok [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok [+]poststarthook/start-legacy-token-tracking-controller ok [+]poststarthook/start-service-ip-repair-controllers ok [+]poststarthook/rbac/bootstrap-roles ok [+]poststarthook/scheduling/bootstrap-system-priority-classes ok [+]poststarthook/priority-and-fairness-config-producer ok [+]poststarthook/bootstrap-controller ok [+]poststarthook/aggregator-reload-proxy-client-cert ok [+]poststarthook/start-kube-aggregator-informers ok [+]poststarthook/apiservice-status-local-available-controller ok [+]poststarthook/apiservice-status-remote-available-controller ok [+]poststarthook/apiservice-registration-controller ok [+]poststarthook/apiservice-wait-for-first-sync ok [+]poststarthook/apiservice-discovery-controller ok [+]poststarthook/kube-apiserver-autoregistration ok [+]autoregister-completion ok [+]poststarthook/apiservice-openapi-controller ok [+]poststarthook/apiservice-openapiv3-controller ok [-]shutdown failed: reason withheld readyz check failed

openshift-kube-apiserver

static-pod-installer

installer-3-master-0

StaticPodInstallerCompleted

Successfully installed revision 3

openshift-kube-apiserver

kubelet

bootstrap-kube-apiserver-master-0

Killing

Stopping container kube-apiserver

default

apiserver

openshift-kube-apiserver

ShutdownInitiated

Received signal to terminate, becoming unready, but keeping serving

openshift-kube-apiserver

kubelet

bootstrap-kube-apiserver-master-0

Unhealthy

Readiness probe failed: HTTP probe failed with statuscode: 500

openshift-kube-apiserver

kubelet

bootstrap-kube-apiserver-master-0

Killing

Stopping container kube-apiserver-insecure-readyz

default

apiserver

openshift-kube-apiserver

TerminationPreShutdownHooksFinished

All pre-shutdown hooks have been finished

default

apiserver

openshift-kube-apiserver

InFlightRequestsDrained

All non long-running request(s) in-flight have drained

default

apiserver

openshift-kube-apiserver

AfterShutdownDelayDuration

The minimal shutdown duration of 0s finished

default

apiserver

openshift-kube-apiserver

HTTPServerStoppedListening

HTTP Server has stopped listening

openshift-kube-apiserver-operator

kube-apiserver-operator-status-controller-statussyncer_kube-apiserver

openshift-kube-apiserver-operator

OperatorVersionChanged

clusteroperator/kube-apiserver version "operator" changed from "" to "4.18.35"

openshift-kube-apiserver-operator

kube-apiserver-operator-status-controller-statussyncer_kube-apiserver

openshift-kube-apiserver-operator

OperatorVersionChanged

clusteroperator/kube-apiserver version "kube-apiserver" changed from "" to "1.31.14"

openshift-kube-apiserver-operator

kube-apiserver-operator-installer-controller

openshift-kube-apiserver-operator

InstallerPodFailed

Failed to create installer pod for revision 3 count 1 on node "master-0": Get "https://172.30.0.1:443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-3-master-0": dial tcp 172.30.0.1:443: connect: connection refused

openshift-kube-apiserver-operator

kube-apiserver-operator-status-controller-statussyncer_kube-apiserver

openshift-kube-apiserver-operator

OperatorStatusChanged

Status for clusteroperator/kube-apiserver changed: status.versions changed from [{"raw-internal" "4.18.35"}] to [{"raw-internal" "4.18.35"} {"kube-apiserver" "1.31.14"} {"operator" "4.18.35"}]

default

apiserver

openshift-kube-apiserver

TerminationGracefulTerminationFinished

All pending requests processed

openshift-kube-apiserver

apiserver

kube-apiserver-master-0

KubeAPIReadyz

readyz=true
(x10)

openshift-machine-config-operator

machineconfigoperator

machine-config

OperatorDegraded: MachineConfigPoolsFailed

Failed to resync 4.18.35 because: Get "https://172.30.0.1:443/apis/machineconfiguration.openshift.io/v1/machineconfigpools/master": dial tcp 172.30.0.1:443: connect: connection refused

default

kubelet

master-0

NodeAllocatableEnforced

Updated Node Allocatable limit across pods

default

kubelet

master-0

Starting

Starting kubelet.
(x3)

default

kubelet

master-0

NodeHasSufficientMemory

Node master-0 status is now: NodeHasSufficientMemory
(x3)

default

kubelet

master-0

NodeHasNoDiskPressure

Node master-0 status is now: NodeHasNoDiskPressure
(x3)

default

kubelet

master-0

NodeHasSufficientPID

Node master-0 status is now: NodeHasSufficientPID

openshift-cloud-controller-manager-operator

kubelet

cluster-cloud-controller-manager-operator-7dff898856-7vxxr

FailedMount

MountVolume.SetUp failed for volume "cloud-controller-manager-operator-tls" : failed to sync secret cache: timed out waiting for the condition

openshift-monitoring

kubelet

openshift-state-metrics-5dc6c74576-j8tmv

FailedMount

MountVolume.SetUp failed for volume "openshift-state-metrics-kube-rbac-proxy-config" : failed to sync secret cache: timed out waiting for the condition

openshift-machine-api

kubelet

machine-api-operator-6fbb6cf6f9-n8tnn

FailedMount

MountVolume.SetUp failed for volume "images" : failed to sync configmap cache: timed out waiting for the condition

openshift-monitoring

kubelet

openshift-state-metrics-5dc6c74576-j8tmv

FailedMount

MountVolume.SetUp failed for volume "metrics-client-ca" : failed to sync configmap cache: timed out waiting for the condition

openshift-monitoring

kubelet

openshift-state-metrics-5dc6c74576-j8tmv

FailedMount

MountVolume.SetUp failed for volume "openshift-state-metrics-kube-rbac-proxy-config" : failed to sync secret cache: timed out waiting for the condition

openshift-cluster-machine-approver

kubelet

machine-approver-5c6485487f-qb94j

FailedMount

MountVolume.SetUp failed for volume "machine-approver-tls" : failed to sync secret cache: timed out waiting for the condition

openshift-service-ca

kubelet

service-ca-79bc6b8d76-72j8t

FailedMount

MountVolume.SetUp failed for volume "signing-cabundle" : failed to sync configmap cache: timed out waiting for the condition

openshift-service-ca

kubelet

service-ca-79bc6b8d76-72j8t

FailedMount

MountVolume.SetUp failed for volume "signing-key" : failed to sync secret cache: timed out waiting for the condition

openshift-cluster-machine-approver

kubelet

machine-approver-5c6485487f-qb94j

FailedMount

MountVolume.SetUp failed for volume "config" : failed to sync configmap cache: timed out waiting for the condition

openshift-cluster-machine-approver

kubelet

machine-approver-5c6485487f-qb94j

FailedMount

MountVolume.SetUp failed for volume "auth-proxy-config" : failed to sync configmap cache: timed out waiting for the condition

openshift-monitoring

kubelet

prometheus-operator-6c8df6d4b-5m857

FailedMount

MountVolume.SetUp failed for volume "prometheus-operator-tls" : failed to sync secret cache: timed out waiting for the condition

openshift-monitoring

kubelet

prometheus-operator-6c8df6d4b-5m857

FailedMount

MountVolume.SetUp failed for volume "metrics-client-ca" : failed to sync configmap cache: timed out waiting for the condition

openshift-monitoring

kubelet

prometheus-operator-6c8df6d4b-5m857

FailedMount

MountVolume.SetUp failed for volume "prometheus-operator-kube-rbac-proxy-config" : failed to sync secret cache: timed out waiting for the condition

openshift-monitoring

kubelet

openshift-state-metrics-5dc6c74576-j8tmv

FailedMount

MountVolume.SetUp failed for volume "openshift-state-metrics-tls" : failed to sync secret cache: timed out waiting for the condition

openshift-monitoring

kubelet

metrics-server-64c67d44c4-s7vfs

FailedMount

MountVolume.SetUp failed for volume "secret-metrics-server-tls" : failed to sync secret cache: timed out waiting for the condition

openshift-monitoring

kubelet

node-exporter-lb4t5

FailedMount

MountVolume.SetUp failed for volume "node-exporter-kube-rbac-proxy-config" : failed to sync secret cache: timed out waiting for the condition

openshift-machine-api

kubelet

machine-api-operator-6fbb6cf6f9-n8tnn

FailedMount

MountVolume.SetUp failed for volume "machine-api-operator-tls" : failed to sync secret cache: timed out waiting for the condition

openshift-machine-config-operator

kubelet

machine-config-server-gj4pm

FailedMount

MountVolume.SetUp failed for volume "certs" : failed to sync secret cache: timed out waiting for the condition

openshift-machine-config-operator

kubelet

machine-config-server-gj4pm

FailedMount

MountVolume.SetUp failed for volume "node-bootstrap-token" : failed to sync secret cache: timed out waiting for the condition

openshift-route-controller-manager

kubelet

route-controller-manager-56f686584b-fdcx5

FailedMount

MountVolume.SetUp failed for volume "client-ca" : failed to sync configmap cache: timed out waiting for the condition

openshift-route-controller-manager

kubelet

route-controller-manager-56f686584b-fdcx5

FailedMount

MountVolume.SetUp failed for volume "serving-cert" : failed to sync secret cache: timed out waiting for the condition

openshift-monitoring

kubelet

metrics-server-64c67d44c4-s7vfs

FailedMount

MountVolume.SetUp failed for volume "secret-metrics-client-certs" : failed to sync secret cache: timed out waiting for the condition

openshift-monitoring

kubelet

metrics-server-64c67d44c4-s7vfs

FailedMount

MountVolume.SetUp failed for volume "metrics-server-audit-profiles" : failed to sync configmap cache: timed out waiting for the condition

openshift-monitoring

kubelet

metrics-server-64c67d44c4-s7vfs

FailedMount

MountVolume.SetUp failed for volume "client-ca-bundle" : failed to sync secret cache: timed out waiting for the condition

openshift-monitoring

kubelet

node-exporter-lb4t5

FailedMount

MountVolume.SetUp failed for volume "node-exporter-tls" : failed to sync secret cache: timed out waiting for the condition

openshift-monitoring

kubelet

metrics-server-64c67d44c4-s7vfs

FailedMount

MountVolume.SetUp failed for volume "configmap-kubelet-serving-ca-bundle" : failed to sync configmap cache: timed out waiting for the condition

openshift-monitoring

kubelet

node-exporter-lb4t5

FailedMount

MountVolume.SetUp failed for volume "metrics-client-ca" : failed to sync configmap cache: timed out waiting for the condition

openshift-cloud-credential-operator

kubelet

cloud-credential-operator-744f9dbf77-r4qvh

FailedMount

MountVolume.SetUp failed for volume "cloud-credential-operator-serving-cert" : failed to sync secret cache: timed out waiting for the condition

openshift-cloud-credential-operator

kubelet

cloud-credential-operator-744f9dbf77-r4qvh

FailedMount

MountVolume.SetUp failed for volume "cco-trusted-ca" : failed to sync configmap cache: timed out waiting for the condition

openshift-dns

kubelet

dns-default-v5h69

FailedMount

MountVolume.SetUp failed for volume "metrics-tls" : failed to sync secret cache: timed out waiting for the condition

openshift-operator-lifecycle-manager

kubelet

packageserver-6c85f64bb9-fmpsg

FailedMount

MountVolume.SetUp failed for volume "apiservice-cert" : failed to sync secret cache: timed out waiting for the condition

openshift-operator-lifecycle-manager

kubelet

packageserver-6c85f64bb9-fmpsg

FailedMount

MountVolume.SetUp failed for volume "webhook-cert" : failed to sync secret cache: timed out waiting for the condition

openshift-monitoring

kubelet

metrics-server-64c67d44c4-s7vfs

FailedMount

MountVolume.SetUp failed for volume "configmap-kubelet-serving-ca-bundle" : failed to sync configmap cache: timed out waiting for the condition

openshift-monitoring

kubelet

metrics-server-64c67d44c4-s7vfs

FailedMount

MountVolume.SetUp failed for volume "client-ca-bundle" : failed to sync secret cache: timed out waiting for the condition

openshift-monitoring

kubelet

metrics-server-64c67d44c4-s7vfs

FailedMount

MountVolume.SetUp failed for volume "metrics-server-audit-profiles" : failed to sync configmap cache: timed out waiting for the condition

openshift-route-controller-manager

kubelet

route-controller-manager-56f686584b-fdcx5

FailedMount

MountVolume.SetUp failed for volume "config" : failed to sync configmap cache: timed out waiting for the condition

openshift-cloud-controller-manager-operator

kubelet

cluster-cloud-controller-manager-operator-7dff898856-7vxxr

FailedMount

MountVolume.SetUp failed for volume "images" : failed to sync configmap cache: timed out waiting for the condition

openshift-monitoring

kubelet

metrics-server-64c67d44c4-s7vfs

FailedMount

MountVolume.SetUp failed for volume "secret-metrics-client-certs" : failed to sync secret cache: timed out waiting for the condition

openshift-cloud-controller-manager-operator

kubelet

cluster-cloud-controller-manager-operator-7dff898856-7vxxr

FailedMount

MountVolume.SetUp failed for volume "auth-proxy-config" : failed to sync configmap cache: timed out waiting for the condition

openshift-machine-config-operator

kubelet

machine-config-daemon-9t8x6

FailedMount

MountVolume.SetUp failed for volume "proxy-tls" : failed to sync secret cache: timed out waiting for the condition

openshift-monitoring

kubelet

node-exporter-lb4t5

FailedMount

MountVolume.SetUp failed for volume "node-exporter-kube-rbac-proxy-config" : failed to sync secret cache: timed out waiting for the condition

openshift-machine-config-operator

kubelet

machine-config-controller-b4f87c5b9-pj7rj

FailedMount

MountVolume.SetUp failed for volume "proxy-tls" : failed to sync secret cache: timed out waiting for the condition

openshift-cluster-samples-operator

kubelet

cluster-samples-operator-85f7577d78-mwfgx

FailedMount

MountVolume.SetUp failed for volume "samples-operator-tls" : failed to sync secret cache: timed out waiting for the condition

openshift-monitoring

kubelet

node-exporter-lb4t5

FailedMount

MountVolume.SetUp failed for volume "node-exporter-tls" : failed to sync secret cache: timed out waiting for the condition

openshift-machine-api

kubelet

machine-api-operator-6fbb6cf6f9-n8tnn

FailedMount

MountVolume.SetUp failed for volume "images" : failed to sync configmap cache: timed out waiting for the condition

openshift-machine-api

kubelet

machine-api-operator-6fbb6cf6f9-n8tnn

FailedMount

MountVolume.SetUp failed for volume "config" : failed to sync configmap cache: timed out waiting for the condition

openshift-machine-api

kubelet

machine-api-operator-6fbb6cf6f9-n8tnn

FailedMount

MountVolume.SetUp failed for volume "machine-api-operator-tls" : failed to sync secret cache: timed out waiting for the condition

openshift-controller-manager

kubelet

controller-manager-fc56bb77c-qd4sn

FailedMount

MountVolume.SetUp failed for volume "serving-cert" : failed to sync secret cache: timed out waiting for the condition

openshift-machine-api

kubelet

control-plane-machine-set-operator-6f97756bc8-7t5qv

FailedMount

MountVolume.SetUp failed for volume "control-plane-machine-set-operator-tls" : failed to sync secret cache: timed out waiting for the condition

openshift-machine-api

kubelet

control-plane-machine-set-operator-6f97756bc8-7t5qv

FailedMount

MountVolume.SetUp failed for volume "control-plane-machine-set-operator-tls" : failed to sync secret cache: timed out waiting for the condition

openshift-monitoring

kubelet

metrics-server-64c67d44c4-s7vfs

FailedMount

MountVolume.SetUp failed for volume "secret-metrics-server-tls" : failed to sync secret cache: timed out waiting for the condition

openshift-monitoring

kubelet

openshift-state-metrics-5dc6c74576-j8tmv

FailedMount

MountVolume.SetUp failed for volume "metrics-client-ca" : failed to sync configmap cache: timed out waiting for the condition

openshift-monitoring

kubelet

openshift-state-metrics-5dc6c74576-j8tmv

FailedMount

MountVolume.SetUp failed for volume "openshift-state-metrics-tls" : failed to sync secret cache: timed out waiting for the condition

openshift-controller-manager

kubelet

controller-manager-fc56bb77c-qd4sn

FailedMount

MountVolume.SetUp failed for volume "client-ca" : failed to sync configmap cache: timed out waiting for the condition

openshift-controller-manager

kubelet

controller-manager-fc56bb77c-qd4sn

FailedMount

MountVolume.SetUp failed for volume "config" : failed to sync configmap cache: timed out waiting for the condition

openshift-monitoring

kubelet

node-exporter-lb4t5

FailedMount

MountVolume.SetUp failed for volume "metrics-client-ca" : failed to sync configmap cache: timed out waiting for the condition

openshift-machine-api

kubelet

cluster-autoscaler-operator-866dc4744-xwxg7

FailedMount

MountVolume.SetUp failed for volume "auth-proxy-config" : failed to sync configmap cache: timed out waiting for the condition

openshift-monitoring

kubelet

prometheus-operator-admission-webhook-69c6b55594-tkc2j

FailedMount

MountVolume.SetUp failed for volume "tls-certificates" : failed to sync secret cache: timed out waiting for the condition

openshift-monitoring

kubelet

kube-state-metrics-7bbc969446-qh6vq

FailedMount

MountVolume.SetUp failed for volume "kube-state-metrics-custom-resource-state-configmap" : failed to sync configmap cache: timed out waiting for the condition

openshift-monitoring

kubelet

kube-state-metrics-7bbc969446-qh6vq

FailedMount

MountVolume.SetUp failed for volume "metrics-client-ca" : failed to sync configmap cache: timed out waiting for the condition

openshift-monitoring

kubelet

kube-state-metrics-7bbc969446-qh6vq

FailedMount

MountVolume.SetUp failed for volume "kube-state-metrics-tls" : failed to sync secret cache: timed out waiting for the condition

openshift-monitoring

kubelet

kube-state-metrics-7bbc969446-qh6vq

FailedMount

MountVolume.SetUp failed for volume "kube-state-metrics-kube-rbac-proxy-config" : failed to sync secret cache: timed out waiting for the condition

openshift-monitoring

kubelet

prometheus-operator-6c8df6d4b-5m857

FailedMount

MountVolume.SetUp failed for volume "prometheus-operator-tls" : failed to sync secret cache: timed out waiting for the condition

openshift-controller-manager

kubelet

controller-manager-fc56bb77c-qd4sn

FailedMount

MountVolume.SetUp failed for volume "proxy-ca-bundles" : failed to sync configmap cache: timed out waiting for the condition

openshift-machine-api

kubelet

cluster-autoscaler-operator-866dc4744-xwxg7

FailedMount

MountVolume.SetUp failed for volume "cert" : failed to sync secret cache: timed out waiting for the condition

openshift-monitoring

kubelet

kube-state-metrics-7bbc969446-qh6vq

FailedMount

MountVolume.SetUp failed for volume "kube-state-metrics-custom-resource-state-configmap" : failed to sync configmap cache: timed out waiting for the condition

openshift-monitoring

kubelet

prometheus-operator-6c8df6d4b-5m857

FailedMount

MountVolume.SetUp failed for volume "metrics-client-ca" : failed to sync configmap cache: timed out waiting for the condition

openshift-monitoring

kubelet

prometheus-operator-6c8df6d4b-5m857

FailedMount

MountVolume.SetUp failed for volume "prometheus-operator-kube-rbac-proxy-config" : failed to sync secret cache: timed out waiting for the condition

openshift-machine-api

kubelet

machine-api-operator-6fbb6cf6f9-n8tnn

FailedMount

MountVolume.SetUp failed for volume "config" : failed to sync configmap cache: timed out waiting for the condition

openshift-cluster-version

kubelet

cluster-version-operator-7d58488df-qmm8h

FailedMount

MountVolume.SetUp failed for volume "serving-cert" : failed to sync secret cache: timed out waiting for the condition

openshift-cluster-version

kubelet

cluster-version-operator-7d58488df-qmm8h

FailedMount

MountVolume.SetUp failed for volume "service-ca" : failed to sync configmap cache: timed out waiting for the condition

openshift-monitoring

kubelet

kube-state-metrics-7bbc969446-qh6vq

FailedMount

MountVolume.SetUp failed for volume "metrics-client-ca" : failed to sync configmap cache: timed out waiting for the condition

openshift-monitoring

kubelet

kube-state-metrics-7bbc969446-qh6vq

FailedMount

MountVolume.SetUp failed for volume "kube-state-metrics-tls" : failed to sync secret cache: timed out waiting for the condition

openshift-monitoring

kubelet

kube-state-metrics-7bbc969446-qh6vq

FailedMount

MountVolume.SetUp failed for volume "kube-state-metrics-kube-rbac-proxy-config" : failed to sync secret cache: timed out waiting for the condition

openshift-insights

kubelet

insights-operator-68bf6ff9d6-mvfn5

FailedMount

MountVolume.SetUp failed for volume "service-ca-bundle" : failed to sync configmap cache: timed out waiting for the condition

openshift-insights

kubelet

insights-operator-68bf6ff9d6-mvfn5

FailedMount

MountVolume.SetUp failed for volume "serving-cert" : failed to sync secret cache: timed out waiting for the condition

openshift-insights

kubelet

insights-operator-68bf6ff9d6-mvfn5

FailedMount

MountVolume.SetUp failed for volume "trusted-ca-bundle" : failed to sync configmap cache: timed out waiting for the condition

openshift-monitoring

kubelet

prometheus-operator-admission-webhook-69c6b55594-tkc2j

FailedMount

MountVolume.SetUp failed for volume "tls-certificates" : failed to sync secret cache: timed out waiting for the condition

openshift-kube-apiserver

cert-regeneration-controller

cert-regeneration-controller-lock

LeaderElection

master-0_38895209-de9f-4d97-9b51-56bb74d04da9 became leader

openshift-machine-api

kubelet

cluster-autoscaler-operator-866dc4744-xwxg7

FailedMount

MountVolume.SetUp failed for volume "cert" : failed to sync secret cache: timed out waiting for the condition

openshift-machine-api

kubelet

cluster-autoscaler-operator-866dc4744-xwxg7

FailedMount

MountVolume.SetUp failed for volume "auth-proxy-config" : failed to sync configmap cache: timed out waiting for the condition

openshift-ingress-operator

kubelet

ingress-operator-66b84d69b-gzg9m

Started

Started container ingress-operator

openshift-ingress-operator

cluster-ingress-operator

ingress-operator

FeatureGatesInitialized

FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{"AWSEFSDriverVolumeMetrics", "AdminNetworkPolicy", "AlibabaPlatform", "AzureWorkloadIdentity", "BareMetalLoadBalancer", "BuildCSIVolumes", "ChunkSizeMiB", "CloudDualStackNodeIPs", "DisableKubeletCloudCredentialProviders", "GCPLabelsTags", "HardwareSpeed", "IngressControllerLBSubnetsAWS", "KMSv1", "ManagedBootImages", "ManagedBootImagesAWS", "MultiArchInstallAWS", "MultiArchInstallGCP", "NetworkDiagnosticsConfig", "NetworkLiveMigration", "NetworkSegmentation", "NewOLM", "NodeDisruptionPolicy", "OnClusterBuild", "PersistentIPsForVirtualization", "PrivateHostedZoneAWS", "SetEIPForNLBIngressController", "StreamingCollectionEncodingToJSON", "StreamingCollectionEncodingToProtobuf", "VSphereControlPlaneMachineSet", "VSphereDriverConfiguration", "VSphereMultiVCenters", "VSphereStaticIPs", "ValidatingAdmissionPolicy"}, Disabled:[]v1.FeatureGateName{"AWSClusterHostedDNS", "AdditionalRoutingCapabilities", "AutomatedEtcdBackup", "BootcNodeManagement", "CSIDriverSharedResource", "ClusterAPIInstall", "ClusterAPIInstallIBMCloud", "ClusterMonitoringConfig", "ConsolePluginContentSecurityPolicy", "DNSNameResolver", "DynamicResourceAllocation", "EtcdBackendQuota", "EventedPLEG", "Example", "ExternalOIDC", "ExternalOIDCWithUIDAndExtraClaimMappings", "GCPClusterHostedDNS", "GatewayAPI", "ImageStreamImportMode", "IngressControllerDynamicConfigurationManager", "InsightsConfig", "InsightsConfigAPI", "InsightsOnDemandDataGather", "InsightsRuntimeExtractor", "MachineAPIMigration", "MachineAPIOperatorDisableMachineHealthCheckController", "MachineAPIProviderOpenStack", "MachineConfigNodes", "MaxUnavailableStatefulSet", "MetricsCollectionProfiles", "MinimumKubeletVersion", "MixedCPUsAllocation", "MultiArchInstallAzure", "NodeSwap", "NutanixMultiSubnets", "OVNObservability", "OpenShiftPodSecurityAdmission", "PinnedImages", "PlatformOperators", "ProcMountType", "RouteAdvertisements", "RouteExternalCertificate", "ServiceAccountTokenNodeBinding", "SignatureStores", "SigstoreImageVerification", "TranslateStreamCloseWebsocketRequests", "UpgradeStatus", "UserNamespacesPodSecurityStandards", "UserNamespacesSupport", "VSphereMultiNetworks", "VolumeAttributesClass", "VolumeGroupSnapshot"}}

openshift-ingress-operator

kubelet

ingress-operator-66b84d69b-gzg9m

Created

Created container: ingress-operator

openshift-ingress-operator

kubelet

ingress-operator-66b84d69b-gzg9m

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:77fff570657d2fa0bfb709b2c8b6665bae0bf90a2be981d8dbca56c674715098" already present on machine

openshift-kube-controller-manager

cluster-policy-controller-namespace-security-allocation-controller

kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openshift-ingress-canary namespace

openshift-ingress-canary

daemonset-controller

ingress-canary

SuccessfulCreate

Created pod: ingress-canary-cst2b

openshift-kube-scheduler

cert-recovery-controller

cert-recovery-controller-lock

LeaderElection

master-0_b8c9eada-6f8e-4585-81ea-2d2a6bf7dc99 became leader

openshift-kube-apiserver

kubelet

kube-apiserver-startup-monitor-master-0

Killing

Stopping container startup-monitor

openshift-kube-apiserver-operator

kube-apiserver-operator-installer-controller

openshift-kube-apiserver-operator

NodeCurrentRevisionChanged

Updated node "master-0" from revision 0 to 3 because static pod is ready

openshift-kube-apiserver-operator

kube-apiserver-operator-status-controller-statussyncer_kube-apiserver

openshift-kube-apiserver-operator

OperatorStatusChanged

Status for clusteroperator/kube-apiserver changed: Progressing changed from True to False ("NodeInstallerProgressing: 1 node is at revision 3"),Available changed from False to True ("StaticPodsAvailable: 1 nodes are active; 1 node is at revision 3")

openshift-kube-controller-manager

cluster-policy-controller-namespace-security-allocation-controller

kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openshift-console namespace

openshift-kube-controller-manager

cluster-policy-controller-namespace-security-allocation-controller

kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openshift-console-user-settings namespace

openshift-cluster-olm-operator

olm-status-controller-statussyncer_olm

cluster-olm-operator

OperatorStatusChanged

Status for clusteroperator/olm changed: Progressing changed from False to True ("CatalogdDeploymentCatalogdControllerManagerProgressing: Waiting for Deployment to deploy pods\nOperatorcontrollerDeploymentOperatorControllerControllerManagerProgressing: Waiting for Deployment to deploy pods"),Available changed from True to False ("CatalogdDeploymentCatalogdControllerManagerAvailable: Waiting for Deployment\nOperatorcontrollerDeploymentOperatorControllerControllerManagerAvailable: Waiting for Deployment")

openshift-kube-controller-manager

cluster-policy-controller-namespace-security-allocation-controller

kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openshift-console-operator namespace
(x2)

openshift-cluster-olm-operator

olm-status-controller-statussyncer_olm

cluster-olm-operator

OperatorStatusChanged

Status for clusteroperator/olm changed: Progressing message changed from "CatalogdDeploymentCatalogdControllerManagerProgressing: Waiting for Deployment to deploy pods\nOperatorcontrollerDeploymentOperatorControllerControllerManagerProgressing: Waiting for Deployment to deploy pods" to "CatalogdDeploymentCatalogdControllerManagerProgressing: Waiting for Deployment to deploy pods",Available message changed from "CatalogdDeploymentCatalogdControllerManagerAvailable: Waiting for Deployment\nOperatorcontrollerDeploymentOperatorControllerControllerManagerAvailable: Waiting for Deployment" to "CatalogdDeploymentCatalogdControllerManagerAvailable: Waiting for Deployment"
(x2)

openshift-cluster-olm-operator

olm-status-controller-statussyncer_olm

cluster-olm-operator

OperatorStatusChanged

Status for clusteroperator/olm changed: Progressing changed from True to False ("All is well"),Available changed from False to True ("All is well")

openshift-console-operator

replicaset-controller

console-operator-76b6568d85

SuccessfulCreate

Created pod: console-operator-76b6568d85-8b8gv

openshift-machine-config-operator

machine-config-operator

machine-config-operator

ConfigMapUpdated

Updated ConfigMap/kube-rbac-proxy -n openshift-machine-config-operator: cause by changes in data.config-file.yaml

openshift-console-operator

deployment-controller

console-operator

ScalingReplicaSet

Scaled up replica set console-operator-76b6568d85 to 1
(x6)

openshift-ingress-canary

kubelet

ingress-canary-cst2b

FailedMount

MountVolume.SetUp failed for volume "cert" : secret "canary-serving-cert" not found

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ServiceAccountCreated

Created ServiceAccount/monitoring-plugin -n openshift-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ServiceCreated

Created Service/monitoring-plugin -n openshift-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ServiceCreated

Created Service/monitoring-plugin -n openshift-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ServiceAccountCreated

Created ServiceAccount/monitoring-plugin -n openshift-monitoring because it was missing

openshift-operator-lifecycle-manager

operator-lifecycle-manager

packageserver

InstallSucceeded

install strategy completed with no errors

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Degraded message changed from "APIServerStaticResourcesDegraded: \"oauth-apiserver/ns.yaml\" (string): Get \"https://172.30.0.1:443/api/v1/namespaces/openshift-oauth-apiserver\": dial tcp 172.30.0.1:443: connect: connection refused\nAPIServerStaticResourcesDegraded: \"oauth-apiserver/apiserver-clusterrolebinding.yaml\" (string): Get \"https://172.30.0.1:443/apis/rbac.authorization.k8s.io/v1/clusterrolebindings/system:openshift:oauth-apiserver\": dial tcp 172.30.0.1:443: connect: connection refused\nAPIServerStaticResourcesDegraded: \nIngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\nOAuthServerServiceEndpointAccessibleControllerDegraded: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready" to "IngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\nOAuthServerServiceEndpointAccessibleControllerDegraded: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready"

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Degraded message changed from "IngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\nOAuthServerServiceEndpointAccessibleControllerDegraded: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready" to "APIServerStaticResourcesDegraded: \"oauth-apiserver/ns.yaml\" (string): Get \"https://172.30.0.1:443/api/v1/namespaces/openshift-oauth-apiserver\": dial tcp 172.30.0.1:443: connect: connection refused\nAPIServerStaticResourcesDegraded: \"oauth-apiserver/apiserver-clusterrolebinding.yaml\" (string): Get \"https://172.30.0.1:443/apis/rbac.authorization.k8s.io/v1/clusterrolebindings/system:openshift:oauth-apiserver\": dial tcp 172.30.0.1:443: connect: connection refused\nAPIServerStaticResourcesDegraded: \nIngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\nOAuthServerServiceEndpointAccessibleControllerDegraded: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready",Available message changed from "APIServicesAvailable: apiservices.apiregistration.k8s.io/v1.oauth.openshift.io: not available: failing or missing response from https://10.128.0.45:8443/apis/oauth.openshift.io/v1: bad status from https://10.128.0.45:8443/apis/oauth.openshift.io/v1: 401\nAPIServicesAvailable: apiservices.apiregistration.k8s.io/v1.user.openshift.io: not available: failing or missing response from https://10.128.0.45:8443/apis/user.openshift.io/v1: bad status from https://10.128.0.45:8443/apis/user.openshift.io/v1: 401\nOAuthServerServiceEndpointAccessibleControllerAvailable: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerAvailable: endpoints \"oauth-openshift\" not found" to "OAuthServerServiceEndpointAccessibleControllerAvailable: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerAvailable: endpoints \"oauth-openshift\" not found"

openshift-monitoring

multus

monitoring-plugin-b7d455597-jgpnw

AddedInterface

Add eth0 [10.128.0.83/23] from ovn-kubernetes

openshift-monitoring

multus

monitoring-plugin-b7d455597-jgpnw

AddedInterface

Add eth0 [10.128.0.83/23] from ovn-kubernetes

openshift-monitoring

deployment-controller

monitoring-plugin

ScalingReplicaSet

Scaled up replica set monitoring-plugin-b7d455597 to 1

openshift-monitoring

replicaset-controller

monitoring-plugin-b7d455597

SuccessfulCreate

Created pod: monitoring-plugin-b7d455597-jgpnw

openshift-monitoring

kubelet

monitoring-plugin-b7d455597-jgpnw

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bddcddc296ce363f3b55783425259057ee0ae6d033c6b4a430d92eacb9830748"

openshift-monitoring

kubelet

monitoring-plugin-b7d455597-jgpnw

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bddcddc296ce363f3b55783425259057ee0ae6d033c6b4a430d92eacb9830748"

openshift-monitoring

replicaset-controller

monitoring-plugin-b7d455597

SuccessfulCreate

Created pod: monitoring-plugin-b7d455597-jgpnw

openshift-monitoring

deployment-controller

monitoring-plugin

ScalingReplicaSet

Scaled up replica set monitoring-plugin-b7d455597 to 1

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-status-controller-statussyncer_kube-scheduler

openshift-kube-scheduler-operator

OperatorStatusChanged

Status for clusteroperator/kube-scheduler changed: Progressing changed from True to False ("NodeInstallerProgressing: 1 node is at revision 3"),Available changed from False to True ("StaticPodsAvailable: 1 nodes are active; 1 node is at revision 3")

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-installer-controller

openshift-kube-scheduler-operator

NodeCurrentRevisionChanged

Updated node "master-0" from revision 0 to 3 because static pod is ready

openshift-monitoring

kubelet

monitoring-plugin-b7d455597-jgpnw

Started

Started container monitoring-plugin

openshift-monitoring

kubelet

monitoring-plugin-b7d455597-jgpnw

Started

Started container monitoring-plugin

openshift-monitoring

kubelet

monitoring-plugin-b7d455597-jgpnw

Created

Created container: monitoring-plugin

openshift-monitoring

kubelet

monitoring-plugin-b7d455597-jgpnw

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bddcddc296ce363f3b55783425259057ee0ae6d033c6b4a430d92eacb9830748" in 1.963s (1.963s including waiting). Image size: 447814986 bytes.

openshift-monitoring

kubelet

monitoring-plugin-b7d455597-jgpnw

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bddcddc296ce363f3b55783425259057ee0ae6d033c6b4a430d92eacb9830748" in 1.963s (1.963s including waiting). Image size: 447814986 bytes.

openshift-monitoring

kubelet

monitoring-plugin-b7d455597-jgpnw

Created

Created container: monitoring-plugin

openshift-cloud-controller-manager-operator

master-0_66f3923e-6738-49bc-a104-f5951752cb61

cluster-cloud-controller-manager-leader

LeaderElection

master-0_66f3923e-6738-49bc-a104-f5951752cb61 became leader

openshift-ingress-canary

kubelet

ingress-canary-cst2b

Started

Started container serve-healthcheck-canary

openshift-ingress-canary

kubelet

ingress-canary-cst2b

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:77fff570657d2fa0bfb709b2c8b6665bae0bf90a2be981d8dbca56c674715098" already present on machine

openshift-ingress-canary

kubelet

ingress-canary-cst2b

Created

Created container: serve-healthcheck-canary

openshift-ingress-canary

multus

ingress-canary-cst2b

AddedInterface

Add eth0 [10.128.0.81/23] from ovn-kubernetes

openshift-authentication-operator

cluster-authentication-operator-payload-config-controller-openshift-authentication-payloadconfig

authentication-operator

SecretCreated

Created Secret/v4-0-config-system-session -n openshift-authentication because it was missing

openshift-machine-config-operator

machine-config-operator

machine-config-operator

ConfigMapUpdated

Updated ConfigMap/kube-rbac-proxy -n openshift-machine-config-operator: cause by changes in data.config-file.yaml

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Degraded message changed from "IngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\nOAuthServerServiceEndpointAccessibleControllerDegraded: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready" to "IngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\nOAuthServerDeploymentDegraded: waiting for the oauth-openshift route to appear: route.route.openshift.io \"oauth-openshift\" not found\nOAuthServerDeploymentDegraded: \nOAuthServerServiceEndpointAccessibleControllerDegraded: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\nOAuthServerWorkloadDegraded: waiting for the oauth-openshift route to appear: route.route.openshift.io \"oauth-openshift\" not found\nOAuthServerWorkloadDegraded: "

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Degraded message changed from "IngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\nOAuthConfigRouteDegraded: The OAuth server route 'openshift-authentication/oauth-openshift' was not found\nOAuthServerDeploymentDegraded: waiting for the oauth-openshift route to contain an admitted ingress: no admitted ingress for route oauth-openshift in namespace openshift-authentication\nOAuthServerDeploymentDegraded: \nOAuthServerServiceEndpointAccessibleControllerDegraded: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\nOAuthServerWorkloadDegraded: waiting for the oauth-openshift route to contain an admitted ingress: no admitted ingress for route oauth-openshift in namespace openshift-authentication\nOAuthServerWorkloadDegraded: " to "IngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\nOAuthConfigRouteDegraded: The OAuth server route 'openshift-authentication/oauth-openshift' was not found\nOAuthServerDeploymentDegraded: waiting for the oauth-openshift route to contain an admitted ingress: no admitted ingress for route oauth-openshift in namespace openshift-authentication\nOAuthServerDeploymentDegraded: \nOAuthServerServiceEndpointAccessibleControllerDegraded: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\nOAuthServerWorkloadDegraded: waiting for the oauth-openshift route to contain an admitted ingress: no admitted ingress for route oauth-openshift in namespace openshift-authentication\nOAuthServerWorkloadDegraded: \nWellKnownReadyControllerDegraded: route.route.openshift.io \"oauth-openshift\" not found"

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Degraded message changed from "IngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\nOAuthServerDeploymentDegraded: waiting for the oauth-openshift route to appear: route.route.openshift.io \"oauth-openshift\" not found\nOAuthServerDeploymentDegraded: \nOAuthServerServiceEndpointAccessibleControllerDegraded: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\nOAuthServerWorkloadDegraded: waiting for the oauth-openshift route to appear: route.route.openshift.io \"oauth-openshift\" not found\nOAuthServerWorkloadDegraded: " to "IngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\nOAuthClientsControllerDegraded: route.route.openshift.io \"oauth-openshift\" not found\nOAuthServerDeploymentDegraded: waiting for the oauth-openshift route to appear: route.route.openshift.io \"oauth-openshift\" not found\nOAuthServerDeploymentDegraded: \nOAuthServerServiceEndpointAccessibleControllerDegraded: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\nOAuthServerWorkloadDegraded: waiting for the oauth-openshift route to appear: route.route.openshift.io \"oauth-openshift\" not found\nOAuthServerWorkloadDegraded: "

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Degraded message changed from "IngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\nOAuthConfigRouteDegraded: The OAuth server route 'openshift-authentication/oauth-openshift' was not found\nOAuthServerDeploymentDegraded: waiting for the oauth-openshift route to contain an admitted ingress: no admitted ingress for route oauth-openshift in namespace openshift-authentication\nOAuthServerDeploymentDegraded: \nOAuthServerServiceEndpointAccessibleControllerDegraded: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\nOAuthServerWorkloadDegraded: waiting for the oauth-openshift route to contain an admitted ingress: no admitted ingress for route oauth-openshift in namespace openshift-authentication\nOAuthServerWorkloadDegraded: \nWellKnownReadyControllerDegraded: route.route.openshift.io \"oauth-openshift\" not found" to "IngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\nOAuthConfigRouteDegraded: The OAuth server route 'openshift-authentication/oauth-openshift' was not found\nOAuthServerDeploymentDegraded: waiting for the oauth-openshift route to contain an admitted ingress: no admitted ingress for route oauth-openshift in namespace openshift-authentication\nOAuthServerDeploymentDegraded: \nOAuthServerRouteEndpointAccessibleControllerDegraded: \"https://oauth-openshift.apps.sno.openstack.lab/healthz\" returned \"503 Service Unavailable\"\nOAuthServerServiceEndpointAccessibleControllerDegraded: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\nOAuthServerWorkloadDegraded: waiting for the oauth-openshift route to contain an admitted ingress: no admitted ingress for route oauth-openshift in namespace openshift-authentication\nOAuthServerWorkloadDegraded: \nWellKnownReadyControllerDegraded: route.route.openshift.io \"oauth-openshift\" not found"

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Available message changed from "OAuthServerRouteEndpointAccessibleControllerAvailable: failed to retrieve route from cache: route.route.openshift.io \"oauth-openshift\" not found\nOAuthServerServiceEndpointAccessibleControllerAvailable: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerAvailable: endpoints \"oauth-openshift\" not found\nWellKnownAvailable: route.route.openshift.io \"oauth-openshift\" not found" to "OAuthServerRouteEndpointAccessibleControllerAvailable: \"https://oauth-openshift.apps.sno.openstack.lab/healthz\" returned \"503 Service Unavailable\"\nOAuthServerServiceEndpointAccessibleControllerAvailable: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerAvailable: endpoints \"oauth-openshift\" not found\nWellKnownAvailable: route.route.openshift.io \"oauth-openshift\" not found"

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Degraded message changed from "IngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\nOAuthClientsControllerDegraded: route.route.openshift.io \"oauth-openshift\" not found\nOAuthServerDeploymentDegraded: waiting for the oauth-openshift route to appear: route.route.openshift.io \"oauth-openshift\" not found\nOAuthServerDeploymentDegraded: \nOAuthServerServiceEndpointAccessibleControllerDegraded: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\nOAuthServerWorkloadDegraded: waiting for the oauth-openshift route to appear: route.route.openshift.io \"oauth-openshift\" not found\nOAuthServerWorkloadDegraded: " to "IngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\nOAuthClientsControllerDegraded: route.route.openshift.io \"oauth-openshift\" not found\nOAuthConfigRouteDegraded: The OAuth server route 'openshift-authentication/oauth-openshift' was not found\nOAuthServerDeploymentDegraded: waiting for the oauth-openshift route to appear: route.route.openshift.io \"oauth-openshift\" not found\nOAuthServerDeploymentDegraded: \nOAuthServerServiceEndpointAccessibleControllerDegraded: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\nOAuthServerWorkloadDegraded: waiting for the oauth-openshift route to appear: route.route.openshift.io \"oauth-openshift\" not found\nOAuthServerWorkloadDegraded: ",Available message changed from "OAuthServerServiceEndpointAccessibleControllerAvailable: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerAvailable: endpoints \"oauth-openshift\" not found" to "OAuthServerRouteEndpointAccessibleControllerAvailable: failed to retrieve route from cache: route.route.openshift.io \"oauth-openshift\" not found\nOAuthServerServiceEndpointAccessibleControllerAvailable: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerAvailable: endpoints \"oauth-openshift\" not found"

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Degraded message changed from "IngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\nOAuthClientsControllerDegraded: route.route.openshift.io \"oauth-openshift\" not found\nOAuthConfigRouteDegraded: The OAuth server route 'openshift-authentication/oauth-openshift' was not found\nOAuthServerDeploymentDegraded: waiting for the oauth-openshift route to contain an admitted ingress: no admitted ingress for route oauth-openshift in namespace openshift-authentication\nOAuthServerDeploymentDegraded: \nOAuthServerServiceEndpointAccessibleControllerDegraded: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\nOAuthServerWorkloadDegraded: waiting for the oauth-openshift route to contain an admitted ingress: no admitted ingress for route oauth-openshift in namespace openshift-authentication\nOAuthServerWorkloadDegraded: " to "IngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\nOAuthConfigRouteDegraded: The OAuth server route 'openshift-authentication/oauth-openshift' was not found\nOAuthServerDeploymentDegraded: waiting for the oauth-openshift route to contain an admitted ingress: no admitted ingress for route oauth-openshift in namespace openshift-authentication\nOAuthServerDeploymentDegraded: \nOAuthServerServiceEndpointAccessibleControllerDegraded: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\nOAuthServerWorkloadDegraded: waiting for the oauth-openshift route to contain an admitted ingress: no admitted ingress for route oauth-openshift in namespace openshift-authentication\nOAuthServerWorkloadDegraded: "

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Degraded message changed from "IngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\nOAuthClientsControllerDegraded: route.route.openshift.io \"oauth-openshift\" not found\nOAuthConfigRouteDegraded: The OAuth server route 'openshift-authentication/oauth-openshift' was not found\nOAuthServerDeploymentDegraded: waiting for the oauth-openshift route to appear: route.route.openshift.io \"oauth-openshift\" not found\nOAuthServerDeploymentDegraded: \nOAuthServerServiceEndpointAccessibleControllerDegraded: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\nOAuthServerWorkloadDegraded: waiting for the oauth-openshift route to appear: route.route.openshift.io \"oauth-openshift\" not found\nOAuthServerWorkloadDegraded: " to "IngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\nOAuthClientsControllerDegraded: route.route.openshift.io \"oauth-openshift\" not found\nOAuthConfigRouteDegraded: The OAuth server route 'openshift-authentication/oauth-openshift' was not found\nOAuthServerDeploymentDegraded: waiting for the oauth-openshift route to contain an admitted ingress: no admitted ingress for route oauth-openshift in namespace openshift-authentication\nOAuthServerDeploymentDegraded: \nOAuthServerServiceEndpointAccessibleControllerDegraded: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\nOAuthServerWorkloadDegraded: waiting for the oauth-openshift route to contain an admitted ingress: no admitted ingress for route oauth-openshift in namespace openshift-authentication\nOAuthServerWorkloadDegraded: "

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Progressing changed from False to True (""),Available message changed from "OAuthServerRouteEndpointAccessibleControllerAvailable: failed to retrieve route from cache: route.route.openshift.io \"oauth-openshift\" not found\nOAuthServerServiceEndpointAccessibleControllerAvailable: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerAvailable: endpoints \"oauth-openshift\" not found" to "OAuthServerRouteEndpointAccessibleControllerAvailable: failed to retrieve route from cache: route.route.openshift.io \"oauth-openshift\" not found\nOAuthServerServiceEndpointAccessibleControllerAvailable: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerAvailable: endpoints \"oauth-openshift\" not found\nWellKnownAvailable: route.route.openshift.io \"oauth-openshift\" not found"

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Available message changed from "OAuthServerRouteEndpointAccessibleControllerAvailable: \"https://oauth-openshift.apps.sno.openstack.lab/healthz\" returned \"503 Service Unavailable\"\nOAuthServerServiceEndpointAccessibleControllerAvailable: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerAvailable: endpoints \"oauth-openshift\" not found\nWellKnownAvailable: The well-known endpoint is not yet available: failed to get oauth metadata from openshift-config-managed/oauth-openshift ConfigMap: configmap \"oauth-openshift\" not found (check authentication operator, it is supposed to create this)" to "OAuthServerRouteEndpointAccessibleControllerAvailable: Get \"https://oauth-openshift.apps.sno.openstack.lab/healthz\": EOF\nOAuthServerServiceEndpointAccessibleControllerAvailable: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerAvailable: endpoints \"oauth-openshift\" not found\nWellKnownAvailable: The well-known endpoint is not yet available: failed to get oauth metadata from openshift-config-managed/oauth-openshift ConfigMap: configmap \"oauth-openshift\" not found (check authentication operator, it is supposed to create this)"

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Degraded message changed from "IngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\nOAuthConfigRouteDegraded: The OAuth server route 'openshift-authentication/oauth-openshift' was not found\nOAuthServerDeploymentDegraded: waiting for the oauth-openshift route to contain an admitted ingress: no admitted ingress for route oauth-openshift in namespace openshift-authentication\nOAuthServerDeploymentDegraded: \nOAuthServerRouteEndpointAccessibleControllerDegraded: \"https://oauth-openshift.apps.sno.openstack.lab/healthz\" returned \"503 Service Unavailable\"\nOAuthServerServiceEndpointAccessibleControllerDegraded: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\nOAuthServerWorkloadDegraded: waiting for the oauth-openshift route to contain an admitted ingress: no admitted ingress for route oauth-openshift in namespace openshift-authentication\nOAuthServerWorkloadDegraded: \nWellKnownReadyControllerDegraded: route.route.openshift.io \"oauth-openshift\" not found" to "IngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\nOAuthConfigRouteDegraded: The OAuth server route 'openshift-authentication/oauth-openshift' was not found\nOAuthServerDeploymentDegraded: waiting for the oauth-openshift route to contain an admitted ingress: no admitted ingress for route oauth-openshift in namespace openshift-authentication\nOAuthServerDeploymentDegraded: \nOAuthServerRouteEndpointAccessibleControllerDegraded: \"https://oauth-openshift.apps.sno.openstack.lab/healthz\" returned \"503 Service Unavailable\"\nOAuthServerServiceEndpointAccessibleControllerDegraded: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\nOAuthServerWorkloadDegraded: waiting for the oauth-openshift route to contain an admitted ingress: no admitted ingress for route oauth-openshift in namespace openshift-authentication\nOAuthServerWorkloadDegraded: \nWellKnownReadyControllerDegraded: failed to get oauth metadata from openshift-config-managed/oauth-openshift ConfigMap: configmap \"oauth-openshift\" not found (check authentication operator, it is supposed to create this)"

openshift-cloud-controller-manager-operator

master-0_c7ce7a94-8f6c-4b0c-aede-80aef3b7f806

cluster-cloud-config-sync-leader

LeaderElection

master-0_c7ce7a94-8f6c-4b0c-aede-80aef3b7f806 became leader

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Degraded message changed from "IngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\nOAuthConfigRouteDegraded: The OAuth server route 'openshift-authentication/oauth-openshift' was not found\nOAuthServerDeploymentDegraded: waiting for the oauth-openshift route to contain an admitted ingress: no admitted ingress for route oauth-openshift in namespace openshift-authentication\nOAuthServerDeploymentDegraded: \nOAuthServerRouteEndpointAccessibleControllerDegraded: \"https://oauth-openshift.apps.sno.openstack.lab/healthz\" returned \"503 Service Unavailable\"\nOAuthServerServiceEndpointAccessibleControllerDegraded: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\nOAuthServerWorkloadDegraded: waiting for the oauth-openshift route to contain an admitted ingress: no admitted ingress for route oauth-openshift in namespace openshift-authentication\nOAuthServerWorkloadDegraded: \nWellKnownReadyControllerDegraded: failed to get oauth metadata from openshift-config-managed/oauth-openshift ConfigMap: configmap \"oauth-openshift\" not found (check authentication operator, it is supposed to create this)" to "IngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\nOAuthConfigRouteDegraded: The OAuth server route 'openshift-authentication/oauth-openshift' was not found\nOAuthServerDeploymentDegraded: waiting for the oauth-openshift route to contain an admitted ingress: no admitted ingress for route oauth-openshift in namespace openshift-authentication\nOAuthServerDeploymentDegraded: \nOAuthServerRouteEndpointAccessibleControllerDegraded: Get \"https://oauth-openshift.apps.sno.openstack.lab/healthz\": EOF\nOAuthServerServiceEndpointAccessibleControllerDegraded: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\nOAuthServerWorkloadDegraded: waiting for the oauth-openshift route to contain an admitted ingress: no admitted ingress for route oauth-openshift in namespace openshift-authentication\nOAuthServerWorkloadDegraded: \nWellKnownReadyControllerDegraded: failed to get oauth metadata from openshift-config-managed/oauth-openshift ConfigMap: configmap \"oauth-openshift\" not found (check authentication operator, it is supposed to create this)"

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Available message changed from "OAuthServerRouteEndpointAccessibleControllerAvailable: \"https://oauth-openshift.apps.sno.openstack.lab/healthz\" returned \"503 Service Unavailable\"\nOAuthServerServiceEndpointAccessibleControllerAvailable: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerAvailable: endpoints \"oauth-openshift\" not found\nWellKnownAvailable: route.route.openshift.io \"oauth-openshift\" not found" to "OAuthServerRouteEndpointAccessibleControllerAvailable: \"https://oauth-openshift.apps.sno.openstack.lab/healthz\" returned \"503 Service Unavailable\"\nOAuthServerServiceEndpointAccessibleControllerAvailable: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerAvailable: endpoints \"oauth-openshift\" not found\nWellKnownAvailable: The well-known endpoint is not yet available: failed to get oauth metadata from openshift-config-managed/oauth-openshift ConfigMap: configmap \"oauth-openshift\" not found (check authentication operator, it is supposed to create this)"

openshift-authentication-operator

cluster-authentication-operator-metadata-controller-openshift-authentication-metadata

authentication-operator

ConfigMapCreated

Created ConfigMap/v4-0-config-system-metadata -n openshift-authentication because it was missing
(x2)

openshift-kube-apiserver-operator

kube-apiserver-operator-config-observer-configobserver

openshift-kube-apiserver-operator

ObservedConfigChanged

Writing updated observed config:   map[string]any{    "admission": map[string]any{"pluginConfig": map[string]any{"PodSecurity": map[string]any{"configuration": map[string]any{"defaults": map[string]any{"audit": string("restricted"), "audit-version": string("latest"), "enforce": string("privileged"), "enforce-version": string("latest"), ...}}}, "network.openshift.io/ExternalIPRanger": map[string]any{"configuration": map[string]any{"allowIngressIP": bool(false), "apiVersion": string("network.openshift.io/v1"), "kind": string("ExternalIPRangerAdmissionConfig")}}, "network.openshift.io/RestrictedEndpointsAdmission": map[string]any{"configuration": map[string]any{"apiVersion": string("network.openshift.io/v1"), "kind": string("RestrictedEndpointsAdmissionConfig"), "restrictedCIDRs": []any{string("10.128.0.0/16"), string("172.30.0.0/16")}}}}},    "apiServerArguments": map[string]any{"api-audiences": []any{string("https://kubernetes.default.svc")}, "authentication-token-webhook-config-file": []any{string("/etc/kubernetes/static-pod-resources/secrets/webhook-authenticat"...)}, "authentication-token-webhook-version": []any{string("v1")}, "etcd-servers": []any{string("https://192.168.32.10:2379"), string("https://localhost:2379")}, ...}, +  "authConfig": map[string]any{ +  "oauthMetadataFile": string("/etc/kubernetes/static-pod-resources/configmaps/oauth-metadata/oauthMetadata"), +  },    "corsAllowedOrigins": []any{string(`//127\.0\.0\.1(:|$)`), string("//localhost(:|$)")},    "gracefulTerminationDuration": string("15"),    ... // 2 identical entries   }
(x2)

openshift-kube-apiserver-operator

kube-apiserver-operator-target-config-controller-targetconfigcontroller

openshift-kube-apiserver-operator

ConfigMapUpdated

Updated ConfigMap/config -n openshift-kube-apiserver: cause by changes in data.config.yaml

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

openshift-kube-apiserver-operator

StartingNewRevision

new revision 4 triggered by "required configmap/config has changed"

openshift-authentication

replicaset-controller

oauth-openshift-86f9447d7

SuccessfulCreate

Created pod: oauth-openshift-86f9447d7-znl8c

openshift-authentication

deployment-controller

oauth-openshift

ScalingReplicaSet

Scaled up replica set oauth-openshift-86f9447d7 to 1

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Degraded message changed from "IngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\nOAuthConfigRouteDegraded: The OAuth server route 'openshift-authentication/oauth-openshift' was not found\nOAuthServerDeploymentDegraded: waiting for the oauth-openshift route to contain an admitted ingress: no admitted ingress for route oauth-openshift in namespace openshift-authentication\nOAuthServerDeploymentDegraded: \nOAuthServerRouteEndpointAccessibleControllerDegraded: Get \"https://oauth-openshift.apps.sno.openstack.lab/healthz\": EOF\nOAuthServerServiceEndpointAccessibleControllerDegraded: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\nOAuthServerWorkloadDegraded: waiting for the oauth-openshift route to contain an admitted ingress: no admitted ingress for route oauth-openshift in namespace openshift-authentication\nOAuthServerWorkloadDegraded: \nWellKnownReadyControllerDegraded: failed to get oauth metadata from openshift-config-managed/oauth-openshift ConfigMap: configmap \"oauth-openshift\" not found (check authentication operator, it is supposed to create this)" to "IngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\nOAuthConfigRouteDegraded: The OAuth server route 'openshift-authentication/oauth-openshift' was not found\nOAuthServerRouteEndpointAccessibleControllerDegraded: Get \"https://oauth-openshift.apps.sno.openstack.lab/healthz\": EOF\nOAuthServerServiceEndpointAccessibleControllerDegraded: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\nWellKnownReadyControllerDegraded: failed to get oauth metadata from openshift-config-managed/oauth-openshift ConfigMap: configmap \"oauth-openshift\" not found (check authentication operator, it is supposed to create this)",Progressing message changed from "" to "OAuthServerDeploymentProgressing: deployment/oauth-openshift.openshift-authentication: observed generation is 0, desired generation is 1.",Available message changed from "OAuthServerRouteEndpointAccessibleControllerAvailable: Get \"https://oauth-openshift.apps.sno.openstack.lab/healthz\": EOF\nOAuthServerServiceEndpointAccessibleControllerAvailable: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerAvailable: endpoints \"oauth-openshift\" not found\nWellKnownAvailable: The well-known endpoint is not yet available: failed to get oauth metadata from openshift-config-managed/oauth-openshift ConfigMap: configmap \"oauth-openshift\" not found (check authentication operator, it is supposed to create this)" to "OAuthServerDeploymentAvailable: no oauth-openshift.openshift-authentication pods available on any node.\nOAuthServerRouteEndpointAccessibleControllerAvailable: Get \"https://oauth-openshift.apps.sno.openstack.lab/healthz\": EOF\nOAuthServerServiceEndpointAccessibleControllerAvailable: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerAvailable: endpoints \"oauth-openshift\" not found\nWellKnownAvailable: The well-known endpoint is not yet available: failed to get oauth metadata from openshift-config-managed/oauth-openshift ConfigMap: configmap \"oauth-openshift\" not found (check authentication operator, it is supposed to create this)"

openshift-authentication-operator

cluster-authentication-operator-oauthserver-workloadworkloadcontroller

authentication-operator

DeploymentCreated

Created Deployment.apps/oauth-openshift -n openshift-authentication because it was missing

openshift-authentication-operator

cluster-authentication-operator-resource-sync-controller-resourcesynccontroller

authentication-operator

ConfigMapCreated

Created ConfigMap/oauth-openshift -n openshift-config-managed because it was missing

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Progressing message changed from "OAuthServerDeploymentProgressing: deployment/oauth-openshift.openshift-authentication: observed generation is 0, desired generation is 1." to "OAuthServerDeploymentProgressing: deployment/oauth-openshift.openshift-authentication: observed generation is 0, desired generation is 1.\nWellKnownReadyControllerProgressing: kube-apiserver oauth endpoint https://192.168.32.10:6443/.well-known/oauth-authorization-server is not yet served and authentication operator keeps waiting (check kube-apiserver operator, and check that instances roll out successfully, which can take several minutes per instance)",Available message changed from "OAuthServerDeploymentAvailable: no oauth-openshift.openshift-authentication pods available on any node.\nOAuthServerRouteEndpointAccessibleControllerAvailable: Get \"https://oauth-openshift.apps.sno.openstack.lab/healthz\": EOF\nOAuthServerServiceEndpointAccessibleControllerAvailable: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerAvailable: endpoints \"oauth-openshift\" not found\nWellKnownAvailable: The well-known endpoint is not yet available: failed to get oauth metadata from openshift-config-managed/oauth-openshift ConfigMap: configmap \"oauth-openshift\" not found (check authentication operator, it is supposed to create this)" to "OAuthServerDeploymentAvailable: no oauth-openshift.openshift-authentication pods available on any node.\nOAuthServerRouteEndpointAccessibleControllerAvailable: Get \"https://oauth-openshift.apps.sno.openstack.lab/healthz\": EOF\nOAuthServerServiceEndpointAccessibleControllerAvailable: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerAvailable: endpoints \"oauth-openshift\" not found\nWellKnownAvailable: The well-known endpoint is not yet available: kube-apiserver oauth endpoint https://192.168.32.10:6443/.well-known/oauth-authorization-server is not yet served and authentication operator keeps waiting (check kube-apiserver operator, and check that instances roll out successfully, which can take several minutes per instance)"

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

openshift-kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/kube-apiserver-pod-4 -n openshift-kube-apiserver because it was missing

openshift-authentication

kubelet

oauth-openshift-86f9447d7-znl8c

FailedMount

MountVolume.SetUp failed for volume "v4-0-config-system-session" : failed to sync secret cache: timed out waiting for the condition

openshift-authentication

kubelet

oauth-openshift-86f9447d7-znl8c

FailedMount

MountVolume.SetUp failed for volume "v4-0-config-system-router-certs" : failed to sync secret cache: timed out waiting for the condition

openshift-authentication

kubelet

oauth-openshift-86f9447d7-znl8c

FailedMount

MountVolume.SetUp failed for volume "kube-api-access-x22rd" : failed to sync configmap cache: timed out waiting for the condition

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Degraded message changed from "IngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\nOAuthConfigRouteDegraded: The OAuth server route 'openshift-authentication/oauth-openshift' was not found\nOAuthServerRouteEndpointAccessibleControllerDegraded: Get \"https://oauth-openshift.apps.sno.openstack.lab/healthz\": EOF\nOAuthServerServiceEndpointAccessibleControllerDegraded: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\nWellKnownReadyControllerDegraded: failed to get oauth metadata from openshift-config-managed/oauth-openshift ConfigMap: configmap \"oauth-openshift\" not found (check authentication operator, it is supposed to create this)" to "IngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\nOAuthConfigRouteDegraded: The OAuth server route 'openshift-authentication/oauth-openshift' was not found\nOAuthServerRouteEndpointAccessibleControllerDegraded: Get \"https://oauth-openshift.apps.sno.openstack.lab/healthz\": EOF\nOAuthServerServiceEndpointAccessibleControllerDegraded: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready"

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

openshift-kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/config-4 -n openshift-kube-apiserver because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-resource-sync-controller-resourcesynccontroller

openshift-kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/oauth-metadata -n openshift-kube-apiserver because it was missing
(x4)

openshift-authentication

kubelet

oauth-openshift-86f9447d7-znl8c

FailedMount

MountVolume.SetUp failed for volume "v4-0-config-system-cliconfig" : configmap "v4-0-config-system-cliconfig" not found

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Degraded message changed from "IngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\nOAuthConfigRouteDegraded: The OAuth server route 'openshift-authentication/oauth-openshift' was not found\nOAuthServerRouteEndpointAccessibleControllerDegraded: Get \"https://oauth-openshift.apps.sno.openstack.lab/healthz\": EOF\nOAuthServerServiceEndpointAccessibleControllerDegraded: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready" to "IngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\nOAuthServerRouteEndpointAccessibleControllerDegraded: Get \"https://oauth-openshift.apps.sno.openstack.lab/healthz\": EOF\nOAuthServerServiceEndpointAccessibleControllerDegraded: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready"

openshift-authentication-operator

cluster-authentication-operator-payload-config-controller-openshift-authentication-payloadconfig

authentication-operator

ConfigMapCreated

Created ConfigMap/v4-0-config-system-cliconfig -n openshift-authentication because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

openshift-kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/kube-apiserver-cert-syncer-kubeconfig-4 -n openshift-kube-apiserver because it was missing

openshift-authentication

deployment-controller

oauth-openshift

ScalingReplicaSet

Scaled up replica set oauth-openshift-5bc4f8b544 to 1 from 0

openshift-authentication

deployment-controller

oauth-openshift

ScalingReplicaSet

Scaled down replica set oauth-openshift-86f9447d7 to 0 from 1

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Progressing message changed from "OAuthServerDeploymentProgressing: deployment/oauth-openshift.openshift-authentication: observed generation is 0, desired generation is 1.\nWellKnownReadyControllerProgressing: kube-apiserver oauth endpoint https://192.168.32.10:6443/.well-known/oauth-authorization-server is not yet served and authentication operator keeps waiting (check kube-apiserver operator, and check that instances roll out successfully, which can take several minutes per instance)" to "OAuthServerDeploymentProgressing: deployment/oauth-openshift.openshift-authentication: observed generation is 1, desired generation is 2.\nWellKnownReadyControllerProgressing: kube-apiserver oauth endpoint https://192.168.32.10:6443/.well-known/oauth-authorization-server is not yet served and authentication operator keeps waiting (check kube-apiserver operator, and check that instances roll out successfully, which can take several minutes per instance)"

openshift-authentication

replicaset-controller

oauth-openshift-86f9447d7

SuccessfulDelete

Deleted pod: oauth-openshift-86f9447d7-znl8c

openshift-authentication

replicaset-controller

oauth-openshift-5bc4f8b544

SuccessfulCreate

Created pod: oauth-openshift-5bc4f8b544-czdmk

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

openshift-kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/oauth-metadata-4 -n openshift-kube-apiserver because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

openshift-kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/bound-sa-token-signing-certs-4 -n openshift-kube-apiserver because it was missing

openshift-authentication

kubelet

oauth-openshift-5bc4f8b544-czdmk

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3fdcbf7be3f90bd080ffb2c75b091d7eef03681e0f90912ff6140ee48c177616"

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

openshift-kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/etcd-serving-ca-4 -n openshift-kube-apiserver because it was missing

openshift-authentication

multus

oauth-openshift-5bc4f8b544-czdmk

AddedInterface

Add eth0 [10.128.0.85/23] from ovn-kubernetes

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

openshift-kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/kube-apiserver-server-ca-4 -n openshift-kube-apiserver because it was missing

openshift-authentication

kubelet

oauth-openshift-5bc4f8b544-czdmk

Created

Created container: oauth-openshift

openshift-authentication

kubelet

oauth-openshift-5bc4f8b544-czdmk

Started

Started container oauth-openshift

openshift-authentication

kubelet

oauth-openshift-5bc4f8b544-czdmk

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3fdcbf7be3f90bd080ffb2c75b091d7eef03681e0f90912ff6140ee48c177616" in 1.995s (1.995s including waiting). Image size: 481463651 bytes.

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Degraded message changed from "IngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\nOAuthServerRouteEndpointAccessibleControllerDegraded: Get \"https://oauth-openshift.apps.sno.openstack.lab/healthz\": EOF\nOAuthServerServiceEndpointAccessibleControllerDegraded: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready" to "IngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\nOAuthServerRouteEndpointAccessibleControllerDegraded: Get \"https://oauth-openshift.apps.sno.openstack.lab/healthz\": EOF\nOAuthServerServiceEndpointAccessibleControllerDegraded: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused"

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Degraded message changed from "IngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\nOAuthServerRouteEndpointAccessibleControllerDegraded: Get \"https://oauth-openshift.apps.sno.openstack.lab/healthz\": EOF" to "IngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server"

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Available message changed from "OAuthServerDeploymentAvailable: no oauth-openshift.openshift-authentication pods available on any node.\nOAuthServerRouteEndpointAccessibleControllerAvailable: Get \"https://oauth-openshift.apps.sno.openstack.lab/healthz\": EOF\nOAuthServerServiceEndpointAccessibleControllerAvailable: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerAvailable: endpoints \"oauth-openshift\" not found\nWellKnownAvailable: The well-known endpoint is not yet available: kube-apiserver oauth endpoint https://192.168.32.10:6443/.well-known/oauth-authorization-server is not yet served and authentication operator keeps waiting (check kube-apiserver operator, and check that instances roll out successfully, which can take several minutes per instance)" to "OAuthServerDeploymentAvailable: no oauth-openshift.openshift-authentication pods available on any node.\nOAuthServerRouteEndpointAccessibleControllerAvailable: Get \"https://oauth-openshift.apps.sno.openstack.lab/healthz\": EOF\nOAuthServerServiceEndpointAccessibleControllerAvailable: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused\nWellKnownAvailable: The well-known endpoint is not yet available: kube-apiserver oauth endpoint https://192.168.32.10:6443/.well-known/oauth-authorization-server is not yet served and authentication operator keeps waiting (check kube-apiserver operator, and check that instances roll out successfully, which can take several minutes per instance)"

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Degraded message changed from "IngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\nOAuthServerRouteEndpointAccessibleControllerDegraded: Get \"https://oauth-openshift.apps.sno.openstack.lab/healthz\": EOF\nOAuthServerServiceEndpointAccessibleControllerDegraded: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused" to "IngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\nOAuthServerRouteEndpointAccessibleControllerDegraded: Get \"https://oauth-openshift.apps.sno.openstack.lab/healthz\": EOF"

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Available message changed from "OAuthServerDeploymentAvailable: no oauth-openshift.openshift-authentication pods available on any node.\nOAuthServerRouteEndpointAccessibleControllerAvailable: Get \"https://oauth-openshift.apps.sno.openstack.lab/healthz\": EOF\nOAuthServerServiceEndpointAccessibleControllerAvailable: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused\nWellKnownAvailable: The well-known endpoint is not yet available: kube-apiserver oauth endpoint https://192.168.32.10:6443/.well-known/oauth-authorization-server is not yet served and authentication operator keeps waiting (check kube-apiserver operator, and check that instances roll out successfully, which can take several minutes per instance)" to "OAuthServerDeploymentAvailable: no oauth-openshift.openshift-authentication pods available on any node.\nOAuthServerRouteEndpointAccessibleControllerAvailable: Get \"https://oauth-openshift.apps.sno.openstack.lab/healthz\": EOF\nWellKnownAvailable: The well-known endpoint is not yet available: kube-apiserver oauth endpoint https://192.168.32.10:6443/.well-known/oauth-authorization-server is not yet served and authentication operator keeps waiting (check kube-apiserver operator, and check that instances roll out successfully, which can take several minutes per instance)"

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

openshift-kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/kubelet-serving-ca-4 -n openshift-kube-apiserver because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

openshift-kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/sa-token-signing-certs-4 -n openshift-kube-apiserver because it was missing

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Degraded changed from True to False ("All is well")

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

openshift-kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/kube-apiserver-audit-policies-4 -n openshift-kube-apiserver because it was missing

openshift-machine-config-operator

machine-config-operator

machine-config-operator

ConfigMapUpdated

Updated ConfigMap/kube-rbac-proxy -n openshift-machine-config-operator: cause by changes in data.config-file.yaml

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

openshift-kube-apiserver-operator

SecretCreated

Created Secret/etcd-client-4 -n openshift-kube-apiserver because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

openshift-kube-apiserver-operator

SecretCreated

Created Secret/localhost-recovery-serving-certkey-4 -n openshift-kube-apiserver because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

openshift-kube-apiserver-operator

SecretCreated

Created Secret/localhost-recovery-client-token-4 -n openshift-kube-apiserver because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

openshift-kube-apiserver-operator

SecretCreated

Created Secret/webhook-authenticator-4 -n openshift-kube-apiserver because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

openshift-kube-apiserver-operator

RevisionTriggered

new revision 4 triggered by "required configmap/config has changed"

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Progressing message changed from "OAuthServerDeploymentProgressing: deployment/oauth-openshift.openshift-authentication: observed generation is 1, desired generation is 2.\nWellKnownReadyControllerProgressing: kube-apiserver oauth endpoint https://192.168.32.10:6443/.well-known/oauth-authorization-server is not yet served and authentication operator keeps waiting (check kube-apiserver operator, and check that instances roll out successfully, which can take several minutes per instance)" to "OAuthServerDeploymentProgressing: deployment/oauth-openshift.openshift-authentication: observed generation is 2, desired generation is 3.\nWellKnownReadyControllerProgressing: kube-apiserver oauth endpoint https://192.168.32.10:6443/.well-known/oauth-authorization-server is not yet served and authentication operator keeps waiting (check kube-apiserver operator, and check that instances roll out successfully, which can take several minutes per instance)",Available message changed from "OAuthServerDeploymentAvailable: no oauth-openshift.openshift-authentication pods available on any node.\nWellKnownAvailable: The well-known endpoint is not yet available: kube-apiserver oauth endpoint https://192.168.32.10:6443/.well-known/oauth-authorization-server is not yet served and authentication operator keeps waiting (check kube-apiserver operator, and check that instances roll out successfully, which can take several minutes per instance)" to "WellKnownAvailable: The well-known endpoint is not yet available: kube-apiserver oauth endpoint https://192.168.32.10:6443/.well-known/oauth-authorization-server is not yet served and authentication operator keeps waiting (check kube-apiserver operator, and check that instances roll out successfully, which can take several minutes per instance)"

openshift-authentication

deployment-controller

oauth-openshift

ScalingReplicaSet

Scaled up replica set oauth-openshift-dcb9594d9 to 1 from 0

openshift-authentication

replicaset-controller

oauth-openshift-5bc4f8b544

SuccessfulDelete

Deleted pod: oauth-openshift-5bc4f8b544-czdmk

openshift-authentication

deployment-controller

oauth-openshift

ScalingReplicaSet

Scaled down replica set oauth-openshift-5bc4f8b544 to 0 from 1

openshift-authentication

kubelet

oauth-openshift-5bc4f8b544-czdmk

Killing

Stopping container oauth-openshift

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Degraded message changed from "All is well" to "OAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready"
(x2)

openshift-authentication-operator

cluster-authentication-operator-oauthserver-workloadworkloadcontroller

authentication-operator

DeploymentUpdated

Updated Deployment.apps/oauth-openshift -n openshift-authentication because it changed

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Available message changed from "WellKnownAvailable: The well-known endpoint is not yet available: kube-apiserver oauth endpoint https://192.168.32.10:6443/.well-known/oauth-authorization-server is not yet served and authentication operator keeps waiting (check kube-apiserver operator, and check that instances roll out successfully, which can take several minutes per instance)" to "OAuthServerRouteEndpointAccessibleControllerAvailable: Get \"https://oauth-openshift.apps.sno.openstack.lab/healthz\": EOF\nWellKnownAvailable: The well-known endpoint is not yet available: kube-apiserver oauth endpoint https://192.168.32.10:6443/.well-known/oauth-authorization-server is not yet served and authentication operator keeps waiting (check kube-apiserver operator, and check that instances roll out successfully, which can take several minutes per instance)"

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Degraded message changed from "OAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready" to "OAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\nOAuthServerRouteEndpointAccessibleControllerDegraded: Get \"https://oauth-openshift.apps.sno.openstack.lab/healthz\": EOF"

openshift-authentication

replicaset-controller

oauth-openshift-dcb9594d9

SuccessfulCreate

Created pod: oauth-openshift-dcb9594d9-wlht7

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Degraded message changed from "OAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\nOAuthServerRouteEndpointAccessibleControllerDegraded: Get \"https://oauth-openshift.apps.sno.openstack.lab/healthz\": EOF" to "OAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\nIngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\nOAuthServerRouteEndpointAccessibleControllerDegraded: Get \"https://oauth-openshift.apps.sno.openstack.lab/healthz\": EOF"

openshift-kube-apiserver-operator

kube-apiserver-operator-status-controller-statussyncer_kube-apiserver

openshift-kube-apiserver-operator

OperatorStatusChanged

Status for clusteroperator/kube-apiserver changed: Progressing changed from False to True ("NodeInstallerProgressing: 1 node is at revision 3; 0 nodes have achieved new revision 4"),Available message changed from "StaticPodsAvailable: 1 nodes are active; 1 node is at revision 3" to "StaticPodsAvailable: 1 nodes are active; 1 node is at revision 3; 0 nodes have achieved new revision 4"

openshift-kube-apiserver-operator

kube-apiserver-operator-installer-controller

openshift-kube-apiserver-operator

NodeTargetRevisionChanged

Updating node "master-0" from revision 3 to 4 because node master-0 with revision 3 is the oldest

openshift-kube-apiserver

multus

installer-4-master-0

AddedInterface

Add eth0 [10.128.0.86/23] from ovn-kubernetes

openshift-image-registry

image-registry-operator

cluster-image-registry-operator

DaemonSetCreated

Created DaemonSet.apps/node-ca -n openshift-image-registry because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-installer-controller

openshift-kube-apiserver-operator

PodCreated

Created Pod/installer-4-master-0 -n openshift-kube-apiserver because it was missing

openshift-image-registry

daemonset-controller

node-ca

SuccessfulCreate

Created pod: node-ca-8ffsg

openshift-kube-apiserver

kubelet

installer-4-master-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c5ce3d1134d6500e2b8528516c1889d7bbc6259aba4981c6983395b0e9eeff65" already present on machine

openshift-kube-apiserver

kubelet

installer-4-master-0

Created

Created container: installer

openshift-image-registry

kubelet

node-ca-8ffsg

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:60637f6eed5e9adc3af1863d0ef311c74b9109f00f464f9ce6cdfd21d0ee4608"

openshift-kube-apiserver

kubelet

installer-4-master-0

Started

Started container installer

openshift-image-registry

kubelet

node-ca-8ffsg

Started

Started container node-ca

openshift-image-registry

kubelet

node-ca-8ffsg

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:60637f6eed5e9adc3af1863d0ef311c74b9109f00f464f9ce6cdfd21d0ee4608" in 2.259s (2.259s including waiting). Image size: 481636992 bytes.

openshift-image-registry

kubelet

node-ca-8ffsg

Created

Created container: node-ca

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Progressing message changed from "OAuthServerDeploymentProgressing: deployment/oauth-openshift.openshift-authentication: observed generation is 2, desired generation is 3.\nWellKnownReadyControllerProgressing: kube-apiserver oauth endpoint https://192.168.32.10:6443/.well-known/oauth-authorization-server is not yet served and authentication operator keeps waiting (check kube-apiserver operator, and check that instances roll out successfully, which can take several minutes per instance)" to "OAuthServerDeploymentProgressing: deployment/oauth-openshift.openshift-authentication: 1/1 pods have been updated to the latest generation and 0/1 pods are available\nWellKnownReadyControllerProgressing: kube-apiserver oauth endpoint https://192.168.32.10:6443/.well-known/oauth-authorization-server is not yet served and authentication operator keeps waiting (check kube-apiserver operator, and check that instances roll out successfully, which can take several minutes per instance)",Available message changed from "OAuthServerRouteEndpointAccessibleControllerAvailable: Get \"https://oauth-openshift.apps.sno.openstack.lab/healthz\": EOF\nWellKnownAvailable: The well-known endpoint is not yet available: kube-apiserver oauth endpoint https://192.168.32.10:6443/.well-known/oauth-authorization-server is not yet served and authentication operator keeps waiting (check kube-apiserver operator, and check that instances roll out successfully, which can take several minutes per instance)" to "OAuthServerDeploymentAvailable: no oauth-openshift.openshift-authentication pods available on any node.\nOAuthServerRouteEndpointAccessibleControllerAvailable: Get \"https://oauth-openshift.apps.sno.openstack.lab/healthz\": EOF\nWellKnownAvailable: The well-known endpoint is not yet available: kube-apiserver oauth endpoint https://192.168.32.10:6443/.well-known/oauth-authorization-server is not yet served and authentication operator keeps waiting (check kube-apiserver operator, and check that instances roll out successfully, which can take several minutes per instance)"

openshift-machine-config-operator

machine-config-operator

machine-config-operator

ConfigMapUpdated

Updated ConfigMap/kube-rbac-proxy -n openshift-machine-config-operator: cause by changes in data.config-file.yaml

openshift-authentication

multus

oauth-openshift-dcb9594d9-wlht7

AddedInterface

Add eth0 [10.128.0.87/23] from ovn-kubernetes

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Degraded message changed from "OAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\nIngressStateEndpointsDegraded: All 1 endpoints for oauth-server are reporting 'not ready'\nOAuthServerRouteEndpointAccessibleControllerDegraded: Get \"https://oauth-openshift.apps.sno.openstack.lab/healthz\": EOF" to "IngressStateEndpointsDegraded: All 1 endpoints for oauth-server are reporting 'not ready'\nOAuthServerRouteEndpointAccessibleControllerDegraded: Get \"https://oauth-openshift.apps.sno.openstack.lab/healthz\": EOF"

openshift-authentication

kubelet

oauth-openshift-dcb9594d9-wlht7

Created

Created container: oauth-openshift

openshift-authentication

kubelet

oauth-openshift-dcb9594d9-wlht7

Started

Started container oauth-openshift

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Degraded message changed from "OAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\nIngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\nOAuthServerRouteEndpointAccessibleControllerDegraded: Get \"https://oauth-openshift.apps.sno.openstack.lab/healthz\": EOF" to "OAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\nIngressStateEndpointsDegraded: All 1 endpoints for oauth-server are reporting 'not ready'\nOAuthServerRouteEndpointAccessibleControllerDegraded: Get \"https://oauth-openshift.apps.sno.openstack.lab/healthz\": EOF"

openshift-authentication

kubelet

oauth-openshift-dcb9594d9-wlht7

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3fdcbf7be3f90bd080ffb2c75b091d7eef03681e0f90912ff6140ee48c177616" already present on machine
(x2)

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Available message changed from "OAuthServerDeploymentAvailable: no oauth-openshift.openshift-authentication pods available on any node.\nOAuthServerRouteEndpointAccessibleControllerAvailable: Get \"https://oauth-openshift.apps.sno.openstack.lab/healthz\": EOF\nWellKnownAvailable: The well-known endpoint is not yet available: kube-apiserver oauth endpoint https://192.168.32.10:6443/.well-known/oauth-authorization-server is not yet served and authentication operator keeps waiting (check kube-apiserver operator, and check that instances roll out successfully, which can take several minutes per instance)" to "OAuthServerDeploymentAvailable: no oauth-openshift.openshift-authentication pods available on any node.\nWellKnownAvailable: The well-known endpoint is not yet available: kube-apiserver oauth endpoint https://192.168.32.10:6443/.well-known/oauth-authorization-server is not yet served and authentication operator keeps waiting (check kube-apiserver operator, and check that instances roll out successfully, which can take several minutes per instance)"

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Degraded message changed from "IngressStateEndpointsDegraded: All 1 endpoints for oauth-server are reporting 'not ready'\nOAuthServerRouteEndpointAccessibleControllerDegraded: Get \"https://oauth-openshift.apps.sno.openstack.lab/healthz\": EOF" to "OAuthServerRouteEndpointAccessibleControllerDegraded: Get \"https://oauth-openshift.apps.sno.openstack.lab/healthz\": EOF"

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Degraded message changed from "OAuthServerRouteEndpointAccessibleControllerDegraded: Get \"https://oauth-openshift.apps.sno.openstack.lab/healthz\": EOF" to "All is well"

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Progressing message changed from "OAuthServerDeploymentProgressing: deployment/oauth-openshift.openshift-authentication: 1/1 pods have been updated to the latest generation and 0/1 pods are available\nWellKnownReadyControllerProgressing: kube-apiserver oauth endpoint https://192.168.32.10:6443/.well-known/oauth-authorization-server is not yet served and authentication operator keeps waiting (check kube-apiserver operator, and check that instances roll out successfully, which can take several minutes per instance)" to "WellKnownReadyControllerProgressing: kube-apiserver oauth endpoint https://192.168.32.10:6443/.well-known/oauth-authorization-server is not yet served and authentication operator keeps waiting (check kube-apiserver operator, and check that instances roll out successfully, which can take several minutes per instance)",Available message changed from "OAuthServerDeploymentAvailable: no oauth-openshift.openshift-authentication pods available on any node.\nWellKnownAvailable: The well-known endpoint is not yet available: kube-apiserver oauth endpoint https://192.168.32.10:6443/.well-known/oauth-authorization-server is not yet served and authentication operator keeps waiting (check kube-apiserver operator, and check that instances roll out successfully, which can take several minutes per instance)" to "WellKnownAvailable: The well-known endpoint is not yet available: kube-apiserver oauth endpoint https://192.168.32.10:6443/.well-known/oauth-authorization-server is not yet served and authentication operator keeps waiting (check kube-apiserver operator, and check that instances roll out successfully, which can take several minutes per instance)"

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: status.versions changed from [{"operator" "4.18.35"} {"oauth-apiserver" "4.18.35"}] to [{"operator" "4.18.35"} {"oauth-apiserver" "4.18.35"} {"oauth-openshift" "4.18.35_openshift"}]

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorVersionChanged

clusteroperator/authentication version "oauth-openshift" changed from "" to "4.18.35_openshift"

openshift-machine-config-operator

machine-config-operator

machine-config-operator

ConfigMapUpdated

Updated ConfigMap/kube-rbac-proxy -n openshift-machine-config-operator: cause by changes in data.config-file.yaml

openshift-kube-apiserver

kubelet

kube-apiserver-master-0

Killing

Stopping container kube-apiserver-insecure-readyz

openshift-kube-apiserver

apiserver

kube-apiserver-master-0

InFlightRequestsDrained

All non long-running request(s) in-flight have drained

openshift-kube-apiserver

kubelet

kube-apiserver-master-0

Killing

Stopping container kube-apiserver

openshift-kube-apiserver

kubelet

kube-apiserver-master-0

Killing

Stopping container kube-apiserver-cert-syncer

openshift-kube-apiserver

kubelet

kube-apiserver-master-0

Killing

Stopping container kube-apiserver-cert-regeneration-controller

openshift-kube-apiserver

apiserver

kube-apiserver-master-0

TerminationPreShutdownHooksFinished

All pre-shutdown hooks have been finished

openshift-kube-apiserver

apiserver

kube-apiserver-master-0

ShutdownInitiated

Received signal to terminate, becoming unready, but keeping serving

openshift-kube-apiserver

apiserver

kube-apiserver-master-0

AfterShutdownDelayDuration

The minimal shutdown duration of 0s finished

openshift-kube-apiserver

kubelet

kube-apiserver-master-0

Killing

Stopping container kube-apiserver-check-endpoints

openshift-kube-apiserver

apiserver

kube-apiserver-master-0

HTTPServerStoppedListening

HTTP Server has stopped listening

openshift-kube-apiserver

kubelet

kube-apiserver-startup-monitor-master-0

Created

Created container: startup-monitor

openshift-kube-apiserver

kubelet

kube-apiserver-startup-monitor-master-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c5ce3d1134d6500e2b8528516c1889d7bbc6259aba4981c6983395b0e9eeff65" already present on machine

openshift-kube-apiserver

kubelet

kube-apiserver-startup-monitor-master-0

Started

Started container startup-monitor

openshift-kube-apiserver

apiserver

kube-apiserver-master-0

TerminationGracefulTerminationFinished

All pending requests processed

openshift-machine-config-operator

machineconfigoperator

machine-config

OperatorDegraded: MachineConfigDaemonFailed

Failed to resync 4.18.35 because: failed to apply machine config daemon manifests: Get "https://172.30.0.1:443/apis/rbac.authorization.k8s.io/v1/namespaces/openshift-machine-config-operator/rolebindings/machine-config-daemon": dial tcp 172.30.0.1:443: connect: connection refused

openshift-kube-apiserver

kubelet

kube-apiserver-master-0

Created

Created container: setup

openshift-kube-apiserver

kubelet

kube-apiserver-master-0

Started

Started container setup

openshift-kube-apiserver

kubelet

kube-apiserver-master-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b23c544d3894e5b31f66a18c554f03b0d29f92c2000c46b57b1c96da7ec25db9" already present on machine

openshift-kube-apiserver

kubelet

kube-apiserver-master-0

Started

Started container kube-apiserver

openshift-kube-apiserver

kubelet

kube-apiserver-master-0

Created

Created container: kube-apiserver

openshift-kube-apiserver

kubelet

kube-apiserver-master-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c5ce3d1134d6500e2b8528516c1889d7bbc6259aba4981c6983395b0e9eeff65" already present on machine

openshift-kube-apiserver

kubelet

kube-apiserver-master-0

Created

Created container: kube-apiserver-cert-syncer

openshift-kube-apiserver

kubelet

kube-apiserver-master-0

Started

Started container kube-apiserver-cert-syncer

openshift-kube-apiserver

kubelet

kube-apiserver-master-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b23c544d3894e5b31f66a18c554f03b0d29f92c2000c46b57b1c96da7ec25db9" already present on machine

openshift-kube-apiserver

kubelet

kube-apiserver-master-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c5ce3d1134d6500e2b8528516c1889d7bbc6259aba4981c6983395b0e9eeff65" already present on machine

openshift-kube-apiserver

kubelet

kube-apiserver-master-0

Started

Started container kube-apiserver-check-endpoints

openshift-kube-apiserver

kubelet

kube-apiserver-master-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c5ce3d1134d6500e2b8528516c1889d7bbc6259aba4981c6983395b0e9eeff65" already present on machine

openshift-kube-apiserver

kubelet

kube-apiserver-master-0

Started

Started container kube-apiserver-insecure-readyz

openshift-kube-apiserver

kubelet

kube-apiserver-master-0

Created

Created container: kube-apiserver-insecure-readyz

openshift-kube-apiserver

kubelet

kube-apiserver-master-0

Created

Created container: kube-apiserver-check-endpoints

openshift-kube-apiserver

apiserver

kube-apiserver-master-0

KubeAPIReadyz

readyz=true

openshift-kube-apiserver

cert-regeneration-controller

cert-regeneration-controller-lock

LeaderElection

master-0_c68072d1-b497-456b-9e58-ea3c9d1c798b became leader

openshift-kube-controller-manager

kubelet

kube-controller-manager-master-0

Unhealthy

Startup probe failed: Get "https://192.168.32.10:10257/healthz": dial tcp 192.168.32.10:10257: connect: connection refused

openshift-kube-controller-manager

kubelet

kube-controller-manager-master-0

ProbeError

Startup probe error: Get "https://192.168.32.10:10257/healthz": dial tcp 192.168.32.10:10257: connect: connection refused body:
(x12)

openshift-machine-config-operator

machineconfigoperator

machine-config

OperatorDegraded: MachineConfigPoolsFailed

Failed to resync 4.18.35 because: Get "https://172.30.0.1:443/apis/machineconfiguration.openshift.io/v1/machineconfigpools/master": dial tcp 172.30.0.1:443: connect: connection refused
(x9)

openshift-console-operator

kubelet

console-operator-76b6568d85-8b8gv

FailedMount

MountVolume.SetUp failed for volume "trusted-ca" : configmap references non-existent config key: ca-bundle.crt

openshift-kube-apiserver

kubelet

kube-apiserver-startup-monitor-master-0

Killing

Stopping container startup-monitor

default

node-controller

master-0

RegisteredNode

Node master-0 event: Registered Node master-0 in Controller

kube-system

kube-controller-manager

kube-controller-manager

LeaderElection

master-0_7598bcec-1838-442f-9a28-61cd6be38502 became leader

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Progressing message changed from "WellKnownReadyControllerProgressing: kube-apiserver oauth endpoint https://192.168.32.10:6443/.well-known/oauth-authorization-server is not yet served and authentication operator keeps waiting (check kube-apiserver operator, and check that instances roll out successfully, which can take several minutes per instance)" to "",Available message changed from "WellKnownAvailable: The well-known endpoint is not yet available: kube-apiserver oauth endpoint https://192.168.32.10:6443/.well-known/oauth-authorization-server is not yet served and authentication operator keeps waiting (check kube-apiserver operator, and check that instances roll out successfully, which can take several minutes per instance)" to "WellKnownAvailable: The well-known endpoint is not yet available: failed to get API server IPs: unable to find kube api server endpointLister port: &v1.Endpoints{TypeMeta:v1.TypeMeta{Kind:\"\", APIVersion:\"\"}, ObjectMeta:v1.ObjectMeta{Name:\"kubernetes\", GenerateName:\"\", Namespace:\"default\", SelfLink:\"\", UID:\"0a8e1475-e015-4961-bc7c-f367a856aa53\", ResourceVersion:\"13212\", Generation:0, CreationTimestamp:time.Date(2026, time.March, 20, 8, 29, 8, 0, time.Local), DeletionTimestamp:<nil>, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{\"endpointslice.kubernetes.io/skip-mirror\":\"true\"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry{v1.ManagedFieldsEntry{Manager:\"kube-apiserver\", Operation:\"Update\", APIVersion:\"v1\", Time:time.Date(2026, time.March, 20, 8, 40, 51, 0, time.Local), FieldsType:\"FieldsV1\", FieldsV1:(*v1.FieldsV1)(0xc00360b890), Subresource:\"\"}}}, Subsets:[]v1.EndpointSubset(nil)} (check kube-apiserver that it deploys correctly)"

openshift-etcd-operator

openshift-cluster-etcd-operator

openshift-cluster-etcd-operator-lock

LeaderElection

etcd-operator-8544cbcf9c-brhw4_e42fcfa3-b977-4af2-8bb8-9088a5707955 became leader

openshift-etcd-operator

openshift-cluster-etcd-operator-status-controller-statussyncer_etcd

etcd-operator

OperatorStatusChanged

Status for clusteroperator/etcd changed: Degraded message changed from "ClusterMemberControllerDegraded: could not get list of unhealthy members: giving up getting a cached client after 3 tries\nClusterMemberRemovalControllerDegraded: IsBootstrapComplete failed to determine bootstrap status: IsBootstrapComplete couldn't list the etcd cluster members: giving up getting a cached client after 3 tries\nEtcdEndpointsDegraded: failed to get member list: giving up getting a cached client after 3 tries\nEtcdMembersControllerDegraded: getting cache client could not retrieve endpoints: configmaps lister not synced" to "ClusterMemberControllerDegraded: could not get list of unhealthy members: giving up getting a cached client after 3 tries\nEtcdEndpointsDegraded: failed to get member list: giving up getting a cached client after 3 tries\nEtcdMembersControllerDegraded: getting cache client could not retrieve endpoints: configmaps lister not synced"

openshift-etcd-operator

openshift-cluster-etcd-operator-status-controller-statussyncer_etcd

etcd-operator

OperatorStatusChanged

Status for clusteroperator/etcd changed: Degraded changed from False to True ("ClusterMemberControllerDegraded: could not get list of unhealthy members: giving up getting a cached client after 3 tries\nClusterMemberRemovalControllerDegraded: IsBootstrapComplete failed to determine bootstrap status: IsBootstrapComplete couldn't list the etcd cluster members: giving up getting a cached client after 3 tries\nEtcdEndpointsDegraded: failed to get member list: giving up getting a cached client after 3 tries\nEtcdMembersControllerDegraded: getting cache client could not retrieve endpoints: configmaps lister not synced")

openshift-etcd-operator

openshift-cluster-etcd-operator-status-controller-statussyncer_etcd

etcd-operator

OperatorStatusChanged

Status for clusteroperator/etcd changed: Degraded message changed from "ClusterMemberControllerDegraded: could not get list of unhealthy members: giving up getting a cached client after 3 tries\nEtcdEndpointsDegraded: failed to get member list: giving up getting a cached client after 3 tries" to "EtcdEndpointsDegraded: failed to get member list: giving up getting a cached client after 3 tries"

openshift-etcd-operator

openshift-cluster-etcd-operator-status-controller-statussyncer_etcd

etcd-operator

OperatorStatusChanged

Status for clusteroperator/etcd changed: Degraded message changed from "ClusterMemberControllerDegraded: could not get list of unhealthy members: giving up getting a cached client after 3 tries\nEtcdEndpointsDegraded: failed to get member list: giving up getting a cached client after 3 tries\nEtcdMembersControllerDegraded: getting cache client could not retrieve endpoints: configmaps lister not synced" to "ClusterMemberControllerDegraded: could not get list of unhealthy members: giving up getting a cached client after 3 tries\nEtcdEndpointsDegraded: failed to get member list: giving up getting a cached client after 3 tries"

openshift-etcd-operator

openshift-cluster-etcd-operator-status-controller-statussyncer_etcd

etcd-operator

OperatorStatusChanged

Status for clusteroperator/etcd changed: Degraded message changed from "EtcdEndpointsDegraded: failed to get member list: giving up getting a cached client after 3 tries" to "EtcdEndpointsDegraded: failed to get member list: giving up getting a cached client after 3 tries\nScriptControllerDegraded: \"configmap/etcd-pod\": missing env var values"

openshift-etcd-operator

openshift-cluster-etcd-operator

etcd-operator

FeatureGatesInitialized

FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{"AWSEFSDriverVolumeMetrics", "AdminNetworkPolicy", "AlibabaPlatform", "AzureWorkloadIdentity", "BareMetalLoadBalancer", "BuildCSIVolumes", "ChunkSizeMiB", "CloudDualStackNodeIPs", "DisableKubeletCloudCredentialProviders", "GCPLabelsTags", "HardwareSpeed", "IngressControllerLBSubnetsAWS", "KMSv1", "ManagedBootImages", "ManagedBootImagesAWS", "MultiArchInstallAWS", "MultiArchInstallGCP", "NetworkDiagnosticsConfig", "NetworkLiveMigration", "NetworkSegmentation", "NewOLM", "NodeDisruptionPolicy", "OnClusterBuild", "PersistentIPsForVirtualization", "PrivateHostedZoneAWS", "SetEIPForNLBIngressController", "StreamingCollectionEncodingToJSON", "StreamingCollectionEncodingToProtobuf", "VSphereControlPlaneMachineSet", "VSphereDriverConfiguration", "VSphereMultiVCenters", "VSphereStaticIPs", "ValidatingAdmissionPolicy"}, Disabled:[]v1.FeatureGateName{"AWSClusterHostedDNS", "AdditionalRoutingCapabilities", "AutomatedEtcdBackup", "BootcNodeManagement", "CSIDriverSharedResource", "ClusterAPIInstall", "ClusterAPIInstallIBMCloud", "ClusterMonitoringConfig", "ConsolePluginContentSecurityPolicy", "DNSNameResolver", "DynamicResourceAllocation", "EtcdBackendQuota", "EventedPLEG", "Example", "ExternalOIDC", "ExternalOIDCWithUIDAndExtraClaimMappings", "GCPClusterHostedDNS", "GatewayAPI", "ImageStreamImportMode", "IngressControllerDynamicConfigurationManager", "InsightsConfig", "InsightsConfigAPI", "InsightsOnDemandDataGather", "InsightsRuntimeExtractor", "MachineAPIMigration", "MachineAPIOperatorDisableMachineHealthCheckController", "MachineAPIProviderOpenStack", "MachineConfigNodes", "MaxUnavailableStatefulSet", "MetricsCollectionProfiles", "MinimumKubeletVersion", "MixedCPUsAllocation", "MultiArchInstallAzure", "NodeSwap", "NutanixMultiSubnets", "OVNObservability", "OpenShiftPodSecurityAdmission", "PinnedImages", "PlatformOperators", "ProcMountType", "RouteAdvertisements", "RouteExternalCertificate", "ServiceAccountTokenNodeBinding", "SignatureStores", "SigstoreImageVerification", "TranslateStreamCloseWebsocketRequests", "UpgradeStatus", "UserNamespacesPodSecurityStandards", "UserNamespacesSupport", "VSphereMultiNetworks", "VolumeAttributesClass", "VolumeGroupSnapshot"}}
(x2)

openshift-etcd-operator

openshift-cluster-etcd-operator-member-observer-controller-etcdmemberscontroller

etcd-operator

ReportEtcdMembersErrorUpdatingStatus

etcds.operator.openshift.io "cluster" not found

openshift-controller-manager-operator

openshift-controller-manager-operator

openshift-controller-manager-operator-lock

LeaderElection

openshift-controller-manager-operator-8c94f4649-p7pt6_97ce7b32-5972-48cf-86c3-7bd935bdd089 became leader

openshift-multus

daemonset-controller

cni-sysctl-allowlist-ds

SuccessfulCreate

Created pod: cni-sysctl-allowlist-ds-x2bwv

openshift-controller-manager-operator

openshift-controller-manager-operator-config-observer-configobserver

openshift-controller-manager-operator

ObservedConfigChanged

Writing updated observed config: map[string]any{ "build": map[string]any{"buildDefaults": map[string]any{"resources": map[string]any{}}, "imageTemplateFormat": map[string]any{"format": string("quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e7030c5cce"...)}}, "controllers": []any{ ... // 8 identical elements string("openshift.io/deploymentconfig"), string("openshift.io/image-import"), strings.Join({ + "-", "openshift.io/image-puller-rolebindings", }, ""), string("openshift.io/image-signature-import"), string("openshift.io/image-trigger"), ... // 2 identical elements string("openshift.io/origin-namespace"), string("openshift.io/serviceaccount"), strings.Join({ + "-", "openshift.io/serviceaccount-pull-secrets", }, ""), string("openshift.io/templateinstance"), string("openshift.io/templateinstancefinalizer"), string("openshift.io/unidling"), }, "deployer": map[string]any{"imageTemplateFormat": map[string]any{"format": string("quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6e5004457a"...)}}, "featureGates": []any{string("BuildCSIVolumes=true")}, "ingress": map[string]any{"ingressIPNetworkCIDR": string("")}, }

openshift-controller-manager-operator

openshift-controller-manager-operator

openshift-controller-manager-operator

FeatureGatesInitialized

FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{"AWSEFSDriverVolumeMetrics", "AdminNetworkPolicy", "AlibabaPlatform", "AzureWorkloadIdentity", "BareMetalLoadBalancer", "BuildCSIVolumes", "ChunkSizeMiB", "CloudDualStackNodeIPs", "DisableKubeletCloudCredentialProviders", "GCPLabelsTags", "HardwareSpeed", "IngressControllerLBSubnetsAWS", "KMSv1", "ManagedBootImages", "ManagedBootImagesAWS", "MultiArchInstallAWS", "MultiArchInstallGCP", "NetworkDiagnosticsConfig", "NetworkLiveMigration", "NetworkSegmentation", "NewOLM", "NodeDisruptionPolicy", "OnClusterBuild", "PersistentIPsForVirtualization", "PrivateHostedZoneAWS", "SetEIPForNLBIngressController", "StreamingCollectionEncodingToJSON", "StreamingCollectionEncodingToProtobuf", "VSphereControlPlaneMachineSet", "VSphereDriverConfiguration", "VSphereMultiVCenters", "VSphereStaticIPs", "ValidatingAdmissionPolicy"}, Disabled:[]v1.FeatureGateName{"AWSClusterHostedDNS", "AdditionalRoutingCapabilities", "AutomatedEtcdBackup", "BootcNodeManagement", "CSIDriverSharedResource", "ClusterAPIInstall", "ClusterAPIInstallIBMCloud", "ClusterMonitoringConfig", "ConsolePluginContentSecurityPolicy", "DNSNameResolver", "DynamicResourceAllocation", "EtcdBackendQuota", "EventedPLEG", "Example", "ExternalOIDC", "ExternalOIDCWithUIDAndExtraClaimMappings", "GCPClusterHostedDNS", "GatewayAPI", "ImageStreamImportMode", "IngressControllerDynamicConfigurationManager", "InsightsConfig", "InsightsConfigAPI", "InsightsOnDemandDataGather", "InsightsRuntimeExtractor", "MachineAPIMigration", "MachineAPIOperatorDisableMachineHealthCheckController", "MachineAPIProviderOpenStack", "MachineConfigNodes", "MaxUnavailableStatefulSet", "MetricsCollectionProfiles", "MinimumKubeletVersion", "MixedCPUsAllocation", "MultiArchInstallAzure", "NodeSwap", "NutanixMultiSubnets", "OVNObservability", "OpenShiftPodSecurityAdmission", "PinnedImages", "PlatformOperators", "ProcMountType", "RouteAdvertisements", "RouteExternalCertificate", "ServiceAccountTokenNodeBinding", "SignatureStores", "SigstoreImageVerification", "TranslateStreamCloseWebsocketRequests", "UpgradeStatus", "UserNamespacesPodSecurityStandards", "UserNamespacesSupport", "VSphereMultiNetworks", "VolumeAttributesClass", "VolumeGroupSnapshot"}}

openshift-multus

daemonset-controller

cni-sysctl-allowlist-ds

SuccessfulCreate

Created pod: cni-sysctl-allowlist-ds-x2bwv

openshift-network-operator

network-operator

network-operator-lock

LeaderElection

master-0_abc2eaef-927c-4383-a06b-da14d06650d5 became leader

openshift-network-operator

cluster-network-operator

network-operator

FeatureGatesInitialized

FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{"AWSEFSDriverVolumeMetrics", "AdminNetworkPolicy", "AlibabaPlatform", "AzureWorkloadIdentity", "BareMetalLoadBalancer", "BuildCSIVolumes", "ChunkSizeMiB", "CloudDualStackNodeIPs", "DisableKubeletCloudCredentialProviders", "GCPLabelsTags", "HardwareSpeed", "IngressControllerLBSubnetsAWS", "KMSv1", "ManagedBootImages", "ManagedBootImagesAWS", "MultiArchInstallAWS", "MultiArchInstallGCP", "NetworkDiagnosticsConfig", "NetworkLiveMigration", "NetworkSegmentation", "NewOLM", "NodeDisruptionPolicy", "OnClusterBuild", "PersistentIPsForVirtualization", "PrivateHostedZoneAWS", "SetEIPForNLBIngressController", "StreamingCollectionEncodingToJSON", "StreamingCollectionEncodingToProtobuf", "VSphereControlPlaneMachineSet", "VSphereDriverConfiguration", "VSphereMultiVCenters", "VSphereStaticIPs", "ValidatingAdmissionPolicy"}, Disabled:[]v1.FeatureGateName{"AWSClusterHostedDNS", "AdditionalRoutingCapabilities", "AutomatedEtcdBackup", "BootcNodeManagement", "CSIDriverSharedResource", "ClusterAPIInstall", "ClusterAPIInstallIBMCloud", "ClusterMonitoringConfig", "ConsolePluginContentSecurityPolicy", "DNSNameResolver", "DynamicResourceAllocation", "EtcdBackendQuota", "EventedPLEG", "Example", "ExternalOIDC", "ExternalOIDCWithUIDAndExtraClaimMappings", "GCPClusterHostedDNS", "GatewayAPI", "ImageStreamImportMode", "IngressControllerDynamicConfigurationManager", "InsightsConfig", "InsightsConfigAPI", "InsightsOnDemandDataGather", "InsightsRuntimeExtractor", "MachineAPIMigration", "MachineAPIOperatorDisableMachineHealthCheckController", "MachineAPIProviderOpenStack", "MachineConfigNodes", "MaxUnavailableStatefulSet", "MetricsCollectionProfiles", "MinimumKubeletVersion", "MixedCPUsAllocation", "MultiArchInstallAzure", "NodeSwap", "NutanixMultiSubnets", "OVNObservability", "OpenShiftPodSecurityAdmission", "PinnedImages", "PlatformOperators", "ProcMountType", "RouteAdvertisements", "RouteExternalCertificate", "ServiceAccountTokenNodeBinding", "SignatureStores", "SigstoreImageVerification", "TranslateStreamCloseWebsocketRequests", "UpgradeStatus", "UserNamespacesPodSecurityStandards", "UserNamespacesSupport", "VSphereMultiNetworks", "VolumeAttributesClass", "VolumeGroupSnapshot"}}

openshift-multus

kubelet

cni-sysctl-allowlist-ds-x2bwv

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5a55ec7ec64efd0f595d8084377b7e463a1807829b7617e5d4a9092dcd924c36" already present on machine

openshift-multus

kubelet

cni-sysctl-allowlist-ds-x2bwv

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5a55ec7ec64efd0f595d8084377b7e463a1807829b7617e5d4a9092dcd924c36" already present on machine

openshift-controller-manager-operator

openshift-controller-manager-operator

openshift-controller-manager-operator

ConfigMapUpdated

Updated ConfigMap/config -n openshift-route-controller-manager: cause by changes in data.config.yaml

openshift-multus

kubelet

cni-sysctl-allowlist-ds-x2bwv

Started

Started container kube-multus-additional-cni-plugins

openshift-multus

kubelet

cni-sysctl-allowlist-ds-x2bwv

Created

Created container: kube-multus-additional-cni-plugins

openshift-etcd-operator

openshift-cluster-etcd-operator-revisioncontroller

etcd-operator

StartingNewRevision

new revision 2 triggered by "required configmap/etcd-endpoints has changed"

openshift-controller-manager-operator

openshift-controller-manager-operator

openshift-controller-manager-operator

ConfigMapUpdated

Updated ConfigMap/config -n openshift-controller-manager: cause by changes in data.config.yaml

openshift-etcd-operator

openshift-cluster-etcd-operator-etcd-endpoints-controller-etcdendpointscontroller

etcd-operator

ConfigMapUpdated

Updated ConfigMap/etcd-endpoints -n openshift-etcd: cause by changes in data.91eb892c5ee87610,data.MTkyLjE2OC4zMi4xMA

openshift-multus

kubelet

cni-sysctl-allowlist-ds-x2bwv

Started

Started container kube-multus-additional-cni-plugins

openshift-multus

kubelet

cni-sysctl-allowlist-ds-x2bwv

Created

Created container: kube-multus-additional-cni-plugins

openshift-route-controller-manager

kubelet

route-controller-manager-56f686584b-fdcx5

Killing

Stopping container route-controller-manager

openshift-controller-manager

kubelet

controller-manager-fc56bb77c-qd4sn

Killing

Stopping container controller-manager

openshift-controller-manager

deployment-controller

controller-manager

ScalingReplicaSet

Scaled up replica set controller-manager-bfcf4df58 to 1 from 0

openshift-controller-manager-operator

openshift-controller-manager-operator

openshift-controller-manager-operator

DeploymentUpdated

Updated Deployment.apps/controller-manager -n openshift-controller-manager because it changed

openshift-controller-manager

replicaset-controller

controller-manager-fc56bb77c

SuccessfulDelete

Deleted pod: controller-manager-fc56bb77c-qd4sn

openshift-route-controller-manager

deployment-controller

route-controller-manager

ScalingReplicaSet

Scaled up replica set route-controller-manager-ffdd4b479 to 1 from 0

openshift-route-controller-manager

deployment-controller

route-controller-manager

ScalingReplicaSet

Scaled down replica set route-controller-manager-56f686584b to 0 from 1

openshift-controller-manager

deployment-controller

controller-manager

ScalingReplicaSet

Scaled down replica set controller-manager-fc56bb77c to 0 from 1

openshift-route-controller-manager

replicaset-controller

route-controller-manager-ffdd4b479

SuccessfulCreate

Created pod: route-controller-manager-ffdd4b479-rhmfx

openshift-route-controller-manager

replicaset-controller

route-controller-manager-56f686584b

SuccessfulDelete

Deleted pod: route-controller-manager-56f686584b-fdcx5

openshift-controller-manager

replicaset-controller

controller-manager-bfcf4df58

SuccessfulCreate

Created pod: controller-manager-bfcf4df58-l6xz7

openshift-controller-manager-operator

openshift-controller-manager-operator

openshift-controller-manager-operator

DeploymentUpdated

Updated Deployment.apps/route-controller-manager -n openshift-route-controller-manager because it changed

openshift-etcd-operator

openshift-cluster-etcd-operator-installer-controller

etcd-operator

NodeCurrentRevisionChanged

Updated node "master-0" from revision 0 to 1 because static pod is ready

openshift-controller-manager-operator

openshift-controller-manager-operator-status-controller-statussyncer_openshift-controller-manager

openshift-controller-manager-operator

OperatorVersionChanged

clusteroperator/openshift-controller-manager version "operator" changed from "" to "4.18.35"

openshift-controller-manager-operator

openshift-controller-manager-operator-status-controller-statussyncer_openshift-controller-manager

openshift-controller-manager-operator

OperatorStatusChanged

Status for clusteroperator/openshift-controller-manager changed: Progressing message changed from "Progressing: deployment/controller-manager: observed generation is 4, desired generation is 5.\nProgressing: deployment/controller-manager: available replicas is 0, desired available replicas > 1\nProgressing: deployment/route-controller-manager: observed generation is 4, desired generation is 5.\nProgressing: deployment/route-controller-manager: available replicas is 0, desired available replicas > 1" to "Progressing: deployment/controller-manager: observed generation is 5, desired generation is 6.\nProgressing: deployment/route-controller-manager: observed generation is 5, desired generation is 6.\nProgressing: openshiftcontrollermanagers.operator.openshift.io/cluster: observed generation is 2, desired generation is 3.",Available changed from False to True ("All is well"),status.versions changed from [] to [{"operator" "4.18.35"}]

openshift-multus

kubelet

cni-sysctl-allowlist-ds-x2bwv

Killing

Stopping container kube-multus-additional-cni-plugins

openshift-controller-manager

openshift-controller-manager

openshift-master-controllers

LeaderElection

controller-manager-bfcf4df58-l6xz7 became leader

openshift-controller-manager

multus

controller-manager-bfcf4df58-l6xz7

AddedInterface

Add eth0 [10.128.0.88/23] from ovn-kubernetes

openshift-route-controller-manager

multus

route-controller-manager-ffdd4b479-rhmfx

AddedInterface

Add eth0 [10.128.0.89/23] from ovn-kubernetes

openshift-multus

kubelet

cni-sysctl-allowlist-ds-x2bwv

Killing

Stopping container kube-multus-additional-cni-plugins

openshift-etcd-operator

openshift-cluster-etcd-operator-revisioncontroller

etcd-operator

ConfigMapCreated

Created ConfigMap/etcd-pod-2 -n openshift-etcd because it was missing

openshift-monitoring

deployment-controller

telemeter-client

ScalingReplicaSet

Scaled up replica set telemeter-client-54684fb888 to 1

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ConfigMapCreated

Created ConfigMap/telemeter-trusted-ca-bundle-8i12ta5c71j38 -n openshift-monitoring because it was missing

openshift-monitoring

deployment-controller

telemeter-client

ScalingReplicaSet

Scaled up replica set telemeter-client-54684fb888 to 1

openshift-monitoring

replicaset-controller

telemeter-client-54684fb888

SuccessfulCreate

Created pod: telemeter-client-54684fb888-p62jk

openshift-route-controller-manager

route-controller-manager

openshift-route-controllers

LeaderElection

route-controller-manager-ffdd4b479-rhmfx_ec40cc88-2b4c-4c34-85dd-9336def7e739 became leader

openshift-monitoring

replicaset-controller

telemeter-client-54684fb888

SuccessfulCreate

Created pod: telemeter-client-54684fb888-p62jk

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ConfigMapCreated

Created ConfigMap/telemeter-trusted-ca-bundle-8i12ta5c71j38 -n openshift-monitoring because it was missing

openshift-monitoring

kubelet

telemeter-client-54684fb888-p62jk

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1be9cf7afc785fbde8c9d5403d13569bc7f7fee8a386d2d8842f2b40758ed430"

openshift-etcd-operator

openshift-cluster-etcd-operator-revisioncontroller

etcd-operator

ConfigMapCreated

Created ConfigMap/etcd-endpoints-2 -n openshift-etcd because it was missing

openshift-monitoring

multus

telemeter-client-54684fb888-p62jk

AddedInterface

Add eth0 [10.128.0.90/23] from ovn-kubernetes

openshift-monitoring

kubelet

telemeter-client-54684fb888-p62jk

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1be9cf7afc785fbde8c9d5403d13569bc7f7fee8a386d2d8842f2b40758ed430"

openshift-monitoring

multus

telemeter-client-54684fb888-p62jk

AddedInterface

Add eth0 [10.128.0.90/23] from ovn-kubernetes

openshift-controller-manager-operator

openshift-controller-manager-operator-status-controller-statussyncer_openshift-controller-manager

openshift-controller-manager-operator

OperatorStatusChanged

Status for clusteroperator/openshift-controller-manager changed: Progressing message changed from "Progressing: deployment/controller-manager: observed generation is 5, desired generation is 6.\nProgressing: deployment/route-controller-manager: observed generation is 5, desired generation is 6.\nProgressing: openshiftcontrollermanagers.operator.openshift.io/cluster: observed generation is 2, desired generation is 3." to "Progressing: deployment/controller-manager: available replicas is 0, desired available replicas > 1\nProgressing: deployment/route-controller-manager: available replicas is 0, desired available replicas > 1",Available changed from True to False ("Available: no pods available on any node.")

openshift-etcd-operator

openshift-cluster-etcd-operator-revisioncontroller

etcd-operator

ConfigMapCreated

Created ConfigMap/etcd-all-bundles-2 -n openshift-etcd because it was missing

openshift-monitoring

kubelet

telemeter-client-54684fb888-p62jk

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1be9cf7afc785fbde8c9d5403d13569bc7f7fee8a386d2d8842f2b40758ed430" in 2.055s (2.055s including waiting). Image size: 480540851 bytes.

openshift-monitoring

kubelet

telemeter-client-54684fb888-p62jk

Started

Started container telemeter-client

openshift-monitoring

kubelet

telemeter-client-54684fb888-p62jk

Created

Created container: telemeter-client

openshift-monitoring

kubelet

telemeter-client-54684fb888-p62jk

Created

Created container: telemeter-client

openshift-monitoring

kubelet

telemeter-client-54684fb888-p62jk

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1be9cf7afc785fbde8c9d5403d13569bc7f7fee8a386d2d8842f2b40758ed430" in 2.055s (2.055s including waiting). Image size: 480540851 bytes.

openshift-monitoring

kubelet

telemeter-client-54684fb888-p62jk

Started

Started container telemeter-client

openshift-monitoring

kubelet

telemeter-client-54684fb888-p62jk

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ea36739b7e81007aec2e901639b356b275434362b254800d4309dd0aa665ca36"

openshift-monitoring

kubelet

telemeter-client-54684fb888-p62jk

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ea36739b7e81007aec2e901639b356b275434362b254800d4309dd0aa665ca36"

openshift-etcd-operator

openshift-cluster-etcd-operator-revisioncontroller

etcd-operator

SecretCreated

Created Secret/etcd-all-certs-2 -n openshift-etcd because it was missing

openshift-monitoring

kubelet

telemeter-client-54684fb888-p62jk

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-monitoring

kubelet

telemeter-client-54684fb888-p62jk

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ea36739b7e81007aec2e901639b356b275434362b254800d4309dd0aa665ca36" in 1.504s (1.504s including waiting). Image size: 437909443 bytes.

openshift-monitoring

kubelet

telemeter-client-54684fb888-p62jk

Created

Created container: reload

openshift-monitoring

kubelet

telemeter-client-54684fb888-p62jk

Started

Started container reload

openshift-multus

deployment-controller

multus-admission-controller

ScalingReplicaSet

Scaled up replica set multus-admission-controller-649577484c to 1

openshift-multus

replicaset-controller

multus-admission-controller-649577484c

SuccessfulCreate

Created pod: multus-admission-controller-649577484c-p72cd

openshift-multus

replicaset-controller

multus-admission-controller-649577484c

SuccessfulCreate

Created pod: multus-admission-controller-649577484c-p72cd

openshift-monitoring

kubelet

telemeter-client-54684fb888-p62jk

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-monitoring

kubelet

telemeter-client-54684fb888-p62jk

Created

Created container: kube-rbac-proxy

openshift-monitoring

kubelet

telemeter-client-54684fb888-p62jk

Started

Started container kube-rbac-proxy

openshift-monitoring

kubelet

telemeter-client-54684fb888-p62jk

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ea36739b7e81007aec2e901639b356b275434362b254800d4309dd0aa665ca36" in 1.504s (1.504s including waiting). Image size: 437909443 bytes.

openshift-monitoring

kubelet

telemeter-client-54684fb888-p62jk

Created

Created container: reload

openshift-monitoring

kubelet

telemeter-client-54684fb888-p62jk

Started

Started container reload

openshift-multus

deployment-controller

multus-admission-controller

ScalingReplicaSet

Scaled up replica set multus-admission-controller-649577484c to 1

openshift-monitoring

kubelet

telemeter-client-54684fb888-p62jk

Created

Created container: kube-rbac-proxy

openshift-monitoring

kubelet

telemeter-client-54684fb888-p62jk

Started

Started container kube-rbac-proxy

openshift-multus

kubelet

multus-admission-controller-649577484c-p72cd

Created

Created container: kube-rbac-proxy

openshift-multus

kubelet

multus-admission-controller-649577484c-p72cd

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-multus

kubelet

multus-admission-controller-649577484c-p72cd

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bcb08821551e9a5b9f82aa794bcea673279cefb93cb47492e19ccac5e2cf18fe" already present on machine

openshift-multus

kubelet

multus-admission-controller-649577484c-p72cd

Created

Created container: multus-admission-controller

openshift-multus

kubelet

multus-admission-controller-649577484c-p72cd

Started

Started container multus-admission-controller

openshift-multus

kubelet

multus-admission-controller-649577484c-p72cd

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-multus

kubelet

multus-admission-controller-649577484c-p72cd

Started

Started container kube-rbac-proxy

openshift-kube-storage-version-migrator-operator

openshift-kube-storage-version-migrator-operator

openshift-kube-storage-version-migrator-operator-lock

LeaderElection

kube-storage-version-migrator-operator-6bb5bfb6fd-64dg5_fe80d8cc-e657-45f0-b6e0-d15e93614167 became leader

openshift-multus

multus

multus-admission-controller-649577484c-p72cd

AddedInterface

Add eth0 [10.128.0.91/23] from ovn-kubernetes

openshift-multus

multus

multus-admission-controller-649577484c-p72cd

AddedInterface

Add eth0 [10.128.0.91/23] from ovn-kubernetes

openshift-multus

kubelet

multus-admission-controller-649577484c-p72cd

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bcb08821551e9a5b9f82aa794bcea673279cefb93cb47492e19ccac5e2cf18fe" already present on machine

openshift-multus

kubelet

multus-admission-controller-649577484c-p72cd

Created

Created container: multus-admission-controller

openshift-multus

kubelet

multus-admission-controller-649577484c-p72cd

Started

Started container multus-admission-controller

openshift-multus

kubelet

multus-admission-controller-649577484c-p72cd

Started

Started container kube-rbac-proxy

openshift-multus

kubelet

multus-admission-controller-649577484c-p72cd

Created

Created container: kube-rbac-proxy

openshift-multus

kubelet

multus-admission-controller-5dbbb8b86f-5rrrh

Killing

Stopping container kube-rbac-proxy

openshift-multus

replicaset-controller

multus-admission-controller-5dbbb8b86f

SuccessfulDelete

Deleted pod: multus-admission-controller-5dbbb8b86f-5rrrh

openshift-multus

kubelet

multus-admission-controller-5dbbb8b86f-5rrrh

Killing

Stopping container multus-admission-controller

openshift-multus

kubelet

multus-admission-controller-5dbbb8b86f-5rrrh

Killing

Stopping container kube-rbac-proxy

openshift-multus

replicaset-controller

multus-admission-controller-5dbbb8b86f

SuccessfulDelete

Deleted pod: multus-admission-controller-5dbbb8b86f-5rrrh

openshift-multus

deployment-controller

multus-admission-controller

ScalingReplicaSet

Scaled down replica set multus-admission-controller-5dbbb8b86f to 0 from 1

openshift-multus

kubelet

multus-admission-controller-5dbbb8b86f-5rrrh

Killing

Stopping container multus-admission-controller

openshift-multus

deployment-controller

multus-admission-controller

ScalingReplicaSet

Scaled down replica set multus-admission-controller-5dbbb8b86f to 0 from 1

openshift-etcd-operator

openshift-cluster-etcd-operator-installer-controller

etcd-operator

NodeTargetRevisionChanged

Updating node "master-0" from revision 1 to 2 because node master-0 with revision 1 is the oldest

openshift-etcd-operator

openshift-cluster-etcd-operator-installer-controller

etcd-operator

PodCreated

Created Pod/installer-2-master-0 -n openshift-etcd because it was missing

openshift-etcd

kubelet

installer-2-master-0

Started

Started container installer

openshift-etcd

kubelet

installer-2-master-0

Created

Created container: installer

openshift-etcd

multus

installer-2-master-0

AddedInterface

Add eth0 [10.128.0.92/23] from ovn-kubernetes

openshift-etcd

kubelet

installer-2-master-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d5a971d5889f167cfe61a64c366424b87c17a6dc141ffcc43406cdcbb50cae2a" already present on machine

openshift-machine-config-operator

machine-config-operator

machine-config-operator

ConfigMapUpdated

Updated ConfigMap/kube-rbac-proxy -n openshift-machine-config-operator: cause by changes in data.config-file.yaml

openshift-controller-manager-operator

openshift-controller-manager-operator-status-controller-statussyncer_openshift-controller-manager

openshift-controller-manager-operator

OperatorStatusChanged

Status for clusteroperator/openshift-controller-manager changed: Progressing changed from True to False ("All is well"),Available changed from False to True ("All is well")

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ConfigMapUpdated

Updated ConfigMap/metrics-client-ca -n openshift-monitoring: cause by changes in data.client-ca.crt

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ConfigMapUpdated

Updated ConfigMap/metrics-client-ca -n openshift-monitoring: cause by changes in data.client-ca.crt

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ValidatingWebhookConfigurationUpdated

Updated ValidatingWebhookConfiguration.admissionregistration.k8s.io/alertmanagerconfigs.openshift.io because it changed

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ValidatingWebhookConfigurationUpdated

Updated ValidatingWebhookConfiguration.admissionregistration.k8s.io/prometheusrules.openshift.io because it changed

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ValidatingWebhookConfigurationUpdated

Updated ValidatingWebhookConfiguration.admissionregistration.k8s.io/alertmanagerconfigs.openshift.io because it changed

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ValidatingWebhookConfigurationUpdated

Updated ValidatingWebhookConfiguration.admissionregistration.k8s.io/prometheusrules.openshift.io because it changed

openshift-kube-controller-manager

cluster-policy-controller-namespace-security-allocation-controller

kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openshift-network-console namespace

openshift-network-console

deployment-controller

networking-console-plugin

ScalingReplicaSet

Scaled up replica set networking-console-plugin-7c6b76c555 to 1

openshift-network-console

replicaset-controller

networking-console-plugin-7c6b76c555

SuccessfulCreate

Created pod: networking-console-plugin-7c6b76c555-zbpk4

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ClusterRoleCreated

Created ClusterRole.rbac.authorization.k8s.io/thanos-querier because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ClusterRoleBindingCreated

Created ClusterRoleBinding.rbac.authorization.k8s.io/alertmanager-main because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ClusterRoleBindingCreated

Created ClusterRoleBinding.rbac.authorization.k8s.io/thanos-querier because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ClusterRoleCreated

Created ClusterRole.rbac.authorization.k8s.io/thanos-querier because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

SecretCreated

Created Secret/thanos-querier-kube-rbac-proxy-web -n openshift-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ServiceAccountCreated

Created ServiceAccount/alertmanager-main -n openshift-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ServiceAccountCreated

Created ServiceAccount/thanos-querier -n openshift-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ClusterRoleCreated

Created ClusterRole.rbac.authorization.k8s.io/alertmanager-main because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

SecretCreated

Created Secret/thanos-querier-kube-rbac-proxy-web -n openshift-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ServiceAccountCreated

Created ServiceAccount/thanos-querier -n openshift-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ServiceAccountCreated

Created ServiceAccount/alertmanager-main -n openshift-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ServiceCreated

Created Service/alertmanager-main -n openshift-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ClusterRoleBindingCreated

Created ClusterRoleBinding.rbac.authorization.k8s.io/alertmanager-main because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ClusterRoleCreated

Created ClusterRole.rbac.authorization.k8s.io/alertmanager-main because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ConfigMapCreated

Created ConfigMap/alertmanager-trusted-ca-bundle -n openshift-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ClusterRoleBindingCreated

Created ClusterRoleBinding.rbac.authorization.k8s.io/thanos-querier because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ServiceCreated

Created Service/alertmanager-main -n openshift-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ConfigMapCreated

Created ConfigMap/alertmanager-trusted-ca-bundle -n openshift-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

RoleCreated

Created Role.rbac.authorization.k8s.io/prometheus-k8s -n kube-system because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

RoleCreated

Created Role.rbac.authorization.k8s.io/prometheus-k8s-config -n openshift-monitoring because it was missing

openshift-monitoring

deployment-controller

thanos-querier

ScalingReplicaSet

Scaled up replica set thanos-querier-5dc898db58 to 1

openshift-monitoring

replicaset-controller

thanos-querier-5dc898db58

SuccessfulCreate

Created pod: thanos-querier-5dc898db58-5m6qg

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

RoleCreated

Created Role.rbac.authorization.k8s.io/prometheus-k8s -n openshift-user-workload-monitoring because it was missing

openshift-monitoring

deployment-controller

thanos-querier

ScalingReplicaSet

Scaled up replica set thanos-querier-5dc898db58 to 1

openshift-monitoring

replicaset-controller

thanos-querier-5dc898db58

SuccessfulCreate

Created pod: thanos-querier-5dc898db58-5m6qg

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

SecretCreated

Created Secret/kube-rbac-proxy -n openshift-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

RoleCreated

Created Role.rbac.authorization.k8s.io/prometheus-k8s -n kube-system because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

SecretCreated

Created Secret/prometheus-k8s-kube-rbac-proxy-web -n openshift-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ServiceAccountCreated

Created ServiceAccount/prometheus-k8s -n openshift-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

RoleBindingCreated

Created RoleBinding.rbac.authorization.k8s.io/alertmanager-prometheusk8s -n openshift-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

RoleCreated

Created Role.rbac.authorization.k8s.io/prometheus-k8s -n openshift-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

SecretCreated

Created Secret/thanos-querier-grpc-tls-anne24rv49795 -n openshift-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

RoleCreated

Created Role.rbac.authorization.k8s.io/prometheus-k8s -n default because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

RoleCreated

Created Role.rbac.authorization.k8s.io/prometheus-k8s -n openshift-user-workload-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

RoleCreated

Created Role.rbac.authorization.k8s.io/prometheus-k8s -n openshift-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

SecretCreated

Created Secret/kube-rbac-proxy -n openshift-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

RoleCreated

Created Role.rbac.authorization.k8s.io/prometheus-k8s -n default because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

SecretCreated

Created Secret/prometheus-k8s-kube-rbac-proxy-web -n openshift-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ServiceAccountCreated

Created ServiceAccount/prometheus-k8s -n openshift-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

RoleBindingCreated

Created RoleBinding.rbac.authorization.k8s.io/alertmanager-prometheusk8s -n openshift-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

SecretCreated

Created Secret/thanos-querier-grpc-tls-anne24rv49795 -n openshift-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

RoleCreated

Created Role.rbac.authorization.k8s.io/prometheus-k8s-config -n openshift-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ServiceCreated

Created Service/prometheus-k8s-thanos-sidecar -n openshift-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

RoleBindingCreated

Created RoleBinding.rbac.authorization.k8s.io/prometheus-k8s -n openshift-user-workload-monitoring because it was missing
(x3)

openshift-multus

kubelet

cni-sysctl-allowlist-ds-x2bwv

Unhealthy

Readiness probe errored: rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ServiceCreated

Created Service/prometheus-k8s-thanos-sidecar -n openshift-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

RoleBindingCreated

Created RoleBinding.rbac.authorization.k8s.io/prometheus-k8s -n default because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

RoleBindingCreated

Created RoleBinding.rbac.authorization.k8s.io/prometheus-k8s -n kube-system because it was missing

openshift-monitoring

multus

thanos-querier-5dc898db58-5m6qg

AddedInterface

Add eth0 [10.128.0.94/23] from ovn-kubernetes

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

RoleBindingCreated

Created RoleBinding.rbac.authorization.k8s.io/prometheus-k8s -n default because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

RoleBindingCreated

Created RoleBinding.rbac.authorization.k8s.io/prometheus-k8s -n openshift-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

RoleBindingCreated

Created RoleBinding.rbac.authorization.k8s.io/prometheus-k8s -n openshift-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

RoleBindingCreated

Created RoleBinding.rbac.authorization.k8s.io/prometheus-k8s-config -n openshift-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ServiceCreated

Created Service/prometheus-k8s -n openshift-monitoring because it was missing

openshift-monitoring

kubelet

thanos-querier-5dc898db58-5m6qg

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf72297fee61ec9950f6868881ad3e84be8692ca08f084b3d155d93a766c0823"

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

RoleBindingCreated

Created RoleBinding.rbac.authorization.k8s.io/prometheus-k8s -n kube-system because it was missing
(x3)

openshift-multus

kubelet

cni-sysctl-allowlist-ds-x2bwv

Unhealthy

Readiness probe errored: rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1

openshift-monitoring

kubelet

thanos-querier-5dc898db58-5m6qg

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf72297fee61ec9950f6868881ad3e84be8692ca08f084b3d155d93a766c0823"

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ServiceCreated

Created Service/prometheus-k8s -n openshift-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

RoleBindingCreated

Created RoleBinding.rbac.authorization.k8s.io/prometheus-k8s-config -n openshift-monitoring because it was missing

openshift-monitoring

multus

thanos-querier-5dc898db58-5m6qg

AddedInterface

Add eth0 [10.128.0.94/23] from ovn-kubernetes

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

RoleBindingCreated

Created RoleBinding.rbac.authorization.k8s.io/prometheus-k8s -n openshift-user-workload-monitoring because it was missing

openshift-monitoring

replicaset-controller

metrics-server-646d59cb8b

SuccessfulCreate

Created pod: metrics-server-646d59cb8b-622bl

openshift-monitoring

deployment-controller

metrics-server

ScalingReplicaSet

Scaled down replica set metrics-server-64c67d44c4 to 0 from 1

openshift-monitoring

deployment-controller

metrics-server

ScalingReplicaSet

Scaled up replica set metrics-server-646d59cb8b to 1

openshift-monitoring

kubelet

metrics-server-64c67d44c4-s7vfs

Killing

Stopping container metrics-server

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

SecretCreated

Created Secret/metrics-server-15j6gm4rb8461 -n openshift-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

SecretCreated

Created Secret/metrics-server-15j6gm4rb8461 -n openshift-monitoring because it was missing

openshift-monitoring

replicaset-controller

metrics-server-646d59cb8b

SuccessfulCreate

Created pod: metrics-server-646d59cb8b-622bl

openshift-monitoring

deployment-controller

metrics-server

ScalingReplicaSet

Scaled up replica set metrics-server-646d59cb8b to 1

openshift-monitoring

replicaset-controller

metrics-server-64c67d44c4

SuccessfulDelete

Deleted pod: metrics-server-64c67d44c4-s7vfs

openshift-monitoring

kubelet

metrics-server-64c67d44c4-s7vfs

Killing

Stopping container metrics-server

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Progressing changed from True to False ("AuthenticatorCertKeyProgressing: All is well"),Available changed from False to True ("All is well")

openshift-monitoring

deployment-controller

metrics-server

ScalingReplicaSet

Scaled down replica set metrics-server-64c67d44c4 to 0 from 1

openshift-monitoring

replicaset-controller

metrics-server-64c67d44c4

SuccessfulDelete

Deleted pod: metrics-server-64c67d44c4-s7vfs

openshift-monitoring

kubelet

thanos-querier-5dc898db58-5m6qg

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:28d99dd1c426021eefd6bdbd01594126623f3473f517f194d39e2a063535147a"

openshift-monitoring

kubelet

thanos-querier-5dc898db58-5m6qg

Created

Created container: kube-rbac-proxy

openshift-monitoring

kubelet

metrics-server-646d59cb8b-622bl

Created

Created container: metrics-server

openshift-monitoring

kubelet

metrics-server-646d59cb8b-622bl

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:39c122c726d1bf520dd481350fee5ad940762d5d4c9f8c012db6bf56b0ca8757" already present on machine

openshift-monitoring

kubelet

thanos-querier-5dc898db58-5m6qg

Started

Started container kube-rbac-proxy

openshift-monitoring

kubelet

thanos-querier-5dc898db58-5m6qg

Created

Created container: kube-rbac-proxy

openshift-monitoring

kubelet

thanos-querier-5dc898db58-5m6qg

Started

Started container thanos-query

openshift-monitoring

kubelet

thanos-querier-5dc898db58-5m6qg

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

SecretCreated

Created Secret/prometheus-k8s-additional-alertmanager-configs -n openshift-monitoring because it was missing

openshift-monitoring

kubelet

thanos-querier-5dc898db58-5m6qg

Started

Started container kube-rbac-proxy-web

openshift-monitoring

kubelet

thanos-querier-5dc898db58-5m6qg

Created

Created container: kube-rbac-proxy-web

openshift-monitoring

multus

metrics-server-646d59cb8b-622bl

AddedInterface

Add eth0 [10.128.0.95/23] from ovn-kubernetes

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

SecretCreated

Created Secret/prometheus-k8s-grpc-tls-fofds5kbvc1lq -n openshift-monitoring because it was missing

openshift-monitoring

kubelet

thanos-querier-5dc898db58-5m6qg

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-monitoring

kubelet

thanos-querier-5dc898db58-5m6qg

Started

Started container thanos-query

openshift-monitoring

kubelet

thanos-querier-5dc898db58-5m6qg

Created

Created container: thanos-query

openshift-monitoring

kubelet

metrics-server-646d59cb8b-622bl

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:39c122c726d1bf520dd481350fee5ad940762d5d4c9f8c012db6bf56b0ca8757" already present on machine

openshift-monitoring

kubelet

thanos-querier-5dc898db58-5m6qg

Created

Created container: thanos-query

openshift-monitoring

kubelet

thanos-querier-5dc898db58-5m6qg

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf72297fee61ec9950f6868881ad3e84be8692ca08f084b3d155d93a766c0823" in 2.834s (2.834s including waiting). Image size: 502712961 bytes.

openshift-monitoring

kubelet

metrics-server-646d59cb8b-622bl

Created

Created container: metrics-server

openshift-monitoring

kubelet

thanos-querier-5dc898db58-5m6qg

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf72297fee61ec9950f6868881ad3e84be8692ca08f084b3d155d93a766c0823" in 2.834s (2.834s including waiting). Image size: 502712961 bytes.

openshift-monitoring

kubelet

thanos-querier-5dc898db58-5m6qg

Created

Created container: kube-rbac-proxy-web

openshift-monitoring

kubelet

thanos-querier-5dc898db58-5m6qg

Started

Started container kube-rbac-proxy-web

openshift-monitoring

multus

metrics-server-646d59cb8b-622bl

AddedInterface

Add eth0 [10.128.0.95/23] from ovn-kubernetes

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

SecretCreated

Created Secret/prometheus-k8s-additional-alertmanager-configs -n openshift-monitoring because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ConfigMapCreated

Created ConfigMap/prometheus-trusted-ca-bundle -n openshift-monitoring because it was missing

openshift-monitoring

kubelet

thanos-querier-5dc898db58-5m6qg

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

SecretCreated

Created Secret/prometheus-k8s-grpc-tls-fofds5kbvc1lq -n openshift-monitoring because it was missing

openshift-monitoring

kubelet

thanos-querier-5dc898db58-5m6qg

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:28d99dd1c426021eefd6bdbd01594126623f3473f517f194d39e2a063535147a"

openshift-monitoring

kubelet

thanos-querier-5dc898db58-5m6qg

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-monitoring

kubelet

metrics-server-646d59cb8b-622bl

Started

Started container metrics-server

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ConfigMapCreated

Created ConfigMap/prometheus-trusted-ca-bundle -n openshift-monitoring because it was missing

openshift-monitoring

kubelet

metrics-server-646d59cb8b-622bl

Started

Started container metrics-server

openshift-monitoring

kubelet

thanos-querier-5dc898db58-5m6qg

Started

Started container kube-rbac-proxy

openshift-monitoring

kubelet

thanos-querier-5dc898db58-5m6qg

Created

Created container: kube-rbac-proxy-metrics

openshift-monitoring

kubelet

thanos-querier-5dc898db58-5m6qg

Started

Started container prom-label-proxy

openshift-monitoring

kubelet

thanos-querier-5dc898db58-5m6qg

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-monitoring

kubelet

thanos-querier-5dc898db58-5m6qg

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-monitoring

kubelet

thanos-querier-5dc898db58-5m6qg

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-monitoring

kubelet

thanos-querier-5dc898db58-5m6qg

Created

Created container: prom-label-proxy

openshift-monitoring

kubelet

thanos-querier-5dc898db58-5m6qg

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:28d99dd1c426021eefd6bdbd01594126623f3473f517f194d39e2a063535147a" in 1.123s (1.123s including waiting). Image size: 413104068 bytes.

openshift-monitoring

kubelet

thanos-querier-5dc898db58-5m6qg

Started

Started container kube-rbac-proxy-rules

openshift-monitoring

kubelet

thanos-querier-5dc898db58-5m6qg

Started

Started container prom-label-proxy

openshift-monitoring

kubelet

thanos-querier-5dc898db58-5m6qg

Created

Created container: kube-rbac-proxy-metrics

openshift-monitoring

kubelet

thanos-querier-5dc898db58-5m6qg

Created

Created container: prom-label-proxy

openshift-monitoring

kubelet

thanos-querier-5dc898db58-5m6qg

Started

Started container kube-rbac-proxy-metrics

openshift-monitoring

kubelet

thanos-querier-5dc898db58-5m6qg

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-monitoring

kubelet

thanos-querier-5dc898db58-5m6qg

Created

Created container: kube-rbac-proxy-rules

openshift-monitoring

kubelet

thanos-querier-5dc898db58-5m6qg

Created

Created container: kube-rbac-proxy-rules

openshift-monitoring

kubelet

thanos-querier-5dc898db58-5m6qg

Started

Started container kube-rbac-proxy-rules

openshift-monitoring

kubelet

thanos-querier-5dc898db58-5m6qg

Started

Started container kube-rbac-proxy-metrics

openshift-monitoring

kubelet

thanos-querier-5dc898db58-5m6qg

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:28d99dd1c426021eefd6bdbd01594126623f3473f517f194d39e2a063535147a" in 1.123s (1.123s including waiting). Image size: 413104068 bytes.

openshift-machine-config-operator

machine-config-operator

machine-config-operator

ConfigMapUpdated

Updated ConfigMap/kube-rbac-proxy -n openshift-machine-config-operator: cause by changes in data.config-file.yaml

openshift-monitoring

statefulset-controller

alertmanager-main

SuccessfulCreate

create Pod alertmanager-main-0 in StatefulSet alertmanager-main successful

openshift-monitoring

statefulset-controller

alertmanager-main

SuccessfulCreate

create Pod alertmanager-main-0 in StatefulSet alertmanager-main successful
(x4)

openshift-monitoring

kubelet

alertmanager-main-0

FailedMount

MountVolume.SetUp failed for volume "secret-alertmanager-main-tls" : secret "alertmanager-main-tls" not found
(x4)

openshift-monitoring

kubelet

alertmanager-main-0

FailedMount

MountVolume.SetUp failed for volume "secret-alertmanager-main-tls" : secret "alertmanager-main-tls" not found
(x6)

openshift-network-console

kubelet

networking-console-plugin-7c6b76c555-zbpk4

FailedMount

MountVolume.SetUp failed for volume "networking-console-plugin-cert" : secret "networking-console-plugin-cert" not found

openshift-monitoring

statefulset-controller

prometheus-k8s

SuccessfulCreate

create Pod prometheus-k8s-0 in StatefulSet prometheus-k8s successful

openshift-monitoring

statefulset-controller

prometheus-k8s

SuccessfulCreate

create Pod prometheus-k8s-0 in StatefulSet prometheus-k8s successful
(x2)

openshift-monitoring

kubelet

prometheus-k8s-0

FailedMount

MountVolume.SetUp failed for volume "secret-prometheus-k8s-thanos-sidecar-tls" : secret "prometheus-k8s-thanos-sidecar-tls" not found
(x2)

openshift-monitoring

kubelet

prometheus-k8s-0

FailedMount

MountVolume.SetUp failed for volume "secret-prometheus-k8s-tls" : secret "prometheus-k8s-tls" not found
(x2)

openshift-monitoring

kubelet

prometheus-k8s-0

FailedMount

MountVolume.SetUp failed for volume "secret-prometheus-k8s-tls" : secret "prometheus-k8s-tls" not found
(x2)

openshift-monitoring

kubelet

prometheus-k8s-0

FailedMount

MountVolume.SetUp failed for volume "secret-prometheus-k8s-thanos-sidecar-tls" : secret "prometheus-k8s-thanos-sidecar-tls" not found

openshift-monitoring

kubelet

alertmanager-main-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ea36739b7e81007aec2e901639b356b275434362b254800d4309dd0aa665ca36" already present on machine

openshift-monitoring

multus

alertmanager-main-0

AddedInterface

Add eth0 [10.128.0.96/23] from ovn-kubernetes

openshift-monitoring

kubelet

alertmanager-main-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ea36739b7e81007aec2e901639b356b275434362b254800d4309dd0aa665ca36" already present on machine

openshift-monitoring

multus

alertmanager-main-0

AddedInterface

Add eth0 [10.128.0.96/23] from ovn-kubernetes

openshift-monitoring

kubelet

prometheus-k8s-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ea36739b7e81007aec2e901639b356b275434362b254800d4309dd0aa665ca36" already present on machine

openshift-monitoring

kubelet

alertmanager-main-0

Created

Created container: init-config-reloader

openshift-monitoring

kubelet

alertmanager-main-0

Started

Started container init-config-reloader

openshift-monitoring

kubelet

alertmanager-main-0

Created

Created container: init-config-reloader

openshift-monitoring

kubelet

prometheus-k8s-0

Started

Started container init-config-reloader

openshift-monitoring

kubelet

prometheus-k8s-0

Created

Created container: init-config-reloader

openshift-monitoring

multus

prometheus-k8s-0

AddedInterface

Add eth0 [10.128.0.97/23] from ovn-kubernetes

openshift-monitoring

kubelet

alertmanager-main-0

Started

Started container init-config-reloader

openshift-monitoring

kubelet

prometheus-k8s-0

Started

Started container init-config-reloader

openshift-monitoring

kubelet

prometheus-k8s-0

Created

Created container: init-config-reloader

openshift-monitoring

kubelet

prometheus-k8s-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ea36739b7e81007aec2e901639b356b275434362b254800d4309dd0aa665ca36" already present on machine

openshift-monitoring

multus

prometheus-k8s-0

AddedInterface

Add eth0 [10.128.0.97/23] from ovn-kubernetes

openshift-etcd

kubelet

etcd-master-0

Killing

Stopping container etcdctl

openshift-monitoring

kubelet

alertmanager-main-0

Created

Created container: config-reloader

openshift-monitoring

kubelet

alertmanager-main-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ea36739b7e81007aec2e901639b356b275434362b254800d4309dd0aa665ca36" already present on machine

openshift-monitoring

kubelet

alertmanager-main-0

Started

Started container alertmanager

openshift-monitoring

kubelet

alertmanager-main-0

Started

Started container config-reloader

openshift-monitoring

kubelet

alertmanager-main-0

Created

Created container: alertmanager

openshift-monitoring

kubelet

prometheus-k8s-0

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f3038df8df25746bb5095296d4e5740f2356f85c1ed8d43f1b3d281e294826e5" in 3.58s (3.58s including waiting). Image size: 605698193 bytes.

openshift-monitoring

kubelet

prometheus-k8s-0

Started

Started container config-reloader

openshift-monitoring

kubelet

alertmanager-main-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-monitoring

kubelet

prometheus-k8s-0

Created

Created container: prometheus

openshift-monitoring

kubelet

prometheus-k8s-0

Started

Started container prometheus

openshift-monitoring

kubelet

prometheus-k8s-0

Started

Started container config-reloader

openshift-monitoring

kubelet

prometheus-k8s-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ea36739b7e81007aec2e901639b356b275434362b254800d4309dd0aa665ca36" already present on machine

openshift-monitoring

kubelet

prometheus-k8s-0

Created

Created container: config-reloader

openshift-monitoring

kubelet

prometheus-k8s-0

Created

Created container: config-reloader

openshift-monitoring

kubelet

alertmanager-main-0

Created

Created container: alertmanager

openshift-monitoring

kubelet

alertmanager-main-0

Started

Started container alertmanager

openshift-monitoring

kubelet

alertmanager-main-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ea36739b7e81007aec2e901639b356b275434362b254800d4309dd0aa665ca36" already present on machine

openshift-monitoring

kubelet

alertmanager-main-0

Created

Created container: config-reloader

openshift-monitoring

kubelet

alertmanager-main-0

Started

Started container config-reloader

openshift-monitoring

kubelet

alertmanager-main-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-monitoring

kubelet

prometheus-k8s-0

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f3038df8df25746bb5095296d4e5740f2356f85c1ed8d43f1b3d281e294826e5" in 3.58s (3.58s including waiting). Image size: 605698193 bytes.

openshift-monitoring

kubelet

prometheus-k8s-0

Created

Created container: prometheus

openshift-monitoring

kubelet

prometheus-k8s-0

Started

Started container prometheus

openshift-monitoring

kubelet

prometheus-k8s-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ea36739b7e81007aec2e901639b356b275434362b254800d4309dd0aa665ca36" already present on machine

openshift-monitoring

kubelet

prometheus-k8s-0

Created

Created container: kube-rbac-proxy-web

openshift-monitoring

kubelet

alertmanager-main-0

Created

Created container: kube-rbac-proxy-web

openshift-monitoring

kubelet

prometheus-k8s-0

Created

Created container: kube-rbac-proxy-thanos

openshift-monitoring

kubelet

prometheus-k8s-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-monitoring

kubelet

prometheus-k8s-0

Started

Started container kube-rbac-proxy

openshift-monitoring

kubelet

prometheus-k8s-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf72297fee61ec9950f6868881ad3e84be8692ca08f084b3d155d93a766c0823" already present on machine

openshift-monitoring

kubelet

prometheus-k8s-0

Created

Created container: thanos-sidecar

openshift-monitoring

kubelet

prometheus-k8s-0

Started

Started container thanos-sidecar

openshift-monitoring

kubelet

prometheus-k8s-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-monitoring

kubelet

prometheus-k8s-0

Created

Created container: kube-rbac-proxy-web

openshift-monitoring

kubelet

prometheus-k8s-0

Started

Started container kube-rbac-proxy-web

openshift-monitoring

kubelet

prometheus-k8s-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-monitoring

kubelet

prometheus-k8s-0

Created

Created container: kube-rbac-proxy

openshift-monitoring

kubelet

prometheus-k8s-0

Started

Started container kube-rbac-proxy

openshift-monitoring

kubelet

alertmanager-main-0

Created

Created container: prom-label-proxy

openshift-monitoring

kubelet

alertmanager-main-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:28d99dd1c426021eefd6bdbd01594126623f3473f517f194d39e2a063535147a" already present on machine

openshift-monitoring

kubelet

alertmanager-main-0

Started

Started container kube-rbac-proxy-metric

openshift-monitoring

kubelet

alertmanager-main-0

Created

Created container: kube-rbac-proxy-metric

openshift-monitoring

kubelet

alertmanager-main-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-monitoring

kubelet

alertmanager-main-0

Started

Started container kube-rbac-proxy

openshift-monitoring

kubelet

alertmanager-main-0

Created

Created container: kube-rbac-proxy

openshift-monitoring

kubelet

alertmanager-main-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-monitoring

kubelet

alertmanager-main-0

Started

Started container kube-rbac-proxy-web

openshift-monitoring

kubelet

prometheus-k8s-0

Created

Created container: kube-rbac-proxy

openshift-monitoring

kubelet

prometheus-k8s-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-monitoring

kubelet

prometheus-k8s-0

Created

Created container: kube-rbac-proxy-thanos

openshift-monitoring

kubelet

alertmanager-main-0

Created

Created container: kube-rbac-proxy-web

openshift-monitoring

kubelet

alertmanager-main-0

Started

Started container kube-rbac-proxy-web

openshift-monitoring

kubelet

alertmanager-main-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-monitoring

kubelet

alertmanager-main-0

Created

Created container: kube-rbac-proxy

openshift-monitoring

kubelet

alertmanager-main-0

Started

Started container kube-rbac-proxy

openshift-monitoring

kubelet

alertmanager-main-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-monitoring

kubelet

alertmanager-main-0

Created

Created container: kube-rbac-proxy-metric

openshift-monitoring

kubelet

alertmanager-main-0

Started

Started container kube-rbac-proxy-metric

openshift-monitoring

kubelet

alertmanager-main-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:28d99dd1c426021eefd6bdbd01594126623f3473f517f194d39e2a063535147a" already present on machine

openshift-monitoring

kubelet

alertmanager-main-0

Created

Created container: prom-label-proxy

openshift-monitoring

kubelet

prometheus-k8s-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf72297fee61ec9950f6868881ad3e84be8692ca08f084b3d155d93a766c0823" already present on machine

openshift-monitoring

kubelet

prometheus-k8s-0

Created

Created container: thanos-sidecar

openshift-monitoring

kubelet

prometheus-k8s-0

Started

Started container thanos-sidecar

openshift-monitoring

kubelet

prometheus-k8s-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-monitoring

kubelet

prometheus-k8s-0

Started

Started container kube-rbac-proxy-web

openshift-monitoring

kubelet

prometheus-k8s-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine
(x4)

openshift-kube-controller-manager

kubelet

kube-controller-manager-master-0

BackOff

Back-off restarting failed container kube-controller-manager in pod kube-controller-manager-master-0_openshift-kube-controller-manager(2028761b8522f874dcebf13c4683d033)
(x2)

openshift-kube-controller-manager

kubelet

kube-controller-manager-master-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b23c544d3894e5b31f66a18c554f03b0d29f92c2000c46b57b1c96da7ec25db9" already present on machine
(x2)

openshift-kube-controller-manager

kubelet

kube-controller-manager-master-0

Created

Created container: kube-controller-manager
(x2)

openshift-kube-controller-manager

kubelet

kube-controller-manager-master-0

Started

Started container kube-controller-manager

openshift-network-node-identity

kubelet

network-node-identity-6t5vb

Started

Started container approver

openshift-network-node-identity

kubelet

network-node-identity-6t5vb

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:2abc1fd79e7781634ed5ed9e8f2b98b9094ea51f40ac3a773c5e5224607bf3d7" already present on machine

openshift-network-node-identity

kubelet

network-node-identity-6t5vb

Created

Created container: approver

openshift-etcd

kubelet

etcd-master-0

Started

Started container setup

openshift-etcd

kubelet

etcd-master-0

Created

Created container: setup

openshift-etcd

kubelet

etcd-master-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e29dc9f042f2d0471171a0611070886cb2f7c57338ab7f112613417bcd33b278" already present on machine

openshift-network-console

kubelet

networking-console-plugin-7c6b76c555-zbpk4

FailedCreatePodSandBox

Failed to create pod sandbox: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_networking-console-plugin-7c6b76c555-zbpk4_openshift-network-console_044870dd-540a-402e-84cb-fa1bf3d6a318_0(e31283cad84b23fb487e1c8a841943f36f80924391a43b0ed4cdfedfd609699e): error adding pod openshift-network-console_networking-console-plugin-7c6b76c555-zbpk4 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"e31283cad84b23fb487e1c8a841943f36f80924391a43b0ed4cdfedfd609699e" Netns:"/var/run/netns/a27c34f3-2026-4282-ad33-a758b28fd8b5" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-network-console;K8S_POD_NAME=networking-console-plugin-7c6b76c555-zbpk4;K8S_POD_INFRA_CONTAINER_ID=e31283cad84b23fb487e1c8a841943f36f80924391a43b0ed4cdfedfd609699e;K8S_POD_UID=044870dd-540a-402e-84cb-fa1bf3d6a318" Path:"" ERRORED: error configuring pod [openshift-network-console/networking-console-plugin-7c6b76c555-zbpk4] networking: Multus: [openshift-network-console/networking-console-plugin-7c6b76c555-zbpk4/044870dd-540a-402e-84cb-fa1bf3d6a318]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod networking-console-plugin-7c6b76c555-zbpk4 in out of cluster comm: SetNetworkStatus: failed to update the pod networking-console-plugin-7c6b76c555-zbpk4 in out of cluster comm: status update failed for pod /: Get "https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-network-console/pods/networking-console-plugin-7c6b76c555-zbpk4?timeout=1m0s": net/http: request canceled (Client.Timeout exceeded while awaiting headers) ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"}

openshift-etcd

kubelet

etcd-master-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e29dc9f042f2d0471171a0611070886cb2f7c57338ab7f112613417bcd33b278" already present on machine

openshift-etcd

kubelet

etcd-master-0

Created

Created container: etcd-ensure-env-vars

openshift-etcd

kubelet

etcd-master-0

Started

Started container etcd-ensure-env-vars

openshift-console-operator

kubelet

console-operator-76b6568d85-8b8gv

FailedCreatePodSandBox

Failed to create pod sandbox: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_console-operator-76b6568d85-8b8gv_openshift-console-operator_348f3880-793f-43e4-9de1-8511626d2552_0(ca8576e8d30903a0ab02eabeb68a25c4d1075706f6eae8f9d699f9069d1ab96b): error adding pod openshift-console-operator_console-operator-76b6568d85-8b8gv to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"ca8576e8d30903a0ab02eabeb68a25c4d1075706f6eae8f9d699f9069d1ab96b" Netns:"/var/run/netns/9c7752cf-7a04-422a-882d-f3da023fd279" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-console-operator;K8S_POD_NAME=console-operator-76b6568d85-8b8gv;K8S_POD_INFRA_CONTAINER_ID=ca8576e8d30903a0ab02eabeb68a25c4d1075706f6eae8f9d699f9069d1ab96b;K8S_POD_UID=348f3880-793f-43e4-9de1-8511626d2552" Path:"" ERRORED: error configuring pod [openshift-console-operator/console-operator-76b6568d85-8b8gv] networking: Multus: [openshift-console-operator/console-operator-76b6568d85-8b8gv/348f3880-793f-43e4-9de1-8511626d2552]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod console-operator-76b6568d85-8b8gv in out of cluster comm: SetNetworkStatus: failed to update the pod console-operator-76b6568d85-8b8gv in out of cluster comm: status update failed for pod /: Get "https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-console-operator/pods/console-operator-76b6568d85-8b8gv?timeout=1m0s": net/http: request canceled (Client.Timeout exceeded while awaiting headers) ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"}

openshift-cloud-controller-manager-operator

kubelet

cluster-cloud-controller-manager-operator-7dff898856-7vxxr

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:112a03f2411f871cdaca5f20daef71024dac710113d5f30897117a5a02f6b6f5" already present on machine

openshift-cloud-controller-manager-operator

kubelet

cluster-cloud-controller-manager-operator-7dff898856-7vxxr

Started

Started container config-sync-controllers

openshift-cloud-controller-manager-operator

kubelet

cluster-cloud-controller-manager-operator-7dff898856-7vxxr

Created

Created container: config-sync-controllers

openshift-catalogd

kubelet

catalogd-controller-manager-6864dc98f7-74mgr

ProbeError

Liveness probe error: Get "http://10.128.0.39:8081/healthz": dial tcp 10.128.0.39:8081: connect: connection refused body:

openshift-catalogd

kubelet

catalogd-controller-manager-6864dc98f7-74mgr

ProbeError

Liveness probe error: Get "http://10.128.0.39:8081/healthz": dial tcp 10.128.0.39:8081: connect: connection refused body:

openshift-catalogd

kubelet

catalogd-controller-manager-6864dc98f7-74mgr

Unhealthy

Liveness probe failed: Get "http://10.128.0.39:8081/healthz": dial tcp 10.128.0.39:8081: connect: connection refused

openshift-catalogd

kubelet

catalogd-controller-manager-6864dc98f7-74mgr

Unhealthy

Liveness probe failed: Get "http://10.128.0.39:8081/healthz": dial tcp 10.128.0.39:8081: connect: connection refused

openshift-catalogd

kubelet

catalogd-controller-manager-6864dc98f7-74mgr

Started

Started container manager

openshift-operator-controller

kubelet

operator-controller-controller-manager-57777556ff-nk2rf

Started

Started container manager

openshift-operator-controller

kubelet

operator-controller-controller-manager-57777556ff-nk2rf

Created

Created container: manager

openshift-operator-controller

kubelet

operator-controller-controller-manager-57777556ff-nk2rf

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:4b5ea1ef4e09b673a0c68c8848ca162ab11d9ac373a377daa52dea702ffa3023" already present on machine

openshift-catalogd

kubelet

catalogd-controller-manager-6864dc98f7-74mgr

Started

Started container manager

openshift-catalogd

kubelet

catalogd-controller-manager-6864dc98f7-74mgr

Created

Created container: manager

openshift-catalogd

kubelet

catalogd-controller-manager-6864dc98f7-74mgr

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3062f6485aec4770e60852b535c69a42527b305161fe856499c8658ead6d1e85" already present on machine

openshift-catalogd

kubelet

catalogd-controller-manager-6864dc98f7-74mgr

Created

Created container: manager

openshift-catalogd

kubelet

catalogd-controller-manager-6864dc98f7-74mgr

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3062f6485aec4770e60852b535c69a42527b305161fe856499c8658ead6d1e85" already present on machine

openshift-cloud-controller-manager-operator

kubelet

cluster-cloud-controller-manager-operator-7dff898856-7vxxr

Created

Created container: cluster-cloud-controller-manager

openshift-cloud-controller-manager-operator

kubelet

cluster-cloud-controller-manager-operator-7dff898856-7vxxr

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:112a03f2411f871cdaca5f20daef71024dac710113d5f30897117a5a02f6b6f5" already present on machine

openshift-cloud-controller-manager-operator

kubelet

cluster-cloud-controller-manager-operator-7dff898856-7vxxr

Started

Started container cluster-cloud-controller-manager

openshift-network-console

kubelet

networking-console-plugin-7c6b76c555-zbpk4

FailedCreatePodSandBox

Failed to create pod sandbox: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_networking-console-plugin-7c6b76c555-zbpk4_openshift-network-console_044870dd-540a-402e-84cb-fa1bf3d6a318_0(7e454bbf635ef1bf6f2b0e78e85135dc4c266557d742057776c7e953ad66d59e): error adding pod openshift-network-console_networking-console-plugin-7c6b76c555-zbpk4 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"7e454bbf635ef1bf6f2b0e78e85135dc4c266557d742057776c7e953ad66d59e" Netns:"/var/run/netns/b764c1f7-e862-40a1-b6b9-db5be997e963" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-network-console;K8S_POD_NAME=networking-console-plugin-7c6b76c555-zbpk4;K8S_POD_INFRA_CONTAINER_ID=7e454bbf635ef1bf6f2b0e78e85135dc4c266557d742057776c7e953ad66d59e;K8S_POD_UID=044870dd-540a-402e-84cb-fa1bf3d6a318" Path:"" ERRORED: error configuring pod [openshift-network-console/networking-console-plugin-7c6b76c555-zbpk4] networking: Multus: [openshift-network-console/networking-console-plugin-7c6b76c555-zbpk4/044870dd-540a-402e-84cb-fa1bf3d6a318]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod networking-console-plugin-7c6b76c555-zbpk4 in out of cluster comm: SetNetworkStatus: failed to update the pod networking-console-plugin-7c6b76c555-zbpk4 in out of cluster comm: status update failed for pod /: Get "https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-network-console/pods/networking-console-plugin-7c6b76c555-zbpk4?timeout=1m0s": net/http: request canceled (Client.Timeout exceeded while awaiting headers) ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"}

openshift-etcd

kubelet

etcd-master-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e29dc9f042f2d0471171a0611070886cb2f7c57338ab7f112613417bcd33b278" already present on machine

openshift-etcd

kubelet

etcd-master-0

Created

Created container: etcd-resources-copy

openshift-etcd

kubelet

etcd-master-0

Started

Started container etcd-resources-copy

openshift-ovn-kubernetes

kubelet

ovnkube-control-plane-57f769d897-z2zpj

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:2abc1fd79e7781634ed5ed9e8f2b98b9094ea51f40ac3a773c5e5224607bf3d7" already present on machine

openshift-ovn-kubernetes

kubelet

ovnkube-control-plane-57f769d897-z2zpj

Created

Created container: ovnkube-cluster-manager

openshift-ovn-kubernetes

kubelet

ovnkube-control-plane-57f769d897-z2zpj

Started

Started container ovnkube-cluster-manager
(x2)

openshift-controller-manager

kubelet

controller-manager-bfcf4df58-l6xz7

Created

Created container: controller-manager
(x2)

openshift-controller-manager

kubelet

controller-manager-bfcf4df58-l6xz7

Started

Started container controller-manager
(x2)

openshift-controller-manager

kubelet

controller-manager-bfcf4df58-l6xz7

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c2dd7a03348212e49876f5359f233d893a541ed9b934df390201a05133a06982" already present on machine

openshift-cluster-machine-approver

kubelet

machine-approver-5c6485487f-qb94j

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:cdd28dfe7132e19af9f013f72cf120d970bc31b6b74693af262f8d2e82a096e1" already present on machine

openshift-cluster-machine-approver

kubelet

machine-approver-5c6485487f-qb94j

Started

Started container machine-approver-controller

openshift-cluster-machine-approver

kubelet

machine-approver-5c6485487f-qb94j

Created

Created container: machine-approver-controller

openshift-kube-scheduler

kubelet

openshift-kube-scheduler-master-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b23c544d3894e5b31f66a18c554f03b0d29f92c2000c46b57b1c96da7ec25db9" already present on machine

openshift-kube-scheduler

kubelet

openshift-kube-scheduler-master-0

Created

Created container: kube-scheduler

openshift-kube-scheduler

kubelet

openshift-kube-scheduler-master-0

Started

Started container kube-scheduler

openshift-kube-scheduler

kubelet

openshift-kube-scheduler-master-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ef199844317b7b012879ed8d29f9b6bc37fad8a6fdb336103cbd5cabc74c4302" already present on machine

openshift-kube-scheduler

kubelet

openshift-kube-scheduler-master-0

Started

Started container kube-scheduler-recovery-controller

openshift-kube-scheduler

kubelet

openshift-kube-scheduler-master-0

Created

Created container: kube-scheduler-recovery-controller

openshift-machine-api

kubelet

control-plane-machine-set-operator-6f97756bc8-7t5qv

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:908eaaf624959bc7645f6d585d160431d1efb070e9a1f37fefed73a3be42b0d3" already present on machine

openshift-machine-api

kubelet

control-plane-machine-set-operator-6f97756bc8-7t5qv

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:908eaaf624959bc7645f6d585d160431d1efb070e9a1f37fefed73a3be42b0d3" already present on machine

openshift-machine-api

kubelet

control-plane-machine-set-operator-6f97756bc8-7t5qv

Created

Created container: control-plane-machine-set-operator

openshift-machine-api

kubelet

control-plane-machine-set-operator-6f97756bc8-7t5qv

Created

Created container: control-plane-machine-set-operator

openshift-machine-api

kubelet

control-plane-machine-set-operator-6f97756bc8-7t5qv

Started

Started container control-plane-machine-set-operator

openshift-machine-api

kubelet

control-plane-machine-set-operator-6f97756bc8-7t5qv

Started

Started container control-plane-machine-set-operator

openshift-console-operator

kubelet

console-operator-76b6568d85-8b8gv

FailedCreatePodSandBox

Failed to create pod sandbox: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_console-operator-76b6568d85-8b8gv_openshift-console-operator_348f3880-793f-43e4-9de1-8511626d2552_0(a7d9b28db0ac22a37d423a7ed15e1230abf1c849668396dd25061a938e1b3f30): error adding pod openshift-console-operator_console-operator-76b6568d85-8b8gv to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"a7d9b28db0ac22a37d423a7ed15e1230abf1c849668396dd25061a938e1b3f30" Netns:"/var/run/netns/0076a471-cb8c-4a76-9e34-57408445194b" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-console-operator;K8S_POD_NAME=console-operator-76b6568d85-8b8gv;K8S_POD_INFRA_CONTAINER_ID=a7d9b28db0ac22a37d423a7ed15e1230abf1c849668396dd25061a938e1b3f30;K8S_POD_UID=348f3880-793f-43e4-9de1-8511626d2552" Path:"" ERRORED: error configuring pod [openshift-console-operator/console-operator-76b6568d85-8b8gv] networking: Multus: [openshift-console-operator/console-operator-76b6568d85-8b8gv/348f3880-793f-43e4-9de1-8511626d2552]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod console-operator-76b6568d85-8b8gv in out of cluster comm: SetNetworkStatus: failed to update the pod console-operator-76b6568d85-8b8gv in out of cluster comm: status update failed for pod /: Get "https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-console-operator/pods/console-operator-76b6568d85-8b8gv?timeout=1m0s": net/http: request canceled (Client.Timeout exceeded while awaiting headers) ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"}
(x3)

openshift-kube-controller-manager

kubelet

kube-controller-manager-master-0

Unhealthy

Startup probe failed: Get "https://localhost:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)

openshift-kube-controller-manager

kubelet

kube-controller-manager-master-0

Killing

Container cluster-policy-controller failed startup probe, will be restarted
(x2)

openshift-kube-controller-manager

kubelet

kube-controller-manager-master-0

Started

Started container cluster-policy-controller
(x3)

openshift-kube-controller-manager

kubelet

kube-controller-manager-master-0

ProbeError

Startup probe error: Get "https://localhost:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) body:
(x2)

openshift-kube-controller-manager

kubelet

kube-controller-manager-master-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1fbbcb390de2563a0177b92fba1b5a65777366e2dc80e2808b61d87c41b47a2d" already present on machine
(x2)

openshift-kube-controller-manager

kubelet

kube-controller-manager-master-0

Created

Created container: cluster-policy-controller

openshift-network-console

kubelet

networking-console-plugin-7c6b76c555-zbpk4

FailedCreatePodSandBox

Failed to create pod sandbox: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_networking-console-plugin-7c6b76c555-zbpk4_openshift-network-console_044870dd-540a-402e-84cb-fa1bf3d6a318_0(4380c2e62b9a68692ac36325cb36833c81d190dc424789b028acbfd6a05362c5): error adding pod openshift-network-console_networking-console-plugin-7c6b76c555-zbpk4 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"4380c2e62b9a68692ac36325cb36833c81d190dc424789b028acbfd6a05362c5" Netns:"/var/run/netns/65006278-e165-47c3-b835-e4c438dca515" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-network-console;K8S_POD_NAME=networking-console-plugin-7c6b76c555-zbpk4;K8S_POD_INFRA_CONTAINER_ID=4380c2e62b9a68692ac36325cb36833c81d190dc424789b028acbfd6a05362c5;K8S_POD_UID=044870dd-540a-402e-84cb-fa1bf3d6a318" Path:"" ERRORED: error configuring pod [openshift-network-console/networking-console-plugin-7c6b76c555-zbpk4] networking: Multus: [openshift-network-console/networking-console-plugin-7c6b76c555-zbpk4/044870dd-540a-402e-84cb-fa1bf3d6a318]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod networking-console-plugin-7c6b76c555-zbpk4 in out of cluster comm: SetNetworkStatus: failed to update the pod networking-console-plugin-7c6b76c555-zbpk4 in out of cluster comm: status update failed for pod /: Get "https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-network-console/pods/networking-console-plugin-7c6b76c555-zbpk4?timeout=1m0s": net/http: request canceled (Client.Timeout exceeded while awaiting headers) ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"}

openshift-kube-controller-manager

kubelet

kube-controller-manager-master-0

ProbeError

Startup probe error: Get "https://localhost:10357/healthz": context deadline exceeded (Client.Timeout exceeded while awaiting headers) body:

openshift-etcd

kubelet

etcd-master-0

Created

Created container: etcd

openshift-etcd

kubelet

etcd-master-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e29dc9f042f2d0471171a0611070886cb2f7c57338ab7f112613417bcd33b278" already present on machine

openshift-etcd

kubelet

etcd-master-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e29dc9f042f2d0471171a0611070886cb2f7c57338ab7f112613417bcd33b278" already present on machine

openshift-etcd

kubelet

etcd-master-0

Created

Created container: etcdctl

openshift-etcd

kubelet

etcd-master-0

Started

Started container etcdctl

openshift-etcd

kubelet

etcd-master-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e29dc9f042f2d0471171a0611070886cb2f7c57338ab7f112613417bcd33b278" already present on machine

openshift-etcd

kubelet

etcd-master-0

Started

Started container etcd

openshift-etcd

kubelet

etcd-master-0

Created

Created container: etcd-metrics

openshift-etcd

kubelet

etcd-master-0

Started

Started container etcd-metrics

openshift-etcd

kubelet

etcd-master-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d5a971d5889f167cfe61a64c366424b87c17a6dc141ffcc43406cdcbb50cae2a" already present on machine

openshift-kube-scheduler

kubelet

openshift-kube-scheduler-master-0

ProbeError

Readiness probe error: Get "https://192.168.32.10:10259/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) body:

openshift-etcd

kubelet

etcd-master-0

Started

Started container etcd-readyz

openshift-etcd

kubelet

etcd-master-0

Created

Created container: etcd-readyz

openshift-etcd

kubelet

etcd-master-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d5a971d5889f167cfe61a64c366424b87c17a6dc141ffcc43406cdcbb50cae2a" already present on machine

openshift-etcd

kubelet

etcd-master-0

Created

Created container: etcd-rev

openshift-etcd

kubelet

etcd-master-0

Started

Started container etcd-rev

openshift-kube-scheduler

kubelet

openshift-kube-scheduler-master-0

Unhealthy

Readiness probe failed: Get "https://192.168.32.10:10259/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)
(x2)

openshift-kube-scheduler

kubelet

openshift-kube-scheduler-master-0

ProbeError

Liveness probe error: Get "https://192.168.32.10:10259/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) body:
(x2)

openshift-kube-scheduler

kubelet

openshift-kube-scheduler-master-0

Unhealthy

Liveness probe failed: Get "https://192.168.32.10:10259/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)

openshift-console-operator

kubelet

console-operator-76b6568d85-8b8gv

FailedCreatePodSandBox

Failed to create pod sandbox: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_console-operator-76b6568d85-8b8gv_openshift-console-operator_348f3880-793f-43e4-9de1-8511626d2552_0(af5af9faf4f8cee86539ef74da22e713d76207507caacf3e4c0d09dc217dfa8f): error adding pod openshift-console-operator_console-operator-76b6568d85-8b8gv to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"af5af9faf4f8cee86539ef74da22e713d76207507caacf3e4c0d09dc217dfa8f" Netns:"/var/run/netns/a8e99ea1-5371-4811-b7a7-4c4fc5dc79ce" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-console-operator;K8S_POD_NAME=console-operator-76b6568d85-8b8gv;K8S_POD_INFRA_CONTAINER_ID=af5af9faf4f8cee86539ef74da22e713d76207507caacf3e4c0d09dc217dfa8f;K8S_POD_UID=348f3880-793f-43e4-9de1-8511626d2552" Path:"" ERRORED: error configuring pod [openshift-console-operator/console-operator-76b6568d85-8b8gv] networking: Multus: [openshift-console-operator/console-operator-76b6568d85-8b8gv/348f3880-793f-43e4-9de1-8511626d2552]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod console-operator-76b6568d85-8b8gv in out of cluster comm: SetNetworkStatus: failed to update the pod console-operator-76b6568d85-8b8gv in out of cluster comm: status update failed for pod /: Get "https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-console-operator/pods/console-operator-76b6568d85-8b8gv?timeout=1m0s": net/http: request canceled (Client.Timeout exceeded while awaiting headers) ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"}
(x2)

openshift-etcd-operator

openshift-cluster-etcd-operator-member-observer-controller-etcdmemberscontroller

etcd-operator

ReportEtcdMembersErrorUpdatingStatus

Timeout: request did not complete within requested timeout - context deadline exceeded
(x2)

openshift-machine-config-operator

machineconfigoperator

machine-config

OperatorDegraded: MachineConfigPoolsFailed

Failed to resync 4.18.35 because: the server was unable to return a response in the time allotted, but may still be processing the request (get machineconfigpools.machineconfiguration.openshift.io master)

openshift-network-console

kubelet

networking-console-plugin-7c6b76c555-zbpk4

FailedCreatePodSandBox

Failed to create pod sandbox: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_networking-console-plugin-7c6b76c555-zbpk4_openshift-network-console_044870dd-540a-402e-84cb-fa1bf3d6a318_0(c2d60fffe56ed279059ff614c498eff53b4dd1933e7096df70ed8391764bdfa7): error adding pod openshift-network-console_networking-console-plugin-7c6b76c555-zbpk4 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"c2d60fffe56ed279059ff614c498eff53b4dd1933e7096df70ed8391764bdfa7" Netns:"/var/run/netns/a037ebf2-fffc-4748-9316-3a236cef8f45" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-network-console;K8S_POD_NAME=networking-console-plugin-7c6b76c555-zbpk4;K8S_POD_INFRA_CONTAINER_ID=c2d60fffe56ed279059ff614c498eff53b4dd1933e7096df70ed8391764bdfa7;K8S_POD_UID=044870dd-540a-402e-84cb-fa1bf3d6a318" Path:"" ERRORED: error configuring pod [openshift-network-console/networking-console-plugin-7c6b76c555-zbpk4] networking: Multus: [openshift-network-console/networking-console-plugin-7c6b76c555-zbpk4/044870dd-540a-402e-84cb-fa1bf3d6a318]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod networking-console-plugin-7c6b76c555-zbpk4 in out of cluster comm: SetNetworkStatus: failed to update the pod networking-console-plugin-7c6b76c555-zbpk4 in out of cluster comm: status update failed for pod /: Get "https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-network-console/pods/networking-console-plugin-7c6b76c555-zbpk4?timeout=1m0s": net/http: request canceled (Client.Timeout exceeded while awaiting headers) ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"}
(x5)

openshift-network-console

multus

networking-console-plugin-7c6b76c555-zbpk4

AddedInterface

Add eth0 [10.128.0.93/23] from ovn-kubernetes
(x12)

openshift-kube-apiserver

kubelet

installer-3-master-0

FailedMount

MountVolume.SetUp failed for volume "kube-api-access" : object "openshift-kube-apiserver"/"kube-root-ca.crt" not registered

openshift-cluster-version

openshift-cluster-version

version

LeaderElection

master-0_d00985cf-7208-44bf-bd7d-3c5284c2a8bf stopped leading

openshift-ovn-kubernetes

ovnk-controlplane

ovn-kubernetes-master

LeaderElection

ovnkube-control-plane-57f769d897-z2zpj became leader
(x6)

openshift-cluster-storage-operator

kubelet

csi-snapshot-controller-64854d9cff-f44gr

BackOff

Back-off restarting failed container snapshot-controller in pod csi-snapshot-controller-64854d9cff-f44gr_openshift-cluster-storage-operator(96de6024-e20f-4b52-9294-b330d65e4153)

openshift-controller-manager

openshift-controller-manager

openshift-master-controllers

LeaderElection

controller-manager-bfcf4df58-l6xz7 became leader

openshift-operator-lifecycle-manager

kubelet

package-server-manager-7b95f86987-2pg77

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6bba3f73c0066e42b24839e0d29f5dce2f36436f0a11f9f5e1029bccc5ed6578" already present on machine

openshift-image-registry

kubelet

cluster-image-registry-operator-5549dc66cb-fmhbq

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7af9f5c5af9d529840233ef4b519120cc0e3f14c4fe28cc43b0823f2c11d8f89" already present on machine
(x2)

openshift-route-controller-manager

kubelet

route-controller-manager-ffdd4b479-rhmfx

Started

Started container route-controller-manager
(x2)

openshift-route-controller-manager

kubelet

route-controller-manager-ffdd4b479-rhmfx

Created

Created container: route-controller-manager

openshift-machine-config-operator

kubelet

machine-config-controller-b4f87c5b9-pj7rj

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:af0fe0ca926422a6471d5bf22fc0e682c36c24fba05496a3bdfac0b7d3733015" already present on machine
(x2)

openshift-route-controller-manager

kubelet

route-controller-manager-ffdd4b479-rhmfx

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:446bedea4916d3c1ee52be94137e484659e9561bd1de95c8189eee279aae984b" already present on machine

openshift-cluster-node-tuning-operator

kubelet

cluster-node-tuning-operator-598fbc5f8f-vxzvg

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:aa5e782406f71c048b1ac3a4bf5d1227ff4be81111114083ad4c7a209c6bfb5a" already present on machine

openshift-cluster-version

kubelet

cluster-version-operator-7d58488df-qmm8h

Pulled

Container image "quay.io/openshift-release-dev/ocp-release@sha256:59727c4b3fef19e5149675cf3350735bbfe2c6588a57654b2e4552dd719f58b1" already present on machine

openshift-machine-config-operator

kubelet

machine-config-operator-84d549f6d5-gm4qr

Created

Created container: machine-config-operator

openshift-cluster-storage-operator

kubelet

cluster-storage-operator-7d87854d6-vlq7h

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:30a2f97d7785ce8b0ea5115e67c4554b64adefbc7856bcf6f4fe6cc7e938a310" already present on machine

openshift-network-operator

kubelet

network-operator-7bd846bfc4-mt454

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ec8fd46dfb35ed10e8f98933166f69ce579c2f35b8db03d21e4c34fc544553e4" already present on machine

openshift-controller-manager-operator

kubelet

openshift-controller-manager-operator-8c94f4649-p7pt6

Started

Started container openshift-controller-manager-operator

openshift-controller-manager-operator

kubelet

openshift-controller-manager-operator-8c94f4649-p7pt6

Created

Created container: openshift-controller-manager-operator

openshift-kube-storage-version-migrator-operator

kubelet

kube-storage-version-migrator-operator-6bb5bfb6fd-64dg5

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5c983016b9ceed0fca1f51bd49c2653243c7e5af91cbf2f478b091db6e028252" already present on machine

openshift-controller-manager-operator

kubelet

openshift-controller-manager-operator-8c94f4649-p7pt6

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:313d1d8ca85e65236a59f058a3316c49436dde691b3a3930d5bc5e3b4b8c8a71" already present on machine

openshift-machine-config-operator

kubelet

machine-config-operator-84d549f6d5-gm4qr

Started

Started container machine-config-operator

openshift-authentication-operator

kubelet

authentication-operator-5885bfd7f4-62zrx

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bfe394b58ec6195de8b8420e781b7630d85a412b9112d892fea903f92b783427" already present on machine

openshift-authentication-operator

kubelet

authentication-operator-5885bfd7f4-62zrx

Created

Created container: authentication-operator

openshift-machine-api

kubelet

machine-api-operator-6fbb6cf6f9-n8tnn

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:de91abd5ad76fb491881a75a0feb4b8ca5600ceb5e15a4b0b687ada01ea0a44c" already present on machine

openshift-service-ca-operator

kubelet

service-ca-operator-b865698dc-qzb2h

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:812819a9d712b9e345ef5f1404b242c281e2518ad724baebc393ec0fd3b3d263" already present on machine

openshift-authentication-operator

kubelet

authentication-operator-5885bfd7f4-62zrx

Started

Started container authentication-operator

openshift-kube-scheduler-operator

kubelet

openshift-kube-scheduler-operator-dddff6458-vx5d7

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ef199844317b7b012879ed8d29f9b6bc37fad8a6fdb336103cbd5cabc74c4302" already present on machine

openshift-cluster-storage-operator

kubelet

csi-snapshot-controller-operator-5f5d689c6b-4dqrh

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c527b4e8239a1f4f4e0a851113e7dd633b7dcb9d75b0e7b21c23d26304abcb3" already present on machine
(x2)

openshift-kube-apiserver

kubelet

kube-apiserver-master-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c5ce3d1134d6500e2b8528516c1889d7bbc6259aba4981c6983395b0e9eeff65" already present on machine

openshift-kube-apiserver-operator

kubelet

kube-apiserver-operator-8b68b9d9b-c6vkz

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c5ce3d1134d6500e2b8528516c1889d7bbc6259aba4981c6983395b0e9eeff65" already present on machine

openshift-machine-api

kubelet

machine-api-operator-6fbb6cf6f9-n8tnn

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:de91abd5ad76fb491881a75a0feb4b8ca5600ceb5e15a4b0b687ada01ea0a44c" already present on machine

openshift-machine-api

kubelet

cluster-autoscaler-operator-866dc4744-xwxg7

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ce8e3088493b4a72dd766b3b5b4ccb83b7d72d514bbf64063a913dfe961273db" already present on machine

openshift-cluster-olm-operator

kubelet

cluster-olm-operator-67dcd4998-c5742

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adb9f6f2fd701863c7caed747df43f83d3569ba9388cfa33ea7219ac6a606b11" already present on machine

openshift-cluster-node-tuning-operator

kubelet

cluster-node-tuning-operator-598fbc5f8f-vxzvg

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:aa5e782406f71c048b1ac3a4bf5d1227ff4be81111114083ad4c7a209c6bfb5a" already present on machine

openshift-etcd-operator

kubelet

etcd-operator-8544cbcf9c-brhw4

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d5a971d5889f167cfe61a64c366424b87c17a6dc141ffcc43406cdcbb50cae2a" already present on machine

openshift-machine-config-operator

kubelet

machine-config-operator-84d549f6d5-gm4qr

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:af0fe0ca926422a6471d5bf22fc0e682c36c24fba05496a3bdfac0b7d3733015" already present on machine

openshift-service-ca

kubelet

service-ca-79bc6b8d76-72j8t

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:812819a9d712b9e345ef5f1404b242c281e2518ad724baebc393ec0fd3b3d263" already present on machine

openshift-machine-api

kubelet

cluster-autoscaler-operator-866dc4744-xwxg7

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ce8e3088493b4a72dd766b3b5b4ccb83b7d72d514bbf64063a913dfe961273db" already present on machine

openshift-machine-api

kubelet

cluster-autoscaler-operator-866dc4744-xwxg7

Started

Started container cluster-autoscaler-operator

openshift-image-registry

kubelet

cluster-image-registry-operator-5549dc66cb-fmhbq

Started

Started container cluster-image-registry-operator

openshift-machine-api

kubelet

machine-api-operator-6fbb6cf6f9-n8tnn

Created

Created container: machine-api-operator

openshift-cluster-storage-operator

kubelet

csi-snapshot-controller-operator-5f5d689c6b-4dqrh

Created

Created container: csi-snapshot-controller-operator

openshift-etcd-operator

kubelet

etcd-operator-8544cbcf9c-brhw4

Created

Created container: etcd-operator

openshift-service-ca

kubelet

service-ca-79bc6b8d76-72j8t

Created

Created container: service-ca-controller

openshift-image-registry

kubelet

cluster-image-registry-operator-5549dc66cb-fmhbq

Created

Created container: cluster-image-registry-operator

openshift-cluster-olm-operator

kubelet

cluster-olm-operator-67dcd4998-c5742

Started

Started container cluster-olm-operator

openshift-service-ca

kubelet

service-ca-79bc6b8d76-72j8t

Started

Started container service-ca-controller

openshift-cluster-node-tuning-operator

kubelet

cluster-node-tuning-operator-598fbc5f8f-vxzvg

Started

Started container cluster-node-tuning-operator

openshift-cluster-node-tuning-operator

kubelet

cluster-node-tuning-operator-598fbc5f8f-vxzvg

Created

Created container: cluster-node-tuning-operator

openshift-etcd-operator

kubelet

etcd-operator-8544cbcf9c-brhw4

Started

Started container etcd-operator

openshift-machine-api

kubelet

cluster-autoscaler-operator-866dc4744-xwxg7

Created

Created container: cluster-autoscaler-operator

openshift-kube-apiserver-operator

kubelet

kube-apiserver-operator-8b68b9d9b-c6vkz

Created

Created container: kube-apiserver-operator

openshift-machine-api

kubelet

machine-api-operator-6fbb6cf6f9-n8tnn

Created

Created container: machine-api-operator

openshift-machine-api

kubelet

machine-api-operator-6fbb6cf6f9-n8tnn

Started

Started container machine-api-operator
(x2)

openshift-kube-apiserver

kubelet

kube-apiserver-master-0

Created

Created container: kube-apiserver-cert-regeneration-controller
(x2)

openshift-kube-apiserver

kubelet

kube-apiserver-master-0

Started

Started container kube-apiserver-cert-regeneration-controller

openshift-kube-scheduler-operator

kubelet

openshift-kube-scheduler-operator-dddff6458-vx5d7

Created

Created container: kube-scheduler-operator-container

openshift-machine-api

kubelet

cluster-autoscaler-operator-866dc4744-xwxg7

Created

Created container: cluster-autoscaler-operator

openshift-kube-apiserver-operator

kubelet

kube-apiserver-operator-8b68b9d9b-c6vkz

Started

Started container kube-apiserver-operator

openshift-machine-api

kubelet

cluster-autoscaler-operator-866dc4744-xwxg7

Started

Started container cluster-autoscaler-operator

openshift-cluster-storage-operator

kubelet

csi-snapshot-controller-operator-5f5d689c6b-4dqrh

Started

Started container csi-snapshot-controller-operator

openshift-kube-scheduler-operator

kubelet

openshift-kube-scheduler-operator-dddff6458-vx5d7

Started

Started container kube-scheduler-operator-container

openshift-operator-lifecycle-manager

kubelet

package-server-manager-7b95f86987-2pg77

Started

Started container package-server-manager

openshift-cluster-olm-operator

kubelet

cluster-olm-operator-67dcd4998-c5742

Created

Created container: cluster-olm-operator

openshift-cluster-storage-operator

kubelet

cluster-storage-operator-7d87854d6-vlq7h

Started

Started container cluster-storage-operator

openshift-network-operator

kubelet

network-operator-7bd846bfc4-mt454

Started

Started container network-operator

openshift-network-operator

kubelet

network-operator-7bd846bfc4-mt454

Created

Created container: network-operator

openshift-cluster-storage-operator

kubelet

cluster-storage-operator-7d87854d6-vlq7h

Created

Created container: cluster-storage-operator

openshift-operator-lifecycle-manager

kubelet

package-server-manager-7b95f86987-2pg77

Created

Created container: package-server-manager

openshift-cluster-version

kubelet

cluster-version-operator-7d58488df-qmm8h

Created

Created container: cluster-version-operator

openshift-cluster-version

kubelet

cluster-version-operator-7d58488df-qmm8h

Started

Started container cluster-version-operator

openshift-machine-api

kubelet

machine-api-operator-6fbb6cf6f9-n8tnn

Started

Started container machine-api-operator

openshift-kube-storage-version-migrator-operator

kubelet

kube-storage-version-migrator-operator-6bb5bfb6fd-64dg5

Created

Created container: kube-storage-version-migrator-operator

openshift-machine-config-operator

kubelet

machine-config-controller-b4f87c5b9-pj7rj

Created

Created container: machine-config-controller

openshift-machine-config-operator

kubelet

machine-config-controller-b4f87c5b9-pj7rj

Started

Started container machine-config-controller

openshift-cluster-node-tuning-operator

kubelet

cluster-node-tuning-operator-598fbc5f8f-vxzvg

Created

Created container: cluster-node-tuning-operator

openshift-cluster-node-tuning-operator

kubelet

cluster-node-tuning-operator-598fbc5f8f-vxzvg

Started

Started container cluster-node-tuning-operator

openshift-service-ca-operator

kubelet

service-ca-operator-b865698dc-qzb2h

Created

Created container: service-ca-operator

openshift-service-ca-operator

kubelet

service-ca-operator-b865698dc-qzb2h

Started

Started container service-ca-operator

openshift-kube-storage-version-migrator-operator

kubelet

kube-storage-version-migrator-operator-6bb5bfb6fd-64dg5

Started

Started container kube-storage-version-migrator-operator

openshift-etcd-operator

kubelet

etcd-operator-8544cbcf9c-brhw4

ProbeError

Liveness probe error: Get "https://10.128.0.28:8443/healthz": dial tcp 10.128.0.28:8443: connect: connection refused body:

openshift-etcd-operator

kubelet

etcd-operator-8544cbcf9c-brhw4

Unhealthy

Liveness probe failed: Get "https://10.128.0.28:8443/healthz": dial tcp 10.128.0.28:8443: connect: connection refused

openshift-kube-scheduler

cert-recovery-controller

openshift-kube-scheduler

ControlPlaneTopology

unable to get control plane topology, using HA cluster values for leader election: the server was unable to return a response in the time allotted, but may still be processing the request (get infrastructures.config.openshift.io cluster)

openshift-config-operator

kubelet

openshift-config-operator-95bf4f4d-25cml

Killing

Container openshift-config-operator failed liveness probe, will be restarted
(x2)

openshift-config-operator

kubelet

openshift-config-operator-95bf4f4d-25cml

Started

Started container openshift-config-operator
(x4)

openshift-cluster-storage-operator

kubelet

csi-snapshot-controller-64854d9cff-f44gr

Started

Started container snapshot-controller
(x4)

openshift-cluster-storage-operator

kubelet

csi-snapshot-controller-64854d9cff-f44gr

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9609c00207cc4db97f0fd6162eb429d7f81654137f020a677e30cba26a887a24" already present on machine
(x4)

openshift-cluster-storage-operator

kubelet

csi-snapshot-controller-64854d9cff-f44gr

Created

Created container: snapshot-controller
(x2)

openshift-config-operator

kubelet

openshift-config-operator-95bf4f4d-25cml

Created

Created container: openshift-config-operator
(x2)

openshift-config-operator

kubelet

openshift-config-operator-95bf4f4d-25cml

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:712d334b7752d95580571059aae2c50e111d879af4fd8ea7cc3dbaf1a8e7dc69" already present on machine

openshift-route-controller-manager

route-controller-manager

openshift-route-controllers

LeaderElection

route-controller-manager-ffdd4b479-rhmfx_59bc68de-98ab-4bcc-b24a-adf02ce5a9fd became leader
(x5)

openshift-console-operator

multus

console-operator-76b6568d85-8b8gv

AddedInterface

Add eth0 [10.128.0.82/23] from ovn-kubernetes

openshift-machine-api

cluster-autoscaler-operator-866dc4744-xwxg7_1530e5d2-791d-4f4f-a1b3-c89513ccaa1a

cluster-autoscaler-operator-leader

LeaderElection

cluster-autoscaler-operator-866dc4744-xwxg7_1530e5d2-791d-4f4f-a1b3-c89513ccaa1a became leader

openshift-console-operator

kubelet

console-operator-76b6568d85-8b8gv

FailedCreatePodSandBox

Failed to create pod sandbox: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_console-operator-76b6568d85-8b8gv_openshift-console-operator_348f3880-793f-43e4-9de1-8511626d2552_0(1d982a2f86882c36c89f1b9b85c981a4dbc3b452811eecb54160515a52a967d4): error adding pod openshift-console-operator_console-operator-76b6568d85-8b8gv to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"1d982a2f86882c36c89f1b9b85c981a4dbc3b452811eecb54160515a52a967d4" Netns:"/var/run/netns/655152b9-33da-4e04-b0ec-a902d2d6b64b" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-console-operator;K8S_POD_NAME=console-operator-76b6568d85-8b8gv;K8S_POD_INFRA_CONTAINER_ID=1d982a2f86882c36c89f1b9b85c981a4dbc3b452811eecb54160515a52a967d4;K8S_POD_UID=348f3880-793f-43e4-9de1-8511626d2552" Path:"" ERRORED: error configuring pod [openshift-console-operator/console-operator-76b6568d85-8b8gv] networking: Multus: [openshift-console-operator/console-operator-76b6568d85-8b8gv/348f3880-793f-43e4-9de1-8511626d2552]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod console-operator-76b6568d85-8b8gv in out of cluster comm: SetNetworkStatus: failed to update the pod console-operator-76b6568d85-8b8gv in out of cluster comm: status update failed for pod /: Get "https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-console-operator/pods/console-operator-76b6568d85-8b8gv?timeout=1m0s": net/http: request canceled (Client.Timeout exceeded while awaiting headers) ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"}

openshift-machine-api

cluster-autoscaler-operator-866dc4744-xwxg7_1530e5d2-791d-4f4f-a1b3-c89513ccaa1a

cluster-autoscaler-operator-leader

LeaderElection

cluster-autoscaler-operator-866dc4744-xwxg7_1530e5d2-791d-4f4f-a1b3-c89513ccaa1a became leader

default

node-controller

master-0

RegisteredNode

Node master-0 event: Registered Node master-0 in Controller

kube-system

kube-controller-manager

kube-controller-manager

LeaderElection

master-0_c8d598b5-64ff-4b78-98db-7338ebda74de became leader
(x5)

openshift-config-operator

kubelet

openshift-config-operator-95bf4f4d-25cml

ProbeError

Readiness probe error: Get "https://10.128.0.17:8443/healthz": dial tcp 10.128.0.17:8443: connect: connection refused body:
(x4)

openshift-config-operator

kubelet

openshift-config-operator-95bf4f4d-25cml

Unhealthy

Liveness probe failed: Get "https://10.128.0.17:8443/healthz": dial tcp 10.128.0.17:8443: connect: connection refused
(x4)

openshift-config-operator

kubelet

openshift-config-operator-95bf4f4d-25cml

ProbeError

Liveness probe error: Get "https://10.128.0.17:8443/healthz": dial tcp 10.128.0.17:8443: connect: connection refused body:
(x5)

openshift-config-operator

kubelet

openshift-config-operator-95bf4f4d-25cml

Unhealthy

Readiness probe failed: Get "https://10.128.0.17:8443/healthz": dial tcp 10.128.0.17:8443: connect: connection refused

openshift-cluster-storage-operator

snapshot-controller-leader/csi-snapshot-controller-64854d9cff-f44gr

snapshot-controller-leader

LeaderElection

csi-snapshot-controller-64854d9cff-f44gr became leader

openshift-network-console

kubelet

networking-console-plugin-7c6b76c555-zbpk4

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a22978e1669cdbaeee6ec0800f83559b56a2344f1c003f8cd60f27fac939680e"

openshift-cluster-node-tuning-operator

performance-profile-controller

cluster-node-tuning-operator

FeatureGatesInitialized

FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{"AWSEFSDriverVolumeMetrics", "AdminNetworkPolicy", "AlibabaPlatform", "AzureWorkloadIdentity", "BareMetalLoadBalancer", "BuildCSIVolumes", "ChunkSizeMiB", "CloudDualStackNodeIPs", "DisableKubeletCloudCredentialProviders", "GCPLabelsTags", "HardwareSpeed", "IngressControllerLBSubnetsAWS", "KMSv1", "ManagedBootImages", "ManagedBootImagesAWS", "MultiArchInstallAWS", "MultiArchInstallGCP", "NetworkDiagnosticsConfig", "NetworkLiveMigration", "NetworkSegmentation", "NewOLM", "NodeDisruptionPolicy", "OnClusterBuild", "PersistentIPsForVirtualization", "PrivateHostedZoneAWS", "SetEIPForNLBIngressController", "StreamingCollectionEncodingToJSON", "StreamingCollectionEncodingToProtobuf", "VSphereControlPlaneMachineSet", "VSphereDriverConfiguration", "VSphereMultiVCenters", "VSphereStaticIPs", "ValidatingAdmissionPolicy"}, Disabled:[]v1.FeatureGateName{"AWSClusterHostedDNS", "AdditionalRoutingCapabilities", "AutomatedEtcdBackup", "BootcNodeManagement", "CSIDriverSharedResource", "ClusterAPIInstall", "ClusterAPIInstallIBMCloud", "ClusterMonitoringConfig", "ConsolePluginContentSecurityPolicy", "DNSNameResolver", "DynamicResourceAllocation", "EtcdBackendQuota", "EventedPLEG", "Example", "ExternalOIDC", "ExternalOIDCWithUIDAndExtraClaimMappings", "GCPClusterHostedDNS", "GatewayAPI", "ImageStreamImportMode", "IngressControllerDynamicConfigurationManager", "InsightsConfig", "InsightsConfigAPI", "InsightsOnDemandDataGather", "InsightsRuntimeExtractor", "MachineAPIMigration", "MachineAPIOperatorDisableMachineHealthCheckController", "MachineAPIProviderOpenStack", "MachineConfigNodes", "MaxUnavailableStatefulSet", "MetricsCollectionProfiles", "MinimumKubeletVersion", "MixedCPUsAllocation", "MultiArchInstallAzure", "NodeSwap", "NutanixMultiSubnets", "OVNObservability", "OpenShiftPodSecurityAdmission", "PinnedImages", "PlatformOperators", "ProcMountType", "RouteAdvertisements", "RouteExternalCertificate", "ServiceAccountTokenNodeBinding", "SignatureStores", "SigstoreImageVerification", "TranslateStreamCloseWebsocketRequests", "UpgradeStatus", "UserNamespacesPodSecurityStandards", "UserNamespacesSupport", "VSphereMultiNetworks", "VolumeAttributesClass", "VolumeGroupSnapshot"}}

openshift-cluster-node-tuning-operator

performance-profile-controller

cluster-node-tuning-operator

FeatureGatesInitialized

FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{"AWSEFSDriverVolumeMetrics", "AdminNetworkPolicy", "AlibabaPlatform", "AzureWorkloadIdentity", "BareMetalLoadBalancer", "BuildCSIVolumes", "ChunkSizeMiB", "CloudDualStackNodeIPs", "DisableKubeletCloudCredentialProviders", "GCPLabelsTags", "HardwareSpeed", "IngressControllerLBSubnetsAWS", "KMSv1", "ManagedBootImages", "ManagedBootImagesAWS", "MultiArchInstallAWS", "MultiArchInstallGCP", "NetworkDiagnosticsConfig", "NetworkLiveMigration", "NetworkSegmentation", "NewOLM", "NodeDisruptionPolicy", "OnClusterBuild", "PersistentIPsForVirtualization", "PrivateHostedZoneAWS", "SetEIPForNLBIngressController", "StreamingCollectionEncodingToJSON", "StreamingCollectionEncodingToProtobuf", "VSphereControlPlaneMachineSet", "VSphereDriverConfiguration", "VSphereMultiVCenters", "VSphereStaticIPs", "ValidatingAdmissionPolicy"}, Disabled:[]v1.FeatureGateName{"AWSClusterHostedDNS", "AdditionalRoutingCapabilities", "AutomatedEtcdBackup", "BootcNodeManagement", "CSIDriverSharedResource", "ClusterAPIInstall", "ClusterAPIInstallIBMCloud", "ClusterMonitoringConfig", "ConsolePluginContentSecurityPolicy", "DNSNameResolver", "DynamicResourceAllocation", "EtcdBackendQuota", "EventedPLEG", "Example", "ExternalOIDC", "ExternalOIDCWithUIDAndExtraClaimMappings", "GCPClusterHostedDNS", "GatewayAPI", "ImageStreamImportMode", "IngressControllerDynamicConfigurationManager", "InsightsConfig", "InsightsConfigAPI", "InsightsOnDemandDataGather", "InsightsRuntimeExtractor", "MachineAPIMigration", "MachineAPIOperatorDisableMachineHealthCheckController", "MachineAPIProviderOpenStack", "MachineConfigNodes", "MaxUnavailableStatefulSet", "MetricsCollectionProfiles", "MinimumKubeletVersion", "MixedCPUsAllocation", "MultiArchInstallAzure", "NodeSwap", "NutanixMultiSubnets", "OVNObservability", "OpenShiftPodSecurityAdmission", "PinnedImages", "PlatformOperators", "ProcMountType", "RouteAdvertisements", "RouteExternalCertificate", "ServiceAccountTokenNodeBinding", "SignatureStores", "SigstoreImageVerification", "TranslateStreamCloseWebsocketRequests", "UpgradeStatus", "UserNamespacesPodSecurityStandards", "UserNamespacesSupport", "VSphereMultiNetworks", "VolumeAttributesClass", "VolumeGroupSnapshot"}}

openshift-machine-api

machineapioperator

machine-api-operator

FeatureGatesInitialized

FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{"AWSEFSDriverVolumeMetrics", "AdminNetworkPolicy", "AlibabaPlatform", "AzureWorkloadIdentity", "BareMetalLoadBalancer", "BuildCSIVolumes", "ChunkSizeMiB", "CloudDualStackNodeIPs", "DisableKubeletCloudCredentialProviders", "GCPLabelsTags", "HardwareSpeed", "IngressControllerLBSubnetsAWS", "KMSv1", "ManagedBootImages", "ManagedBootImagesAWS", "MultiArchInstallAWS", "MultiArchInstallGCP", "NetworkDiagnosticsConfig", "NetworkLiveMigration", "NetworkSegmentation", "NewOLM", "NodeDisruptionPolicy", "OnClusterBuild", "PersistentIPsForVirtualization", "PrivateHostedZoneAWS", "SetEIPForNLBIngressController", "StreamingCollectionEncodingToJSON", "StreamingCollectionEncodingToProtobuf", "VSphereControlPlaneMachineSet", "VSphereDriverConfiguration", "VSphereMultiVCenters", "VSphereStaticIPs", "ValidatingAdmissionPolicy"}, Disabled:[]v1.FeatureGateName{"AWSClusterHostedDNS", "AdditionalRoutingCapabilities", "AutomatedEtcdBackup", "BootcNodeManagement", "CSIDriverSharedResource", "ClusterAPIInstall", "ClusterAPIInstallIBMCloud", "ClusterMonitoringConfig", "ConsolePluginContentSecurityPolicy", "DNSNameResolver", "DynamicResourceAllocation", "EtcdBackendQuota", "EventedPLEG", "Example", "ExternalOIDC", "ExternalOIDCWithUIDAndExtraClaimMappings", "GCPClusterHostedDNS", "GatewayAPI", "ImageStreamImportMode", "IngressControllerDynamicConfigurationManager", "InsightsConfig", "InsightsConfigAPI", "InsightsOnDemandDataGather", "InsightsRuntimeExtractor", "MachineAPIMigration", "MachineAPIOperatorDisableMachineHealthCheckController", "MachineAPIProviderOpenStack", "MachineConfigNodes", "MaxUnavailableStatefulSet", "MetricsCollectionProfiles", "MinimumKubeletVersion", "MixedCPUsAllocation", "MultiArchInstallAzure", "NodeSwap", "NutanixMultiSubnets", "OVNObservability", "OpenShiftPodSecurityAdmission", "PinnedImages", "PlatformOperators", "ProcMountType", "RouteAdvertisements", "RouteExternalCertificate", "ServiceAccountTokenNodeBinding", "SignatureStores", "SigstoreImageVerification", "TranslateStreamCloseWebsocketRequests", "UpgradeStatus", "UserNamespacesPodSecurityStandards", "UserNamespacesSupport", "VSphereMultiNetworks", "VolumeAttributesClass", "VolumeGroupSnapshot"}}

openshift-kube-apiserver-operator

kube-apiserver-operator-audit-policy-controller-auditpolicycontroller

kube-apiserver-operator

FastControllerResync

Controller "auditPolicyController" resync interval is set to 10s which might lead to client request throttling

openshift-kube-storage-version-migrator-operator

openshift-kube-storage-version-migrator-operator

openshift-kube-storage-version-migrator-operator-lock

LeaderElection

kube-storage-version-migrator-operator-6bb5bfb6fd-64dg5_519c444d-af41-4dfc-b922-6d94c65bb53e became leader

openshift-kube-apiserver-operator

kube-apiserver-operator

kube-apiserver-operator-lock

LeaderElection

kube-apiserver-operator-8b68b9d9b-c6vkz_565932cc-94c6-40fb-af28-f7906f6e1613 became leader
(x3)

openshift-machine-api

kubelet

cluster-baremetal-operator-6f69995874-dv6cd

BackOff

Back-off restarting failed container cluster-baremetal-operator in pod cluster-baremetal-operator-6f69995874-dv6cd_openshift-machine-api(f53bc282-5937-49ac-ac98-2ee37ccb268d)

openshift-machine-api

machineapioperator

machine-api-operator

FeatureGatesInitialized

FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{"AWSEFSDriverVolumeMetrics", "AdminNetworkPolicy", "AlibabaPlatform", "AzureWorkloadIdentity", "BareMetalLoadBalancer", "BuildCSIVolumes", "ChunkSizeMiB", "CloudDualStackNodeIPs", "DisableKubeletCloudCredentialProviders", "GCPLabelsTags", "HardwareSpeed", "IngressControllerLBSubnetsAWS", "KMSv1", "ManagedBootImages", "ManagedBootImagesAWS", "MultiArchInstallAWS", "MultiArchInstallGCP", "NetworkDiagnosticsConfig", "NetworkLiveMigration", "NetworkSegmentation", "NewOLM", "NodeDisruptionPolicy", "OnClusterBuild", "PersistentIPsForVirtualization", "PrivateHostedZoneAWS", "SetEIPForNLBIngressController", "StreamingCollectionEncodingToJSON", "StreamingCollectionEncodingToProtobuf", "VSphereControlPlaneMachineSet", "VSphereDriverConfiguration", "VSphereMultiVCenters", "VSphereStaticIPs", "ValidatingAdmissionPolicy"}, Disabled:[]v1.FeatureGateName{"AWSClusterHostedDNS", "AdditionalRoutingCapabilities", "AutomatedEtcdBackup", "BootcNodeManagement", "CSIDriverSharedResource", "ClusterAPIInstall", "ClusterAPIInstallIBMCloud", "ClusterMonitoringConfig", "ConsolePluginContentSecurityPolicy", "DNSNameResolver", "DynamicResourceAllocation", "EtcdBackendQuota", "EventedPLEG", "Example", "ExternalOIDC", "ExternalOIDCWithUIDAndExtraClaimMappings", "GCPClusterHostedDNS", "GatewayAPI", "ImageStreamImportMode", "IngressControllerDynamicConfigurationManager", "InsightsConfig", "InsightsConfigAPI", "InsightsOnDemandDataGather", "InsightsRuntimeExtractor", "MachineAPIMigration", "MachineAPIOperatorDisableMachineHealthCheckController", "MachineAPIProviderOpenStack", "MachineConfigNodes", "MaxUnavailableStatefulSet", "MetricsCollectionProfiles", "MinimumKubeletVersion", "MixedCPUsAllocation", "MultiArchInstallAzure", "NodeSwap", "NutanixMultiSubnets", "OVNObservability", "OpenShiftPodSecurityAdmission", "PinnedImages", "PlatformOperators", "ProcMountType", "RouteAdvertisements", "RouteExternalCertificate", "ServiceAccountTokenNodeBinding", "SignatureStores", "SigstoreImageVerification", "TranslateStreamCloseWebsocketRequests", "UpgradeStatus", "UserNamespacesPodSecurityStandards", "UserNamespacesSupport", "VSphereMultiNetworks", "VolumeAttributesClass", "VolumeGroupSnapshot"}}

openshift-controller-manager-operator

openshift-controller-manager-operator

openshift-controller-manager-operator

FeatureGatesInitialized

FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{"AWSEFSDriverVolumeMetrics", "AdminNetworkPolicy", "AlibabaPlatform", "AzureWorkloadIdentity", "BareMetalLoadBalancer", "BuildCSIVolumes", "ChunkSizeMiB", "CloudDualStackNodeIPs", "DisableKubeletCloudCredentialProviders", "GCPLabelsTags", "HardwareSpeed", "IngressControllerLBSubnetsAWS", "KMSv1", "ManagedBootImages", "ManagedBootImagesAWS", "MultiArchInstallAWS", "MultiArchInstallGCP", "NetworkDiagnosticsConfig", "NetworkLiveMigration", "NetworkSegmentation", "NewOLM", "NodeDisruptionPolicy", "OnClusterBuild", "PersistentIPsForVirtualization", "PrivateHostedZoneAWS", "SetEIPForNLBIngressController", "StreamingCollectionEncodingToJSON", "StreamingCollectionEncodingToProtobuf", "VSphereControlPlaneMachineSet", "VSphereDriverConfiguration", "VSphereMultiVCenters", "VSphereStaticIPs", "ValidatingAdmissionPolicy"}, Disabled:[]v1.FeatureGateName{"AWSClusterHostedDNS", "AdditionalRoutingCapabilities", "AutomatedEtcdBackup", "BootcNodeManagement", "CSIDriverSharedResource", "ClusterAPIInstall", "ClusterAPIInstallIBMCloud", "ClusterMonitoringConfig", "ConsolePluginContentSecurityPolicy", "DNSNameResolver", "DynamicResourceAllocation", "EtcdBackendQuota", "EventedPLEG", "Example", "ExternalOIDC", "ExternalOIDCWithUIDAndExtraClaimMappings", "GCPClusterHostedDNS", "GatewayAPI", "ImageStreamImportMode", "IngressControllerDynamicConfigurationManager", "InsightsConfig", "InsightsConfigAPI", "InsightsOnDemandDataGather", "InsightsRuntimeExtractor", "MachineAPIMigration", "MachineAPIOperatorDisableMachineHealthCheckController", "MachineAPIProviderOpenStack", "MachineConfigNodes", "MaxUnavailableStatefulSet", "MetricsCollectionProfiles", "MinimumKubeletVersion", "MixedCPUsAllocation", "MultiArchInstallAzure", "NodeSwap", "NutanixMultiSubnets", "OVNObservability", "OpenShiftPodSecurityAdmission", "PinnedImages", "PlatformOperators", "ProcMountType", "RouteAdvertisements", "RouteExternalCertificate", "ServiceAccountTokenNodeBinding", "SignatureStores", "SigstoreImageVerification", "TranslateStreamCloseWebsocketRequests", "UpgradeStatus", "UserNamespacesPodSecurityStandards", "UserNamespacesSupport", "VSphereMultiNetworks", "VolumeAttributesClass", "VolumeGroupSnapshot"}}

openshift-controller-manager-operator

openshift-controller-manager-operator

openshift-controller-manager-operator-lock

LeaderElection

openshift-controller-manager-operator-8c94f4649-p7pt6_98bcf1c0-0854-415a-9e13-3e85e0ba13c0 became leader

openshift-kube-apiserver

cert-regeneration-controller

cert-regeneration-controller-lock

LeaderElection

master-0_0a176ba4-d550-4788-a00d-1a7579e6996f became leader

openshift-console-operator

kubelet

console-operator-76b6568d85-8b8gv

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98bf5467a01195e20aeea7d6f0b130ddacc00b73bc5312253b8c34e7208538f8"

openshift-kube-apiserver-operator

kube-apiserver-operator

kube-apiserver-operator

FeatureGatesInitialized

FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{"AWSEFSDriverVolumeMetrics", "AdminNetworkPolicy", "AlibabaPlatform", "AzureWorkloadIdentity", "BareMetalLoadBalancer", "BuildCSIVolumes", "ChunkSizeMiB", "CloudDualStackNodeIPs", "DisableKubeletCloudCredentialProviders", "GCPLabelsTags", "HardwareSpeed", "IngressControllerLBSubnetsAWS", "KMSv1", "ManagedBootImages", "ManagedBootImagesAWS", "MultiArchInstallAWS", "MultiArchInstallGCP", "NetworkDiagnosticsConfig", "NetworkLiveMigration", "NetworkSegmentation", "NewOLM", "NodeDisruptionPolicy", "OnClusterBuild", "PersistentIPsForVirtualization", "PrivateHostedZoneAWS", "SetEIPForNLBIngressController", "StreamingCollectionEncodingToJSON", "StreamingCollectionEncodingToProtobuf", "VSphereControlPlaneMachineSet", "VSphereDriverConfiguration", "VSphereMultiVCenters", "VSphereStaticIPs", "ValidatingAdmissionPolicy"}, Disabled:[]v1.FeatureGateName{"AWSClusterHostedDNS", "AdditionalRoutingCapabilities", "AutomatedEtcdBackup", "BootcNodeManagement", "CSIDriverSharedResource", "ClusterAPIInstall", "ClusterAPIInstallIBMCloud", "ClusterMonitoringConfig", "ConsolePluginContentSecurityPolicy", "DNSNameResolver", "DynamicResourceAllocation", "EtcdBackendQuota", "EventedPLEG", "Example", "ExternalOIDC", "ExternalOIDCWithUIDAndExtraClaimMappings", "GCPClusterHostedDNS", "GatewayAPI", "ImageStreamImportMode", "IngressControllerDynamicConfigurationManager", "InsightsConfig", "InsightsConfigAPI", "InsightsOnDemandDataGather", "InsightsRuntimeExtractor", "MachineAPIMigration", "MachineAPIOperatorDisableMachineHealthCheckController", "MachineAPIProviderOpenStack", "MachineConfigNodes", "MaxUnavailableStatefulSet", "MetricsCollectionProfiles", "MinimumKubeletVersion", "MixedCPUsAllocation", "MultiArchInstallAzure", "NodeSwap", "NutanixMultiSubnets", "OVNObservability", "OpenShiftPodSecurityAdmission", "PinnedImages", "PlatformOperators", "ProcMountType", "RouteAdvertisements", "RouteExternalCertificate", "ServiceAccountTokenNodeBinding", "SignatureStores", "SigstoreImageVerification", "TranslateStreamCloseWebsocketRequests", "UpgradeStatus", "UserNamespacesPodSecurityStandards", "UserNamespacesSupport", "VSphereMultiNetworks", "VolumeAttributesClass", "VolumeGroupSnapshot"}}
(x3)

openshift-machine-api

kubelet

cluster-baremetal-operator-6f69995874-dv6cd

BackOff

Back-off restarting failed container cluster-baremetal-operator in pod cluster-baremetal-operator-6f69995874-dv6cd_openshift-machine-api(f53bc282-5937-49ac-ac98-2ee37ccb268d)

openshift-service-ca-operator

service-ca-operator

service-ca-operator-lock

LeaderElection

service-ca-operator-b865698dc-qzb2h_1e0c1a2c-82ba-446c-9d0e-4a663c131080 became leader

openshift-etcd-operator

openshift-cluster-etcd-operator-status-controller-statussyncer_etcd

etcd-operator

OperatorStatusChanged

Status for clusteroperator/etcd changed: Degraded message changed from "NodeControllerDegraded: All master nodes are ready\nEtcdMembersDegraded: No unhealthy members found" to "EtcdMembersControllerDegraded: getting cache client could not retrieve endpoints: node lister not synced\nNodeControllerDegraded: All master nodes are ready\nEtcdMembersDegraded: No unhealthy members found"

openshift-etcd-operator

openshift-cluster-etcd-operator-status-controller-statussyncer_etcd

etcd-operator

OperatorStatusChanged

Status for clusteroperator/etcd changed: Degraded message changed from "EtcdMembersControllerDegraded: getting cache client could not retrieve endpoints: node lister not synced\nNodeControllerDegraded: All master nodes are ready\nEtcdMembersDegraded: No unhealthy members found" to "EtcdMembersControllerDegraded: getting cache client could not retrieve endpoints: configmaps lister not synced\nNodeControllerDegraded: All master nodes are ready\nEtcdMembersDegraded: No unhealthy members found"

openshift-etcd-operator

openshift-cluster-etcd-operator

etcd-operator

FeatureGatesInitialized

FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{"AWSEFSDriverVolumeMetrics", "AdminNetworkPolicy", "AlibabaPlatform", "AzureWorkloadIdentity", "BareMetalLoadBalancer", "BuildCSIVolumes", "ChunkSizeMiB", "CloudDualStackNodeIPs", "DisableKubeletCloudCredentialProviders", "GCPLabelsTags", "HardwareSpeed", "IngressControllerLBSubnetsAWS", "KMSv1", "ManagedBootImages", "ManagedBootImagesAWS", "MultiArchInstallAWS", "MultiArchInstallGCP", "NetworkDiagnosticsConfig", "NetworkLiveMigration", "NetworkSegmentation", "NewOLM", "NodeDisruptionPolicy", "OnClusterBuild", "PersistentIPsForVirtualization", "PrivateHostedZoneAWS", "SetEIPForNLBIngressController", "StreamingCollectionEncodingToJSON", "StreamingCollectionEncodingToProtobuf", "VSphereControlPlaneMachineSet", "VSphereDriverConfiguration", "VSphereMultiVCenters", "VSphereStaticIPs", "ValidatingAdmissionPolicy"}, Disabled:[]v1.FeatureGateName{"AWSClusterHostedDNS", "AdditionalRoutingCapabilities", "AutomatedEtcdBackup", "BootcNodeManagement", "CSIDriverSharedResource", "ClusterAPIInstall", "ClusterAPIInstallIBMCloud", "ClusterMonitoringConfig", "ConsolePluginContentSecurityPolicy", "DNSNameResolver", "DynamicResourceAllocation", "EtcdBackendQuota", "EventedPLEG", "Example", "ExternalOIDC", "ExternalOIDCWithUIDAndExtraClaimMappings", "GCPClusterHostedDNS", "GatewayAPI", "ImageStreamImportMode", "IngressControllerDynamicConfigurationManager", "InsightsConfig", "InsightsConfigAPI", "InsightsOnDemandDataGather", "InsightsRuntimeExtractor", "MachineAPIMigration", "MachineAPIOperatorDisableMachineHealthCheckController", "MachineAPIProviderOpenStack", "MachineConfigNodes", "MaxUnavailableStatefulSet", "MetricsCollectionProfiles", "MinimumKubeletVersion", "MixedCPUsAllocation", "MultiArchInstallAzure", "NodeSwap", "NutanixMultiSubnets", "OVNObservability", "OpenShiftPodSecurityAdmission", "PinnedImages", "PlatformOperators", "ProcMountType", "RouteAdvertisements", "RouteExternalCertificate", "ServiceAccountTokenNodeBinding", "SignatureStores", "SigstoreImageVerification", "TranslateStreamCloseWebsocketRequests", "UpgradeStatus", "UserNamespacesPodSecurityStandards", "UserNamespacesSupport", "VSphereMultiNetworks", "VolumeAttributesClass", "VolumeGroupSnapshot"}}

openshift-authentication-operator

cluster-authentication-operator

cluster-authentication-operator-lock

LeaderElection

authentication-operator-5885bfd7f4-62zrx_b354b6a2-b0dc-40f4-8515-acb3e891b696 became leader

openshift-network-operator

cluster-network-operator

network-operator

FeatureGatesInitialized

FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{"AWSEFSDriverVolumeMetrics", "AdminNetworkPolicy", "AlibabaPlatform", "AzureWorkloadIdentity", "BareMetalLoadBalancer", "BuildCSIVolumes", "ChunkSizeMiB", "CloudDualStackNodeIPs", "DisableKubeletCloudCredentialProviders", "GCPLabelsTags", "HardwareSpeed", "IngressControllerLBSubnetsAWS", "KMSv1", "ManagedBootImages", "ManagedBootImagesAWS", "MultiArchInstallAWS", "MultiArchInstallGCP", "NetworkDiagnosticsConfig", "NetworkLiveMigration", "NetworkSegmentation", "NewOLM", "NodeDisruptionPolicy", "OnClusterBuild", "PersistentIPsForVirtualization", "PrivateHostedZoneAWS", "SetEIPForNLBIngressController", "StreamingCollectionEncodingToJSON", "StreamingCollectionEncodingToProtobuf", "VSphereControlPlaneMachineSet", "VSphereDriverConfiguration", "VSphereMultiVCenters", "VSphereStaticIPs", "ValidatingAdmissionPolicy"}, Disabled:[]v1.FeatureGateName{"AWSClusterHostedDNS", "AdditionalRoutingCapabilities", "AutomatedEtcdBackup", "BootcNodeManagement", "CSIDriverSharedResource", "ClusterAPIInstall", "ClusterAPIInstallIBMCloud", "ClusterMonitoringConfig", "ConsolePluginContentSecurityPolicy", "DNSNameResolver", "DynamicResourceAllocation", "EtcdBackendQuota", "EventedPLEG", "Example", "ExternalOIDC", "ExternalOIDCWithUIDAndExtraClaimMappings", "GCPClusterHostedDNS", "GatewayAPI", "ImageStreamImportMode", "IngressControllerDynamicConfigurationManager", "InsightsConfig", "InsightsConfigAPI", "InsightsOnDemandDataGather", "InsightsRuntimeExtractor", "MachineAPIMigration", "MachineAPIOperatorDisableMachineHealthCheckController", "MachineAPIProviderOpenStack", "MachineConfigNodes", "MaxUnavailableStatefulSet", "MetricsCollectionProfiles", "MinimumKubeletVersion", "MixedCPUsAllocation", "MultiArchInstallAzure", "NodeSwap", "NutanixMultiSubnets", "OVNObservability", "OpenShiftPodSecurityAdmission", "PinnedImages", "PlatformOperators", "ProcMountType", "RouteAdvertisements", "RouteExternalCertificate", "ServiceAccountTokenNodeBinding", "SignatureStores", "SigstoreImageVerification", "TranslateStreamCloseWebsocketRequests", "UpgradeStatus", "UserNamespacesPodSecurityStandards", "UserNamespacesSupport", "VSphereMultiNetworks", "VolumeAttributesClass", "VolumeGroupSnapshot"}}

openshift-etcd-operator

openshift-cluster-etcd-operator

openshift-cluster-etcd-operator-lock

LeaderElection

etcd-operator-8544cbcf9c-brhw4_67745e68-6f0f-48d3-a703-3551419bb818 became leader

openshift-network-operator

network-operator

network-operator-lock

LeaderElection

master-0_992f3801-f9f2-43a4-825c-72321fcd0c91 became leader

openshift-etcd-operator

openshift-cluster-etcd-operator-status-controller-statussyncer_etcd

etcd-operator

OperatorStatusChanged

Status for clusteroperator/etcd changed: Degraded message changed from "EtcdMembersControllerDegraded: getting cache client could not retrieve endpoints: configmaps lister not synced\nNodeControllerDegraded: All master nodes are ready\nScriptControllerDegraded: \"configmap/etcd-pod\": missing env var values\nEtcdMembersDegraded: No unhealthy members found" to "EtcdMembersControllerDegraded: getting cache client could not retrieve endpoints: configmaps lister not synced\nNodeControllerDegraded: All master nodes are ready\nMissingStaticPodControllerDegraded: static pod lifecycle failure - static pod: \"etcd\" in namespace: \"openshift-etcd\" for revision: 2 on node: \"master-0\" didn't show up, waited: 3m30s\nScriptControllerDegraded: \"configmap/etcd-pod\": missing env var values\nEtcdMembersDegraded: No unhealthy members found"

openshift-etcd-operator

openshift-cluster-etcd-operator-missingstaticpodcontroller

etcd-operator

MissingStaticPod

static pod lifecycle failure - static pod: "etcd" in namespace: "openshift-etcd" for revision: 2 on node: "master-0" didn't show up, waited: 3m30s

openshift-multus

daemonset-controller

cni-sysctl-allowlist-ds

SuccessfulCreate

Created pod: cni-sysctl-allowlist-ds-8zskc

openshift-network-console

kubelet

networking-console-plugin-7c6b76c555-zbpk4

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a22978e1669cdbaeee6ec0800f83559b56a2344f1c003f8cd60f27fac939680e" in 2.909s (2.909s including waiting). Image size: 446952788 bytes.

openshift-multus

daemonset-controller

cni-sysctl-allowlist-ds

SuccessfulCreate

Created pod: cni-sysctl-allowlist-ds-8zskc

openshift-etcd-operator

openshift-cluster-etcd-operator-status-controller-statussyncer_etcd

etcd-operator

OperatorStatusChanged

Status for clusteroperator/etcd changed: Degraded message changed from "EtcdMembersControllerDegraded: getting cache client could not retrieve endpoints: configmaps lister not synced\nNodeControllerDegraded: All master nodes are ready\nEtcdMembersDegraded: No unhealthy members found" to "EtcdMembersControllerDegraded: getting cache client could not retrieve endpoints: configmaps lister not synced\nNodeControllerDegraded: All master nodes are ready\nScriptControllerDegraded: \"configmap/etcd-pod\": missing env var values\nEtcdMembersDegraded: No unhealthy members found"

openshift-kube-controller-manager

cluster-policy-controller

kube-controller-manager-master-0

ControlPlaneTopology

unable to get control plane topology, using HA cluster values for leader election: infrastructures.config.openshift.io "cluster" is forbidden: User "system:kube-controller-manager" cannot get resource "infrastructures" in API group "config.openshift.io" at the cluster scope

openshift-network-console

kubelet

networking-console-plugin-7c6b76c555-zbpk4

Started

Started container networking-console-plugin

openshift-network-console

kubelet

networking-console-plugin-7c6b76c555-zbpk4

Created

Created container: networking-console-plugin

openshift-console-operator

kubelet

console-operator-76b6568d85-8b8gv

Started

Started container console-operator

openshift-etcd-operator

openshift-cluster-etcd-operator-status-controller-statussyncer_etcd

etcd-operator

OperatorStatusChanged

Status for clusteroperator/etcd changed: Degraded message changed from "EtcdMembersControllerDegraded: getting cache client could not retrieve endpoints: configmaps lister not synced\nNodeControllerDegraded: All master nodes are ready\nMissingStaticPodControllerDegraded: static pod lifecycle failure - static pod: \"etcd\" in namespace: \"openshift-etcd\" for revision: 2 on node: \"master-0\" didn't show up, waited: 3m30s\nScriptControllerDegraded: \"configmap/etcd-pod\": missing env var values\nEtcdMembersDegraded: No unhealthy members found" to "NodeControllerDegraded: All master nodes are ready\nMissingStaticPodControllerDegraded: static pod lifecycle failure - static pod: \"etcd\" in namespace: \"openshift-etcd\" for revision: 2 on node: \"master-0\" didn't show up, waited: 3m30s\nScriptControllerDegraded: \"configmap/etcd-pod\": missing env var values\nEtcdMembersDegraded: No unhealthy members found"

openshift-console-operator

console-operator

console-operator

FeatureGatesInitialized

FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{"AWSEFSDriverVolumeMetrics", "AdminNetworkPolicy", "AlibabaPlatform", "AzureWorkloadIdentity", "BareMetalLoadBalancer", "BuildCSIVolumes", "ChunkSizeMiB", "CloudDualStackNodeIPs", "DisableKubeletCloudCredentialProviders", "GCPLabelsTags", "HardwareSpeed", "IngressControllerLBSubnetsAWS", "KMSv1", "ManagedBootImages", "ManagedBootImagesAWS", "MultiArchInstallAWS", "MultiArchInstallGCP", "NetworkDiagnosticsConfig", "NetworkLiveMigration", "NetworkSegmentation", "NewOLM", "NodeDisruptionPolicy", "OnClusterBuild", "PersistentIPsForVirtualization", "PrivateHostedZoneAWS", "SetEIPForNLBIngressController", "StreamingCollectionEncodingToJSON", "StreamingCollectionEncodingToProtobuf", "VSphereControlPlaneMachineSet", "VSphereDriverConfiguration", "VSphereMultiVCenters", "VSphereStaticIPs", "ValidatingAdmissionPolicy"}, Disabled:[]v1.FeatureGateName{"AWSClusterHostedDNS", "AdditionalRoutingCapabilities", "AutomatedEtcdBackup", "BootcNodeManagement", "CSIDriverSharedResource", "ClusterAPIInstall", "ClusterAPIInstallIBMCloud", "ClusterMonitoringConfig", "ConsolePluginContentSecurityPolicy", "DNSNameResolver", "DynamicResourceAllocation", "EtcdBackendQuota", "EventedPLEG", "Example", "ExternalOIDC", "ExternalOIDCWithUIDAndExtraClaimMappings", "GCPClusterHostedDNS", "GatewayAPI", "ImageStreamImportMode", "IngressControllerDynamicConfigurationManager", "InsightsConfig", "InsightsConfigAPI", "InsightsOnDemandDataGather", "InsightsRuntimeExtractor", "MachineAPIMigration", "MachineAPIOperatorDisableMachineHealthCheckController", "MachineAPIProviderOpenStack", "MachineConfigNodes", "MaxUnavailableStatefulSet", "MetricsCollectionProfiles", "MinimumKubeletVersion", "MixedCPUsAllocation", "MultiArchInstallAzure", "NodeSwap", "NutanixMultiSubnets", "OVNObservability", "OpenShiftPodSecurityAdmission", "PinnedImages", "PlatformOperators", "ProcMountType", "RouteAdvertisements", "RouteExternalCertificate", "ServiceAccountTokenNodeBinding", "SignatureStores", "SigstoreImageVerification", "TranslateStreamCloseWebsocketRequests", "UpgradeStatus", "UserNamespacesPodSecurityStandards", "UserNamespacesSupport", "VSphereMultiNetworks", "VolumeAttributesClass", "VolumeGroupSnapshot"}}

openshift-console-operator

console-operator

console-operator-lock

LeaderElection

console-operator-76b6568d85-8b8gv_f9796dad-dc12-48be-9097-40b9bdcafeae became leader

openshift-authentication-operator

oauth-apiserver-audit-policy-controller-auditpolicycontroller

authentication-operator

FastControllerResync

Controller "auditPolicyController" resync interval is set to 10s which might lead to client request throttling

openshift-console-operator

kubelet

console-operator-76b6568d85-8b8gv

Created

Created container: console-operator

openshift-console-operator

kubelet

console-operator-76b6568d85-8b8gv

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98bf5467a01195e20aeea7d6f0b130ddacc00b73bc5312253b8c34e7208538f8" in 4.052s (4.052s including waiting). Image size: 512235769 bytes.

openshift-authentication-operator

oauth-apiserver-apiservice-openshift-apiserver-controller-openshift-apiserver-apiservice

authentication-operator

FastControllerResync

Controller "openshift-apiserver-APIService" resync interval is set to 10s which might lead to client request throttling

openshift-apiserver-operator

openshift-apiserver-operator

openshift-apiserver-operator-lock

LeaderElection

openshift-apiserver-operator-d65958b8-th2vj_fea5a335-ccfa-4dca-8ee9-6c9ecdcaa854 became leader

openshift-apiserver-operator

openshift-apiserver-operator-audit-policy-controller-auditpolicycontroller

openshift-apiserver-operator

FastControllerResync

Controller "auditPolicyController" resync interval is set to 10s which might lead to client request throttling

openshift-apiserver-operator

openshift-apiserver-operator

openshift-apiserver-operator

FeatureGatesInitialized

FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{"AWSEFSDriverVolumeMetrics", "AdminNetworkPolicy", "AlibabaPlatform", "AzureWorkloadIdentity", "BareMetalLoadBalancer", "BuildCSIVolumes", "ChunkSizeMiB", "CloudDualStackNodeIPs", "DisableKubeletCloudCredentialProviders", "GCPLabelsTags", "HardwareSpeed", "IngressControllerLBSubnetsAWS", "KMSv1", "ManagedBootImages", "ManagedBootImagesAWS", "MultiArchInstallAWS", "MultiArchInstallGCP", "NetworkDiagnosticsConfig", "NetworkLiveMigration", "NetworkSegmentation", "NewOLM", "NodeDisruptionPolicy", "OnClusterBuild", "PersistentIPsForVirtualization", "PrivateHostedZoneAWS", "SetEIPForNLBIngressController", "StreamingCollectionEncodingToJSON", "StreamingCollectionEncodingToProtobuf", "VSphereControlPlaneMachineSet", "VSphereDriverConfiguration", "VSphereMultiVCenters", "VSphereStaticIPs", "ValidatingAdmissionPolicy"}, Disabled:[]v1.FeatureGateName{"AWSClusterHostedDNS", "AdditionalRoutingCapabilities", "AutomatedEtcdBackup", "BootcNodeManagement", "CSIDriverSharedResource", "ClusterAPIInstall", "ClusterAPIInstallIBMCloud", "ClusterMonitoringConfig", "ConsolePluginContentSecurityPolicy", "DNSNameResolver", "DynamicResourceAllocation", "EtcdBackendQuota", "EventedPLEG", "Example", "ExternalOIDC", "ExternalOIDCWithUIDAndExtraClaimMappings", "GCPClusterHostedDNS", "GatewayAPI", "ImageStreamImportMode", "IngressControllerDynamicConfigurationManager", "InsightsConfig", "InsightsConfigAPI", "InsightsOnDemandDataGather", "InsightsRuntimeExtractor", "MachineAPIMigration", "MachineAPIOperatorDisableMachineHealthCheckController", "MachineAPIProviderOpenStack", "MachineConfigNodes", "MaxUnavailableStatefulSet", "MetricsCollectionProfiles", "MinimumKubeletVersion", "MixedCPUsAllocation", "MultiArchInstallAzure", "NodeSwap", "NutanixMultiSubnets", "OVNObservability", "OpenShiftPodSecurityAdmission", "PinnedImages", "PlatformOperators", "ProcMountType", "RouteAdvertisements", "RouteExternalCertificate", "ServiceAccountTokenNodeBinding", "SignatureStores", "SigstoreImageVerification", "TranslateStreamCloseWebsocketRequests", "UpgradeStatus", "UserNamespacesPodSecurityStandards", "UserNamespacesSupport", "VSphereMultiNetworks", "VolumeAttributesClass", "VolumeGroupSnapshot"}}

openshift-apiserver-operator

openshift-apiserver-operator-status-controller-statussyncer_openshift-apiserver

openshift-apiserver-operator

OperatorStatusChanged

Status for clusteroperator/openshift-apiserver changed: Available changed from False to True ("All is well")

openshift-apiserver-operator

openshift-apiserver-operator-apiservice-openshift-apiserver-controller-openshift-apiserver-apiservice

openshift-apiserver-operator

FastControllerResync

Controller "openshift-apiserver-APIService" resync interval is set to 10s which might lead to client request throttling
(x3)

openshift-machine-api

kubelet

cluster-baremetal-operator-6f69995874-dv6cd

Started

Started container cluster-baremetal-operator
(x3)

openshift-machine-api

kubelet

cluster-baremetal-operator-6f69995874-dv6cd

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f933312f49083e8746fc41ab5e46a9a757b448374f14971e256ebcb36f11dd97" already present on machine
(x3)

openshift-machine-api

kubelet

cluster-baremetal-operator-6f69995874-dv6cd

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f933312f49083e8746fc41ab5e46a9a757b448374f14971e256ebcb36f11dd97" already present on machine
(x3)

openshift-machine-api

kubelet

cluster-baremetal-operator-6f69995874-dv6cd

Created

Created container: cluster-baremetal-operator
(x3)

openshift-machine-api

kubelet

cluster-baremetal-operator-6f69995874-dv6cd

Created

Created container: cluster-baremetal-operator
(x3)

openshift-machine-api

kubelet

cluster-baremetal-operator-6f69995874-dv6cd

Started

Started container cluster-baremetal-operator

openshift-multus

replicaset-controller

multus-admission-controller-69c949f678

SuccessfulCreate

Created pod: multus-admission-controller-69c949f678-v7wb6

openshift-multus

deployment-controller

multus-admission-controller

ScalingReplicaSet

Scaled up replica set multus-admission-controller-69c949f678 to 1

openshift-kube-apiserver-operator

kube-apiserver-operator-status-controller-statussyncer_kube-apiserver

kube-apiserver-operator

OperatorStatusChanged

Status for clusteroperator/kube-apiserver changed: Progressing changed from True to False ("NodeInstallerProgressing: 1 node is at revision 4"),Available message changed from "StaticPodsAvailable: 1 nodes are active; 1 node is at revision 3; 0 nodes have achieved new revision 4" to "StaticPodsAvailable: 1 nodes are active; 1 node is at revision 4"

openshift-multus

replicaset-controller

multus-admission-controller-69c949f678

SuccessfulCreate

Created pod: multus-admission-controller-69c949f678-v7wb6

openshift-multus

deployment-controller

multus-admission-controller

ScalingReplicaSet

Scaled up replica set multus-admission-controller-69c949f678 to 1

openshift-console-operator

console-operator-downloads-pdb-controller-poddisruptionbudgetcontroller

console-operator

PodDisruptionBudgetCreated

Created PodDisruptionBudget.policy/downloads -n openshift-console because it was missing
(x2)

openshift-console

controllermanager

console

NoPods

No matching pods found

openshift-console-operator

console-operator-status-controller-statussyncer_console

console-operator

OperatorStatusChanged

Status for clusteroperator/console changed: Degraded message changed from "All is well" to "SyncLoopRefreshDegraded: route.route.openshift.io \"console\" not found"

openshift-console-operator

console-operator-console-pdb-controller-poddisruptionbudgetcontroller

console-operator

PodDisruptionBudgetCreated

Created PodDisruptionBudget.policy/console -n openshift-console because it was missing

openshift-console-operator

console-operator-resource-sync-controller-resourcesynccontroller

console-operator

ConfigMapCreated

Created ConfigMap/oauth-serving-cert -n openshift-console because it was missing
(x2)

openshift-console-operator

console-operator-status-controller-statussyncer_console

console-operator

OperatorVersionChanged

clusteroperator/console version "operator" changed from "" to "4.18.35"

openshift-console-operator

console-operator-status-controller-statussyncer_console

console-operator

OperatorStatusChanged

Status for clusteroperator/console changed: Degraded set to False ("All is well"),Progressing set to False ("All is well"),Available set to Unknown (""),Upgradeable set to Unknown (""),EvaluationConditionsDetected set to Unknown (""),status.relatedObjects changed from [] to [{"console.openshift.io" "consoleplugins" "" "monitoring-plugin"} {"console.openshift.io" "consoleplugins" "" "networking-console-plugin"} {"operator.openshift.io" "consoles" "" "cluster"} {"config.openshift.io" "consoles" "" "cluster"} {"config.openshift.io" "infrastructures" "" "cluster"} {"config.openshift.io" "proxies" "" "cluster"} {"config.openshift.io" "oauths" "" "cluster"} {"oauth.openshift.io" "oauthclients" "" "console"} {"" "namespaces" "" "openshift-console-operator"} {"" "namespaces" "" "openshift-console"} {"" "configmaps" "openshift-config-managed" "console-public"}],status.versions changed from [] to [{"operator" "4.18.35"}]

openshift-console-operator

console-operator-health-check-controller-healthcheckcontroller

console-operator

FastControllerResync

Controller "HealthCheckController" resync interval is set to 30s which might lead to client request throttling

openshift-console

deployment-controller

downloads

ScalingReplicaSet

Scaled up replica set downloads-66b8ffb895 to 1

openshift-console

controllermanager

downloads

NoPods

No matching pods found

openshift-console

replicaset-controller

downloads-66b8ffb895

SuccessfulCreate

Created pod: downloads-66b8ffb895-dgq8g

openshift-console-operator

console-operator-console-downloads-deployment-controller-consoledownloadsdeploymentsynccontroller

console-operator

DeploymentCreated

Created Deployment.apps/downloads -n openshift-console because it was missing

openshift-console-operator

console-operator-console-service-controller-consoleservicecontroller

console-operator

ServiceCreated

Created Service/downloads -n openshift-console because it was missing

openshift-console-operator

console-operator-console-service-controller-consoleservicecontroller

console-operator

ServiceCreated

Created Service/console -n openshift-console because it was missing

openshift-console-operator

console-operator-resource-sync-controller-resourcesynccontroller

console-operator

ConfigMapCreated

Created ConfigMap/default-ingress-cert -n openshift-console because it was missing

openshift-console-operator

console-operator-status-controller-statussyncer_console

console-operator

OperatorStatusChanged

Status for clusteroperator/console changed: Degraded message changed from "SyncLoopRefreshDegraded: route.route.openshift.io \"console\" not found" to "SyncLoopRefreshDegraded: route.route.openshift.io \"console\" not found\nOAuthClientsControllerDegraded: route.route.openshift.io \"console\" not found"

openshift-console-operator

console-operator-oauthclient-secret-controller-oauthclientsecretcontroller

console-operator

SecretCreated

Created Secret/console-oauth-config -n openshift-console because it was missing

openshift-console-operator

console-operator-status-controller-statussyncer_console

console-operator

OperatorStatusChanged

Status for clusteroperator/console changed: Degraded message changed from "SyncLoopRefreshDegraded: route.route.openshift.io \"console\" not found\nOAuthClientsControllerDegraded: route.route.openshift.io \"console\" not found" to "SyncLoopRefreshDegraded: route.route.openshift.io \"console\" not found\nOAuthClientsControllerDegraded: secret \"console-oauth-config\" not found"

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ConfigMapCreated

Created ConfigMap/monitoring-shared-config -n openshift-config-managed because it was missing

openshift-monitoring

cluster-monitoring-operator

cluster-monitoring-operator

ConfigMapCreated

Created ConfigMap/monitoring-shared-config -n openshift-config-managed because it was missing

openshift-console-operator

console-operator-status-controller-statussyncer_console

console-operator

OperatorStatusChanged

Status for clusteroperator/console changed: Degraded message changed from "SyncLoopRefreshDegraded: route.route.openshift.io \"console\" not found\nOAuthClientsControllerDegraded: secret \"console-oauth-config\" not found" to "SyncLoopRefreshDegraded: route.route.openshift.io \"console\" not found"

openshift-console-operator

console-operator-console-operator-consoleoperator

console-operator

DeploymentCreated

Created Deployment.apps/console -n openshift-console because it was missing

openshift-console-operator

console-operator-console-operator-consoleoperator

console-operator

ConfigMapCreated

Created ConfigMap/console-config -n openshift-console because it was missing

openshift-console

replicaset-controller

console-ffb45cdb5

SuccessfulCreate

Created pod: console-ffb45cdb5-7r4lm

openshift-console

deployment-controller

console

ScalingReplicaSet

Scaled up replica set console-ffb45cdb5 to 1

openshift-console-operator

console-operator-console-operator-consoleoperator

console-operator

ConfigMapCreated

Created ConfigMap/console-public -n openshift-config-managed because it was missing

openshift-authentication-operator

cluster-authentication-operator-config-observer-configobserver

authentication-operator

ObservedConfigChanged

Writing updated section ("oauthServer") of observed config: "\u00a0\u00a0map[string]any{\n\u00a0\u00a0\t\"corsAllowedOrigins\": []any{string(`//127\\.0\\.0\\.1(:|$)`), string(\"//localhost(:|$)\")},\n\u00a0\u00a0\t\"oauthConfig\": map[string]any{\n-\u00a0\t\t\"assetPublicURL\": string(\"\"),\n+\u00a0\t\t\"assetPublicURL\": string(\"https://console-openshift-console.apps.sno.openstack.lab\"),\n\u00a0\u00a0\t\t\"loginURL\": string(\"https://api.sno.openstack.lab:6443\"),\n\u00a0\u00a0\t\t\"templates\": map[string]any{\"error\": string(\"/var/config/system/secrets/v4-0-config-system-ocp-branding-templ\"...), \"login\": string(\"/var/config/system/secrets/v4-0-config-system-ocp-branding-templ\"...), \"providerSelection\": string(\"/var/config/system/secrets/v4-0-config-system-ocp-branding-templ\"...)},\n\u00a0\u00a0\t\t\"tokenConfig\": map[string]any{\"accessTokenMaxAgeSeconds\": float64(86400), \"authorizeTokenMaxAgeSeconds\": float64(300)},\n\u00a0\u00a0\t},\n\u00a0\u00a0\t\"serverArguments\": map[string]any{\"audit-log-format\": []any{string(\"json\")}, \"audit-log-maxbackup\": []any{string(\"10\")}, \"audit-log-maxsize\": []any{string(\"100\")}, \"audit-log-path\": []any{string(\"/var/log/oauth-server/audit.log\")}, ...},\n\u00a0\u00a0\t\"servingInfo\": map[string]any{\"cipherSuites\": []any{string(\"TLS_AES_128_GCM_SHA256\"), string(\"TLS_AES_256_GCM_SHA384\"), string(\"TLS_CHACHA20_POLY1305_SHA256\"), string(\"TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256\"), ...}, \"minTLSVersion\": string(\"VersionTLS12\"), \"namedCertificates\": []any{map[string]any{\"certFile\": string(\"/var/config/system/secrets/v4-0-config-system-router-certs/apps.\"...), \"keyFile\": string(\"/var/config/system/secrets/v4-0-config-system-router-certs/apps.\"...), \"names\": []any{string(\"*.apps.sno.openstack.lab\")}}}},\n\u00a0\u00a0\t\"volumesToMount\": map[string]any{\"identityProviders\": string(\"{}\")},\n\u00a0\u00a0}\n"

openshift-authentication-operator

cluster-authentication-operator-config-observer-configobserver

authentication-operator

ObserveConsoleURL

assetPublicURL changed from to https://console-openshift-console.apps.sno.openstack.lab

openshift-console-operator

console-operator-status-controller-statussyncer_console

console-operator

OperatorStatusChanged

Status for clusteroperator/console changed: Upgradeable changed from Unknown to True ("All is well")

openshift-authentication-operator

cluster-authentication-operator-payload-config-controller-openshift-authentication-payloadconfig

authentication-operator

ConfigMapUpdated

Updated ConfigMap/v4-0-config-system-cliconfig -n openshift-authentication: cause by changes in data.v4-0-config-system-cliconfig

openshift-console-operator

console-operator-status-controller-statussyncer_console

console-operator

OperatorStatusChanged

Status for clusteroperator/console changed: Degraded message changed from "SyncLoopRefreshDegraded: route.route.openshift.io \"console\" not found" to "SyncLoopRefreshDegraded: route.route.openshift.io \"console\" not found\nRouteHealthDegraded: route not yet available, https://console-openshift-console.apps.sno.openstack.lab returns '503 Service Unavailable'",Available changed from Unknown to False ("RouteHealthAvailable: route not yet available, https://console-openshift-console.apps.sno.openstack.lab returns '503 Service Unavailable'")

openshift-console

deployment-controller

console

ScalingReplicaSet

Scaled up replica set console-58857c5dc9 to 1

openshift-console

replicaset-controller

console-58857c5dc9

SuccessfulCreate

Created pod: console-58857c5dc9-9wf6k
(x3)

openshift-console-operator

console-operator-console-downloads-deployment-controller-consoledownloadsdeploymentsynccontroller

console-operator

DeploymentUpdated

Updated Deployment.apps/downloads -n openshift-console because it changed

openshift-kube-controller-manager-operator

kube-controller-manager-operator

kube-controller-manager-operator-lock

LeaderElection

kube-controller-manager-operator-ff989d6cc-qcpb4_6573d4cd-f713-43df-a45e-214415bb64a2 became leader

openshift-kube-controller-manager-operator

kube-controller-manager-operator

kube-controller-manager-operator

FeatureGatesInitialized

FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{"AWSEFSDriverVolumeMetrics", "AdminNetworkPolicy", "AlibabaPlatform", "AzureWorkloadIdentity", "BareMetalLoadBalancer", "BuildCSIVolumes", "ChunkSizeMiB", "CloudDualStackNodeIPs", "DisableKubeletCloudCredentialProviders", "GCPLabelsTags", "HardwareSpeed", "IngressControllerLBSubnetsAWS", "KMSv1", "ManagedBootImages", "ManagedBootImagesAWS", "MultiArchInstallAWS", "MultiArchInstallGCP", "NetworkDiagnosticsConfig", "NetworkLiveMigration", "NetworkSegmentation", "NewOLM", "NodeDisruptionPolicy", "OnClusterBuild", "PersistentIPsForVirtualization", "PrivateHostedZoneAWS", "SetEIPForNLBIngressController", "StreamingCollectionEncodingToJSON", "StreamingCollectionEncodingToProtobuf", "VSphereControlPlaneMachineSet", "VSphereDriverConfiguration", "VSphereMultiVCenters", "VSphereStaticIPs", "ValidatingAdmissionPolicy"}, Disabled:[]v1.FeatureGateName{"AWSClusterHostedDNS", "AdditionalRoutingCapabilities", "AutomatedEtcdBackup", "BootcNodeManagement", "CSIDriverSharedResource", "ClusterAPIInstall", "ClusterAPIInstallIBMCloud", "ClusterMonitoringConfig", "ConsolePluginContentSecurityPolicy", "DNSNameResolver", "DynamicResourceAllocation", "EtcdBackendQuota", "EventedPLEG", "Example", "ExternalOIDC", "ExternalOIDCWithUIDAndExtraClaimMappings", "GCPClusterHostedDNS", "GatewayAPI", "ImageStreamImportMode", "IngressControllerDynamicConfigurationManager", "InsightsConfig", "InsightsConfigAPI", "InsightsOnDemandDataGather", "InsightsRuntimeExtractor", "MachineAPIMigration", "MachineAPIOperatorDisableMachineHealthCheckController", "MachineAPIProviderOpenStack", "MachineConfigNodes", "MaxUnavailableStatefulSet", "MetricsCollectionProfiles", "MinimumKubeletVersion", "MixedCPUsAllocation", "MultiArchInstallAzure", "NodeSwap", "NutanixMultiSubnets", "OVNObservability", "OpenShiftPodSecurityAdmission", "PinnedImages", "PlatformOperators", "ProcMountType", "RouteAdvertisements", "RouteExternalCertificate", "ServiceAccountTokenNodeBinding", "SignatureStores", "SigstoreImageVerification", "TranslateStreamCloseWebsocketRequests", "UpgradeStatus", "UserNamespacesPodSecurityStandards", "UserNamespacesSupport", "VSphereMultiNetworks", "VolumeAttributesClass", "VolumeGroupSnapshot"}}

openshift-kube-controller-manager-operator

kube-controller-manager-operator-revisioncontroller

kube-controller-manager-operator

StartingNewRevision

new revision 3 triggered by "required secret/localhost-recovery-client-token has changed"

openshift-kube-controller-manager-operator

kube-controller-manager-operator-status-controller-statussyncer_kube-controller-manager

kube-controller-manager-operator

OperatorVersionChanged

clusteroperator/kube-controller-manager version "operator" changed from "" to "4.18.35"

openshift-kube-controller-manager-operator

kube-controller-manager-operator-status-controller-statussyncer_kube-controller-manager

kube-controller-manager-operator

OperatorVersionChanged

clusteroperator/kube-controller-manager version "kube-controller-manager" changed from "" to "1.31.14"

openshift-kube-controller-manager-operator

kube-controller-manager-operator-status-controller-statussyncer_kube-controller-manager

kube-controller-manager-operator

OperatorStatusChanged

Status for clusteroperator/kube-controller-manager changed: status.versions changed from [{"raw-internal" "4.18.35"}] to [{"raw-internal" "4.18.35"} {"operator" "4.18.35"} {"kube-controller-manager" "1.31.14"}]

openshift-kube-controller-manager-operator

kube-controller-manager-operator-satokensignercontroller

kube-controller-manager-operator

SATokenSignerControllerOK

found expected kube-apiserver endpoints

openshift-console-operator

console-operator-status-controller-statussyncer_console

console-operator

OperatorStatusChanged

Status for clusteroperator/console changed: Degraded message changed from "SyncLoopRefreshDegraded: route.route.openshift.io \"console\" not found\nRouteHealthDegraded: route not yet available, https://console-openshift-console.apps.sno.openstack.lab returns '503 Service Unavailable'" to "RouteHealthDegraded: route not yet available, https://console-openshift-console.apps.sno.openstack.lab returns '503 Service Unavailable'",Progressing changed from False to True ("SyncLoopRefreshProgressing: changes made during sync updates, additional sync expected"),Available message changed from "RouteHealthAvailable: route not yet available, https://console-openshift-console.apps.sno.openstack.lab returns '503 Service Unavailable'" to "DeploymentAvailable: 0 replicas available for console deployment\nRouteHealthAvailable: route not yet available, https://console-openshift-console.apps.sno.openstack.lab returns '503 Service Unavailable'"

openshift-console-operator

console-operator-status-controller-statussyncer_console

console-operator

OperatorStatusChanged

Status for clusteroperator/console changed: Progressing message changed from "SyncLoopRefreshProgressing: changes made during sync updates, additional sync expected" to "SyncLoopRefreshProgressing: working toward version 4.18.35, 0 replicas available"

openshift-kube-controller-manager-operator

kube-controller-manager-operator-revisioncontroller

kube-controller-manager-operator

ConfigMapCreated

Created ConfigMap/kube-controller-manager-pod-3 -n openshift-kube-controller-manager because it was missing

openshift-kube-controller-manager-operator

kube-controller-manager-operator-satokensignercontroller

kube-controller-manager-operator

SecretCreated

Created Secret/next-service-account-private-key -n openshift-kube-controller-manager-operator because it was missing

openshift-kube-controller-manager-operator

kube-controller-manager-operator-revisioncontroller

kube-controller-manager-operator

ConfigMapCreated

Created ConfigMap/config-3 -n openshift-kube-controller-manager because it was missing

openshift-kube-controller-manager-operator

kube-controller-manager-operator-status-controller-statussyncer_kube-controller-manager

kube-controller-manager-operator

OperatorStatusChanged

Status for clusteroperator/kube-controller-manager changed: Progressing changed from True to False ("NodeInstallerProgressing: 1 node is at revision 2"),Available changed from False to True ("StaticPodsAvailable: 1 nodes are active; 1 node is at revision 2")

openshift-kube-controller-manager-operator

kube-controller-manager-operator-installer-controller

kube-controller-manager-operator

NodeCurrentRevisionChanged

Updated node "master-0" from revision 0 to 2 because static pod is ready

openshift-kube-controller-manager-operator

kube-controller-manager-operator-satokensignercontroller

kube-controller-manager-operator

ConfigMapUpdated

Updated ConfigMap/sa-token-signing-certs -n openshift-config-managed: cause by changes in data.service-account-002.pub

openshift-kube-controller-manager-operator

kube-controller-manager-operator-revisioncontroller

kube-controller-manager-operator

ConfigMapCreated

Created ConfigMap/cluster-policy-controller-config-3 -n openshift-kube-controller-manager because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/kube-apiserver-pod-5 -n openshift-kube-apiserver because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

kube-apiserver-operator

StartingNewRevision

new revision 5 triggered by "required configmap/sa-token-signing-certs has changed"

openshift-kube-apiserver-operator

kube-apiserver-operator-resource-sync-controller-resourcesynccontroller

kube-apiserver-operator

ConfigMapUpdated

Updated ConfigMap/sa-token-signing-certs -n openshift-kube-apiserver: cause by changes in data.service-account-002.pub

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/config-5 -n openshift-kube-apiserver because it was missing

openshift-kube-controller-manager-operator

kube-controller-manager-operator-revisioncontroller

kube-controller-manager-operator

ConfigMapCreated

Created ConfigMap/controller-manager-kubeconfig-3 -n openshift-kube-controller-manager because it was missing

openshift-kube-controller-manager-operator

kube-controller-manager-operator-revisioncontroller

kube-controller-manager-operator

ConfigMapCreated

Created ConfigMap/kube-controller-cert-syncer-kubeconfig-3 -n openshift-kube-controller-manager because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/kube-apiserver-cert-syncer-kubeconfig-5 -n openshift-kube-apiserver because it was missing

openshift-monitoring

kubelet

alertmanager-main-0

Killing

Stopping container alertmanager

openshift-monitoring

kubelet

alertmanager-main-0

Killing

Stopping container alertmanager

openshift-monitoring

statefulset-controller

alertmanager-main

SuccessfulDelete

delete Pod alertmanager-main-0 in StatefulSet alertmanager-main successful

openshift-kube-controller-manager-operator

kube-controller-manager-operator-revisioncontroller

kube-controller-manager-operator

ConfigMapCreated

Created ConfigMap/serviceaccount-ca-3 -n openshift-kube-controller-manager because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/oauth-metadata-5 -n openshift-kube-apiserver because it was missing

openshift-monitoring

statefulset-controller

alertmanager-main

SuccessfulDelete

delete Pod alertmanager-main-0 in StatefulSet alertmanager-main successful

openshift-kube-controller-manager-operator

kube-controller-manager-operator-revisioncontroller

kube-controller-manager-operator

ConfigMapCreated

Created ConfigMap/recycler-config-3 -n openshift-kube-controller-manager because it was missing

openshift-kube-controller-manager-operator

kube-controller-manager-operator-revisioncontroller

kube-controller-manager-operator

ConfigMapCreated

Created ConfigMap/service-ca-3 -n openshift-kube-controller-manager because it was missing

openshift-monitoring

statefulset-controller

alertmanager-main

SuccessfulCreate

create Pod alertmanager-main-0 in StatefulSet alertmanager-main successful

openshift-monitoring

statefulset-controller

alertmanager-main

SuccessfulCreate

create Pod alertmanager-main-0 in StatefulSet alertmanager-main successful

openshift-kube-controller-manager-operator

kube-controller-manager-operator-revisioncontroller

kube-controller-manager-operator

SecretCreated

Created Secret/service-account-private-key-3 -n openshift-kube-controller-manager because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/bound-sa-token-signing-certs-5 -n openshift-kube-apiserver because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/etcd-serving-ca-5 -n openshift-kube-apiserver because it was missing

openshift-kube-controller-manager-operator

kube-controller-manager-operator-revisioncontroller

kube-controller-manager-operator

SecretCreated

Created Secret/serving-cert-3 -n openshift-kube-controller-manager because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/kube-apiserver-server-ca-5 -n openshift-kube-apiserver because it was missing

openshift-kube-controller-manager-operator

kube-controller-manager-operator-revisioncontroller

kube-controller-manager-operator

SecretCreated

Created Secret/localhost-recovery-client-token-3 -n openshift-kube-controller-manager because it was missing

openshift-monitoring

kubelet

prometheus-k8s-0

Killing

Stopping container prometheus

openshift-monitoring

statefulset-controller

prometheus-k8s

SuccessfulDelete

delete Pod prometheus-k8s-0 in StatefulSet prometheus-k8s successful

openshift-monitoring

statefulset-controller

prometheus-k8s

SuccessfulDelete

delete Pod prometheus-k8s-0 in StatefulSet prometheus-k8s successful

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/kubelet-serving-ca-5 -n openshift-kube-apiserver because it was missing

openshift-monitoring

kubelet

prometheus-k8s-0

Killing

Stopping container prometheus

openshift-kube-controller-manager-operator

kube-controller-manager-operator-revisioncontroller

kube-controller-manager-operator

RevisionTriggered

new revision 3 triggered by "required secret/localhost-recovery-client-token has changed"

openshift-monitoring

statefulset-controller

prometheus-k8s

SuccessfulCreate

create Pod prometheus-k8s-0 in StatefulSet prometheus-k8s successful

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/sa-token-signing-certs-5 -n openshift-kube-apiserver because it was missing

openshift-monitoring

statefulset-controller

prometheus-k8s

SuccessfulCreate

create Pod prometheus-k8s-0 in StatefulSet prometheus-k8s successful

openshift-kube-controller-manager-operator

kube-controller-manager-operator-status-controller-statussyncer_kube-controller-manager

kube-controller-manager-operator

OperatorStatusChanged

Status for clusteroperator/kube-controller-manager changed: Progressing changed from False to True ("NodeInstallerProgressing: 1 node is at revision 2; 0 nodes have achieved new revision 3"),Available message changed from "StaticPodsAvailable: 1 nodes are active; 1 node is at revision 2" to "StaticPodsAvailable: 1 nodes are active; 1 node is at revision 2; 0 nodes have achieved new revision 3"

openshift-kube-controller-manager-operator

kube-controller-manager-operator-installer-controller

kube-controller-manager-operator

NodeTargetRevisionChanged

Updating node "master-0" from revision 2 to 3 because node master-0 with revision 2 is the oldest

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

kube-apiserver-operator

ConfigMapCreated

Created ConfigMap/kube-apiserver-audit-policies-5 -n openshift-kube-apiserver because it was missing

openshift-etcd-operator

openshift-cluster-etcd-operator-installer-controller

etcd-operator

NodeCurrentRevisionChanged

Updated node "master-0" from revision 1 to 2 because static pod is ready

openshift-kube-controller-manager-operator

kube-controller-manager-operator-installer-controller

kube-controller-manager-operator

PodCreated

Created Pod/installer-3-master-0 -n openshift-kube-controller-manager because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

kube-apiserver-operator

SecretCreated

Created Secret/etcd-client-5 -n openshift-kube-apiserver because it was missing

openshift-kube-controller-manager

kubelet

installer-3-master-0

Created

Created container: installer

openshift-kube-controller-manager

multus

installer-3-master-0

AddedInterface

Add eth0 [10.128.0.98/23] from ovn-kubernetes

openshift-kube-controller-manager

kubelet

installer-3-master-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c032f87ae61d6f757ff3ce52620a70a43516591987731f25da77aba152f17458" already present on machine

openshift-kube-controller-manager

kubelet

installer-3-master-0

Started

Started container installer

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

kube-apiserver-operator

SecretCreated

Created Secret/localhost-recovery-serving-certkey-5 -n openshift-kube-apiserver because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

kube-apiserver-operator

SecretCreated

Created Secret/localhost-recovery-client-token-5 -n openshift-kube-apiserver because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

kube-apiserver-operator

SecretCreated

Created Secret/webhook-authenticator-5 -n openshift-kube-apiserver because it was missing

openshift-kube-apiserver-operator

kube-apiserver-operator-revisioncontroller

kube-apiserver-operator

RevisionTriggered

new revision 5 triggered by "required configmap/sa-token-signing-certs has changed"

openshift-kube-apiserver-operator

kube-apiserver-operator-status-controller-statussyncer_kube-apiserver

kube-apiserver-operator

OperatorStatusChanged

Status for clusteroperator/kube-apiserver changed: Progressing changed from False to True ("NodeInstallerProgressing: 1 node is at revision 4; 0 nodes have achieved new revision 5"),Available message changed from "StaticPodsAvailable: 1 nodes are active; 1 node is at revision 4" to "StaticPodsAvailable: 1 nodes are active; 1 node is at revision 4; 0 nodes have achieved new revision 5"

openshift-kube-apiserver-operator

kube-apiserver-operator-installer-controller

kube-apiserver-operator

NodeTargetRevisionChanged

Updating node "master-0" from revision 4 to 5 because node master-0 with revision 4 is the oldest

openshift-network-node-identity

master-0_185a04fc-5098-48d7-89e2-b8e5f598619b

ovnkube-identity

LeaderElection

master-0_185a04fc-5098-48d7-89e2-b8e5f598619b became leader

openshift-kube-apiserver

kubelet

installer-5-master-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c5ce3d1134d6500e2b8528516c1889d7bbc6259aba4981c6983395b0e9eeff65" already present on machine

openshift-kube-apiserver

multus

installer-5-master-0

AddedInterface

Add eth0 [10.128.0.99/23] from ovn-kubernetes

openshift-kube-apiserver-operator

kube-apiserver-operator-installer-controller

kube-apiserver-operator

PodCreated

Created Pod/installer-5-master-0 -n openshift-kube-apiserver because it was missing

openshift-kube-apiserver

kubelet

installer-5-master-0

Started

Started container installer

openshift-kube-apiserver

kubelet

installer-5-master-0

Created

Created container: installer

openshift-kube-controller-manager

static-pod-installer

installer-3-master-0

StaticPodInstallerCompleted

Successfully installed revision 3

openshift-authentication-operator

cluster-authentication-operator-oauthserver-workloadworkloadcontroller

authentication-operator

DeploymentUpdated

Updated Deployment.apps/oauth-openshift -n openshift-authentication because it changed

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Progressing changed from False to True ("OAuthServerDeploymentProgressing: deployment/oauth-openshift.openshift-authentication: observed generation is 3, desired generation is 4.")

openshift-kube-controller-manager

kubelet

kube-controller-manager-master-0

Started

Started container cluster-policy-controller

openshift-kube-controller-manager

kubelet

kube-controller-manager-master-0

Created

Created container: kube-controller-manager

openshift-kube-controller-manager

kubelet

kube-controller-manager-master-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1fbbcb390de2563a0177b92fba1b5a65777366e2dc80e2808b61d87c41b47a2d" already present on machine

openshift-kube-controller-manager

kubelet

kube-controller-manager-master-0

Started

Started container kube-controller-manager

openshift-kube-controller-manager

kubelet

kube-controller-manager-master-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c032f87ae61d6f757ff3ce52620a70a43516591987731f25da77aba152f17458" already present on machine

openshift-kube-controller-manager

kubelet

kube-controller-manager-master-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b23c544d3894e5b31f66a18c554f03b0d29f92c2000c46b57b1c96da7ec25db9" already present on machine

openshift-kube-controller-manager

kubelet

kube-controller-manager-master-0

Created

Created container: cluster-policy-controller

openshift-kube-controller-manager

kubelet

kube-controller-manager-master-0

Created

Created container: kube-controller-manager-recovery-controller

openshift-kube-controller-manager

kubelet

kube-controller-manager-master-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c032f87ae61d6f757ff3ce52620a70a43516591987731f25da77aba152f17458" already present on machine

openshift-kube-controller-manager

cluster-policy-controller

kube-controller-manager-master-0

ControlPlaneTopology

unable to get control plane topology, using HA cluster values for leader election: infrastructures.config.openshift.io "cluster" is forbidden: User "system:kube-controller-manager" cannot get resource "infrastructures" in API group "config.openshift.io" at the cluster scope

openshift-kube-controller-manager

kubelet

kube-controller-manager-master-0

Started

Started container kube-controller-manager-cert-syncer

openshift-kube-controller-manager

kubelet

kube-controller-manager-master-0

Started

Started container kube-controller-manager-recovery-controller

openshift-kube-controller-manager

kubelet

kube-controller-manager-master-0

Created

Created container: kube-controller-manager-cert-syncer

openshift-kube-controller-manager

cert-recovery-controller

cert-recovery-controller-lock

LeaderElection

master-0_2b495287-c060-4510-b2d3-4dd63483d710 became leader

openshift-kube-apiserver

kubelet

kube-apiserver-master-0

Killing

Stopping container kube-apiserver-cert-syncer

openshift-kube-apiserver

kubelet

kube-apiserver-startup-monitor-master-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c5ce3d1134d6500e2b8528516c1889d7bbc6259aba4981c6983395b0e9eeff65" already present on machine

openshift-kube-apiserver

apiserver

kube-apiserver-master-0

InFlightRequestsDrained

All non long-running request(s) in-flight have drained

openshift-kube-apiserver

kubelet

kube-apiserver-master-0

Killing

Stopping container kube-apiserver

openshift-kube-apiserver

apiserver

kube-apiserver-master-0

TerminationPreShutdownHooksFinished

All pre-shutdown hooks have been finished

openshift-kube-apiserver

apiserver

kube-apiserver-master-0

AfterShutdownDelayDuration

The minimal shutdown duration of 0s finished

openshift-kube-apiserver

apiserver

kube-apiserver-master-0

ShutdownInitiated

Received signal to terminate, becoming unready, but keeping serving

openshift-kube-apiserver

kubelet

kube-apiserver-master-0

Killing

Stopping container kube-apiserver-cert-regeneration-controller

openshift-kube-apiserver

kubelet

kube-apiserver-master-0

Killing

Stopping container kube-apiserver-insecure-readyz

openshift-kube-apiserver

kubelet

kube-apiserver-master-0

Killing

Stopping container kube-apiserver-check-endpoints

openshift-kube-apiserver

apiserver

kube-apiserver-master-0

HTTPServerStoppedListening

HTTP Server has stopped listening

openshift-kube-apiserver

kubelet

kube-apiserver-startup-monitor-master-0

Created

Created container: startup-monitor

openshift-kube-apiserver

kubelet

kube-apiserver-startup-monitor-master-0

Started

Started container startup-monitor

openshift-kube-apiserver

apiserver

kube-apiserver-master-0

TerminationGracefulTerminationFinished

All pending requests processed

openshift-kube-apiserver

kubelet

kube-apiserver-master-0

Created

Created container: setup

openshift-kube-apiserver

kubelet

kube-apiserver-master-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b23c544d3894e5b31f66a18c554f03b0d29f92c2000c46b57b1c96da7ec25db9" already present on machine

openshift-kube-apiserver

kubelet

kube-apiserver-master-0

Started

Started container setup

openshift-kube-apiserver

kubelet

kube-apiserver-master-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b23c544d3894e5b31f66a18c554f03b0d29f92c2000c46b57b1c96da7ec25db9" already present on machine

openshift-kube-apiserver

kubelet

kube-apiserver-master-0

Started

Started container kube-apiserver

openshift-kube-apiserver

kubelet

kube-apiserver-master-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c5ce3d1134d6500e2b8528516c1889d7bbc6259aba4981c6983395b0e9eeff65" already present on machine

openshift-kube-apiserver

kubelet

kube-apiserver-master-0

Started

Started container kube-apiserver-cert-regeneration-controller

openshift-kube-apiserver

kubelet

kube-apiserver-master-0

Created

Created container: kube-apiserver-cert-regeneration-controller

openshift-kube-apiserver

kubelet

kube-apiserver-master-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c5ce3d1134d6500e2b8528516c1889d7bbc6259aba4981c6983395b0e9eeff65" already present on machine

openshift-kube-apiserver

kubelet

kube-apiserver-master-0

Started

Started container kube-apiserver-cert-syncer

openshift-kube-apiserver

kubelet

kube-apiserver-master-0

Created

Created container: kube-apiserver-cert-syncer

openshift-kube-apiserver

kubelet

kube-apiserver-master-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c5ce3d1134d6500e2b8528516c1889d7bbc6259aba4981c6983395b0e9eeff65" already present on machine

openshift-kube-apiserver

kubelet

kube-apiserver-master-0

Created

Created container: kube-apiserver

openshift-kube-apiserver

kubelet

kube-apiserver-master-0

Created

Created container: kube-apiserver-check-endpoints

openshift-kube-apiserver

kubelet

kube-apiserver-master-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c5ce3d1134d6500e2b8528516c1889d7bbc6259aba4981c6983395b0e9eeff65" already present on machine

openshift-kube-apiserver

kubelet

kube-apiserver-master-0

Created

Created container: kube-apiserver-insecure-readyz

openshift-kube-apiserver

kubelet

kube-apiserver-master-0

Started

Started container kube-apiserver-check-endpoints

openshift-kube-apiserver

kubelet

kube-apiserver-master-0

Started

Started container kube-apiserver-insecure-readyz

openshift-kube-apiserver

apiserver

kube-apiserver-master-0

KubeAPIReadyz

readyz=true

kube-system

kube-controller-manager

kube-controller-manager

LeaderElection

master-0_ac368a31-d6e3-4879-9353-2110d182a25b became leader

openshift-kube-apiserver

cert-regeneration-controller

cert-regeneration-controller-lock

LeaderElection

master-0_e63406b8-776e-4162-987e-fcce1aac4d18 became leader
(x5)

openshift-etcd-operator

openshift-cluster-etcd-operator-script-controller-scriptcontroller

etcd-operator

ScriptControllerErrorUpdatingStatus

Put "https://172.30.0.1:443/apis/operator.openshift.io/v1/etcds/cluster/status": dial tcp 172.30.0.1:443: connect: connection refused

openshift-authentication

replicaset-controller

oauth-openshift-8684f7dbf

SuccessfulCreate

Created pod: oauth-openshift-8684f7dbf-tqtg9

openshift-authentication

deployment-controller

oauth-openshift

ScalingReplicaSet

Scaled down replica set oauth-openshift-dcb9594d9 to 0 from 1

openshift-authentication

deployment-controller

oauth-openshift

ScalingReplicaSet

Scaled up replica set oauth-openshift-8684f7dbf to 1 from 0

openshift-authentication

replicaset-controller

oauth-openshift-dcb9594d9

SuccessfulDelete

Deleted pod: oauth-openshift-dcb9594d9-wlht7

default

node-controller

master-0

RegisteredNode

Node master-0 event: Registered Node master-0 in Controller

openshift-cloud-controller-manager-operator

master-0_a6d4365e-d459-4ca1-a1b8-0931312e163c

cluster-cloud-controller-manager-leader

LeaderElection

master-0_a6d4365e-d459-4ca1-a1b8-0931312e163c became leader
(x5)

openshift-etcd-operator

openshift-cluster-etcd-operator-etcd-endpoints-controller-etcdendpointscontroller

etcd-operator

EtcdEndpointsErrorUpdatingStatus

Put "https://172.30.0.1:443/apis/operator.openshift.io/v1/etcds/cluster/status": dial tcp 172.30.0.1:443: connect: connection refused

openshift-authentication

kubelet

oauth-openshift-dcb9594d9-wlht7

Killing

Stopping container oauth-openshift

openshift-operator-controller

operator-controller-controller-manager-57777556ff-nk2rf_a58a9f38-17e8-4170-80cd-628af4bf6798

9c4404e7.operatorframework.io

LeaderElection

operator-controller-controller-manager-57777556ff-nk2rf_a58a9f38-17e8-4170-80cd-628af4bf6798 became leader

openshift-cluster-machine-approver

master-0_c89779ef-148b-4e25-8a42-8c5cf7b2eb58

cluster-machine-approver-leader

LeaderElection

master-0_c89779ef-148b-4e25-8a42-8c5cf7b2eb58 became leader

openshift-kube-apiserver

kubelet

kube-apiserver-startup-monitor-master-0

Killing

Stopping container startup-monitor

openshift-machine-api

cluster-baremetal-operator-6f69995874-dv6cd_bab88e76-81fd-4f40-9f4e-966136b03e80

cluster-baremetal-operator

LeaderElection

cluster-baremetal-operator-6f69995874-dv6cd_bab88e76-81fd-4f40-9f4e-966136b03e80 became leader

openshift-machine-api

cluster-baremetal-operator-6f69995874-dv6cd_bab88e76-81fd-4f40-9f4e-966136b03e80

cluster-baremetal-operator

LeaderElection

cluster-baremetal-operator-6f69995874-dv6cd_bab88e76-81fd-4f40-9f4e-966136b03e80 became leader

openshift-catalogd

catalogd-controller-manager-6864dc98f7-74mgr_9f1b3da1-8af0-41a7-808b-9da85e395556

catalogd-operator-lock

LeaderElection

catalogd-controller-manager-6864dc98f7-74mgr_9f1b3da1-8af0-41a7-808b-9da85e395556 became leader

openshift-catalogd

catalogd-controller-manager-6864dc98f7-74mgr_9f1b3da1-8af0-41a7-808b-9da85e395556

catalogd-operator-lock

LeaderElection

catalogd-controller-manager-6864dc98f7-74mgr_9f1b3da1-8af0-41a7-808b-9da85e395556 became leader

openshift-cloud-controller-manager-operator

master-0_0fd914f9-106c-49dd-ab32-1465a2b05927

cluster-cloud-config-sync-leader

LeaderElection

master-0_0fd914f9-106c-49dd-ab32-1465a2b05927 became leader

openshift-machine-api

control-plane-machine-set-operator-6f97756bc8-7t5qv_664c7774-3200-4a94-b155-f30085fa2cfa

control-plane-machine-set-leader

LeaderElection

control-plane-machine-set-operator-6f97756bc8-7t5qv_664c7774-3200-4a94-b155-f30085fa2cfa became leader

openshift-machine-api

control-plane-machine-set-operator-6f97756bc8-7t5qv_664c7774-3200-4a94-b155-f30085fa2cfa

control-plane-machine-set-leader

LeaderElection

control-plane-machine-set-operator-6f97756bc8-7t5qv_664c7774-3200-4a94-b155-f30085fa2cfa became leader

openshift-kube-scheduler

cert-recovery-controller

cert-recovery-controller-lock

LeaderElection

master-0_69a3f525-66e6-477c-8908-4e4b5d367ef3 became leader

openshift-marketplace

kubelet

marketplace-operator-89ccd998f-mvn4t

ProbeError

Liveness probe error: Get "http://10.128.0.20:8080/healthz": dial tcp 10.128.0.20:8080: connect: connection refused body:

openshift-marketplace

kubelet

marketplace-operator-89ccd998f-mvn4t

Unhealthy

Liveness probe failed: Get "http://10.128.0.20:8080/healthz": dial tcp 10.128.0.20:8080: connect: connection refused
(x2)

openshift-marketplace

kubelet

marketplace-operator-89ccd998f-mvn4t

BackOff

Back-off restarting failed container marketplace-operator in pod marketplace-operator-89ccd998f-mvn4t_openshift-marketplace(acb704a9-6c8d-4378-ae93-e7095b1fce85)
(x2)

openshift-marketplace

kubelet

marketplace-operator-89ccd998f-mvn4t

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:632e80bba5077068ecca05fddb95aedebad4493af6f36152c01c6ae490975b62" already present on machine
(x2)

openshift-marketplace

kubelet

marketplace-operator-89ccd998f-mvn4t

Created

Created container: marketplace-operator
(x2)

openshift-marketplace

kubelet

marketplace-operator-89ccd998f-mvn4t

Started

Started container marketplace-operator

openshift-console-operator

console-operator-status-controller-statussyncer_console

console-operator

OperatorStatusChanged

Status for clusteroperator/console changed: Degraded changed from False to True ("RouteHealthDegraded: route not yet available, https://console-openshift-console.apps.sno.openstack.lab returns '503 Service Unavailable'")

openshift-kube-controller-manager-operator

kube-controller-manager-operator-status-controller-statussyncer_kube-controller-manager

kube-controller-manager-operator

OperatorStatusChanged

Status for clusteroperator/kube-controller-manager changed: Degraded message changed from "NodeControllerDegraded: All master nodes are ready" to "NodeControllerDegraded: All master nodes are ready\nKubeControllerManagerStaticResourcesDegraded: \"assets/kube-controller-manager/ns.yaml\" (string): Get \"https://172.30.0.1:443/api/v1/namespaces/openshift-kube-controller-manager\": dial tcp 172.30.0.1:443: connect: connection refused\nKubeControllerManagerStaticResourcesDegraded: \"assets/kube-controller-manager/leader-election-rolebinding.yaml\" (string): Get \"https://172.30.0.1:443/apis/rbac.authorization.k8s.io/v1/namespaces/kube-system/rolebindings/system:openshift:leader-locking-kube-controller-manager\": dial tcp 172.30.0.1:443: connect: connection refused\nKubeControllerManagerStaticResourcesDegraded: \"assets/kube-controller-manager/leader-election-cluster-policy-controller-role.yaml\" (string): Get \"https://172.30.0.1:443/apis/rbac.authorization.k8s.io/v1/namespaces/openshift-kube-controller-manager/roles/system:openshift:leader-election-lock-cluster-policy-controller\": dial tcp 172.30.0.1:443: connect: connection refused\nKubeControllerManagerStaticResourcesDegraded: \"assets/kube-controller-manager/leader-election-cluster-policy-controller-rolebinding.yaml\" (string): Get \"https://172.30.0.1:443/apis/rbac.authorization.k8s.io/v1/namespaces/openshift-kube-controller-manager/rolebindings/system:openshift:leader-election-lock-cluster-policy-controller\": dial tcp 172.30.0.1:443: connect: connection refused\nKubeControllerManagerStaticResourcesDegraded: \"assets/kube-controller-manager/namespace-security-allocation-controller-clusterrole.yaml\" (string): Get \"https://172.30.0.1:443/apis/rbac.authorization.k8s.io/v1/clusterroles/system:openshift:controller:namespace-security-allocation-controller\": dial tcp 172.30.0.1:443: connect: connection refused\nKubeControllerManagerStaticResourcesDegraded: \"assets/kube-controller-manager/namespace-security-allocation-controller-clusterrolebinding.yaml\" (string): Get \"https://172.30.0.1:443/apis/rbac.authorization.k8s.io/v1/clusterrolebindings/system:openshift:controller:namespace-security-allocation-controller\": dial tcp 172.30.0.1:443: connect: connection refused\nKubeControllerManagerStaticResourcesDegraded: \"assets/kube-controller-manager/podsecurity-admission-label-syncer-controller-clusterrole.yaml\" (string): Get \"https://172.30.0.1:443/apis/rbac.authorization.k8s.io/v1/clusterroles/system:openshift:controller:podsecurity-admission-label-syncer-controller\": dial tcp 172.30.0.1:443: connect: connection refused\nKubeControllerManagerStaticResourcesDegraded: \"assets/kube-controller-manager/podsecurity-admission-label-syncer-controller-clusterrolebinding.yaml\" (string): Get \"https://172.30.0.1:443/apis/rbac.authorization.k8s.io/v1/clusterrolebindings/system:openshift:controller:podsecurity-admission-label-syncer-controller\": dial tcp 172.30.0.1:443: connect: connection refused\nKubeControllerManagerStaticResourcesDegraded: \"assets/kube-controller-manager/podsecurity-admission-label-privileged-namespaces-syncer-controller-clusterrole.yaml\" (string): Get \"https://172.30.0.1:443/apis/rbac.authorization.k8s.io/v1/clusterroles/system:openshift:controller:privileged-namespaces-psa-label-syncer\": dial tcp 172.30.0.1:443: connect: connection refused\nKubeControllerManagerStaticResourcesDegraded: \"assets/kube-controller-manager/podsecurity-admission-label-privileged-namespaces-syncer-controller-clusterrolebinding.yaml\" (string): Get \"https://172.30.0.1:443/apis/rbac.authorization.k8s.io/v1/clusterrolebindings/system:openshift:controller:privileged-namespaces-psa-label-syncer\": dial tcp 172.30.0.1:443: connect: connection refused\nKubeControllerManagerStaticResourcesDegraded: \"assets/kube-controller-manager/namespace-openshift-infra.yaml\" (string): Get \"https://172.30.0.1:443/api/v1/namespaces/openshift-infra\": dial tcp 172.30.0.1:443: connect: connection refused\nKubeControllerManagerStaticResourcesDegraded: \"assets/kube-controller-manager/svc.yaml\" (string): Get \"https://172.30.0.1:443/api/v1/namespaces/openshift-kube-controller-manager/services/kube-controller-manager\": dial tcp 172.30.0.1:443: connect: connection refused\nKubeControllerManagerStaticResourcesDegraded: \"assets/kube-controller-manager/sa.yaml\" (string): Get \"https://172.30.0.1:443/api/v1/namespaces/openshift-kube-controller-manager/serviceaccounts/kube-controller-manager-sa\": dial tcp 172.30.0.1:443: connect: connection refused\nKubeControllerManagerStaticResourcesDegraded: \"assets/kube-controller-manager/recycler-sa.yaml\" (string): Get \"https://172.30.0.1:443/api/v1/namespaces/openshift-infra/serviceaccounts/pv-recycler-controller\": dial tcp 172.30.0.1:443: connect: connection refused\nKubeControllerManagerStaticResourcesDegraded: \"assets/kube-controller-manager/localhost-recovery-client-crb.yaml\" (string): Get \"https://172.30.0.1:443/apis/rbac.authorization.k8s.io/v1/clusterrolebindings/system:openshift:operator:kube-controller-manager-recovery\": dial tcp 172.30.0.1:443: connect: connection refused\nKubeControllerManagerStaticResourcesDegraded: \"assets/kube-controller-manager/localhost-recovery-sa.yaml\" (string): Get \"https://172.30.0.1:443/api/v1/namespaces/openshift-kube-controller-manager/serviceaccounts/localhost-recovery-client\": dial tcp 172.30.0.1:443: connect: connection refused\nKubeControllerManagerStaticResourcesDegraded: \"assets/kube-controller-manager/csr_approver_clusterrole.yaml\" (string): Get \"https://172.30.0.1:443/apis/rbac.authorization.k8s.io/v1/clusterroles/system:openshift:controller:cluster-csr-approver-controller\": dial tcp 172.30.0.1:443: connect: connection refused\nKubeControllerManagerStaticResourcesDegraded: \"assets/kube-controller-manager/csr_approver_clusterrolebinding.yaml\" (string): Get \"https://172.30.0.1:443/apis/rbac.authorization.k8s.io/v1/clusterrolebindings/system:openshift:controller:cluster-csr-approver-controller\": dial tcp 172.30.0.1:443: connect: connection refused\nKubeControllerManagerStaticResourcesDegraded: \"assets/kube-controller-manager/vsphere/legacy-cloud-provider-sa.yaml\" (string): Delete \"https://172.30.0.1:443/api/v1/namespaces/kube-system/serviceaccounts/vsphere-legacy-cloud-provider\": dial tcp 172.30.0.1:443: connect: connection refused\nKubeControllerManagerStaticResourcesDegraded: "

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Degraded message changed from "All is well" to "OAuthServerServiceEndpointAccessibleControllerDegraded: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused\nIngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\nOAuthServerRouteEndpointAccessibleControllerDegraded: Get \"https://oauth-openshift.apps.sno.openstack.lab/healthz\": EOF\nWellKnownReadyControllerDegraded: failed to get API server IPs: unable to find kube api server endpointLister port: &v1.Endpoints{TypeMeta:v1.TypeMeta{Kind:\"\", APIVersion:\"\"}, ObjectMeta:v1.ObjectMeta{Name:\"kubernetes\", GenerateName:\"\", Namespace:\"default\", SelfLink:\"\", UID:\"0a8e1475-e015-4961-bc7c-f367a856aa53\", ResourceVersion:\"15724\", Generation:0, CreationTimestamp:time.Date(2026, time.March, 20, 8, 29, 8, 0, time.Local), DeletionTimestamp:<nil>, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{\"endpointslice.kubernetes.io/skip-mirror\":\"true\"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry{v1.ManagedFieldsEntry{Manager:\"kube-apiserver\", Operation:\"Update\", APIVersion:\"v1\", Time:time.Date(2026, time.March, 20, 8, 43, 6, 0, time.Local), FieldsType:\"FieldsV1\", FieldsV1:(*v1.FieldsV1)(0xc00155bbd8), Subresource:\"\"}}}, Subsets:[]v1.EndpointSubset(nil)} (check kube-apiserver that it deploys correctly)",Progressing message changed from "OAuthServerDeploymentProgressing: deployment/oauth-openshift.openshift-authentication: observed generation is 3, desired generation is 4." to "OAuthServerDeploymentProgressing: deployment/oauth-openshift.openshift-authentication: 1/1 pods have been updated to the latest generation and 0/1 pods are available",Available changed from True to False ("OAuthServerDeploymentAvailable: no oauth-openshift.openshift-authentication pods available on any node.\nOAuthServerRouteEndpointAccessibleControllerAvailable: Get \"https://oauth-openshift.apps.sno.openstack.lab/healthz\": EOF\nOAuthServerServiceEndpointAccessibleControllerAvailable: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused\nWellKnownAvailable: The well-known endpoint is not yet available: failed to get API server IPs: unable to find kube api server endpointLister port: &v1.Endpoints{TypeMeta:v1.TypeMeta{Kind:\"\", APIVersion:\"\"}, ObjectMeta:v1.ObjectMeta{Name:\"kubernetes\", GenerateName:\"\", Namespace:\"default\", SelfLink:\"\", UID:\"0a8e1475-e015-4961-bc7c-f367a856aa53\", ResourceVersion:\"15724\", Generation:0, CreationTimestamp:time.Date(2026, time.March, 20, 8, 29, 8, 0, time.Local), DeletionTimestamp:<nil>, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{\"endpointslice.kubernetes.io/skip-mirror\":\"true\"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry{v1.ManagedFieldsEntry{Manager:\"kube-apiserver\", Operation:\"Update\", APIVersion:\"v1\", Time:time.Date(2026, time.March, 20, 8, 43, 6, 0, time.Local), FieldsType:\"FieldsV1\", FieldsV1:(*v1.FieldsV1)(0xc00155bbd8), Subresource:\"\"}}}, Subsets:[]v1.EndpointSubset(nil)} (check kube-apiserver that it deploys correctly)")

openshift-kube-controller-manager-operator

kube-controller-manager-operator-status-controller-statussyncer_kube-controller-manager

kube-controller-manager-operator

OperatorStatusChanged

Status for clusteroperator/kube-controller-manager changed: Progressing changed from True to False ("NodeInstallerProgressing: 1 node is at revision 3"),Available message changed from "StaticPodsAvailable: 1 nodes are active; 1 node is at revision 2; 0 nodes have achieved new revision 3" to "StaticPodsAvailable: 1 nodes are active; 1 node is at revision 3"

openshift-kube-controller-manager-operator

kube-controller-manager-operator-installer-controller

kube-controller-manager-operator

NodeCurrentRevisionChanged

Updated node "master-0" from revision 2 to 3 because static pod is ready

openshift-kube-scheduler

default-scheduler

kube-scheduler

LeaderElection

master-0_3b91f3c2-25cf-493b-bc07-580d86edee40 became leader

openshift-authentication

multus

oauth-openshift-8684f7dbf-tqtg9

AddedInterface

Add eth0 [10.128.0.103/23] from ovn-kubernetes

openshift-authentication

kubelet

oauth-openshift-8684f7dbf-tqtg9

Started

Started container oauth-openshift

openshift-monitoring

kubelet

prometheus-k8s-0

Started

Started container init-config-reloader

openshift-monitoring

kubelet

prometheus-k8s-0

Created

Created container: init-config-reloader

openshift-monitoring

kubelet

prometheus-k8s-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ea36739b7e81007aec2e901639b356b275434362b254800d4309dd0aa665ca36" already present on machine

openshift-monitoring

multus

prometheus-k8s-0

AddedInterface

Add eth0 [10.128.0.102/23] from ovn-kubernetes

openshift-monitoring

kubelet

alertmanager-main-0

Started

Started container init-config-reloader

openshift-monitoring

kubelet

alertmanager-main-0

Created

Created container: init-config-reloader

openshift-monitoring

kubelet

alertmanager-main-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ea36739b7e81007aec2e901639b356b275434362b254800d4309dd0aa665ca36" already present on machine

openshift-monitoring

multus

alertmanager-main-0

AddedInterface

Add eth0 [10.128.0.104/23] from ovn-kubernetes

openshift-multus

kubelet

multus-admission-controller-69c949f678-v7wb6

Started

Started container kube-rbac-proxy

openshift-multus

multus

multus-admission-controller-69c949f678-v7wb6

AddedInterface

Add eth0 [10.128.0.106/23] from ovn-kubernetes

openshift-multus

kubelet

multus-admission-controller-69c949f678-v7wb6

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bcb08821551e9a5b9f82aa794bcea673279cefb93cb47492e19ccac5e2cf18fe" already present on machine

openshift-monitoring

multus

alertmanager-main-0

AddedInterface

Add eth0 [10.128.0.104/23] from ovn-kubernetes

openshift-monitoring

kubelet

alertmanager-main-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ea36739b7e81007aec2e901639b356b275434362b254800d4309dd0aa665ca36" already present on machine

openshift-monitoring

kubelet

alertmanager-main-0

Created

Created container: init-config-reloader

openshift-monitoring

kubelet

alertmanager-main-0

Started

Started container init-config-reloader

openshift-multus

kubelet

multus-admission-controller-69c949f678-v7wb6

Created

Created container: multus-admission-controller

openshift-multus

kubelet

multus-admission-controller-69c949f678-v7wb6

Started

Started container multus-admission-controller

openshift-multus

kubelet

multus-admission-controller-69c949f678-v7wb6

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-multus

kubelet

multus-admission-controller-69c949f678-v7wb6

Created

Created container: kube-rbac-proxy

openshift-console

kubelet

downloads-66b8ffb895-dgq8g

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ddc5283caf2ced75a94ddf0e8a43c431889692007e8a875a187b25c35b45a9e2"

openshift-console

multus

downloads-66b8ffb895-dgq8g

AddedInterface

Add eth0 [10.128.0.100/23] from ovn-kubernetes

openshift-multus

multus

multus-admission-controller-69c949f678-v7wb6

AddedInterface

Add eth0 [10.128.0.106/23] from ovn-kubernetes

openshift-multus

kubelet

multus-admission-controller-69c949f678-v7wb6

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bcb08821551e9a5b9f82aa794bcea673279cefb93cb47492e19ccac5e2cf18fe" already present on machine

openshift-monitoring

kubelet

prometheus-k8s-0

Started

Started container init-config-reloader

openshift-monitoring

kubelet

prometheus-k8s-0

Created

Created container: init-config-reloader

openshift-monitoring

kubelet

prometheus-k8s-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ea36739b7e81007aec2e901639b356b275434362b254800d4309dd0aa665ca36" already present on machine

openshift-monitoring

multus

prometheus-k8s-0

AddedInterface

Add eth0 [10.128.0.102/23] from ovn-kubernetes

openshift-multus

kubelet

multus-admission-controller-69c949f678-v7wb6

Created

Created container: multus-admission-controller

openshift-multus

kubelet

multus-admission-controller-69c949f678-v7wb6

Started

Started container multus-admission-controller

openshift-multus

kubelet

multus-admission-controller-69c949f678-v7wb6

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-multus

kubelet

multus-admission-controller-69c949f678-v7wb6

Created

Created container: kube-rbac-proxy

openshift-multus

kubelet

multus-admission-controller-69c949f678-v7wb6

Started

Started container kube-rbac-proxy

openshift-authentication

kubelet

oauth-openshift-8684f7dbf-tqtg9

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3fdcbf7be3f90bd080ffb2c75b091d7eef03681e0f90912ff6140ee48c177616" already present on machine

openshift-authentication

kubelet

oauth-openshift-8684f7dbf-tqtg9

Created

Created container: oauth-openshift

openshift-multus

kubelet

multus-admission-controller-649577484c-p72cd

Killing

Stopping container kube-rbac-proxy

openshift-monitoring

kubelet

alertmanager-main-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ea36739b7e81007aec2e901639b356b275434362b254800d4309dd0aa665ca36" already present on machine

openshift-multus

replicaset-controller

multus-admission-controller-649577484c

SuccessfulDelete

Deleted pod: multus-admission-controller-649577484c-p72cd

openshift-multus

kubelet

multus-admission-controller-649577484c-p72cd

Killing

Stopping container multus-admission-controller

openshift-monitoring

kubelet

alertmanager-main-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:0be5d73579621976f063d98db555f3bceee2f5a91b14422481ce30561438712c" already present on machine

openshift-monitoring

kubelet

alertmanager-main-0

Created

Created container: alertmanager

openshift-multus

deployment-controller

multus-admission-controller

ScalingReplicaSet

Scaled down replica set multus-admission-controller-649577484c to 0 from 1

openshift-monitoring

kubelet

prometheus-k8s-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ea36739b7e81007aec2e901639b356b275434362b254800d4309dd0aa665ca36" already present on machine

openshift-monitoring

kubelet

alertmanager-main-0

Started

Started container alertmanager

openshift-monitoring

kubelet

alertmanager-main-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:0be5d73579621976f063d98db555f3bceee2f5a91b14422481ce30561438712c" already present on machine

openshift-monitoring

kubelet

prometheus-k8s-0

Started

Started container prometheus

openshift-monitoring

kubelet

alertmanager-main-0

Created

Created container: alertmanager

openshift-monitoring

kubelet

alertmanager-main-0

Started

Started container alertmanager

openshift-monitoring

kubelet

alertmanager-main-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ea36739b7e81007aec2e901639b356b275434362b254800d4309dd0aa665ca36" already present on machine

openshift-multus

kubelet

multus-admission-controller-649577484c-p72cd

Killing

Stopping container multus-admission-controller

openshift-multus

kubelet

multus-admission-controller-649577484c-p72cd

Killing

Stopping container kube-rbac-proxy

openshift-multus

replicaset-controller

multus-admission-controller-649577484c

SuccessfulDelete

Deleted pod: multus-admission-controller-649577484c-p72cd

openshift-monitoring

kubelet

prometheus-k8s-0

Created

Created container: prometheus

openshift-monitoring

kubelet

prometheus-k8s-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ea36739b7e81007aec2e901639b356b275434362b254800d4309dd0aa665ca36" already present on machine

openshift-monitoring

kubelet

prometheus-k8s-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f3038df8df25746bb5095296d4e5740f2356f85c1ed8d43f1b3d281e294826e5" already present on machine

openshift-monitoring

kubelet

prometheus-k8s-0

Started

Started container prometheus

openshift-multus

deployment-controller

multus-admission-controller

ScalingReplicaSet

Scaled down replica set multus-admission-controller-649577484c to 0 from 1

openshift-kube-controller-manager-operator

kube-controller-manager-operator-status-controller-statussyncer_kube-controller-manager

kube-controller-manager-operator

OperatorStatusChanged

Status for clusteroperator/kube-controller-manager changed: Degraded message changed from "NodeControllerDegraded: All master nodes are ready\nKubeControllerManagerStaticResourcesDegraded: \"assets/kube-controller-manager/ns.yaml\" (string): Get \"https://172.30.0.1:443/api/v1/namespaces/openshift-kube-controller-manager\": dial tcp 172.30.0.1:443: connect: connection refused\nKubeControllerManagerStaticResourcesDegraded: \"assets/kube-controller-manager/leader-election-rolebinding.yaml\" (string): Get \"https://172.30.0.1:443/apis/rbac.authorization.k8s.io/v1/namespaces/kube-system/rolebindings/system:openshift:leader-locking-kube-controller-manager\": dial tcp 172.30.0.1:443: connect: connection refused\nKubeControllerManagerStaticResourcesDegraded: \"assets/kube-controller-manager/leader-election-cluster-policy-controller-role.yaml\" (string): Get \"https://172.30.0.1:443/apis/rbac.authorization.k8s.io/v1/namespaces/openshift-kube-controller-manager/roles/system:openshift:leader-election-lock-cluster-policy-controller\": dial tcp 172.30.0.1:443: connect: connection refused\nKubeControllerManagerStaticResourcesDegraded: \"assets/kube-controller-manager/leader-election-cluster-policy-controller-rolebinding.yaml\" (string): Get \"https://172.30.0.1:443/apis/rbac.authorization.k8s.io/v1/namespaces/openshift-kube-controller-manager/rolebindings/system:openshift:leader-election-lock-cluster-policy-controller\": dial tcp 172.30.0.1:443: connect: connection refused\nKubeControllerManagerStaticResourcesDegraded: \"assets/kube-controller-manager/namespace-security-allocation-controller-clusterrole.yaml\" (string): Get \"https://172.30.0.1:443/apis/rbac.authorization.k8s.io/v1/clusterroles/system:openshift:controller:namespace-security-allocation-controller\": dial tcp 172.30.0.1:443: connect: connection refused\nKubeControllerManagerStaticResourcesDegraded: \"assets/kube-controller-manager/namespace-security-allocation-controller-clusterrolebinding.yaml\" (string): Get \"https://172.30.0.1:443/apis/rbac.authorization.k8s.io/v1/clusterrolebindings/system:openshift:controller:namespace-security-allocation-controller\": dial tcp 172.30.0.1:443: connect: connection refused\nKubeControllerManagerStaticResourcesDegraded: \"assets/kube-controller-manager/podsecurity-admission-label-syncer-controller-clusterrole.yaml\" (string): Get \"https://172.30.0.1:443/apis/rbac.authorization.k8s.io/v1/clusterroles/system:openshift:controller:podsecurity-admission-label-syncer-controller\": dial tcp 172.30.0.1:443: connect: connection refused\nKubeControllerManagerStaticResourcesDegraded: \"assets/kube-controller-manager/podsecurity-admission-label-syncer-controller-clusterrolebinding.yaml\" (string): Get \"https://172.30.0.1:443/apis/rbac.authorization.k8s.io/v1/clusterrolebindings/system:openshift:controller:podsecurity-admission-label-syncer-controller\": dial tcp 172.30.0.1:443: connect: connection refused\nKubeControllerManagerStaticResourcesDegraded: \"assets/kube-controller-manager/podsecurity-admission-label-privileged-namespaces-syncer-controller-clusterrole.yaml\" (string): Get \"https://172.30.0.1:443/apis/rbac.authorization.k8s.io/v1/clusterroles/system:openshift:controller:privileged-namespaces-psa-label-syncer\": dial tcp 172.30.0.1:443: connect: connection refused\nKubeControllerManagerStaticResourcesDegraded: \"assets/kube-controller-manager/podsecurity-admission-label-privileged-namespaces-syncer-controller-clusterrolebinding.yaml\" (string): Get \"https://172.30.0.1:443/apis/rbac.authorization.k8s.io/v1/clusterrolebindings/system:openshift:controller:privileged-namespaces-psa-label-syncer\": dial tcp 172.30.0.1:443: connect: connection refused\nKubeControllerManagerStaticResourcesDegraded: \"assets/kube-controller-manager/namespace-openshift-infra.yaml\" (string): Get \"https://172.30.0.1:443/api/v1/namespaces/openshift-infra\": dial tcp 172.30.0.1:443: connect: connection refused\nKubeControllerManagerStaticResourcesDegraded: \"assets/kube-controller-manager/svc.yaml\" (string): Get \"https://172.30.0.1:443/api/v1/namespaces/openshift-kube-controller-manager/services/kube-controller-manager\": dial tcp 172.30.0.1:443: connect: connection refused\nKubeControllerManagerStaticResourcesDegraded: \"assets/kube-controller-manager/sa.yaml\" (string): Get \"https://172.30.0.1:443/api/v1/namespaces/openshift-kube-controller-manager/serviceaccounts/kube-controller-manager-sa\": dial tcp 172.30.0.1:443: connect: connection refused\nKubeControllerManagerStaticResourcesDegraded: \"assets/kube-controller-manager/recycler-sa.yaml\" (string): Get \"https://172.30.0.1:443/api/v1/namespaces/openshift-infra/serviceaccounts/pv-recycler-controller\": dial tcp 172.30.0.1:443: connect: connection refused\nKubeControllerManagerStaticResourcesDegraded: \"assets/kube-controller-manager/localhost-recovery-client-crb.yaml\" (string): Get \"https://172.30.0.1:443/apis/rbac.authorization.k8s.io/v1/clusterrolebindings/system:openshift:operator:kube-controller-manager-recovery\": dial tcp 172.30.0.1:443: connect: connection refused\nKubeControllerManagerStaticResourcesDegraded: \"assets/kube-controller-manager/localhost-recovery-sa.yaml\" (string): Get \"https://172.30.0.1:443/api/v1/namespaces/openshift-kube-controller-manager/serviceaccounts/localhost-recovery-client\": dial tcp 172.30.0.1:443: connect: connection refused\nKubeControllerManagerStaticResourcesDegraded: \"assets/kube-controller-manager/csr_approver_clusterrole.yaml\" (string): Get \"https://172.30.0.1:443/apis/rbac.authorization.k8s.io/v1/clusterroles/system:openshift:controller:cluster-csr-approver-controller\": dial tcp 172.30.0.1:443: connect: connection refused\nKubeControllerManagerStaticResourcesDegraded: \"assets/kube-controller-manager/csr_approver_clusterrolebinding.yaml\" (string): Get \"https://172.30.0.1:443/apis/rbac.authorization.k8s.io/v1/clusterrolebindings/system:openshift:controller:cluster-csr-approver-controller\": dial tcp 172.30.0.1:443: connect: connection refused\nKubeControllerManagerStaticResourcesDegraded: \"assets/kube-controller-manager/vsphere/legacy-cloud-provider-sa.yaml\" (string): Delete \"https://172.30.0.1:443/api/v1/namespaces/kube-system/serviceaccounts/vsphere-legacy-cloud-provider\": dial tcp 172.30.0.1:443: connect: connection refused\nKubeControllerManagerStaticResourcesDegraded: " to "NodeControllerDegraded: All master nodes are ready"

openshift-monitoring

kubelet

prometheus-k8s-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f3038df8df25746bb5095296d4e5740f2356f85c1ed8d43f1b3d281e294826e5" already present on machine

openshift-monitoring

kubelet

prometheus-k8s-0

Created

Created container: prometheus

openshift-monitoring

kubelet

prometheus-k8s-0

Created

Created container: kube-rbac-proxy

openshift-monitoring

kubelet

alertmanager-main-0

Started

Started container kube-rbac-proxy

openshift-monitoring

kubelet

prometheus-k8s-0

Started

Started container kube-rbac-proxy-web

openshift-monitoring

kubelet

prometheus-k8s-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-monitoring

kubelet

alertmanager-main-0

Created

Created container: kube-rbac-proxy-web

openshift-monitoring

kubelet

alertmanager-main-0

Started

Started container kube-rbac-proxy-web

openshift-monitoring

kubelet

alertmanager-main-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-monitoring

kubelet

alertmanager-main-0

Started

Started container config-reloader

openshift-monitoring

kubelet

alertmanager-main-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-monitoring

kubelet

alertmanager-main-0

Created

Created container: kube-rbac-proxy

openshift-monitoring

kubelet

alertmanager-main-0

Started

Started container kube-rbac-proxy

openshift-monitoring

kubelet

alertmanager-main-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-monitoring

kubelet

alertmanager-main-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-monitoring

kubelet

prometheus-k8s-0

Started

Started container thanos-sidecar

openshift-monitoring

kubelet

prometheus-k8s-0

Created

Created container: thanos-sidecar

openshift-monitoring

kubelet

alertmanager-main-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:28d99dd1c426021eefd6bdbd01594126623f3473f517f194d39e2a063535147a" already present on machine

openshift-monitoring

kubelet

prometheus-k8s-0

Created

Created container: kube-rbac-proxy

openshift-monitoring

kubelet

alertmanager-main-0

Started

Started container kube-rbac-proxy-metric

openshift-monitoring

kubelet

alertmanager-main-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:28d99dd1c426021eefd6bdbd01594126623f3473f517f194d39e2a063535147a" already present on machine

openshift-monitoring

kubelet

prometheus-k8s-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf72297fee61ec9950f6868881ad3e84be8692ca08f084b3d155d93a766c0823" already present on machine

openshift-monitoring

kubelet

prometheus-k8s-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf72297fee61ec9950f6868881ad3e84be8692ca08f084b3d155d93a766c0823" already present on machine

openshift-monitoring

kubelet

alertmanager-main-0

Started

Started container kube-rbac-proxy-metric

openshift-monitoring

kubelet

prometheus-k8s-0

Created

Created container: kube-rbac-proxy-web

openshift-monitoring

kubelet

prometheus-k8s-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-monitoring

kubelet

prometheus-k8s-0

Started

Started container thanos-sidecar

openshift-monitoring

kubelet

alertmanager-main-0

Started

Started container kube-rbac-proxy-web

openshift-monitoring

kubelet

prometheus-k8s-0

Started

Started container config-reloader

openshift-monitoring

kubelet

alertmanager-main-0

Created

Created container: kube-rbac-proxy-metric

openshift-monitoring

kubelet

alertmanager-main-0

Started

Started container config-reloader

openshift-monitoring

kubelet

alertmanager-main-0

Created

Created container: config-reloader

openshift-monitoring

kubelet

alertmanager-main-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-monitoring

kubelet

alertmanager-main-0

Created

Created container: kube-rbac-proxy-web

openshift-monitoring

kubelet

prometheus-k8s-0

Started

Started container kube-rbac-proxy

openshift-monitoring

kubelet

alertmanager-main-0

Created

Created container: kube-rbac-proxy-metric

openshift-monitoring

kubelet

prometheus-k8s-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-monitoring

kubelet

alertmanager-main-0

Created

Created container: config-reloader

openshift-monitoring

kubelet

alertmanager-main-0

Created

Created container: kube-rbac-proxy

openshift-monitoring

kubelet

prometheus-k8s-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-monitoring

kubelet

prometheus-k8s-0

Created

Created container: config-reloader

openshift-monitoring

kubelet

prometheus-k8s-0

Started

Started container config-reloader

openshift-monitoring

kubelet

alertmanager-main-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-monitoring

kubelet

prometheus-k8s-0

Started

Started container kube-rbac-proxy

openshift-monitoring

kubelet

prometheus-k8s-0

Created

Created container: thanos-sidecar

openshift-monitoring

kubelet

prometheus-k8s-0

Created

Created container: config-reloader

openshift-monitoring

kubelet

prometheus-k8s-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-monitoring

kubelet

prometheus-k8s-0

Started

Started container kube-rbac-proxy-web

openshift-monitoring

kubelet

prometheus-k8s-0

Created

Created container: kube-rbac-proxy-web

openshift-monitoring

kubelet

prometheus-k8s-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d12d0dc7eb86bbedf6b2d7689a28fd51f0d928f720e4a6783744304297c661ed" already present on machine

openshift-monitoring

kubelet

alertmanager-main-0

Started

Started container prom-label-proxy

openshift-monitoring

kubelet

alertmanager-main-0

Created

Created container: prom-label-proxy

openshift-monitoring

kubelet

prometheus-k8s-0

Started

Started container kube-rbac-proxy-thanos

openshift-monitoring

kubelet

prometheus-k8s-0

Created

Created container: kube-rbac-proxy-thanos

openshift-monitoring

kubelet

prometheus-k8s-0

Created

Created container: kube-rbac-proxy-thanos

openshift-monitoring

kubelet

alertmanager-main-0

Created

Created container: prom-label-proxy

openshift-monitoring

kubelet

alertmanager-main-0

Started

Started container prom-label-proxy

openshift-monitoring

kubelet

prometheus-k8s-0

Started

Started container kube-rbac-proxy-thanos

openshift-console-operator

console-operator-status-controller-statussyncer_console

console-operator

OperatorStatusChanged

Status for clusteroperator/console changed: Degraded message changed from "RouteHealthDegraded: route not yet available, https://console-openshift-console.apps.sno.openstack.lab returns '503 Service Unavailable'" to "ConfigMapSyncDegraded: Get \"https://172.30.0.1:443/apis/operators.coreos.com/v1/olmconfigs/cluster\": dial tcp 172.30.0.1:443: connect: connection refused\nPDBSyncDegraded: Get \"https://172.30.0.1:443/apis/policy/v1/namespaces/openshift-console/poddisruptionbudgets/downloads\": dial tcp 172.30.0.1:443: connect: connection refused\nRouteHealthDegraded: route not yet available, https://console-openshift-console.apps.sno.openstack.lab returns '503 Service Unavailable'\nServiceSyncDegraded: Get \"https://172.30.0.1:443/api/v1/namespaces/openshift-console/services/console\": dial tcp 172.30.0.1:443: connect: connection refused",Progressing changed from True to False ("All is well")

openshift-console-operator

console-operator-status-controller-statussyncer_console

console-operator

OperatorStatusChanged

Status for clusteroperator/console changed: Degraded message changed from "ConfigMapSyncDegraded: Get \"https://172.30.0.1:443/apis/operators.coreos.com/v1/olmconfigs/cluster\": dial tcp 172.30.0.1:443: connect: connection refused\nPDBSyncDegraded: Get \"https://172.30.0.1:443/apis/policy/v1/namespaces/openshift-console/poddisruptionbudgets/downloads\": dial tcp 172.30.0.1:443: connect: connection refused\nRouteHealthDegraded: route not yet available, https://console-openshift-console.apps.sno.openstack.lab returns '503 Service Unavailable'\nServiceSyncDegraded: Get \"https://172.30.0.1:443/api/v1/namespaces/openshift-console/services/console\": dial tcp 172.30.0.1:443: connect: connection refused" to "PDBSyncDegraded: Get \"https://172.30.0.1:443/apis/policy/v1/namespaces/openshift-console/poddisruptionbudgets/downloads\": dial tcp 172.30.0.1:443: connect: connection refused\nRouteHealthDegraded: route not yet available, https://console-openshift-console.apps.sno.openstack.lab returns '503 Service Unavailable'\nServiceSyncDegraded: Get \"https://172.30.0.1:443/api/v1/namespaces/openshift-console/services/console\": dial tcp 172.30.0.1:443: connect: connection refused",Progressing changed from False to True ("SyncLoopRefreshProgressing: working toward version 4.18.35, 0 replicas available")

openshift-console-operator

console-operator-status-controller-statussyncer_console

console-operator

OperatorStatusChanged

Status for clusteroperator/console changed: Degraded message changed from "PDBSyncDegraded: Get \"https://172.30.0.1:443/apis/policy/v1/namespaces/openshift-console/poddisruptionbudgets/downloads\": dial tcp 172.30.0.1:443: connect: connection refused\nRouteHealthDegraded: route not yet available, https://console-openshift-console.apps.sno.openstack.lab returns '503 Service Unavailable'\nServiceSyncDegraded: Get \"https://172.30.0.1:443/api/v1/namespaces/openshift-console/services/console\": dial tcp 172.30.0.1:443: connect: connection refused" to "RouteHealthDegraded: route not yet available, https://console-openshift-console.apps.sno.openstack.lab returns '503 Service Unavailable'"

openshift-authentication-operator

oauth-apiserver-status-controller-statussyncer_authentication

authentication-operator

OperatorStatusChanged

Status for clusteroperator/authentication changed: Degraded message changed from "OAuthServerServiceEndpointAccessibleControllerDegraded: Get \"https://172.30.89.113:443/healthz\": dial tcp 172.30.89.113:443: connect: connection refused\nIngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\nOAuthServerRouteEndpointAccessibleControllerDegraded: Get \"https://oauth-openshift.apps.sno.openstack.lab/healthz\": EOF\nWellKnownReadyControllerDegraded: failed to get API server IPs: unable to find kube api server endpointLister port: &v1.Endpoints{TypeMeta:v1.TypeMeta{Kind:\"\", APIVersion:\"\"}, ObjectMeta:v1.ObjectMeta{Name:\"kubernetes\", GenerateName:\"\", Namespace:\"default\", SelfLink:\"\", UID:\"0a8e1475-e015-4961-bc7c-f367a856aa53\", ResourceVersion:\"15724\", Generation:0, CreationTimestamp:time.Date(2026, time.March, 20, 8, 29, 8, 0, time.Local), DeletionTimestamp:<nil>, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{\"endpointslice.kubernetes.io/skip-mirror\":\"true\"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry{v1.ManagedFieldsEntry{Manager:\"kube-apiserver\", Operation:\"Update\", APIVersion:\"v1\", Time:time.Date(2026, time.March, 20, 8, 43, 6, 0, time.Local), FieldsType:\"FieldsV1\", FieldsV1:(*v1.FieldsV1)(0xc00155bbd8), Subresource:\"\"}}}, Subsets:[]v1.EndpointSubset(nil)} (check kube-apiserver that it deploys correctly)" to "All is well",Progressing changed from True to False ("AuthenticatorCertKeyProgressing: All is well"),Available changed from False to True ("All is well")

openshift-console

kubelet

downloads-66b8ffb895-dgq8g

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ddc5283caf2ced75a94ddf0e8a43c431889692007e8a875a187b25c35b45a9e2" in 40.971s (40.971s including waiting). Image size: 2895807090 bytes.

openshift-console

kubelet

downloads-66b8ffb895-dgq8g

Created

Created container: download-server

openshift-console

kubelet

downloads-66b8ffb895-dgq8g

Started

Started container download-server
(x2)

openshift-console

kubelet

downloads-66b8ffb895-dgq8g

Unhealthy

Readiness probe failed: Get "http://10.128.0.100:8080/": dial tcp 10.128.0.100:8080: connect: connection refused
(x2)

openshift-console

kubelet

downloads-66b8ffb895-dgq8g

ProbeError

Readiness probe error: Get "http://10.128.0.100:8080/": dial tcp 10.128.0.100:8080: connect: connection refused body:

openshift-kube-apiserver-operator

kube-apiserver-operator-status-controller-statussyncer_kube-apiserver

kube-apiserver-operator

OperatorStatusChanged

Status for clusteroperator/kube-apiserver changed: Progressing changed from True to False ("NodeInstallerProgressing: 1 node is at revision 5"),Available message changed from "StaticPodsAvailable: 1 nodes are active; 1 node is at revision 4; 0 nodes have achieved new revision 5" to "StaticPodsAvailable: 1 nodes are active; 1 node is at revision 5"

openshift-kube-controller-manager

cluster-policy-controller

cluster-policy-controller-lock

LeaderElection

master-0_7f3eb034-9aef-4025-9f94-4d8a19e4567e became leader

openshift-cluster-storage-operator

csi-snapshot-controller-operator-status-controller-statussyncer_csi-snapshot-controller

csi-snapshot-controller-operator

OperatorStatusChanged

Status for clusteroperator/csi-snapshot-controller changed: Degraded message changed from "WebhookRemovalControllerDegraded: csisnapshotcontrollers.operator.openshift.io \"cluster\" not found" to "All is well"

openshift-cluster-storage-operator

csi-snapshot-controller-operator-status-controller-statussyncer_csi-snapshot-controller

csi-snapshot-controller-operator

OperatorStatusChanged

Status for clusteroperator/csi-snapshot-controller changed: Degraded message changed from "All is well" to "WebhookRemovalControllerDegraded: csisnapshotcontrollers.operator.openshift.io \"cluster\" not found"

openshift-cluster-storage-operator

csi-snapshot-controller-operator

csi-snapshot-controller-operator

FeatureGatesInitialized

FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{"AWSEFSDriverVolumeMetrics", "AdminNetworkPolicy", "AlibabaPlatform", "AzureWorkloadIdentity", "BareMetalLoadBalancer", "BuildCSIVolumes", "ChunkSizeMiB", "CloudDualStackNodeIPs", "DisableKubeletCloudCredentialProviders", "GCPLabelsTags", "HardwareSpeed", "IngressControllerLBSubnetsAWS", "KMSv1", "ManagedBootImages", "ManagedBootImagesAWS", "MultiArchInstallAWS", "MultiArchInstallGCP", "NetworkDiagnosticsConfig", "NetworkLiveMigration", "NetworkSegmentation", "NewOLM", "NodeDisruptionPolicy", "OnClusterBuild", "PersistentIPsForVirtualization", "PrivateHostedZoneAWS", "SetEIPForNLBIngressController", "StreamingCollectionEncodingToJSON", "StreamingCollectionEncodingToProtobuf", "VSphereControlPlaneMachineSet", "VSphereDriverConfiguration", "VSphereMultiVCenters", "VSphereStaticIPs", "ValidatingAdmissionPolicy"}, Disabled:[]v1.FeatureGateName{"AWSClusterHostedDNS", "AdditionalRoutingCapabilities", "AutomatedEtcdBackup", "BootcNodeManagement", "CSIDriverSharedResource", "ClusterAPIInstall", "ClusterAPIInstallIBMCloud", "ClusterMonitoringConfig", "ConsolePluginContentSecurityPolicy", "DNSNameResolver", "DynamicResourceAllocation", "EtcdBackendQuota", "EventedPLEG", "Example", "ExternalOIDC", "ExternalOIDCWithUIDAndExtraClaimMappings", "GCPClusterHostedDNS", "GatewayAPI", "ImageStreamImportMode", "IngressControllerDynamicConfigurationManager", "InsightsConfig", "InsightsConfigAPI", "InsightsOnDemandDataGather", "InsightsRuntimeExtractor", "MachineAPIMigration", "MachineAPIOperatorDisableMachineHealthCheckController", "MachineAPIProviderOpenStack", "MachineConfigNodes", "MaxUnavailableStatefulSet", "MetricsCollectionProfiles", "MinimumKubeletVersion", "MixedCPUsAllocation", "MultiArchInstallAzure", "NodeSwap", "NutanixMultiSubnets", "OVNObservability", "OpenShiftPodSecurityAdmission", "PinnedImages", "PlatformOperators", "ProcMountType", "RouteAdvertisements", "RouteExternalCertificate", "ServiceAccountTokenNodeBinding", "SignatureStores", "SigstoreImageVerification", "TranslateStreamCloseWebsocketRequests", "UpgradeStatus", "UserNamespacesPodSecurityStandards", "UserNamespacesSupport", "VSphereMultiNetworks", "VolumeAttributesClass", "VolumeGroupSnapshot"}}

openshift-cluster-storage-operator

csi-snapshot-controller-operator

csi-snapshot-controller-operator-lock

LeaderElection

csi-snapshot-controller-operator-5f5d689c6b-4dqrh_6277f246-8fc7-419a-9ded-0e24ca29f7e4 became leader

openshift-cluster-version

openshift-cluster-version

version

LeaderElection

master-0_42f7cff3-f955-46e8-9b8c-f5e4de56ee7e became leader

openshift-operator-lifecycle-manager

package-server-manager-7b95f86987-2pg77_ae91c2dc-e361-4f7d-8d5d-22a64ddcbcb5

packageserver-controller-lock

LeaderElection

package-server-manager-7b95f86987-2pg77_ae91c2dc-e361-4f7d-8d5d-22a64ddcbcb5 became leader

openshift-cluster-node-tuning-operator

cluster-node-tuning-operator-598fbc5f8f-vxzvg_a19f64dd-8d40-4b19-9143-6d9dd471172c

node-tuning-operator-lock

LeaderElection

cluster-node-tuning-operator-598fbc5f8f-vxzvg_a19f64dd-8d40-4b19-9143-6d9dd471172c became leader

openshift-cluster-node-tuning-operator

cluster-node-tuning-operator-598fbc5f8f-vxzvg_a19f64dd-8d40-4b19-9143-6d9dd471172c

node-tuning-operator-lock

LeaderElection

cluster-node-tuning-operator-598fbc5f8f-vxzvg_a19f64dd-8d40-4b19-9143-6d9dd471172c became leader

openshift-cluster-olm-operator

cluster-olm-operator

cluster-olm-operator-lock

LeaderElection

cluster-olm-operator-67dcd4998-c5742_fc70de31-f27c-4c3e-9443-e175cef0c9b6 became leader

openshift-cluster-version

openshift-cluster-version

version

LoadPayload

Loading payload version="4.18.35" image="quay.io/openshift-release-dev/ocp-release@sha256:59727c4b3fef19e5149675cf3350735bbfe2c6588a57654b2e4552dd719f58b1"

openshift-cluster-version

openshift-cluster-version

version

RetrievePayload

Retrieving and verifying payload version="4.18.35" image="quay.io/openshift-release-dev/ocp-release@sha256:59727c4b3fef19e5149675cf3350735bbfe2c6588a57654b2e4552dd719f58b1"

openshift-cluster-olm-operator

CatalogdStaticResources-catalogdstaticresources-catalogdstaticresources-staticresources

cluster-olm-operator

MutatingWebhookConfigurationUpdated

Updated MutatingWebhookConfiguration.admissionregistration.k8s.io/catalogd-mutating-webhook-configuration because it changed

openshift-cluster-version

openshift-cluster-version

version

PayloadLoaded

Payload loaded version="4.18.35" image="quay.io/openshift-release-dev/ocp-release@sha256:59727c4b3fef19e5149675cf3350735bbfe2c6588a57654b2e4552dd719f58b1" architecture="amd64"

openshift-config-operator

config-operator-configoperatorcontroller

openshift-config-operator

FastControllerResync

Controller "ConfigOperatorController" resync interval is set to 10s which might lead to client request throttling

openshift-config-operator

config-operator

config-operator-lock

LeaderElection

openshift-config-operator-95bf4f4d-25cml_6a43176e-59eb-4a6a-98fa-82bf47e13bfd became leader

openshift-machine-config-operator

machine-config-operator

master-0

FeatureGatesInitialized

FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{"AWSEFSDriverVolumeMetrics", "AdminNetworkPolicy", "AlibabaPlatform", "AzureWorkloadIdentity", "BareMetalLoadBalancer", "BuildCSIVolumes", "ChunkSizeMiB", "CloudDualStackNodeIPs", "DisableKubeletCloudCredentialProviders", "GCPLabelsTags", "HardwareSpeed", "IngressControllerLBSubnetsAWS", "KMSv1", "ManagedBootImages", "ManagedBootImagesAWS", "MultiArchInstallAWS", "MultiArchInstallGCP", "NetworkDiagnosticsConfig", "NetworkLiveMigration", "NetworkSegmentation", "NewOLM", "NodeDisruptionPolicy", "OnClusterBuild", "PersistentIPsForVirtualization", "PrivateHostedZoneAWS", "SetEIPForNLBIngressController", "StreamingCollectionEncodingToJSON", "StreamingCollectionEncodingToProtobuf", "VSphereControlPlaneMachineSet", "VSphereDriverConfiguration", "VSphereMultiVCenters", "VSphereStaticIPs", "ValidatingAdmissionPolicy"}, Disabled:[]v1.FeatureGateName{"AWSClusterHostedDNS", "AdditionalRoutingCapabilities", "AutomatedEtcdBackup", "BootcNodeManagement", "CSIDriverSharedResource", "ClusterAPIInstall", "ClusterAPIInstallIBMCloud", "ClusterMonitoringConfig", "ConsolePluginContentSecurityPolicy", "DNSNameResolver", "DynamicResourceAllocation", "EtcdBackendQuota", "EventedPLEG", "Example", "ExternalOIDC", "ExternalOIDCWithUIDAndExtraClaimMappings", "GCPClusterHostedDNS", "GatewayAPI", "ImageStreamImportMode", "IngressControllerDynamicConfigurationManager", "InsightsConfig", "InsightsConfigAPI", "InsightsOnDemandDataGather", "InsightsRuntimeExtractor", "MachineAPIMigration", "MachineAPIOperatorDisableMachineHealthCheckController", "MachineAPIProviderOpenStack", "MachineConfigNodes", "MaxUnavailableStatefulSet", "MetricsCollectionProfiles", "MinimumKubeletVersion", "MixedCPUsAllocation", "MultiArchInstallAzure", "NodeSwap", "NutanixMultiSubnets", "OVNObservability", "OpenShiftPodSecurityAdmission", "PinnedImages", "PlatformOperators", "ProcMountType", "RouteAdvertisements", "RouteExternalCertificate", "ServiceAccountTokenNodeBinding", "SignatureStores", "SigstoreImageVerification", "TranslateStreamCloseWebsocketRequests", "UpgradeStatus", "UserNamespacesPodSecurityStandards", "UserNamespacesSupport", "VSphereMultiNetworks", "VolumeAttributesClass", "VolumeGroupSnapshot"}}

openshift-machine-config-operator

machine-config-operator

machine-config-operator

ConfigMapUpdated

Updated ConfigMap/kube-rbac-proxy -n openshift-machine-config-operator: cause by changes in data.config-file.yaml
(x9)

openshift-console

kubelet

console-ffb45cdb5-7r4lm

FailedMount

MountVolume.SetUp failed for volume "console-serving-cert" : secret "console-serving-cert" not found
(x9)

openshift-console

kubelet

console-58857c5dc9-9wf6k

FailedMount

MountVolume.SetUp failed for volume "console-serving-cert" : secret "console-serving-cert" not found

openshift-cluster-storage-operator

cluster-storage-operator

cluster-storage-operator

FeatureGatesInitialized

FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{"AWSEFSDriverVolumeMetrics", "AdminNetworkPolicy", "AlibabaPlatform", "AzureWorkloadIdentity", "BareMetalLoadBalancer", "BuildCSIVolumes", "ChunkSizeMiB", "CloudDualStackNodeIPs", "DisableKubeletCloudCredentialProviders", "GCPLabelsTags", "HardwareSpeed", "IngressControllerLBSubnetsAWS", "KMSv1", "ManagedBootImages", "ManagedBootImagesAWS", "MultiArchInstallAWS", "MultiArchInstallGCP", "NetworkDiagnosticsConfig", "NetworkLiveMigration", "NetworkSegmentation", "NewOLM", "NodeDisruptionPolicy", "OnClusterBuild", "PersistentIPsForVirtualization", "PrivateHostedZoneAWS", "SetEIPForNLBIngressController", "StreamingCollectionEncodingToJSON", "StreamingCollectionEncodingToProtobuf", "VSphereControlPlaneMachineSet", "VSphereDriverConfiguration", "VSphereMultiVCenters", "VSphereStaticIPs", "ValidatingAdmissionPolicy"}, Disabled:[]v1.FeatureGateName{"AWSClusterHostedDNS", "AdditionalRoutingCapabilities", "AutomatedEtcdBackup", "BootcNodeManagement", "CSIDriverSharedResource", "ClusterAPIInstall", "ClusterAPIInstallIBMCloud", "ClusterMonitoringConfig", "ConsolePluginContentSecurityPolicy", "DNSNameResolver", "DynamicResourceAllocation", "EtcdBackendQuota", "EventedPLEG", "Example", "ExternalOIDC", "ExternalOIDCWithUIDAndExtraClaimMappings", "GCPClusterHostedDNS", "GatewayAPI", "ImageStreamImportMode", "IngressControllerDynamicConfigurationManager", "InsightsConfig", "InsightsConfigAPI", "InsightsOnDemandDataGather", "InsightsRuntimeExtractor", "MachineAPIMigration", "MachineAPIOperatorDisableMachineHealthCheckController", "MachineAPIProviderOpenStack", "MachineConfigNodes", "MaxUnavailableStatefulSet", "MetricsCollectionProfiles", "MinimumKubeletVersion", "MixedCPUsAllocation", "MultiArchInstallAzure", "NodeSwap", "NutanixMultiSubnets", "OVNObservability", "OpenShiftPodSecurityAdmission", "PinnedImages", "PlatformOperators", "ProcMountType", "RouteAdvertisements", "RouteExternalCertificate", "ServiceAccountTokenNodeBinding", "SignatureStores", "SigstoreImageVerification", "TranslateStreamCloseWebsocketRequests", "UpgradeStatus", "UserNamespacesPodSecurityStandards", "UserNamespacesSupport", "VSphereMultiNetworks", "VolumeAttributesClass", "VolumeGroupSnapshot"}}

openshift-cluster-storage-operator

cluster-storage-operator

cluster-storage-operator-lock

LeaderElection

cluster-storage-operator-7d87854d6-vlq7h_aa8ce179-7f19-4d64-b44e-8a7fdc487101 became leader

openshift-kube-controller-manager-operator

kube-controller-manager-operator-revisioncontroller

kube-controller-manager-operator

StartingNewRevision

new revision 4 triggered by "required secret/service-account-private-key has changed"

openshift-kube-controller-manager-operator

kube-controller-manager-operator-revisioncontroller

kube-controller-manager-operator

ConfigMapCreated

Created ConfigMap/kube-controller-manager-pod-4 -n openshift-kube-controller-manager because it was missing

openshift-kube-controller-manager-operator

kube-controller-manager-operator-revisioncontroller

kube-controller-manager-operator

ConfigMapCreated

Created ConfigMap/config-4 -n openshift-kube-controller-manager because it was missing

openshift-kube-controller-manager-operator

kube-controller-manager-operator-satokensignercontroller

kube-controller-manager-operator

SecretUpdated

Updated Secret/service-account-private-key -n openshift-kube-controller-manager because it changed

openshift-kube-controller-manager-operator

kube-controller-manager-operator-revisioncontroller

kube-controller-manager-operator

ConfigMapCreated

Created ConfigMap/cluster-policy-controller-config-4 -n openshift-kube-controller-manager because it was missing

openshift-kube-controller-manager-operator

kube-controller-manager-operator-revisioncontroller

kube-controller-manager-operator

ConfigMapCreated

Created ConfigMap/controller-manager-kubeconfig-4 -n openshift-kube-controller-manager because it was missing

openshift-kube-controller-manager-operator

kube-controller-manager-operator-revisioncontroller

kube-controller-manager-operator

ConfigMapCreated

Created ConfigMap/kube-controller-cert-syncer-kubeconfig-4 -n openshift-kube-controller-manager because it was missing

openshift-kube-controller-manager-operator

kube-controller-manager-operator-revisioncontroller

kube-controller-manager-operator

ConfigMapCreated

Created ConfigMap/serviceaccount-ca-4 -n openshift-kube-controller-manager because it was missing

openshift-kube-controller-manager-operator

kube-controller-manager-operator-revisioncontroller

kube-controller-manager-operator

ConfigMapCreated

Created ConfigMap/service-ca-4 -n openshift-kube-controller-manager because it was missing

openshift-kube-controller-manager-operator

kube-controller-manager-operator-revisioncontroller

kube-controller-manager-operator

ConfigMapCreated

Created ConfigMap/recycler-config-4 -n openshift-kube-controller-manager because it was missing

openshift-kube-controller-manager-operator

kube-controller-manager-operator-revisioncontroller

kube-controller-manager-operator

SecretCreated

Created Secret/service-account-private-key-4 -n openshift-kube-controller-manager because it was missing

openshift-kube-controller-manager-operator

kube-controller-manager-operator-revisioncontroller

kube-controller-manager-operator

SecretCreated

Created Secret/serving-cert-4 -n openshift-kube-controller-manager because it was missing

openshift-kube-controller-manager-operator

kube-controller-manager-operator-revisioncontroller

kube-controller-manager-operator

RevisionTriggered

new revision 4 triggered by "required secret/service-account-private-key has changed"

openshift-kube-controller-manager-operator

kube-controller-manager-operator-revisioncontroller

kube-controller-manager-operator

SecretCreated

Created Secret/localhost-recovery-client-token-4 -n openshift-kube-controller-manager because it was missing

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator-lock

LeaderElection

openshift-kube-scheduler-operator-dddff6458-vx5d7_3c8537c5-9dc0-4424-8441-ed999c2f74de became leader

openshift-kube-scheduler-operator

openshift-cluster-kube-scheduler-operator

openshift-kube-scheduler-operator

FeatureGatesInitialized

FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{"AWSEFSDriverVolumeMetrics", "AdminNetworkPolicy", "AlibabaPlatform", "AzureWorkloadIdentity", "BareMetalLoadBalancer", "BuildCSIVolumes", "ChunkSizeMiB", "CloudDualStackNodeIPs", "DisableKubeletCloudCredentialProviders", "GCPLabelsTags", "HardwareSpeed", "IngressControllerLBSubnetsAWS", "KMSv1", "ManagedBootImages", "ManagedBootImagesAWS", "MultiArchInstallAWS", "MultiArchInstallGCP", "NetworkDiagnosticsConfig", "NetworkLiveMigration", "NetworkSegmentation", "NewOLM", "NodeDisruptionPolicy", "OnClusterBuild", "PersistentIPsForVirtualization", "PrivateHostedZoneAWS", "SetEIPForNLBIngressController", "StreamingCollectionEncodingToJSON", "StreamingCollectionEncodingToProtobuf", "VSphereControlPlaneMachineSet", "VSphereDriverConfiguration", "VSphereMultiVCenters", "VSphereStaticIPs", "ValidatingAdmissionPolicy"}, Disabled:[]v1.FeatureGateName{"AWSClusterHostedDNS", "AdditionalRoutingCapabilities", "AutomatedEtcdBackup", "BootcNodeManagement", "CSIDriverSharedResource", "ClusterAPIInstall", "ClusterAPIInstallIBMCloud", "ClusterMonitoringConfig", "ConsolePluginContentSecurityPolicy", "DNSNameResolver", "DynamicResourceAllocation", "EtcdBackendQuota", "EventedPLEG", "Example", "ExternalOIDC", "ExternalOIDCWithUIDAndExtraClaimMappings", "GCPClusterHostedDNS", "GatewayAPI", "ImageStreamImportMode", "IngressControllerDynamicConfigurationManager", "InsightsConfig", "InsightsConfigAPI", "InsightsOnDemandDataGather", "InsightsRuntimeExtractor", "MachineAPIMigration", "MachineAPIOperatorDisableMachineHealthCheckController", "MachineAPIProviderOpenStack", "MachineConfigNodes", "MaxUnavailableStatefulSet", "MetricsCollectionProfiles", "MinimumKubeletVersion", "MixedCPUsAllocation", "MultiArchInstallAzure", "NodeSwap", "NutanixMultiSubnets", "OVNObservability", "OpenShiftPodSecurityAdmission", "PinnedImages", "PlatformOperators", "ProcMountType", "RouteAdvertisements", "RouteExternalCertificate", "ServiceAccountTokenNodeBinding", "SignatureStores", "SigstoreImageVerification", "TranslateStreamCloseWebsocketRequests", "UpgradeStatus", "UserNamespacesPodSecurityStandards", "UserNamespacesSupport", "VSphereMultiNetworks", "VolumeAttributesClass", "VolumeGroupSnapshot"}}

openshift-machine-config-operator

machine-config-operator

master-0

FeatureGatesInitialized

FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{"AWSEFSDriverVolumeMetrics", "AdminNetworkPolicy", "AlibabaPlatform", "AzureWorkloadIdentity", "BareMetalLoadBalancer", "BuildCSIVolumes", "ChunkSizeMiB", "CloudDualStackNodeIPs", "DisableKubeletCloudCredentialProviders", "GCPLabelsTags", "HardwareSpeed", "IngressControllerLBSubnetsAWS", "KMSv1", "ManagedBootImages", "ManagedBootImagesAWS", "MultiArchInstallAWS", "MultiArchInstallGCP", "NetworkDiagnosticsConfig", "NetworkLiveMigration", "NetworkSegmentation", "NewOLM", "NodeDisruptionPolicy", "OnClusterBuild", "PersistentIPsForVirtualization", "PrivateHostedZoneAWS", "SetEIPForNLBIngressController", "StreamingCollectionEncodingToJSON", "StreamingCollectionEncodingToProtobuf", "VSphereControlPlaneMachineSet", "VSphereDriverConfiguration", "VSphereMultiVCenters", "VSphereStaticIPs", "ValidatingAdmissionPolicy"}, Disabled:[]v1.FeatureGateName{"AWSClusterHostedDNS", "AdditionalRoutingCapabilities", "AutomatedEtcdBackup", "BootcNodeManagement", "CSIDriverSharedResource", "ClusterAPIInstall", "ClusterAPIInstallIBMCloud", "ClusterMonitoringConfig", "ConsolePluginContentSecurityPolicy", "DNSNameResolver", "DynamicResourceAllocation", "EtcdBackendQuota", "EventedPLEG", "Example", "ExternalOIDC", "ExternalOIDCWithUIDAndExtraClaimMappings", "GCPClusterHostedDNS", "GatewayAPI", "ImageStreamImportMode", "IngressControllerDynamicConfigurationManager", "InsightsConfig", "InsightsConfigAPI", "InsightsOnDemandDataGather", "InsightsRuntimeExtractor", "MachineAPIMigration", "MachineAPIOperatorDisableMachineHealthCheckController", "MachineAPIProviderOpenStack", "MachineConfigNodes", "MaxUnavailableStatefulSet", "MetricsCollectionProfiles", "MinimumKubeletVersion", "MixedCPUsAllocation", "MultiArchInstallAzure", "NodeSwap", "NutanixMultiSubnets", "OVNObservability", "OpenShiftPodSecurityAdmission", "PinnedImages", "PlatformOperators", "ProcMountType", "RouteAdvertisements", "RouteExternalCertificate", "ServiceAccountTokenNodeBinding", "SignatureStores", "SigstoreImageVerification", "TranslateStreamCloseWebsocketRequests", "UpgradeStatus", "UserNamespacesPodSecurityStandards", "UserNamespacesSupport", "VSphereMultiNetworks", "VolumeAttributesClass", "VolumeGroupSnapshot"}}

openshift-kube-controller-manager-operator

kube-controller-manager-operator-installer-controller

kube-controller-manager-operator

NodeTargetRevisionChanged

Updating node "master-0" from revision 3 to 4 because node master-0 with revision 3 is the oldest

openshift-kube-controller-manager-operator

kube-controller-manager-operator-status-controller-statussyncer_kube-controller-manager

kube-controller-manager-operator

OperatorStatusChanged

Status for clusteroperator/kube-controller-manager changed: Progressing changed from False to True ("NodeInstallerProgressing: 1 node is at revision 3; 0 nodes have achieved new revision 4"),Available message changed from "StaticPodsAvailable: 1 nodes are active; 1 node is at revision 3" to "StaticPodsAvailable: 1 nodes are active; 1 node is at revision 3; 0 nodes have achieved new revision 4"

openshift-kube-controller-manager

kubelet

installer-4-master-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c032f87ae61d6f757ff3ce52620a70a43516591987731f25da77aba152f17458" already present on machine

openshift-kube-controller-manager

multus

installer-4-master-0

AddedInterface

Add eth0 [10.128.0.107/23] from ovn-kubernetes

openshift-kube-controller-manager-operator

kube-controller-manager-operator-installer-controller

kube-controller-manager-operator

PodCreated

Created Pod/installer-4-master-0 -n openshift-kube-controller-manager because it was missing

openshift-kube-controller-manager

kubelet

installer-4-master-0

Started

Started container installer

openshift-kube-controller-manager

kubelet

installer-4-master-0

Created

Created container: installer

openshift-image-registry

image-registry-operator

cluster-image-registry-operator

FeatureGatesInitialized

FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{"AWSEFSDriverVolumeMetrics", "AdminNetworkPolicy", "AlibabaPlatform", "AzureWorkloadIdentity", "BareMetalLoadBalancer", "BuildCSIVolumes", "ChunkSizeMiB", "CloudDualStackNodeIPs", "DisableKubeletCloudCredentialProviders", "GCPLabelsTags", "HardwareSpeed", "IngressControllerLBSubnetsAWS", "KMSv1", "ManagedBootImages", "ManagedBootImagesAWS", "MultiArchInstallAWS", "MultiArchInstallGCP", "NetworkDiagnosticsConfig", "NetworkLiveMigration", "NetworkSegmentation", "NewOLM", "NodeDisruptionPolicy", "OnClusterBuild", "PersistentIPsForVirtualization", "PrivateHostedZoneAWS", "SetEIPForNLBIngressController", "StreamingCollectionEncodingToJSON", "StreamingCollectionEncodingToProtobuf", "VSphereControlPlaneMachineSet", "VSphereDriverConfiguration", "VSphereMultiVCenters", "VSphereStaticIPs", "ValidatingAdmissionPolicy"}, Disabled:[]v1.FeatureGateName{"AWSClusterHostedDNS", "AdditionalRoutingCapabilities", "AutomatedEtcdBackup", "BootcNodeManagement", "CSIDriverSharedResource", "ClusterAPIInstall", "ClusterAPIInstallIBMCloud", "ClusterMonitoringConfig", "ConsolePluginContentSecurityPolicy", "DNSNameResolver", "DynamicResourceAllocation", "EtcdBackendQuota", "EventedPLEG", "Example", "ExternalOIDC", "ExternalOIDCWithUIDAndExtraClaimMappings", "GCPClusterHostedDNS", "GatewayAPI", "ImageStreamImportMode", "IngressControllerDynamicConfigurationManager", "InsightsConfig", "InsightsConfigAPI", "InsightsOnDemandDataGather", "InsightsRuntimeExtractor", "MachineAPIMigration", "MachineAPIOperatorDisableMachineHealthCheckController", "MachineAPIProviderOpenStack", "MachineConfigNodes", "MaxUnavailableStatefulSet", "MetricsCollectionProfiles", "MinimumKubeletVersion", "MixedCPUsAllocation", "MultiArchInstallAzure", "NodeSwap", "NutanixMultiSubnets", "OVNObservability", "OpenShiftPodSecurityAdmission", "PinnedImages", "PlatformOperators", "ProcMountType", "RouteAdvertisements", "RouteExternalCertificate", "ServiceAccountTokenNodeBinding", "SignatureStores", "SigstoreImageVerification", "TranslateStreamCloseWebsocketRequests", "UpgradeStatus", "UserNamespacesPodSecurityStandards", "UserNamespacesSupport", "VSphereMultiNetworks", "VolumeAttributesClass", "VolumeGroupSnapshot"}}

openshift-image-registry

image-registry-operator

openshift-master-controllers

LeaderElection

cluster-image-registry-operator-5549dc66cb-fmhbq_5e222a84-6954-4c8c-a728-c74e1e35a457 became leader

openshift-console-operator

console-operator-status-controller-statussyncer_console

console-operator

OperatorStatusChanged

Status for clusteroperator/console changed: Degraded message changed from "RouteHealthDegraded: route not yet available, https://console-openshift-console.apps.sno.openstack.lab returns '503 Service Unavailable'" to "DeploymentSyncDegraded: Operation cannot be fulfilled on deployments.apps \"console\": the object has been modified; please apply your changes to the latest version and try again\nRouteHealthDegraded: route not yet available, https://console-openshift-console.apps.sno.openstack.lab returns '503 Service Unavailable'",Progressing changed from True to False ("All is well")

openshift-console-operator

console-operator-console-operator-consoleoperator

console-operator

DeploymentUpdateFailed

Failed to update Deployment.apps/console -n openshift-console: Operation cannot be fulfilled on deployments.apps "console": the object has been modified; please apply your changes to the latest version and try again

openshift-console

replicaset-controller

console-ffb45cdb5

SuccessfulDelete

Deleted pod: console-ffb45cdb5-7r4lm

openshift-console

replicaset-controller

console-d4699dbdf

SuccessfulCreate

Created pod: console-d4699dbdf-6z95l

openshift-service-ca

service-ca-controller

service-ca-controller-lock

LeaderElection

service-ca-79bc6b8d76-72j8t_cf6adfaf-2313-4c05-9956-ca2b5d387637 became leader

openshift-console

deployment-controller

console

ScalingReplicaSet

Scaled up replica set console-d4699dbdf to 1 from 0

openshift-console-operator

console-operator-status-controller-statussyncer_console

console-operator

OperatorStatusChanged

Status for clusteroperator/console changed: Progressing message changed from "SyncLoopRefreshProgressing: working toward version 4.18.35, 0 replicas available" to "SyncLoopRefreshProgressing: changes made during sync updates, additional sync expected"

openshift-console

deployment-controller

console

ScalingReplicaSet

Scaled down replica set console-ffb45cdb5 to 0 from 1

openshift-console

kubelet

console-d4699dbdf-6z95l

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5bbb8535e2496de8389585ebbe696e7d7b9bad2b27785ad8a30a0fc683b0a22d"

openshift-console

multus

console-d4699dbdf-6z95l

AddedInterface

Add eth0 [10.128.0.108/23] from ovn-kubernetes

openshift-console-operator

console-operator-status-controller-statussyncer_console

console-operator

OperatorStatusChanged

Status for clusteroperator/console changed: Degraded message changed from "DeploymentSyncDegraded: Operation cannot be fulfilled on deployments.apps \"console\": the object has been modified; please apply your changes to the latest version and try again\nRouteHealthDegraded: route not yet available, https://console-openshift-console.apps.sno.openstack.lab returns '503 Service Unavailable'" to "RouteHealthDegraded: route not yet available, https://console-openshift-console.apps.sno.openstack.lab returns '503 Service Unavailable'",Progressing changed from False to True ("SyncLoopRefreshProgressing: working toward version 4.18.35, 0 replicas available")

openshift-machine-config-operator

machine-config-operator

machine-config-operator

ConfigMapUpdated

Updated ConfigMap/kube-rbac-proxy -n openshift-machine-config-operator: cause by changes in data.config-file.yaml

openshift-console

kubelet

console-d4699dbdf-6z95l

Started

Started container console

openshift-console

kubelet

console-d4699dbdf-6z95l

Created

Created container: console

openshift-console

kubelet

console-d4699dbdf-6z95l

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5bbb8535e2496de8389585ebbe696e7d7b9bad2b27785ad8a30a0fc683b0a22d" in 4.868s (4.868s including waiting). Image size: 633877280 bytes.

openshift-console

replicaset-controller

console-58857c5dc9

SuccessfulDelete

Deleted pod: console-58857c5dc9-9wf6k

openshift-console

deployment-controller

console

ScalingReplicaSet

Scaled down replica set console-58857c5dc9 to 0 from 1

openshift-console-operator

console-operator-status-controller-statussyncer_console

console-operator

OperatorStatusChanged

Status for clusteroperator/console changed: Progressing message changed from "SyncLoopRefreshProgressing: working toward version 4.18.35, 0 replicas available" to "SyncLoopRefreshProgressing: working toward version 4.18.35, 1 replicas available",Available message changed from "DeploymentAvailable: 0 replicas available for console deployment\nRouteHealthAvailable: route not yet available, https://console-openshift-console.apps.sno.openstack.lab returns '503 Service Unavailable'" to "RouteHealthAvailable: route not yet available, https://console-openshift-console.apps.sno.openstack.lab returns '503 Service Unavailable'"

openshift-console-operator

console-operator-status-controller-statussyncer_console

console-operator

OperatorStatusChanged

Status for clusteroperator/console changed: Degraded changed from True to False ("All is well"),Available changed from False to True ("All is well")

openshift-kube-controller-manager

kubelet

kube-controller-manager-master-0

Killing

Stopping container cluster-policy-controller

openshift-kube-controller-manager

kubelet

kube-controller-manager-master-0

Killing

Stopping container kube-controller-manager-cert-syncer

openshift-kube-controller-manager

static-pod-installer

installer-4-master-0

StaticPodInstallerCompleted

Successfully installed revision 4

openshift-kube-controller-manager

kubelet

kube-controller-manager-master-0

Killing

Stopping container kube-controller-manager

openshift-kube-controller-manager

kubelet

kube-controller-manager-master-0

Killing

Stopping container kube-controller-manager-recovery-controller

openshift-machine-config-operator

machine-config-operator

machine-config-operator

ConfigMapUpdated

Updated ConfigMap/kube-rbac-proxy -n openshift-machine-config-operator: cause by changes in data.config-file.yaml

openshift-kube-apiserver-operator

kube-apiserver-operator-connectivity-check-controller-connectivitycheckcontroller

kube-apiserver-operator

CustomResourceDefinitionCreateFailed

Failed to create CustomResourceDefinition.apiextensions.k8s.io/podnetworkconnectivitychecks.controlplane.operator.openshift.io: customresourcedefinitions.apiextensions.k8s.io "podnetworkconnectivitychecks.controlplane.operator.openshift.io" already exists

openshift-apiserver-operator

openshift-apiserver-operator-connectivity-check-controller-connectivitycheckcontroller

openshift-apiserver-operator

CustomResourceDefinitionCreated

Created CustomResourceDefinition.apiextensions.k8s.io/podnetworkconnectivitychecks.controlplane.operator.openshift.io because it was missing

openshift-kube-controller-manager

kubelet

kube-controller-manager-master-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c032f87ae61d6f757ff3ce52620a70a43516591987731f25da77aba152f17458" already present on machine

openshift-kube-controller-manager

kubelet

kube-controller-manager-master-0

Started

Started container kube-controller-manager-cert-syncer

openshift-kube-controller-manager

kubelet

kube-controller-manager-master-0

Created

Created container: kube-controller-manager-cert-syncer

openshift-kube-controller-manager

kubelet

kube-controller-manager-master-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c032f87ae61d6f757ff3ce52620a70a43516591987731f25da77aba152f17458" already present on machine

openshift-kube-controller-manager

kubelet

kube-controller-manager-master-0

Started

Started container cluster-policy-controller

openshift-kube-controller-manager

kubelet

kube-controller-manager-master-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1fbbcb390de2563a0177b92fba1b5a65777366e2dc80e2808b61d87c41b47a2d" already present on machine

openshift-kube-controller-manager

kubelet

kube-controller-manager-master-0

Created

Created container: cluster-policy-controller

openshift-kube-controller-manager

cert-recovery-controller

cert-recovery-controller-lock

LeaderElection

master-0_9c7fc7b1-8757-4d14-9c8f-4c8e3edbb80e became leader

openshift-kube-controller-manager

cluster-policy-controller

kube-controller-manager-master-0

ControlPlaneTopology

unable to get control plane topology, using HA cluster values for leader election: infrastructures.config.openshift.io "cluster" is forbidden: User "system:kube-controller-manager" cannot get resource "infrastructures" in API group "config.openshift.io" at the cluster scope

openshift-kube-controller-manager

cluster-policy-controller

cluster-policy-controller-lock

LeaderElection

master-0_74c17ed7-df0c-46bf-9dea-49debad05d86 became leader

openshift-kube-controller-manager

kubelet

kube-controller-manager-master-0

Started

Started container kube-controller-manager-recovery-controller

openshift-kube-controller-manager

kubelet

kube-controller-manager-master-0

Created

Created container: kube-controller-manager-recovery-controller

openshift-kube-controller-manager-operator

kube-controller-manager-operator-status-controller-statussyncer_kube-controller-manager

kube-controller-manager-operator

OperatorStatusChanged

Status for clusteroperator/kube-controller-manager changed: Progressing changed from True to False ("NodeInstallerProgressing: 1 node is at revision 4"),Available message changed from "StaticPodsAvailable: 1 nodes are active; 1 node is at revision 3; 0 nodes have achieved new revision 4" to "StaticPodsAvailable: 1 nodes are active; 1 node is at revision 4"

openshift-kube-controller-manager-operator

kube-controller-manager-operator-installer-controller

kube-controller-manager-operator

NodeCurrentRevisionChanged

Updated node "master-0" from revision 3 to 4 because static pod is ready

kube-system

kube-controller-manager

kube-controller-manager

LeaderElection

master-0_366b1b29-ca44-40e7-ab58-72f26f2e489a became leader

default

node-controller

master-0

RegisteredNode

Node master-0 event: Registered Node master-0 in Controller

openshift-kube-controller-manager

cluster-policy-controller-namespace-security-allocation-controller

kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for sushy-emulator namespace

sushy-emulator

deployment-controller

sushy-emulator

ScalingReplicaSet

Scaled up replica set sushy-emulator-59477995f9 to 1

sushy-emulator

replicaset-controller

sushy-emulator-59477995f9

SuccessfulCreate

Created pod: sushy-emulator-59477995f9-mk4kq

sushy-emulator

kubelet

sushy-emulator-59477995f9-mk4kq

Pulling

Pulling image "quay.io/rhn_gps_hjensas/sushy-tools:dev-1773400388"

sushy-emulator

multus

sushy-emulator-59477995f9-mk4kq

AddedInterface

Add eth0 [10.128.0.109/23] from ovn-kubernetes

sushy-emulator

kubelet

sushy-emulator-59477995f9-mk4kq

Pulled

Successfully pulled image "quay.io/rhn_gps_hjensas/sushy-tools:dev-1773400388" in 8.489s (8.489s including waiting). Image size: 326085552 bytes.

sushy-emulator

kubelet

sushy-emulator-59477995f9-mk4kq

Created

Created container: sushy-emulator

sushy-emulator

kubelet

sushy-emulator-59477995f9-mk4kq

Started

Started container sushy-emulator

sushy-emulator

deployment-controller

nova-console-poller

ScalingReplicaSet

Scaled up replica set nova-console-poller-557b677b49 to 1

sushy-emulator

replicaset-controller

nova-console-poller-557b677b49

SuccessfulCreate

Created pod: nova-console-poller-557b677b49-vslxq

sushy-emulator

kubelet

nova-console-poller-557b677b49-vslxq

Pulling

Pulling image "quay.io/rhn_gps_hjensas/nova-console-poller:latest"

sushy-emulator

multus

nova-console-poller-557b677b49-vslxq

AddedInterface

Add eth0 [10.128.0.110/23] from ovn-kubernetes

sushy-emulator

kubelet

nova-console-poller-557b677b49-vslxq

Pulled

Successfully pulled image "quay.io/rhn_gps_hjensas/nova-console-poller:latest" in 5.536s (5.536s including waiting). Image size: 202633582 bytes.

sushy-emulator

kubelet

nova-console-poller-557b677b49-vslxq

Pulled

Successfully pulled image "quay.io/rhn_gps_hjensas/nova-console-poller:latest" in 510ms (510ms including waiting). Image size: 202633582 bytes.

sushy-emulator

kubelet

nova-console-poller-557b677b49-vslxq

Started

Started container console-poller-97cc6af5-97d5-4509-be1e-24890ef6c8f3

sushy-emulator

kubelet

nova-console-poller-557b677b49-vslxq

Started

Started container console-poller-70bded0b-b6c9-4262-bfd9-3176e3774c04

sushy-emulator

kubelet

nova-console-poller-557b677b49-vslxq

Pulling

Pulling image "quay.io/rhn_gps_hjensas/nova-console-poller:latest"

sushy-emulator

kubelet

nova-console-poller-557b677b49-vslxq

Created

Created container: console-poller-70bded0b-b6c9-4262-bfd9-3176e3774c04

sushy-emulator

kubelet

nova-console-poller-557b677b49-vslxq

Created

Created container: console-poller-97cc6af5-97d5-4509-be1e-24890ef6c8f3

sushy-emulator

deployment-controller

nova-console-recorder

ScalingReplicaSet

Scaled up replica set nova-console-recorder-6766d57b69 to 1

sushy-emulator

replicaset-controller

nova-console-recorder-6766d57b69

SuccessfulCreate

Created pod: nova-console-recorder-6766d57b69-2wgj4

sushy-emulator

multus

nova-console-recorder-6766d57b69-2wgj4

AddedInterface

Add eth0 [10.128.0.111/23] from ovn-kubernetes

sushy-emulator

kubelet

nova-console-recorder-6766d57b69-2wgj4

Pulling

Pulling image "quay.io/rhn_gps_hjensas/nova-console-recorder:latest"

sushy-emulator

kubelet

nova-console-recorder-6766d57b69-2wgj4

Started

Started container console-recorder-97cc6af5-97d5-4509-be1e-24890ef6c8f3

sushy-emulator

kubelet

nova-console-recorder-6766d57b69-2wgj4

Pulling

Pulling image "quay.io/rhn_gps_hjensas/nova-console-recorder:latest"

sushy-emulator

kubelet

nova-console-recorder-6766d57b69-2wgj4

Created

Created container: console-recorder-97cc6af5-97d5-4509-be1e-24890ef6c8f3

sushy-emulator

kubelet

nova-console-recorder-6766d57b69-2wgj4

Pulled

Successfully pulled image "quay.io/rhn_gps_hjensas/nova-console-recorder:latest" in 8.337s (8.337s including waiting). Image size: 664134874 bytes.

sushy-emulator

kubelet

nova-console-recorder-6766d57b69-2wgj4

Created

Created container: console-recorder-70bded0b-b6c9-4262-bfd9-3176e3774c04

sushy-emulator

kubelet

nova-console-recorder-6766d57b69-2wgj4

Started

Started container console-recorder-70bded0b-b6c9-4262-bfd9-3176e3774c04

sushy-emulator

kubelet

nova-console-recorder-6766d57b69-2wgj4

Pulled

Successfully pulled image "quay.io/rhn_gps_hjensas/nova-console-recorder:latest" in 492ms (492ms including waiting). Image size: 664134874 bytes.

openshift-machine-api

replicaset-controller

metal3-c957566cd

SuccessfulCreate

Created pod: metal3-c957566cd-lgms8

openshift-machine-api

deployment-controller

metal3

ScalingReplicaSet

Scaled up replica set metal3-c957566cd to 1

openshift-machine-api

replicaset-controller

metal3-c957566cd

SuccessfulCreate

Created pod: metal3-c957566cd-lgms8

openshift-machine-api

deployment-controller

metal3

ScalingReplicaSet

Scaled up replica set metal3-c957566cd to 1

openshift-machine-api

kubelet

metal3-c957566cd-lgms8

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:00db3efdb8113f49d0cf5fac1ce22ba738b29fb7ec51faa94e235d09dcfac70b"

openshift-machine-api

deployment-controller

metal3-baremetal-operator

ScalingReplicaSet

Scaled up replica set metal3-baremetal-operator-78474bdc48 to 1

openshift-machine-api

kubelet

metal3-c957566cd-lgms8

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:00db3efdb8113f49d0cf5fac1ce22ba738b29fb7ec51faa94e235d09dcfac70b"

openshift-machine-api

deployment-controller

metal3-baremetal-operator

ScalingReplicaSet

Scaled up replica set metal3-baremetal-operator-78474bdc48 to 1

openshift-machine-api

replicaset-controller

metal3-baremetal-operator-78474bdc48

SuccessfulCreate

Created pod: metal3-baremetal-operator-78474bdc48-ndj7c

openshift-machine-api

replicaset-controller

metal3-baremetal-operator-78474bdc48

SuccessfulCreate

Created pod: metal3-baremetal-operator-78474bdc48-ndj7c
(x2)

openshift-machine-api

kubelet

metal3-baremetal-operator-78474bdc48-ndj7c

FailedMount

MountVolume.SetUp failed for volume "cert" : secret "baremetal-operator-webhook-server-cert" not found
(x2)

openshift-machine-api

kubelet

metal3-baremetal-operator-78474bdc48-ndj7c

FailedMount

MountVolume.SetUp failed for volume "cert" : secret "baremetal-operator-webhook-server-cert" not found

openshift-machine-api

multus

metal3-baremetal-operator-78474bdc48-ndj7c

AddedInterface

Add eth0 [10.128.0.112/23] from ovn-kubernetes

openshift-machine-api

kubelet

metal3-baremetal-operator-78474bdc48-ndj7c

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3160b9c4d5f4c3af05c6a073a1c590b9679be82d06193a819aaed0a2914e27f7"

openshift-machine-api

kubelet

metal3-baremetal-operator-78474bdc48-ndj7c

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3160b9c4d5f4c3af05c6a073a1c590b9679be82d06193a819aaed0a2914e27f7"

openshift-machine-api

multus

metal3-baremetal-operator-78474bdc48-ndj7c

AddedInterface

Add eth0 [10.128.0.112/23] from ovn-kubernetes

openshift-machine-api

replicaset-controller

metal3-image-customization-7dd5d8c865

SuccessfulCreate

Created pod: metal3-image-customization-7dd5d8c865-5ppkq

openshift-machine-api

multus

metal3-image-customization-7dd5d8c865-5ppkq

AddedInterface

Add eth0 [10.128.0.113/23] from ovn-kubernetes

openshift-machine-api

deployment-controller

metal3-image-customization

ScalingReplicaSet

Scaled up replica set metal3-image-customization-7dd5d8c865 to 1

openshift-machine-api

deployment-controller

metal3-image-customization

ScalingReplicaSet

Scaled up replica set metal3-image-customization-7dd5d8c865 to 1

openshift-machine-api

replicaset-controller

metal3-image-customization-7dd5d8c865

SuccessfulCreate

Created pod: metal3-image-customization-7dd5d8c865-5ppkq

openshift-machine-api

multus

metal3-image-customization-7dd5d8c865-5ppkq

AddedInterface

Add eth0 [10.128.0.113/23] from ovn-kubernetes

openshift-machine-api

kubelet

metal3-image-customization-7dd5d8c865-5ppkq

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:00db3efdb8113f49d0cf5fac1ce22ba738b29fb7ec51faa94e235d09dcfac70b"

openshift-machine-api

daemonset-controller

ironic-proxy

SuccessfulCreate

Created pod: ironic-proxy-28zbq

openshift-machine-api

kubelet

ironic-proxy-28zbq

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8484f65d4a84230f93c986362dde19aff9b77de01b50e5af1948748b51382001"

openshift-machine-api

kubelet

ironic-proxy-28zbq

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8484f65d4a84230f93c986362dde19aff9b77de01b50e5af1948748b51382001"

openshift-machine-api

kubelet

metal3-image-customization-7dd5d8c865-5ppkq

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:00db3efdb8113f49d0cf5fac1ce22ba738b29fb7ec51faa94e235d09dcfac70b"

openshift-machine-api

daemonset-controller

ironic-proxy

SuccessfulCreate

Created pod: ironic-proxy-28zbq

openshift-machine-api

kubelet

metal3-baremetal-operator-78474bdc48-ndj7c

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3160b9c4d5f4c3af05c6a073a1c590b9679be82d06193a819aaed0a2914e27f7" in 3.097s (3.097s including waiting). Image size: 512688600 bytes.

openshift-machine-api

kubelet

metal3-baremetal-operator-78474bdc48-ndj7c

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3160b9c4d5f4c3af05c6a073a1c590b9679be82d06193a819aaed0a2914e27f7" in 3.097s (3.097s including waiting). Image size: 512688600 bytes.

openshift-machine-api

kubelet

metal3-baremetal-operator-78474bdc48-ndj7c

Created

Created container: metal3-baremetal-operator

openshift-machine-api

kubelet

metal3-baremetal-operator-78474bdc48-ndj7c

Started

Started container metal3-baremetal-operator

openshift-machine-api

kubelet

metal3-baremetal-operator-78474bdc48-ndj7c

Created

Created container: metal3-baremetal-operator

openshift-machine-api

metal3-baremetal-operator-78474bdc48-ndj7c_6043fef3-8721-4c84-96de-0dcbca06c1f0

baremetal-operator

LeaderElection

metal3-baremetal-operator-78474bdc48-ndj7c_6043fef3-8721-4c84-96de-0dcbca06c1f0 became leader

openshift-machine-api

metal3-baremetal-operator-78474bdc48-ndj7c_6043fef3-8721-4c84-96de-0dcbca06c1f0

baremetal-operator

LeaderElection

metal3-baremetal-operator-78474bdc48-ndj7c_6043fef3-8721-4c84-96de-0dcbca06c1f0 became leader

openshift-machine-api

kubelet

metal3-baremetal-operator-78474bdc48-ndj7c

Started

Started container metal3-baremetal-operator

openshift-machine-api

kubelet

ironic-proxy-28zbq

Created

Created container: ironic-proxy

openshift-machine-api

kubelet

ironic-proxy-28zbq

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8484f65d4a84230f93c986362dde19aff9b77de01b50e5af1948748b51382001" in 10.091s (10.091s including waiting). Image size: 710479311 bytes.

openshift-machine-api

kubelet

ironic-proxy-28zbq

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8484f65d4a84230f93c986362dde19aff9b77de01b50e5af1948748b51382001" in 10.091s (10.091s including waiting). Image size: 710479311 bytes.

openshift-machine-api

kubelet

ironic-proxy-28zbq

Created

Created container: ironic-proxy

openshift-machine-api

kubelet

ironic-proxy-28zbq

Started

Started container ironic-proxy

openshift-machine-api

kubelet

ironic-proxy-28zbq

Started

Started container ironic-proxy

openshift-machine-api

kubelet

metal3-image-customization-7dd5d8c865-5ppkq

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:00db3efdb8113f49d0cf5fac1ce22ba738b29fb7ec51faa94e235d09dcfac70b" in 30.715s (30.715s including waiting). Image size: 1738357022 bytes.

openshift-machine-api

kubelet

metal3-image-customization-7dd5d8c865-5ppkq

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:00db3efdb8113f49d0cf5fac1ce22ba738b29fb7ec51faa94e235d09dcfac70b" in 30.715s (30.715s including waiting). Image size: 1738357022 bytes.

openshift-machine-api

kubelet

metal3-c957566cd-lgms8

Started

Started container machine-os-images

openshift-machine-api

kubelet

metal3-c957566cd-lgms8

Created

Created container: machine-os-images

openshift-machine-api

kubelet

metal3-c957566cd-lgms8

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:00db3efdb8113f49d0cf5fac1ce22ba738b29fb7ec51faa94e235d09dcfac70b" in 34.252s (34.252s including waiting). Image size: 1738357022 bytes.

openshift-machine-api

kubelet

metal3-c957566cd-lgms8

Started

Started container machine-os-images

openshift-machine-api

kubelet

metal3-c957566cd-lgms8

Created

Created container: machine-os-images

openshift-machine-api

kubelet

metal3-c957566cd-lgms8

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:00db3efdb8113f49d0cf5fac1ce22ba738b29fb7ec51faa94e235d09dcfac70b" in 34.252s (34.252s including waiting). Image size: 1738357022 bytes.

openshift-machine-api

kubelet

metal3-c957566cd-lgms8

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8484f65d4a84230f93c986362dde19aff9b77de01b50e5af1948748b51382001" already present on machine

openshift-machine-api

kubelet

metal3-c957566cd-lgms8

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8484f65d4a84230f93c986362dde19aff9b77de01b50e5af1948748b51382001" already present on machine

openshift-machine-api

kubelet

metal3-c957566cd-lgms8

Created

Created container: metal3-httpd

openshift-machine-api

kubelet

metal3-c957566cd-lgms8

Created

Created container: metal3-httpd

openshift-machine-api

kubelet

metal3-c957566cd-lgms8

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8484f65d4a84230f93c986362dde19aff9b77de01b50e5af1948748b51382001" already present on machine

openshift-machine-api

kubelet

metal3-c957566cd-lgms8

Started

Started container metal3-httpd

openshift-machine-api

kubelet

metal3-c957566cd-lgms8

Started

Started container metal3-httpd

openshift-machine-api

kubelet

metal3-c957566cd-lgms8

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8484f65d4a84230f93c986362dde19aff9b77de01b50e5af1948748b51382001" already present on machine

openshift-machine-api

kubelet

metal3-c957566cd-lgms8

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8484f65d4a84230f93c986362dde19aff9b77de01b50e5af1948748b51382001" already present on machine

openshift-machine-api

kubelet

metal3-c957566cd-lgms8

Started

Started container metal3-ironic

openshift-machine-api

kubelet

metal3-c957566cd-lgms8

Created

Created container: metal3-ironic

openshift-machine-api

kubelet

metal3-c957566cd-lgms8

Started

Started container metal3-ironic

openshift-machine-api

kubelet

metal3-c957566cd-lgms8

Created

Created container: metal3-ironic

openshift-machine-api

kubelet

metal3-c957566cd-lgms8

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8484f65d4a84230f93c986362dde19aff9b77de01b50e5af1948748b51382001" already present on machine

openshift-machine-api

kubelet

metal3-image-customization-7dd5d8c865-5ppkq

BackOff

Back-off restarting failed container machine-os-images in pod metal3-image-customization-7dd5d8c865-5ppkq_openshift-machine-api(a78acf22-ffe5-4b13-9ec6-a4b212930d6a)

openshift-machine-api

kubelet

metal3-image-customization-7dd5d8c865-5ppkq

BackOff

Back-off restarting failed container machine-os-images in pod metal3-image-customization-7dd5d8c865-5ppkq_openshift-machine-api(a78acf22-ffe5-4b13-9ec6-a4b212930d6a)

openshift-machine-config-operator

machine-config-operator

machine-config-operator

ConfigMapUpdated

Updated ConfigMap/kube-rbac-proxy -n openshift-machine-config-operator: cause by changes in data.config-file.yaml

openshift-machine-api

kubelet

metal3-c957566cd-lgms8

Created

Created container: metal3-ramdisk-logs

openshift-machine-api

kubelet

metal3-c957566cd-lgms8

Created

Created container: metal3-ramdisk-logs

openshift-machine-api

kubelet

metal3-c957566cd-lgms8

Started

Started container metal3-ramdisk-logs

openshift-machine-api

kubelet

metal3-c957566cd-lgms8

Started

Started container metal3-ramdisk-logs

openshift-kube-controller-manager

cluster-policy-controller-namespace-security-allocation-controller

kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openshift-storage namespace
(x3)

openshift-machine-api

kubelet

metal3-image-customization-7dd5d8c865-5ppkq

Created

Created container: machine-os-images
(x3)

openshift-machine-api

kubelet

metal3-image-customization-7dd5d8c865-5ppkq

Started

Started container machine-os-images
(x3)

openshift-machine-api

kubelet

metal3-image-customization-7dd5d8c865-5ppkq

Started

Started container machine-os-images
(x3)

openshift-machine-api

kubelet

metal3-image-customization-7dd5d8c865-5ppkq

Created

Created container: machine-os-images
(x2)

openshift-machine-api

kubelet

metal3-image-customization-7dd5d8c865-5ppkq

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:00db3efdb8113f49d0cf5fac1ce22ba738b29fb7ec51faa94e235d09dcfac70b" already present on machine

openshift-marketplace

job-controller

7f6062bfcf66f08711c4d599873349559e66916847a22b4b74a32f97d40dd54

SuccessfulCreate

Created pod: 7f6062bfcf66f08711c4d599873349559e66916847a22b4b74a32f97d4gkzrg
(x2)

openshift-machine-api

kubelet

metal3-image-customization-7dd5d8c865-5ppkq

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:00db3efdb8113f49d0cf5fac1ce22ba738b29fb7ec51faa94e235d09dcfac70b" already present on machine

openshift-machine-api

kubelet

metal3-image-customization-7dd5d8c865-5ppkq

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f28c3488ff36d9ed2ab6da459a8bead5f5949a4216e12b83f22a26bc502faed1"

openshift-marketplace

kubelet

7f6062bfcf66f08711c4d599873349559e66916847a22b4b74a32f97d4gkzrg

Created

Created container: util

openshift-marketplace

multus

7f6062bfcf66f08711c4d599873349559e66916847a22b4b74a32f97d4gkzrg

AddedInterface

Add eth0 [10.128.0.114/23] from ovn-kubernetes

openshift-marketplace

kubelet

7f6062bfcf66f08711c4d599873349559e66916847a22b4b74a32f97d4gkzrg

Started

Started container util

openshift-machine-api

kubelet

metal3-image-customization-7dd5d8c865-5ppkq

Pulling

Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f28c3488ff36d9ed2ab6da459a8bead5f5949a4216e12b83f22a26bc502faed1"

openshift-marketplace

kubelet

7f6062bfcf66f08711c4d599873349559e66916847a22b4b74a32f97d4gkzrg

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6bba3f73c0066e42b24839e0d29f5dce2f36436f0a11f9f5e1029bccc5ed6578" already present on machine

openshift-marketplace

kubelet

7f6062bfcf66f08711c4d599873349559e66916847a22b4b74a32f97d4gkzrg

Pulling

Pulling image "registry.redhat.io/lvms4/lvms-operator-bundle@sha256:82bcaba7f44d28e8529915af868c847107932a6f8cc9d2eaf34796c578c7a5ba"

openshift-marketplace

kubelet

7f6062bfcf66f08711c4d599873349559e66916847a22b4b74a32f97d4gkzrg

Pulled

Successfully pulled image "registry.redhat.io/lvms4/lvms-operator-bundle@sha256:82bcaba7f44d28e8529915af868c847107932a6f8cc9d2eaf34796c578c7a5ba" in 1.138s (1.138s including waiting). Image size: 108204 bytes.

openshift-marketplace

kubelet

7f6062bfcf66f08711c4d599873349559e66916847a22b4b74a32f97d4gkzrg

Created

Created container: pull

openshift-marketplace

kubelet

7f6062bfcf66f08711c4d599873349559e66916847a22b4b74a32f97d4gkzrg

Started

Started container pull

openshift-machine-api

kubelet

metal3-image-customization-7dd5d8c865-5ppkq

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f28c3488ff36d9ed2ab6da459a8bead5f5949a4216e12b83f22a26bc502faed1" in 3.782s (3.782s including waiting). Image size: 538654221 bytes.

openshift-marketplace

kubelet

7f6062bfcf66f08711c4d599873349559e66916847a22b4b74a32f97d4gkzrg

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b00c42562d477ef44d51f35950253a26d7debc7de86e53270831aafef5795c1" already present on machine

openshift-marketplace

kubelet

7f6062bfcf66f08711c4d599873349559e66916847a22b4b74a32f97d4gkzrg

Created

Created container: extract

openshift-marketplace

kubelet

7f6062bfcf66f08711c4d599873349559e66916847a22b4b74a32f97d4gkzrg

Started

Started container extract

openshift-machine-api

kubelet

metal3-image-customization-7dd5d8c865-5ppkq

Pulled

Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f28c3488ff36d9ed2ab6da459a8bead5f5949a4216e12b83f22a26bc502faed1" in 3.782s (3.782s including waiting). Image size: 538654221 bytes.

openshift-machine-api

kubelet

metal3-image-customization-7dd5d8c865-5ppkq

Started

Started container machine-image-customization-controller

openshift-machine-api

kubelet

metal3-image-customization-7dd5d8c865-5ppkq

Created

Created container: machine-image-customization-controller

openshift-machine-api

kubelet

metal3-image-customization-7dd5d8c865-5ppkq

Created

Created container: machine-image-customization-controller

openshift-machine-api

kubelet

metal3-image-customization-7dd5d8c865-5ppkq

Started

Started container machine-image-customization-controller

openshift-marketplace

job-controller

7f6062bfcf66f08711c4d599873349559e66916847a22b4b74a32f97d40dd54

Completed

Job completed

openshift-storage

operator-lifecycle-manager

lvms-operator.v4.18.4

RequirementsUnknown

requirements not yet checked

openshift-storage

operator-lifecycle-manager

lvms-operator.v4.18.4

RequirementsUnknown

requirements not yet checked

openshift-storage

operator-lifecycle-manager

lvms-operator.v4.18.4

RequirementsNotMet

one or more requirements couldn't be found

openshift-storage

operator-lifecycle-manager

lvms-operator.v4.18.4

RequirementsNotMet

one or more requirements couldn't be found
(x2)

openshift-storage

operator-lifecycle-manager

lvms-operator.v4.18.4

AllRequirementsMet

all requirements found, attempting install
(x2)

openshift-storage

operator-lifecycle-manager

lvms-operator.v4.18.4

AllRequirementsMet

all requirements found, attempting install

openshift-storage

replicaset-controller

lvms-operator-dd6b59969

SuccessfulCreate

Created pod: lvms-operator-dd6b59969-zzdqx

openshift-storage

multus

lvms-operator-dd6b59969-zzdqx

AddedInterface

Add eth0 [10.128.0.115/23] from ovn-kubernetes

openshift-storage

deployment-controller

lvms-operator

ScalingReplicaSet

Scaled up replica set lvms-operator-dd6b59969 to 1
(x2)

openshift-storage

operator-lifecycle-manager

lvms-operator.v4.18.4

InstallSucceeded

waiting for install components to report healthy
(x2)

openshift-storage

operator-lifecycle-manager

lvms-operator.v4.18.4

InstallWaiting

installing: waiting for deployment lvms-operator to become ready: deployment "lvms-operator" not available: Deployment does not have minimum availability.
(x2)

openshift-storage

operator-lifecycle-manager

lvms-operator.v4.18.4

InstallWaiting

installing: waiting for deployment lvms-operator to become ready: deployment "lvms-operator" not available: Deployment does not have minimum availability.
(x2)

openshift-storage

operator-lifecycle-manager

lvms-operator.v4.18.4

InstallSucceeded

waiting for install components to report healthy

openshift-storage

replicaset-controller

lvms-operator-dd6b59969

SuccessfulCreate

Created pod: lvms-operator-dd6b59969-zzdqx

openshift-storage

multus

lvms-operator-dd6b59969-zzdqx

AddedInterface

Add eth0 [10.128.0.115/23] from ovn-kubernetes

openshift-storage

deployment-controller

lvms-operator

ScalingReplicaSet

Scaled up replica set lvms-operator-dd6b59969 to 1

openshift-storage

kubelet

lvms-operator-dd6b59969-zzdqx

Pulling

Pulling image "registry.redhat.io/lvms4/lvms-rhel9-operator@sha256:be49109ec5be53a783b2de9bc6529db99930a69021577c05cfa0bf1312e06d69"

openshift-storage

kubelet

lvms-operator-dd6b59969-zzdqx

Pulling

Pulling image "registry.redhat.io/lvms4/lvms-rhel9-operator@sha256:be49109ec5be53a783b2de9bc6529db99930a69021577c05cfa0bf1312e06d69"

openshift-storage

kubelet

lvms-operator-dd6b59969-zzdqx

Created

Created container: manager

openshift-storage

kubelet

lvms-operator-dd6b59969-zzdqx

Pulled

Successfully pulled image "registry.redhat.io/lvms4/lvms-rhel9-operator@sha256:be49109ec5be53a783b2de9bc6529db99930a69021577c05cfa0bf1312e06d69" in 4.735s (4.735s including waiting). Image size: 238305644 bytes.

openshift-storage

kubelet

lvms-operator-dd6b59969-zzdqx

Started

Started container manager

openshift-storage

kubelet

lvms-operator-dd6b59969-zzdqx

Started

Started container manager

openshift-storage

kubelet

lvms-operator-dd6b59969-zzdqx

Pulled

Successfully pulled image "registry.redhat.io/lvms4/lvms-rhel9-operator@sha256:be49109ec5be53a783b2de9bc6529db99930a69021577c05cfa0bf1312e06d69" in 4.735s (4.735s including waiting). Image size: 238305644 bytes.

openshift-storage

kubelet

lvms-operator-dd6b59969-zzdqx

Created

Created container: manager

openshift-storage

operator-lifecycle-manager

lvms-operator.v4.18.4

InstallSucceeded

install strategy completed with no errors

openshift-storage

operator-lifecycle-manager

lvms-operator.v4.18.4

InstallSucceeded

install strategy completed with no errors

openshift-kube-controller-manager

cluster-policy-controller-namespace-security-allocation-controller

kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for metallb-system namespace

openshift-kube-controller-manager

cluster-policy-controller-namespace-security-allocation-controller

kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for cert-manager-operator namespace

openshift-kube-controller-manager

cluster-policy-controller-namespace-security-allocation-controller

kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openshift-nmstate namespace

openshift-marketplace

job-controller

925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e56eb0c

SuccessfulCreate

Created pod: 925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5kp5jf

openshift-marketplace

kubelet

925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5kp5jf

Created

Created container: util

openshift-marketplace

multus

925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5kp5jf

AddedInterface

Add eth0 [10.128.0.116/23] from ovn-kubernetes

openshift-marketplace

kubelet

925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5kp5jf

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6bba3f73c0066e42b24839e0d29f5dce2f36436f0a11f9f5e1029bccc5ed6578" already present on machine

openshift-marketplace

kubelet

925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5kp5jf

Started

Started container util

openshift-marketplace

kubelet

925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5kp5jf

Pulling

Pulling image "registry.redhat.io/cert-manager/cert-manager-operator-bundle@sha256:e4e3f81062da90a9cfcdce27085f0624952374a9aec5fbdd5796a09d24f83908"

openshift-marketplace

job-controller

1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874832f3

SuccessfulCreate

Created pod: 1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874g29cm

openshift-marketplace

job-controller

2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c166a6a

SuccessfulCreate

Created pod: 2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1v54q9

openshift-marketplace

kubelet

2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1v54q9

Started

Started container util

openshift-marketplace

kubelet

2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1v54q9

Created

Created container: util

openshift-marketplace

multus

1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874g29cm

AddedInterface

Add eth0 [10.128.0.118/23] from ovn-kubernetes

openshift-marketplace

kubelet

1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874g29cm

Started

Started container util

openshift-marketplace

kubelet

1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874g29cm

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6bba3f73c0066e42b24839e0d29f5dce2f36436f0a11f9f5e1029bccc5ed6578" already present on machine

openshift-marketplace

kubelet

2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1v54q9

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6bba3f73c0066e42b24839e0d29f5dce2f36436f0a11f9f5e1029bccc5ed6578" already present on machine

openshift-marketplace

kubelet

1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874g29cm

Created

Created container: util

openshift-marketplace

multus

2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1v54q9

AddedInterface

Add eth0 [10.128.0.117/23] from ovn-kubernetes

openshift-marketplace

kubelet

2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1v54q9

Pulling

Pulling image "registry.redhat.io/openshift4/ose-metallb-operator-bundle@sha256:8d089fd8dd2786d76c87bd470470abb86f06587c447a3b309efe4116911aa11c"

openshift-marketplace

kubelet

1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874g29cm

Pulling

Pulling image "registry.redhat.io/openshift4/kubernetes-nmstate-operator-bundle@sha256:0a730171e8f18a8286180b7514213248748be998b454d1053b10d047ca51ae1e"

openshift-marketplace

kubelet

925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5kp5jf

Pulled

Successfully pulled image "registry.redhat.io/cert-manager/cert-manager-operator-bundle@sha256:e4e3f81062da90a9cfcdce27085f0624952374a9aec5fbdd5796a09d24f83908" in 4.856s (4.856s including waiting). Image size: 108352841 bytes.

openshift-marketplace

kubelet

925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5kp5jf

Created

Created container: pull

openshift-marketplace

kubelet

925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5kp5jf

Started

Started container pull

openshift-marketplace

job-controller

93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726f148f

SuccessfulCreate

Created pod: 93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726rptjh

openshift-marketplace

kubelet

925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5kp5jf

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b00c42562d477ef44d51f35950253a26d7debc7de86e53270831aafef5795c1" already present on machine

openshift-marketplace

kubelet

1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874g29cm

Pulled

Successfully pulled image "registry.redhat.io/openshift4/kubernetes-nmstate-operator-bundle@sha256:0a730171e8f18a8286180b7514213248748be998b454d1053b10d047ca51ae1e" in 3.7s (3.7s including waiting). Image size: 255829 bytes.

openshift-marketplace

kubelet

925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5kp5jf

Started

Started container extract

openshift-marketplace

kubelet

925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5kp5jf

Created

Created container: extract

openshift-marketplace

kubelet

1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874g29cm

Started

Started container pull

openshift-marketplace

kubelet

1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874g29cm

Started

Started container extract

openshift-marketplace

kubelet

93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726rptjh

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6bba3f73c0066e42b24839e0d29f5dce2f36436f0a11f9f5e1029bccc5ed6578" already present on machine

openshift-marketplace

multus

93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726rptjh

AddedInterface

Add eth0 [10.128.0.119/23] from ovn-kubernetes

openshift-marketplace

kubelet

1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874g29cm

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b00c42562d477ef44d51f35950253a26d7debc7de86e53270831aafef5795c1" already present on machine

openshift-marketplace

kubelet

1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874g29cm

Created

Created container: pull

openshift-marketplace

kubelet

1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874g29cm

Created

Created container: extract

openshift-marketplace

kubelet

93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726rptjh

Pulling

Pulling image "registry.redhat.io/cluster-observability-operator/cluster-observability-operator-bundle@sha256:0415e8263a185c51897bcd5d3ac2f5fe68e4818282a2f9dc89f215ee3b9dd1ed"

openshift-marketplace

kubelet

2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1v54q9

Pulled

Successfully pulled image "registry.redhat.io/openshift4/ose-metallb-operator-bundle@sha256:8d089fd8dd2786d76c87bd470470abb86f06587c447a3b309efe4116911aa11c" in 5.088s (5.088s including waiting). Image size: 408540 bytes.

openshift-marketplace

kubelet

93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726rptjh

Started

Started container util

openshift-marketplace

kubelet

93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726rptjh

Created

Created container: util

openshift-marketplace

kubelet

2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1v54q9

Created

Created container: pull

openshift-marketplace

kubelet

2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1v54q9

Started

Started container pull

openshift-marketplace

kubelet

2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1v54q9

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b00c42562d477ef44d51f35950253a26d7debc7de86e53270831aafef5795c1" already present on machine

openshift-marketplace

kubelet

2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1v54q9

Started

Started container extract

openshift-marketplace

kubelet

2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1v54q9

Created

Created container: extract

openshift-marketplace

job-controller

925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e56eb0c

Completed

Job completed

openshift-marketplace

job-controller

1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874832f3

Completed

Job completed

openshift-marketplace

kubelet

93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726rptjh

Started

Started container pull

openshift-marketplace

kubelet

93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726rptjh

Pulled

Successfully pulled image "registry.redhat.io/cluster-observability-operator/cluster-observability-operator-bundle@sha256:0415e8263a185c51897bcd5d3ac2f5fe68e4818282a2f9dc89f215ee3b9dd1ed" in 3.265s (3.265s including waiting). Image size: 5243975 bytes.

openshift-marketplace

kubelet

93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726rptjh

Created

Created container: pull

openshift-marketplace

kubelet

93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726rptjh

Created

Created container: extract

openshift-marketplace

kubelet

93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726rptjh

Started

Started container extract

openshift-marketplace

job-controller

2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c166a6a

Completed

Job completed

openshift-marketplace

kubelet

93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726rptjh

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b00c42562d477ef44d51f35950253a26d7debc7de86e53270831aafef5795c1" already present on machine

openshift-marketplace

job-controller

93d662022be5376a0ed3676a120a68427f47e4653a19a985adf9239726f148f

Completed

Job completed

metallb-system

operator-lifecycle-manager

metallb-operator.v4.18.0-202603040208

RequirementsUnknown

requirements not yet checked

metallb-system

operator-lifecycle-manager

metallb-operator.v4.18.0-202603040208

RequirementsUnknown

requirements not yet checked

cert-manager

deployment-controller

cert-manager-cainjector

ScalingReplicaSet

Scaled up replica set cert-manager-cainjector-5545bd876 to 1

cert-manager

deployment-controller

cert-manager-cainjector

ScalingReplicaSet

Scaled up replica set cert-manager-cainjector-5545bd876 to 1

cert-manager

deployment-controller

cert-manager-webhook

ScalingReplicaSet

Scaled up replica set cert-manager-webhook-6888856db4 to 1

default

cert-manager-istio-csr-controller

ControllerStarted

controller is starting

metallb-system

operator-lifecycle-manager

metallb-operator.v4.18.0-202603040208

RequirementsNotMet

one or more requirements couldn't be found

cert-manager

deployment-controller

cert-manager-webhook

ScalingReplicaSet

Scaled up replica set cert-manager-webhook-6888856db4 to 1

metallb-system

operator-lifecycle-manager

metallb-operator.v4.18.0-202603040208

RequirementsNotMet

one or more requirements couldn't be found

openshift-kube-controller-manager

cluster-policy-controller-namespace-security-allocation-controller

kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for cert-manager namespace
(x9)

cert-manager

replicaset-controller

cert-manager-webhook-6888856db4

FailedCreate

Error creating: pods "cert-manager-webhook-6888856db4-" is forbidden: error looking up service account cert-manager/cert-manager-webhook: serviceaccount "cert-manager-webhook" not found
(x9)

cert-manager

replicaset-controller

cert-manager-webhook-6888856db4

FailedCreate

Error creating: pods "cert-manager-webhook-6888856db4-" is forbidden: error looking up service account cert-manager/cert-manager-webhook: serviceaccount "cert-manager-webhook" not found

cert-manager

multus

cert-manager-webhook-6888856db4-kkc9m

AddedInterface

Add eth0 [10.128.0.121/23] from ovn-kubernetes
(x10)

cert-manager

replicaset-controller

cert-manager-cainjector-5545bd876

FailedCreate

Error creating: pods "cert-manager-cainjector-5545bd876-" is forbidden: error looking up service account cert-manager/cert-manager-cainjector: serviceaccount "cert-manager-cainjector" not found

cert-manager

replicaset-controller

cert-manager-webhook-6888856db4

SuccessfulCreate

Created pod: cert-manager-webhook-6888856db4-kkc9m
(x10)

cert-manager

replicaset-controller

cert-manager-cainjector-5545bd876

FailedCreate

Error creating: pods "cert-manager-cainjector-5545bd876-" is forbidden: error looking up service account cert-manager/cert-manager-cainjector: serviceaccount "cert-manager-cainjector" not found

cert-manager

multus

cert-manager-webhook-6888856db4-kkc9m

AddedInterface

Add eth0 [10.128.0.121/23] from ovn-kubernetes

cert-manager

replicaset-controller

cert-manager-webhook-6888856db4

SuccessfulCreate

Created pod: cert-manager-webhook-6888856db4-kkc9m

cert-manager

kubelet

cert-manager-webhook-6888856db4-kkc9m

Pulling

Pulling image "registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:903ce74138b1ffc735846a7c5fcdf62bbe82ca29568a6b38caec2656f6637671"

cert-manager

deployment-controller

cert-manager

ScalingReplicaSet

Scaled up replica set cert-manager-545d4d4674 to 1

cert-manager

deployment-controller

cert-manager

ScalingReplicaSet

Scaled up replica set cert-manager-545d4d4674 to 1

cert-manager

kubelet

cert-manager-webhook-6888856db4-kkc9m

Pulling

Pulling image "registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:903ce74138b1ffc735846a7c5fcdf62bbe82ca29568a6b38caec2656f6637671"

openshift-nmstate

operator-lifecycle-manager

kubernetes-nmstate-operator.4.18.0-202603041813

RequirementsNotMet

one or more requirements couldn't be found

openshift-nmstate

operator-lifecycle-manager

kubernetes-nmstate-operator.4.18.0-202603041813

RequirementsUnknown

requirements not yet checked

openshift-nmstate

operator-lifecycle-manager

kubernetes-nmstate-operator.4.18.0-202603041813

RequirementsNotMet

one or more requirements couldn't be found

cert-manager

replicaset-controller

cert-manager-cainjector-5545bd876

SuccessfulCreate

Created pod: cert-manager-cainjector-5545bd876-mtx97

openshift-nmstate

operator-lifecycle-manager

kubernetes-nmstate-operator.4.18.0-202603041813

RequirementsUnknown

requirements not yet checked

cert-manager

replicaset-controller

cert-manager-cainjector-5545bd876

SuccessfulCreate

Created pod: cert-manager-cainjector-5545bd876-mtx97

cert-manager

kubelet

cert-manager-cainjector-5545bd876-mtx97

Pulling

Pulling image "registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:903ce74138b1ffc735846a7c5fcdf62bbe82ca29568a6b38caec2656f6637671"

openshift-nmstate

operator-lifecycle-manager

kubernetes-nmstate-operator.4.18.0-202603041813

AllRequirementsMet

all requirements found, attempting install

cert-manager

multus

cert-manager-cainjector-5545bd876-mtx97

AddedInterface

Add eth0 [10.128.0.122/23] from ovn-kubernetes

openshift-nmstate

operator-lifecycle-manager

kubernetes-nmstate-operator.4.18.0-202603041813

AllRequirementsMet

all requirements found, attempting install

cert-manager

multus

cert-manager-cainjector-5545bd876-mtx97

AddedInterface

Add eth0 [10.128.0.122/23] from ovn-kubernetes

cert-manager

kubelet

cert-manager-cainjector-5545bd876-mtx97

Pulling

Pulling image "registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:903ce74138b1ffc735846a7c5fcdf62bbe82ca29568a6b38caec2656f6637671"

openshift-nmstate

replicaset-controller

nmstate-operator-796d4cfff4

SuccessfulCreate

Created pod: nmstate-operator-796d4cfff4-q2cgp
(x2)

openshift-nmstate

operator-lifecycle-manager

kubernetes-nmstate-operator.4.18.0-202603041813

InstallSucceeded

waiting for install components to report healthy

openshift-nmstate

multus

nmstate-operator-796d4cfff4-q2cgp

AddedInterface

Add eth0 [10.128.0.123/23] from ovn-kubernetes

openshift-nmstate

deployment-controller

nmstate-operator

ScalingReplicaSet

Scaled up replica set nmstate-operator-796d4cfff4 to 1

openshift-nmstate

multus

nmstate-operator-796d4cfff4-q2cgp

AddedInterface

Add eth0 [10.128.0.123/23] from ovn-kubernetes

openshift-nmstate

replicaset-controller

nmstate-operator-796d4cfff4

SuccessfulCreate

Created pod: nmstate-operator-796d4cfff4-q2cgp

openshift-nmstate

deployment-controller

nmstate-operator

ScalingReplicaSet

Scaled up replica set nmstate-operator-796d4cfff4 to 1
(x2)

openshift-nmstate

operator-lifecycle-manager

kubernetes-nmstate-operator.4.18.0-202603041813

InstallSucceeded

waiting for install components to report healthy
(x2)

openshift-nmstate

operator-lifecycle-manager

kubernetes-nmstate-operator.4.18.0-202603041813

InstallWaiting

installing: waiting for deployment nmstate-operator to become ready: deployment "nmstate-operator" not available: Deployment does not have minimum availability.
(x2)

openshift-nmstate

operator-lifecycle-manager

kubernetes-nmstate-operator.4.18.0-202603041813

InstallWaiting

installing: waiting for deployment nmstate-operator to become ready: deployment "nmstate-operator" not available: Deployment does not have minimum availability.
(x11)

cert-manager

replicaset-controller

cert-manager-545d4d4674

FailedCreate

Error creating: pods "cert-manager-545d4d4674-" is forbidden: error looking up service account cert-manager/cert-manager: serviceaccount "cert-manager" not found
(x11)

cert-manager

replicaset-controller

cert-manager-545d4d4674

FailedCreate

Error creating: pods "cert-manager-545d4d4674-" is forbidden: error looking up service account cert-manager/cert-manager: serviceaccount "cert-manager" not found

cert-manager

kubelet

cert-manager-webhook-6888856db4-kkc9m

Pulled

Successfully pulled image "registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:903ce74138b1ffc735846a7c5fcdf62bbe82ca29568a6b38caec2656f6637671" in 6.247s (6.247s including waiting). Image size: 319887149 bytes.

cert-manager

kubelet

cert-manager-cainjector-5545bd876-mtx97

Created

Created container: cert-manager-cainjector

cert-manager

kubelet

cert-manager-cainjector-5545bd876-mtx97

Pulled

Successfully pulled image "registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:903ce74138b1ffc735846a7c5fcdf62bbe82ca29568a6b38caec2656f6637671" in 3.985s (3.985s including waiting). Image size: 319887149 bytes.

cert-manager

kubelet

cert-manager-cainjector-5545bd876-mtx97

Created

Created container: cert-manager-cainjector

cert-manager

kubelet

cert-manager-cainjector-5545bd876-mtx97

Started

Started container cert-manager-cainjector

cert-manager

kubelet

cert-manager-webhook-6888856db4-kkc9m

Created

Created container: cert-manager-webhook

cert-manager

kubelet

cert-manager-webhook-6888856db4-kkc9m

Started

Started container cert-manager-webhook

cert-manager

kubelet

cert-manager-cainjector-5545bd876-mtx97

Pulled

Successfully pulled image "registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:903ce74138b1ffc735846a7c5fcdf62bbe82ca29568a6b38caec2656f6637671" in 3.985s (3.985s including waiting). Image size: 319887149 bytes.

cert-manager

kubelet

cert-manager-webhook-6888856db4-kkc9m

Pulled

Successfully pulled image "registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:903ce74138b1ffc735846a7c5fcdf62bbe82ca29568a6b38caec2656f6637671" in 6.247s (6.247s including waiting). Image size: 319887149 bytes.

cert-manager

kubelet

cert-manager-webhook-6888856db4-kkc9m

Created

Created container: cert-manager-webhook

cert-manager

kubelet

cert-manager-webhook-6888856db4-kkc9m

Started

Started container cert-manager-webhook

openshift-nmstate

kubelet

nmstate-operator-796d4cfff4-q2cgp

Pulling

Pulling image "registry.redhat.io/openshift4/kubernetes-nmstate-rhel9-operator@sha256:60ec3d3da1ba06551932e9ebf8f98bd2cdf5e18c0b4b05c124847b7672458094"

openshift-nmstate

kubelet

nmstate-operator-796d4cfff4-q2cgp

Pulling

Pulling image "registry.redhat.io/openshift4/kubernetes-nmstate-rhel9-operator@sha256:60ec3d3da1ba06551932e9ebf8f98bd2cdf5e18c0b4b05c124847b7672458094"

kube-system

cert-manager-cainjector-5545bd876-mtx97_ecc00c39-5491-4d69-a19c-4d36aeaccbf1

cert-manager-cainjector-leader-election

LeaderElection

cert-manager-cainjector-5545bd876-mtx97_ecc00c39-5491-4d69-a19c-4d36aeaccbf1 became leader

cert-manager

kubelet

cert-manager-cainjector-5545bd876-mtx97

Started

Started container cert-manager-cainjector

openshift-nmstate

kubelet

nmstate-operator-796d4cfff4-q2cgp

Pulled

Successfully pulled image "registry.redhat.io/openshift4/kubernetes-nmstate-rhel9-operator@sha256:60ec3d3da1ba06551932e9ebf8f98bd2cdf5e18c0b4b05c124847b7672458094" in 3.608s (3.608s including waiting). Image size: 451496534 bytes.

openshift-nmstate

kubelet

nmstate-operator-796d4cfff4-q2cgp

Pulled

Successfully pulled image "registry.redhat.io/openshift4/kubernetes-nmstate-rhel9-operator@sha256:60ec3d3da1ba06551932e9ebf8f98bd2cdf5e18c0b4b05c124847b7672458094" in 3.608s (3.608s including waiting). Image size: 451496534 bytes.

openshift-nmstate

kubelet

nmstate-operator-796d4cfff4-q2cgp

Started

Started container nmstate-operator

openshift-nmstate

kubelet

nmstate-operator-796d4cfff4-q2cgp

Created

Created container: nmstate-operator

openshift-operators

operator-lifecycle-manager

cluster-observability-operator.v1.4.0

RequirementsUnknown

requirements not yet checked

openshift-operators

operator-lifecycle-manager

cluster-observability-operator.v1.4.0

RequirementsUnknown

requirements not yet checked

openshift-nmstate

kubelet

nmstate-operator-796d4cfff4-q2cgp

Started

Started container nmstate-operator

openshift-nmstate

kubelet

nmstate-operator-796d4cfff4-q2cgp

Created

Created container: nmstate-operator

cert-manager

replicaset-controller

cert-manager-545d4d4674

SuccessfulCreate

Created pod: cert-manager-545d4d4674-6ncn2
(x2)

openshift-operators

controllermanager

obo-prometheus-operator-admission-webhook

NoPods

No matching pods found

cert-manager

replicaset-controller

cert-manager-545d4d4674

SuccessfulCreate

Created pod: cert-manager-545d4d4674-6ncn2

openshift-operators

operator-lifecycle-manager

cluster-observability-operator.v1.4.0

RequirementsNotMet

one or more requirements couldn't be found

openshift-operators

operator-lifecycle-manager

cluster-observability-operator.v1.4.0

RequirementsNotMet

one or more requirements couldn't be found
(x2)

openshift-operators

controllermanager

obo-prometheus-operator-admission-webhook

NoPods

No matching pods found

cert-manager

kubelet

cert-manager-545d4d4674-6ncn2

Pulled

Container image "registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:903ce74138b1ffc735846a7c5fcdf62bbe82ca29568a6b38caec2656f6637671" already present on machine

cert-manager

multus

cert-manager-545d4d4674-6ncn2

AddedInterface

Add eth0 [10.128.0.124/23] from ovn-kubernetes

metallb-system

replicaset-controller

metallb-operator-controller-manager-64ff997757

SuccessfulCreate

Created pod: metallb-operator-controller-manager-64ff997757-xgs48

cert-manager

kubelet

cert-manager-545d4d4674-6ncn2

Pulled

Container image "registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:903ce74138b1ffc735846a7c5fcdf62bbe82ca29568a6b38caec2656f6637671" already present on machine

metallb-system

deployment-controller

metallb-operator-controller-manager

ScalingReplicaSet

Scaled up replica set metallb-operator-controller-manager-64ff997757 to 1

openshift-nmstate

operator-lifecycle-manager

kubernetes-nmstate-operator.4.18.0-202603041813

InstallSucceeded

install strategy completed with no errors

metallb-system

deployment-controller

metallb-operator-webhook-server

ScalingReplicaSet

Scaled up replica set metallb-operator-webhook-server-6dc87c8bf8 to 1

metallb-system

deployment-controller

metallb-operator-webhook-server

ScalingReplicaSet

Scaled up replica set metallb-operator-webhook-server-6dc87c8bf8 to 1

cert-manager

multus

cert-manager-545d4d4674-6ncn2

AddedInterface

Add eth0 [10.128.0.124/23] from ovn-kubernetes

metallb-system

deployment-controller

metallb-operator-controller-manager

ScalingReplicaSet

Scaled up replica set metallb-operator-controller-manager-64ff997757 to 1

metallb-system

replicaset-controller

metallb-operator-controller-manager-64ff997757

SuccessfulCreate

Created pod: metallb-operator-controller-manager-64ff997757-xgs48

openshift-nmstate

operator-lifecycle-manager

kubernetes-nmstate-operator.4.18.0-202603041813

InstallSucceeded

install strategy completed with no errors

cert-manager

kubelet

cert-manager-545d4d4674-6ncn2

Started

Started container cert-manager-controller

metallb-system

multus

metallb-operator-controller-manager-64ff997757-xgs48

AddedInterface

Add eth0 [10.128.0.125/23] from ovn-kubernetes

cert-manager

kubelet

cert-manager-545d4d4674-6ncn2

Created

Created container: cert-manager-controller

cert-manager

kubelet

cert-manager-545d4d4674-6ncn2

Started

Started container cert-manager-controller

metallb-system

replicaset-controller

metallb-operator-webhook-server-6dc87c8bf8

SuccessfulCreate

Created pod: metallb-operator-webhook-server-6dc87c8bf8-srnwn

metallb-system

replicaset-controller

metallb-operator-webhook-server-6dc87c8bf8

SuccessfulCreate

Created pod: metallb-operator-webhook-server-6dc87c8bf8-srnwn

metallb-system

kubelet

metallb-operator-controller-manager-64ff997757-xgs48

Pulling

Pulling image "registry.redhat.io/openshift4/metallb-rhel9-operator@sha256:9d74242d31d5f83bb8207d71e2a766ce9ababf218795d5c6fbb50450af5c29e8"

metallb-system

multus

metallb-operator-controller-manager-64ff997757-xgs48

AddedInterface

Add eth0 [10.128.0.125/23] from ovn-kubernetes

metallb-system

kubelet

metallb-operator-controller-manager-64ff997757-xgs48

Pulling

Pulling image "registry.redhat.io/openshift4/metallb-rhel9-operator@sha256:9d74242d31d5f83bb8207d71e2a766ce9ababf218795d5c6fbb50450af5c29e8"

cert-manager

kubelet

cert-manager-545d4d4674-6ncn2

Created

Created container: cert-manager-controller

metallb-system

kubelet

metallb-operator-webhook-server-6dc87c8bf8-srnwn

Pulling

Pulling image "registry.redhat.io/openshift4/metallb-rhel9@sha256:2db2c546af02ea7593f9c55d648f055c042800b55e3bfa13f7f43029aa9c6592"

metallb-system

kubelet

metallb-operator-webhook-server-6dc87c8bf8-srnwn

Pulling

Pulling image "registry.redhat.io/openshift4/metallb-rhel9@sha256:2db2c546af02ea7593f9c55d648f055c042800b55e3bfa13f7f43029aa9c6592"

metallb-system

multus

metallb-operator-webhook-server-6dc87c8bf8-srnwn

AddedInterface

Add eth0 [10.128.0.126/23] from ovn-kubernetes

metallb-system

multus

metallb-operator-webhook-server-6dc87c8bf8-srnwn

AddedInterface

Add eth0 [10.128.0.126/23] from ovn-kubernetes

metallb-system

kubelet

metallb-operator-controller-manager-64ff997757-xgs48

Pulled

Successfully pulled image "registry.redhat.io/openshift4/metallb-rhel9-operator@sha256:9d74242d31d5f83bb8207d71e2a766ce9ababf218795d5c6fbb50450af5c29e8" in 5.323s (5.323s including waiting). Image size: 462537291 bytes.

metallb-system

kubelet

metallb-operator-controller-manager-64ff997757-xgs48

Pulled

Successfully pulled image "registry.redhat.io/openshift4/metallb-rhel9-operator@sha256:9d74242d31d5f83bb8207d71e2a766ce9ababf218795d5c6fbb50450af5c29e8" in 5.323s (5.323s including waiting). Image size: 462537291 bytes.

metallb-system

metallb-operator-controller-manager-64ff997757-xgs48_e019798f-0b11-42cb-ba93-5097eb484906

metallb.io.metallboperator

LeaderElection

metallb-operator-controller-manager-64ff997757-xgs48_e019798f-0b11-42cb-ba93-5097eb484906 became leader

metallb-system

metallb-operator-controller-manager-64ff997757-xgs48_e019798f-0b11-42cb-ba93-5097eb484906

metallb.io.metallboperator

LeaderElection

metallb-operator-controller-manager-64ff997757-xgs48_e019798f-0b11-42cb-ba93-5097eb484906 became leader

openshift-operators

operator-lifecycle-manager

cluster-observability-operator.v1.4.0

AllRequirementsMet

all requirements found, attempting install

openshift-operators

operator-lifecycle-manager

cluster-observability-operator.v1.4.0

AllRequirementsMet

all requirements found, attempting install

metallb-system

kubelet

metallb-operator-webhook-server-6dc87c8bf8-srnwn

Started

Started container webhook-server

openshift-operators

deployment-controller

obo-prometheus-operator

ScalingReplicaSet

Scaled up replica set obo-prometheus-operator-8ff7d675 to 1

openshift-operators

replicaset-controller

obo-prometheus-operator-8ff7d675

SuccessfulCreate

Created pod: obo-prometheus-operator-8ff7d675-wvbr6

openshift-operators

replicaset-controller

obo-prometheus-operator-8ff7d675

SuccessfulCreate

Created pod: obo-prometheus-operator-8ff7d675-wvbr6

metallb-system

kubelet

metallb-operator-webhook-server-6dc87c8bf8-srnwn

Created

Created container: webhook-server

metallb-system

kubelet

metallb-operator-webhook-server-6dc87c8bf8-srnwn

Pulled

Successfully pulled image "registry.redhat.io/openshift4/metallb-rhel9@sha256:2db2c546af02ea7593f9c55d648f055c042800b55e3bfa13f7f43029aa9c6592" in 7.079s (7.079s including waiting). Image size: 555122396 bytes.

metallb-system

kubelet

metallb-operator-webhook-server-6dc87c8bf8-srnwn

Pulled

Successfully pulled image "registry.redhat.io/openshift4/metallb-rhel9@sha256:2db2c546af02ea7593f9c55d648f055c042800b55e3bfa13f7f43029aa9c6592" in 7.079s (7.079s including waiting). Image size: 555122396 bytes.

metallb-system

kubelet

metallb-operator-webhook-server-6dc87c8bf8-srnwn

Created

Created container: webhook-server

metallb-system

kubelet

metallb-operator-webhook-server-6dc87c8bf8-srnwn

Started

Started container webhook-server

openshift-operators

deployment-controller

obo-prometheus-operator

ScalingReplicaSet

Scaled up replica set obo-prometheus-operator-8ff7d675 to 1

openshift-operators

multus

obo-prometheus-operator-8ff7d675-wvbr6

AddedInterface

Add eth0 [10.128.0.127/23] from ovn-kubernetes

openshift-operators

kubelet

obo-prometheus-operator-8ff7d675-wvbr6

Pulling

Pulling image "registry.redhat.io/cluster-observability-operator/obo-prometheus-rhel9-operator@sha256:161082f81c8c77471a421b3b4bcb8a47ca64aa08a5dd1abf27e7f2f964b35a2a"

openshift-operators

multus

obo-prometheus-operator-8ff7d675-wvbr6

AddedInterface

Add eth0 [10.128.0.127/23] from ovn-kubernetes

openshift-operators

kubelet

obo-prometheus-operator-8ff7d675-wvbr6

Pulling

Pulling image "registry.redhat.io/cluster-observability-operator/obo-prometheus-rhel9-operator@sha256:161082f81c8c77471a421b3b4bcb8a47ca64aa08a5dd1abf27e7f2f964b35a2a"

openshift-operators

replicaset-controller

observability-operator-6dd7dd855f

SuccessfulCreate

Created pod: observability-operator-6dd7dd855f-m27sp

openshift-operators

replicaset-controller

obo-prometheus-operator-admission-webhook-6976bc9f87

SuccessfulCreate

Created pod: obo-prometheus-operator-admission-webhook-6976bc9f87-xwq6f

openshift-operators

replicaset-controller

obo-prometheus-operator-admission-webhook-6976bc9f87

SuccessfulCreate

Created pod: obo-prometheus-operator-admission-webhook-6976bc9f87-ngfr6

openshift-operators

deployment-controller

observability-operator

ScalingReplicaSet

Scaled up replica set observability-operator-6dd7dd855f to 1

openshift-operators

replicaset-controller

observability-operator-6dd7dd855f

SuccessfulCreate

Created pod: observability-operator-6dd7dd855f-m27sp

openshift-operators

replicaset-controller

obo-prometheus-operator-admission-webhook-6976bc9f87

SuccessfulCreate

Created pod: obo-prometheus-operator-admission-webhook-6976bc9f87-ngfr6

openshift-operators

deployment-controller

observability-operator

ScalingReplicaSet

Scaled up replica set observability-operator-6dd7dd855f to 1

openshift-operators

deployment-controller

obo-prometheus-operator-admission-webhook

ScalingReplicaSet

Scaled up replica set obo-prometheus-operator-admission-webhook-6976bc9f87 to 2

openshift-operators

deployment-controller

obo-prometheus-operator-admission-webhook

ScalingReplicaSet

Scaled up replica set obo-prometheus-operator-admission-webhook-6976bc9f87 to 2

openshift-operators

replicaset-controller

obo-prometheus-operator-admission-webhook-6976bc9f87

SuccessfulCreate

Created pod: obo-prometheus-operator-admission-webhook-6976bc9f87-xwq6f

openshift-operators

kubelet

obo-prometheus-operator-admission-webhook-6976bc9f87-xwq6f

Pulling

Pulling image "registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:fb1030480e5a55ead0d9748615a2e4b9228522f14b77a782f44407883c24ba93"

openshift-operators

multus

obo-prometheus-operator-admission-webhook-6976bc9f87-xwq6f

AddedInterface

Add eth0 [10.128.0.129/23] from ovn-kubernetes

openshift-operators

replicaset-controller

perses-operator-777f7bbbbb

SuccessfulCreate

Created pod: perses-operator-777f7bbbbb-d5qjm

openshift-operators

multus

obo-prometheus-operator-admission-webhook-6976bc9f87-ngfr6

AddedInterface

Add eth0 [10.128.0.128/23] from ovn-kubernetes

openshift-operators

deployment-controller

perses-operator

ScalingReplicaSet

Scaled up replica set perses-operator-777f7bbbbb to 1

openshift-operators

replicaset-controller

perses-operator-777f7bbbbb

SuccessfulCreate

Created pod: perses-operator-777f7bbbbb-d5qjm

openshift-operators

multus

obo-prometheus-operator-admission-webhook-6976bc9f87-xwq6f

AddedInterface

Add eth0 [10.128.0.129/23] from ovn-kubernetes

openshift-operators

kubelet

obo-prometheus-operator-admission-webhook-6976bc9f87-ngfr6

Pulling

Pulling image "registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:fb1030480e5a55ead0d9748615a2e4b9228522f14b77a782f44407883c24ba93"

openshift-operators

kubelet

obo-prometheus-operator-admission-webhook-6976bc9f87-ngfr6

Pulling

Pulling image "registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:fb1030480e5a55ead0d9748615a2e4b9228522f14b77a782f44407883c24ba93"

openshift-operators

kubelet

observability-operator-6dd7dd855f-m27sp

Pulling

Pulling image "registry.redhat.io/cluster-observability-operator/cluster-observability-rhel9-operator@sha256:29ffc7689432fad53f18d3e12a1b335b69d49dbdcb7d8b4a77078bc7f79f941f"

openshift-operators

multus

obo-prometheus-operator-admission-webhook-6976bc9f87-ngfr6

AddedInterface

Add eth0 [10.128.0.128/23] from ovn-kubernetes

openshift-operators

multus

observability-operator-6dd7dd855f-m27sp

AddedInterface

Add eth0 [10.128.0.130/23] from ovn-kubernetes

openshift-operators

multus

observability-operator-6dd7dd855f-m27sp

AddedInterface

Add eth0 [10.128.0.130/23] from ovn-kubernetes

openshift-operators

kubelet

obo-prometheus-operator-admission-webhook-6976bc9f87-xwq6f

Pulling

Pulling image "registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:fb1030480e5a55ead0d9748615a2e4b9228522f14b77a782f44407883c24ba93"

openshift-operators

kubelet

observability-operator-6dd7dd855f-m27sp

Pulling

Pulling image "registry.redhat.io/cluster-observability-operator/cluster-observability-rhel9-operator@sha256:29ffc7689432fad53f18d3e12a1b335b69d49dbdcb7d8b4a77078bc7f79f941f"

openshift-operators

deployment-controller

perses-operator

ScalingReplicaSet

Scaled up replica set perses-operator-777f7bbbbb to 1

openshift-operators

multus

perses-operator-777f7bbbbb-d5qjm

AddedInterface

Add eth0 [10.128.0.131/23] from ovn-kubernetes

openshift-operators

multus

perses-operator-777f7bbbbb-d5qjm

AddedInterface

Add eth0 [10.128.0.131/23] from ovn-kubernetes

openshift-operators

kubelet

perses-operator-777f7bbbbb-d5qjm

Pulling

Pulling image "registry.redhat.io/cluster-observability-operator/perses-rhel9-operator@sha256:f78b160ba3b815f53d6a72425f3f3a9d7946795177bd68c7c614fa84f97be630"

openshift-operators

kubelet

perses-operator-777f7bbbbb-d5qjm

Pulling

Pulling image "registry.redhat.io/cluster-observability-operator/perses-rhel9-operator@sha256:f78b160ba3b815f53d6a72425f3f3a9d7946795177bd68c7c614fa84f97be630"
(x2)

metallb-system

operator-lifecycle-manager

install-n68kq

AppliedWithWarnings

1 warning(s) generated during installation of operator "metallb-operator.v4.18.0-202603040208" (CustomResourceDefinition "bgppeers.metallb.io"): v1beta1 is deprecated, please use v1beta2
(x2)

metallb-system

operator-lifecycle-manager

install-n68kq

AppliedWithWarnings

1 warning(s) generated during installation of operator "metallb-operator.v4.18.0-202603040208" (CustomResourceDefinition "bgppeers.metallb.io"): v1beta1 is deprecated, please use v1beta2

metallb-system

operator-lifecycle-manager

metallb-operator.v4.18.0-202603040208

NeedsReinstall

calculated deployment install is bad

metallb-system

operator-lifecycle-manager

metallb-operator.v4.18.0-202603040208

NeedsReinstall

calculated deployment install is bad
(x2)

openshift-operators

operator-lifecycle-manager

cluster-observability-operator.v1.4.0

InstallWaiting

installing: waiting for deployment obo-prometheus-operator to become ready: deployment "obo-prometheus-operator" not available: Deployment does not have minimum availability.
(x2)

openshift-operators

operator-lifecycle-manager

cluster-observability-operator.v1.4.0

InstallWaiting

installing: waiting for deployment obo-prometheus-operator to become ready: deployment "obo-prometheus-operator" not available: Deployment does not have minimum availability.
(x2)

metallb-system

operator-lifecycle-manager

metallb-operator.v4.18.0-202603040208

InstallWaiting

installing: waiting for deployment metallb-operator-controller-manager to become ready: deployment "metallb-operator-controller-manager" not available: Deployment does not have minimum availability.
(x2)

metallb-system

operator-lifecycle-manager

metallb-operator.v4.18.0-202603040208

InstallWaiting

installing: waiting for deployment metallb-operator-controller-manager to become ready: deployment "metallb-operator-controller-manager" not available: Deployment does not have minimum availability.

kube-system

cert-manager-leader-election

cert-manager-controller

LeaderElection

cert-manager-545d4d4674-6ncn2-external-cert-manager-controller became leader

openshift-operators

kubelet

perses-operator-777f7bbbbb-d5qjm

Pulled

Successfully pulled image "registry.redhat.io/cluster-observability-operator/perses-rhel9-operator@sha256:f78b160ba3b815f53d6a72425f3f3a9d7946795177bd68c7c614fa84f97be630" in 10.345s (10.345s including waiting). Image size: 175801363 bytes.

openshift-operators

kubelet

observability-operator-6dd7dd855f-m27sp

Pulled

Successfully pulled image "registry.redhat.io/cluster-observability-operator/cluster-observability-rhel9-operator@sha256:29ffc7689432fad53f18d3e12a1b335b69d49dbdcb7d8b4a77078bc7f79f941f" in 11.066s (11.066s including waiting). Image size: 343063302 bytes.

openshift-operators

kubelet

observability-operator-6dd7dd855f-m27sp

Pulled

Successfully pulled image "registry.redhat.io/cluster-observability-operator/cluster-observability-rhel9-operator@sha256:29ffc7689432fad53f18d3e12a1b335b69d49dbdcb7d8b4a77078bc7f79f941f" in 11.066s (11.066s including waiting). Image size: 343063302 bytes.
(x2)

openshift-operators

operator-lifecycle-manager

install-c7kwm

AppliedWithWarnings

1 warning(s) generated during installation of operator "cluster-observability-operator.v1.4.0" (CustomResourceDefinition "persesdashboards.perses.dev"): perses.dev/v1alpha1 is deprecated; use perses.dev/v1alpha2

openshift-operators

kubelet

obo-prometheus-operator-8ff7d675-wvbr6

Pulled

Successfully pulled image "registry.redhat.io/cluster-observability-operator/obo-prometheus-rhel9-operator@sha256:161082f81c8c77471a421b3b4bcb8a47ca64aa08a5dd1abf27e7f2f964b35a2a" in 12.659s (12.659s including waiting). Image size: 204104155 bytes.

openshift-operators

kubelet

obo-prometheus-operator-admission-webhook-6976bc9f87-ngfr6

Pulled

Successfully pulled image "registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:fb1030480e5a55ead0d9748615a2e4b9228522f14b77a782f44407883c24ba93" in 11.26s (11.26s including waiting). Image size: 151317463 bytes.
(x2)

openshift-operators

operator-lifecycle-manager

install-c7kwm

AppliedWithWarnings

1 warning(s) generated during installation of operator "cluster-observability-operator.v1.4.0" (CustomResourceDefinition "persesdashboards.perses.dev"): perses.dev/v1alpha1 is deprecated; use perses.dev/v1alpha2

openshift-operators

kubelet

obo-prometheus-operator-8ff7d675-wvbr6

Pulled

Successfully pulled image "registry.redhat.io/cluster-observability-operator/obo-prometheus-rhel9-operator@sha256:161082f81c8c77471a421b3b4bcb8a47ca64aa08a5dd1abf27e7f2f964b35a2a" in 12.659s (12.659s including waiting). Image size: 204104155 bytes.

openshift-operators

kubelet

obo-prometheus-operator-admission-webhook-6976bc9f87-xwq6f

Pulled

Successfully pulled image "registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:fb1030480e5a55ead0d9748615a2e4b9228522f14b77a782f44407883c24ba93" in 11.1s (11.1s including waiting). Image size: 151317463 bytes.

openshift-operators

kubelet

obo-prometheus-operator-admission-webhook-6976bc9f87-ngfr6

Pulled

Successfully pulled image "registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:fb1030480e5a55ead0d9748615a2e4b9228522f14b77a782f44407883c24ba93" in 11.26s (11.26s including waiting). Image size: 151317463 bytes.

openshift-operators

kubelet

obo-prometheus-operator-admission-webhook-6976bc9f87-xwq6f

Pulled

Successfully pulled image "registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:fb1030480e5a55ead0d9748615a2e4b9228522f14b77a782f44407883c24ba93" in 11.1s (11.1s including waiting). Image size: 151317463 bytes.

openshift-operators

kubelet

perses-operator-777f7bbbbb-d5qjm

Pulled

Successfully pulled image "registry.redhat.io/cluster-observability-operator/perses-rhel9-operator@sha256:f78b160ba3b815f53d6a72425f3f3a9d7946795177bd68c7c614fa84f97be630" in 10.345s (10.345s including waiting). Image size: 175801363 bytes.

openshift-operators

kubelet

obo-prometheus-operator-8ff7d675-wvbr6

Created

Created container: prometheus-operator

openshift-operators

kubelet

obo-prometheus-operator-admission-webhook-6976bc9f87-xwq6f

Started

Started container prometheus-operator-admission-webhook

openshift-operators

kubelet

observability-operator-6dd7dd855f-m27sp

Started

Started container operator

openshift-operators

kubelet

obo-prometheus-operator-8ff7d675-wvbr6

Started

Started container prometheus-operator

openshift-operators

kubelet

obo-prometheus-operator-admission-webhook-6976bc9f87-xwq6f

Created

Created container: prometheus-operator-admission-webhook

openshift-operators

kubelet

observability-operator-6dd7dd855f-m27sp

Created

Created container: operator

openshift-operators

kubelet

obo-prometheus-operator-admission-webhook-6976bc9f87-ngfr6

Created

Created container: prometheus-operator-admission-webhook

openshift-operators

kubelet

observability-operator-6dd7dd855f-m27sp

Started

Started container operator

openshift-operators

kubelet

observability-operator-6dd7dd855f-m27sp

Created

Created container: operator

openshift-operators

kubelet

obo-prometheus-operator-admission-webhook-6976bc9f87-ngfr6

Created

Created container: prometheus-operator-admission-webhook

openshift-operators

kubelet

perses-operator-777f7bbbbb-d5qjm

Started

Started container perses-operator

openshift-operators

kubelet

obo-prometheus-operator-admission-webhook-6976bc9f87-ngfr6

Started

Started container prometheus-operator-admission-webhook

openshift-operators

kubelet

obo-prometheus-operator-8ff7d675-wvbr6

Started

Started container prometheus-operator

openshift-operators

kubelet

perses-operator-777f7bbbbb-d5qjm

Started

Started container perses-operator

openshift-operators

kubelet

perses-operator-777f7bbbbb-d5qjm

Created

Created container: perses-operator

openshift-operators

kubelet

perses-operator-777f7bbbbb-d5qjm

Created

Created container: perses-operator

openshift-operators

kubelet

obo-prometheus-operator-admission-webhook-6976bc9f87-ngfr6

Started

Started container prometheus-operator-admission-webhook

openshift-operators

kubelet

obo-prometheus-operator-8ff7d675-wvbr6

Created

Created container: prometheus-operator

openshift-operators

kubelet

obo-prometheus-operator-admission-webhook-6976bc9f87-xwq6f

Started

Started container prometheus-operator-admission-webhook

openshift-operators

kubelet

obo-prometheus-operator-admission-webhook-6976bc9f87-xwq6f

Created

Created container: prometheus-operator-admission-webhook
(x2)

openshift-operators

operator-lifecycle-manager

install-c7kwm

AppliedWithWarnings

1 warning(s) generated during installation of operator "cluster-observability-operator.v1.4.0" (CustomResourceDefinition "perses.perses.dev"): perses.dev/v1alpha1 is deprecated; use perses.dev/v1alpha2
(x2)

openshift-operators

operator-lifecycle-manager

install-c7kwm

AppliedWithWarnings

1 warning(s) generated during installation of operator "cluster-observability-operator.v1.4.0" (CustomResourceDefinition "perses.perses.dev"): perses.dev/v1alpha1 is deprecated; use perses.dev/v1alpha2

openshift-operators

operator-lifecycle-manager

cluster-observability-operator.v1.4.0

InstallWaiting

installing: waiting for deployment perses-operator to become ready: deployment "perses-operator" not available: Deployment does not have minimum availability.

openshift-operators

operator-lifecycle-manager

cluster-observability-operator.v1.4.0

InstallWaiting

installing: waiting for deployment perses-operator to become ready: deployment "perses-operator" not available: Deployment does not have minimum availability.
(x3)

openshift-operators

operator-lifecycle-manager

cluster-observability-operator.v1.4.0

InstallWaiting

Webhook install failed: conversionWebhook not ready
(x3)

openshift-operators

operator-lifecycle-manager

cluster-observability-operator.v1.4.0

InstallWaiting

Webhook install failed: conversionWebhook not ready
(x2)

openshift-operators

operator-lifecycle-manager

install-c7kwm

AppliedWithWarnings

1 warning(s) generated during installation of operator "cluster-observability-operator.v1.4.0" (CustomResourceDefinition "persesdatasources.perses.dev"): perses.dev/v1alpha1 is deprecated; use perses.dev/v1alpha2
(x4)

openshift-operators

operator-lifecycle-manager

cluster-observability-operator.v1.4.0

InstallSucceeded

waiting for install components to report healthy
(x4)

openshift-operators

operator-lifecycle-manager

cluster-observability-operator.v1.4.0

InstallSucceeded

waiting for install components to report healthy
(x2)

openshift-operators

operator-lifecycle-manager

install-c7kwm

AppliedWithWarnings

1 warning(s) generated during installation of operator "cluster-observability-operator.v1.4.0" (CustomResourceDefinition "persesdatasources.perses.dev"): perses.dev/v1alpha1 is deprecated; use perses.dev/v1alpha2

openshift-operators

operator-lifecycle-manager

cluster-observability-operator.v1.4.0

InstallSucceeded

install strategy completed with no errors

openshift-operators

operator-lifecycle-manager

cluster-observability-operator.v1.4.0

InstallSucceeded

install strategy completed with no errors

metallb-system

kubelet

frr-k8s-pp5nf

FailedMount

MountVolume.SetUp failed for volume "metrics-certs" : secret "frr-k8s-certs-secret" not found

default

garbage-collector-controller

frr-k8s-validating-webhook-configuration

OwnerRefInvalidNamespace

ownerRef [metallb.io/v1beta1/MetalLB, namespace: , name: metallb, uid: 1c3f6435-b6df-436c-b702-585794b6aeb6] does not exist in namespace ""

metallb-system

replicaset-controller

frr-k8s-webhook-server-bcc4b6f68

SuccessfulCreate

Created pod: frr-k8s-webhook-server-bcc4b6f68-lrv7p

metallb-system

deployment-controller

frr-k8s-webhook-server

ScalingReplicaSet

Scaled up replica set frr-k8s-webhook-server-bcc4b6f68 to 1

metallb-system

replicaset-controller

controller-7bb4cc7c98

SuccessfulCreate

Created pod: controller-7bb4cc7c98-77wvd

metallb-system

kubelet

speaker-6cndp

FailedMount

MountVolume.SetUp failed for volume "metrics-certs" : secret "speaker-certs-secret" not found

metallb-system

kubelet

controller-7bb4cc7c98-77wvd

FailedMount

MountVolume.SetUp failed for volume "metrics-certs" : secret "controller-certs-secret" not found

metallb-system

daemonset-controller

speaker

SuccessfulCreate

Created pod: speaker-6cndp

metallb-system

deployment-controller

controller

ScalingReplicaSet

Scaled up replica set controller-7bb4cc7c98 to 1

metallb-system

kubelet

controller-7bb4cc7c98-77wvd

FailedMount

MountVolume.SetUp failed for volume "metrics-certs" : secret "controller-certs-secret" not found

metallb-system

replicaset-controller

frr-k8s-webhook-server-bcc4b6f68

SuccessfulCreate

Created pod: frr-k8s-webhook-server-bcc4b6f68-lrv7p

metallb-system

replicaset-controller

controller-7bb4cc7c98

SuccessfulCreate

Created pod: controller-7bb4cc7c98-77wvd

metallb-system

deployment-controller

controller

ScalingReplicaSet

Scaled up replica set controller-7bb4cc7c98 to 1

metallb-system

daemonset-controller

speaker

SuccessfulCreate

Created pod: speaker-6cndp

metallb-system

kubelet

frr-k8s-pp5nf

FailedMount

MountVolume.SetUp failed for volume "metrics-certs" : secret "frr-k8s-certs-secret" not found

metallb-system

daemonset-controller

frr-k8s

SuccessfulCreate

Created pod: frr-k8s-pp5nf

metallb-system

daemonset-controller

frr-k8s

SuccessfulCreate

Created pod: frr-k8s-pp5nf

metallb-system

kubelet

speaker-6cndp

FailedMount

MountVolume.SetUp failed for volume "metrics-certs" : secret "speaker-certs-secret" not found

metallb-system

deployment-controller

frr-k8s-webhook-server

ScalingReplicaSet

Scaled up replica set frr-k8s-webhook-server-bcc4b6f68 to 1

metallb-system

kubelet

controller-7bb4cc7c98-77wvd

Pulled

Container image "registry.redhat.io/openshift4/metallb-rhel9@sha256:2db2c546af02ea7593f9c55d648f055c042800b55e3bfa13f7f43029aa9c6592" already present on machine

metallb-system

multus

frr-k8s-webhook-server-bcc4b6f68-lrv7p

AddedInterface

Add eth0 [10.128.0.132/23] from ovn-kubernetes

metallb-system

multus

frr-k8s-webhook-server-bcc4b6f68-lrv7p

AddedInterface

Add eth0 [10.128.0.132/23] from ovn-kubernetes

metallb-system

kubelet

frr-k8s-webhook-server-bcc4b6f68-lrv7p

Pulling

Pulling image "registry.redhat.io/openshift4/frr-rhel9@sha256:2157d8b664937a8c3871c12e9a4ee90e7da1a3db2b240bdd320b5dc619b9b8d4"

metallb-system

kubelet

frr-k8s-pp5nf

Pulling

Pulling image "registry.redhat.io/openshift4/frr-rhel9@sha256:2157d8b664937a8c3871c12e9a4ee90e7da1a3db2b240bdd320b5dc619b9b8d4"

metallb-system

kubelet

frr-k8s-pp5nf

Pulling

Pulling image "registry.redhat.io/openshift4/frr-rhel9@sha256:2157d8b664937a8c3871c12e9a4ee90e7da1a3db2b240bdd320b5dc619b9b8d4"

metallb-system

kubelet

frr-k8s-webhook-server-bcc4b6f68-lrv7p

Pulling

Pulling image "registry.redhat.io/openshift4/frr-rhel9@sha256:2157d8b664937a8c3871c12e9a4ee90e7da1a3db2b240bdd320b5dc619b9b8d4"

metallb-system

multus

controller-7bb4cc7c98-77wvd

AddedInterface

Add eth0 [10.128.0.133/23] from ovn-kubernetes

metallb-system

multus

controller-7bb4cc7c98-77wvd

AddedInterface

Add eth0 [10.128.0.133/23] from ovn-kubernetes

metallb-system

kubelet

controller-7bb4cc7c98-77wvd

Pulled

Container image "registry.redhat.io/openshift4/metallb-rhel9@sha256:2db2c546af02ea7593f9c55d648f055c042800b55e3bfa13f7f43029aa9c6592" already present on machine

metallb-system

kubelet

controller-7bb4cc7c98-77wvd

Pulling

Pulling image "registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:4fdd6da66aba2523d2c21cef306b7650659926bbadb96dedd000d2b8c0229078"

openshift-nmstate

deployment-controller

nmstate-console-plugin

ScalingReplicaSet

Scaled up replica set nmstate-console-plugin-86f58fcf4 to 1

openshift-console-operator

console-operator-console-operator-consoleoperator

console-operator

ConfigMapUpdated

Updated ConfigMap/console-config -n openshift-console: cause by changes in data.console-config.yaml

openshift-nmstate

deployment-controller

nmstate-webhook

ScalingReplicaSet

Scaled up replica set nmstate-webhook-5f558f5558 to 1

openshift-nmstate

replicaset-controller

nmstate-webhook-5f558f5558

SuccessfulCreate

Created pod: nmstate-webhook-5f558f5558-w8sq6

openshift-nmstate

kubelet

nmstate-webhook-5f558f5558-w8sq6

FailedMount

MountVolume.SetUp failed for volume "tls-key-pair" : secret "openshift-nmstate-webhook" not found
(x4)

openshift-console-operator

console-operator-console-operator-consoleoperator

console-operator

DeploymentUpdated

Updated Deployment.apps/console -n openshift-console because it changed
(x3)

metallb-system

kubelet

speaker-6cndp

FailedMount

MountVolume.SetUp failed for volume "memberlist" : secret "metallb-memberlist" not found

openshift-nmstate

deployment-controller

nmstate-metrics

ScalingReplicaSet

Scaled up replica set nmstate-metrics-9b8c8685d to 1

openshift-console-operator

console-operator-status-controller-statussyncer_console

console-operator

OperatorStatusChanged

Status for clusteroperator/console changed: Progressing changed from False to True ("SyncLoopRefreshProgressing: changes made during sync updates, additional sync expected")

openshift-nmstate

replicaset-controller

nmstate-metrics-9b8c8685d

SuccessfulCreate

Created pod: nmstate-metrics-9b8c8685d-p28fn

openshift-console

deployment-controller

console

ScalingReplicaSet

Scaled up replica set console-5df6fc5897 to 1

openshift-console

replicaset-controller

console-5df6fc5897

SuccessfulCreate

Created pod: console-5df6fc5897-d9ws4

openshift-nmstate

daemonset-controller

nmstate-handler

SuccessfulCreate

Created pod: nmstate-handler-8ffsp

openshift-nmstate

kubelet

nmstate-webhook-5f558f5558-w8sq6

FailedMount

MountVolume.SetUp failed for volume "tls-key-pair" : secret "openshift-nmstate-webhook" not found

openshift-nmstate

kubelet

nmstate-console-plugin-86f58fcf4-5wfxh

FailedMount

MountVolume.SetUp failed for volume "plugin-serving-cert" : secret "plugin-serving-cert" not found

metallb-system

kubelet

controller-7bb4cc7c98-77wvd

Pulling

Pulling image "registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:4fdd6da66aba2523d2c21cef306b7650659926bbadb96dedd000d2b8c0229078"

metallb-system

kubelet

controller-7bb4cc7c98-77wvd

Started

Started container controller

metallb-system

kubelet

controller-7bb4cc7c98-77wvd

Created

Created container: controller

metallb-system

kubelet

controller-7bb4cc7c98-77wvd

Created

Created container: controller

metallb-system

kubelet

controller-7bb4cc7c98-77wvd

Started

Started container controller

openshift-nmstate

deployment-controller

nmstate-metrics

ScalingReplicaSet

Scaled up replica set nmstate-metrics-9b8c8685d to 1

openshift-nmstate

replicaset-controller

nmstate-metrics-9b8c8685d

SuccessfulCreate

Created pod: nmstate-metrics-9b8c8685d-p28fn

openshift-nmstate

deployment-controller

nmstate-webhook

ScalingReplicaSet

Scaled up replica set nmstate-webhook-5f558f5558 to 1

openshift-nmstate

replicaset-controller

nmstate-webhook-5f558f5558

SuccessfulCreate

Created pod: nmstate-webhook-5f558f5558-w8sq6
(x3)

metallb-system

kubelet

speaker-6cndp

FailedMount

MountVolume.SetUp failed for volume "memberlist" : secret "metallb-memberlist" not found

openshift-nmstate

kubelet

nmstate-handler-8ffsp

Pulling

Pulling image "registry.redhat.io/openshift4/ose-kubernetes-nmstate-handler-rhel9@sha256:b1744b2b84d6e23d83f465f450d2621a86bfec595d64373438b2e7ce5331e82e"

openshift-console-operator

console-operator-status-controller-statussyncer_console

console-operator

OperatorStatusChanged

Status for clusteroperator/console changed: status.relatedObjects changed from [{"console.openshift.io" "consoleplugins" "" "monitoring-plugin"} {"console.openshift.io" "consoleplugins" "" "networking-console-plugin"} {"operator.openshift.io" "consoles" "" "cluster"} {"config.openshift.io" "consoles" "" "cluster"} {"config.openshift.io" "infrastructures" "" "cluster"} {"config.openshift.io" "proxies" "" "cluster"} {"config.openshift.io" "oauths" "" "cluster"} {"oauth.openshift.io" "oauthclients" "" "console"} {"" "namespaces" "" "openshift-console-operator"} {"" "namespaces" "" "openshift-console"} {"" "configmaps" "openshift-config-managed" "console-public"}] to [{"console.openshift.io" "consoleplugins" "" "monitoring-plugin"} {"console.openshift.io" "consoleplugins" "" "networking-console-plugin"} {"console.openshift.io" "consoleplugins" "" "nmstate-console-plugin"} {"operator.openshift.io" "consoles" "" "cluster"} {"config.openshift.io" "consoles" "" "cluster"} {"config.openshift.io" "infrastructures" "" "cluster"} {"config.openshift.io" "proxies" "" "cluster"} {"config.openshift.io" "oauths" "" "cluster"} {"oauth.openshift.io" "oauthclients" "" "console"} {"" "namespaces" "" "openshift-console-operator"} {"" "namespaces" "" "openshift-console"} {"" "configmaps" "openshift-config-managed" "console-public"}]

openshift-nmstate

replicaset-controller

nmstate-console-plugin-86f58fcf4

SuccessfulCreate

Created pod: nmstate-console-plugin-86f58fcf4-5wfxh

openshift-nmstate

kubelet

nmstate-console-plugin-86f58fcf4-5wfxh

FailedMount

MountVolume.SetUp failed for volume "plugin-serving-cert" : secret "plugin-serving-cert" not found

openshift-nmstate

daemonset-controller

nmstate-handler

SuccessfulCreate

Created pod: nmstate-handler-8ffsp

openshift-nmstate

replicaset-controller

nmstate-console-plugin-86f58fcf4

SuccessfulCreate

Created pod: nmstate-console-plugin-86f58fcf4-5wfxh

openshift-nmstate

deployment-controller

nmstate-console-plugin

ScalingReplicaSet

Scaled up replica set nmstate-console-plugin-86f58fcf4 to 1

openshift-nmstate

kubelet

nmstate-handler-8ffsp

Pulling

Pulling image "registry.redhat.io/openshift4/ose-kubernetes-nmstate-handler-rhel9@sha256:b1744b2b84d6e23d83f465f450d2621a86bfec595d64373438b2e7ce5331e82e"

openshift-nmstate

multus

nmstate-console-plugin-86f58fcf4-5wfxh

AddedInterface

Add eth0 [10.128.0.136/23] from ovn-kubernetes

openshift-nmstate

multus

nmstate-metrics-9b8c8685d-p28fn

AddedInterface

Add eth0 [10.128.0.134/23] from ovn-kubernetes

openshift-nmstate

kubelet

nmstate-metrics-9b8c8685d-p28fn

Pulling

Pulling image "registry.redhat.io/openshift4/ose-kubernetes-nmstate-handler-rhel9@sha256:b1744b2b84d6e23d83f465f450d2621a86bfec595d64373438b2e7ce5331e82e"

openshift-console

multus

console-5df6fc5897-d9ws4

AddedInterface

Add eth0 [10.128.0.137/23] from ovn-kubernetes

openshift-console

kubelet

console-5df6fc5897-d9ws4

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5bbb8535e2496de8389585ebbe696e7d7b9bad2b27785ad8a30a0fc683b0a22d" already present on machine

metallb-system

kubelet

controller-7bb4cc7c98-77wvd

Started

Started container kube-rbac-proxy

metallb-system

kubelet

controller-7bb4cc7c98-77wvd

Created

Created container: kube-rbac-proxy

metallb-system

kubelet

controller-7bb4cc7c98-77wvd

Pulled

Successfully pulled image "registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:4fdd6da66aba2523d2c21cef306b7650659926bbadb96dedd000d2b8c0229078" in 1.655s (1.655s including waiting). Image size: 465090934 bytes.

openshift-console-operator

console-operator-status-controller-statussyncer_console

console-operator

OperatorStatusChanged

Status for clusteroperator/console changed: Progressing message changed from "SyncLoopRefreshProgressing: changes made during sync updates, additional sync expected" to "SyncLoopRefreshProgressing: working toward version 4.18.35, 1 replicas available"

openshift-console

kubelet

console-5df6fc5897-d9ws4

Created

Created container: console

openshift-console

kubelet

console-5df6fc5897-d9ws4

Started

Started container console

openshift-nmstate

multus

nmstate-metrics-9b8c8685d-p28fn

AddedInterface

Add eth0 [10.128.0.134/23] from ovn-kubernetes

openshift-nmstate

kubelet

nmstate-metrics-9b8c8685d-p28fn

Pulling

Pulling image "registry.redhat.io/openshift4/ose-kubernetes-nmstate-handler-rhel9@sha256:b1744b2b84d6e23d83f465f450d2621a86bfec595d64373438b2e7ce5331e82e"

openshift-nmstate

kubelet

nmstate-webhook-5f558f5558-w8sq6

Pulling

Pulling image "registry.redhat.io/openshift4/ose-kubernetes-nmstate-handler-rhel9@sha256:b1744b2b84d6e23d83f465f450d2621a86bfec595d64373438b2e7ce5331e82e"

metallb-system

kubelet

controller-7bb4cc7c98-77wvd

Pulled

Successfully pulled image "registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:4fdd6da66aba2523d2c21cef306b7650659926bbadb96dedd000d2b8c0229078" in 1.655s (1.655s including waiting). Image size: 465090934 bytes.

metallb-system

kubelet

controller-7bb4cc7c98-77wvd

Created

Created container: kube-rbac-proxy

metallb-system

kubelet

controller-7bb4cc7c98-77wvd

Started

Started container kube-rbac-proxy

openshift-nmstate

multus

nmstate-console-plugin-86f58fcf4-5wfxh

AddedInterface

Add eth0 [10.128.0.136/23] from ovn-kubernetes

openshift-nmstate

multus

nmstate-webhook-5f558f5558-w8sq6

AddedInterface

Add eth0 [10.128.0.135/23] from ovn-kubernetes

openshift-nmstate

kubelet

nmstate-webhook-5f558f5558-w8sq6

Pulling

Pulling image "registry.redhat.io/openshift4/ose-kubernetes-nmstate-handler-rhel9@sha256:b1744b2b84d6e23d83f465f450d2621a86bfec595d64373438b2e7ce5331e82e"

openshift-nmstate

multus

nmstate-webhook-5f558f5558-w8sq6

AddedInterface

Add eth0 [10.128.0.135/23] from ovn-kubernetes

metallb-system

kubelet

speaker-6cndp

Created

Created container: speaker

openshift-nmstate

kubelet

nmstate-console-plugin-86f58fcf4-5wfxh

Pulling

Pulling image "registry.redhat.io/openshift4/nmstate-console-plugin-rhel9@sha256:dcf6081eab6e9ce9595482d29ae143452dfc76682cc40354a9a64c8e3284c83a"

openshift-nmstate

kubelet

nmstate-console-plugin-86f58fcf4-5wfxh

Pulling

Pulling image "registry.redhat.io/openshift4/nmstate-console-plugin-rhel9@sha256:dcf6081eab6e9ce9595482d29ae143452dfc76682cc40354a9a64c8e3284c83a"

metallb-system

kubelet

speaker-6cndp

Pulled

Container image "registry.redhat.io/openshift4/metallb-rhel9@sha256:2db2c546af02ea7593f9c55d648f055c042800b55e3bfa13f7f43029aa9c6592" already present on machine

metallb-system

kubelet

speaker-6cndp

Created

Created container: speaker

metallb-system

kubelet

speaker-6cndp

Started

Started container speaker

metallb-system

kubelet

speaker-6cndp

Pulled

Container image "registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:4fdd6da66aba2523d2c21cef306b7650659926bbadb96dedd000d2b8c0229078" already present on machine

metallb-system

kubelet

speaker-6cndp

Started

Started container speaker

metallb-system

kubelet

speaker-6cndp

Pulled

Container image "registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:4fdd6da66aba2523d2c21cef306b7650659926bbadb96dedd000d2b8c0229078" already present on machine

metallb-system

kubelet

speaker-6cndp

Pulled

Container image "registry.redhat.io/openshift4/metallb-rhel9@sha256:2db2c546af02ea7593f9c55d648f055c042800b55e3bfa13f7f43029aa9c6592" already present on machine

metallb-system

kubelet

speaker-6cndp

Started

Started container kube-rbac-proxy

metallb-system

kubelet

speaker-6cndp

Created

Created container: kube-rbac-proxy

metallb-system

kubelet

speaker-6cndp

Started

Started container kube-rbac-proxy

metallb-system

kubelet

speaker-6cndp

Created

Created container: kube-rbac-proxy

openshift-nmstate

kubelet

nmstate-webhook-5f558f5558-w8sq6

Pulled

Successfully pulled image "registry.redhat.io/openshift4/ose-kubernetes-nmstate-handler-rhel9@sha256:b1744b2b84d6e23d83f465f450d2621a86bfec595d64373438b2e7ce5331e82e" in 5.973s (5.973s including waiting). Image size: 489111276 bytes.

openshift-nmstate

kubelet

nmstate-webhook-5f558f5558-w8sq6

Pulled

Successfully pulled image "registry.redhat.io/openshift4/ose-kubernetes-nmstate-handler-rhel9@sha256:b1744b2b84d6e23d83f465f450d2621a86bfec595d64373438b2e7ce5331e82e" in 5.973s (5.973s including waiting). Image size: 489111276 bytes.

openshift-nmstate

kubelet

nmstate-handler-8ffsp

Pulled

Successfully pulled image "registry.redhat.io/openshift4/ose-kubernetes-nmstate-handler-rhel9@sha256:b1744b2b84d6e23d83f465f450d2621a86bfec595d64373438b2e7ce5331e82e" in 6.874s (6.874s including waiting). Image size: 489111276 bytes.

metallb-system

kubelet

frr-k8s-webhook-server-bcc4b6f68-lrv7p

Pulled

Successfully pulled image "registry.redhat.io/openshift4/frr-rhel9@sha256:2157d8b664937a8c3871c12e9a4ee90e7da1a3db2b240bdd320b5dc619b9b8d4" in 8.658s (8.658s including waiting). Image size: 662223062 bytes.

openshift-nmstate

kubelet

nmstate-console-plugin-86f58fcf4-5wfxh

Pulled

Successfully pulled image "registry.redhat.io/openshift4/nmstate-console-plugin-rhel9@sha256:dcf6081eab6e9ce9595482d29ae143452dfc76682cc40354a9a64c8e3284c83a" in 5.736s (5.736s including waiting). Image size: 453916031 bytes.

openshift-nmstate

kubelet

nmstate-metrics-9b8c8685d-p28fn

Pulled

Successfully pulled image "registry.redhat.io/openshift4/ose-kubernetes-nmstate-handler-rhel9@sha256:b1744b2b84d6e23d83f465f450d2621a86bfec595d64373438b2e7ce5331e82e" in 6.503s (6.503s including waiting). Image size: 489111276 bytes.

openshift-nmstate

kubelet

nmstate-handler-8ffsp

Pulled

Successfully pulled image "registry.redhat.io/openshift4/ose-kubernetes-nmstate-handler-rhel9@sha256:b1744b2b84d6e23d83f465f450d2621a86bfec595d64373438b2e7ce5331e82e" in 6.874s (6.874s including waiting). Image size: 489111276 bytes.

openshift-nmstate

kubelet

nmstate-metrics-9b8c8685d-p28fn

Pulled

Successfully pulled image "registry.redhat.io/openshift4/ose-kubernetes-nmstate-handler-rhel9@sha256:b1744b2b84d6e23d83f465f450d2621a86bfec595d64373438b2e7ce5331e82e" in 6.503s (6.503s including waiting). Image size: 489111276 bytes.

metallb-system

kubelet

frr-k8s-pp5nf

Pulled

Successfully pulled image "registry.redhat.io/openshift4/frr-rhel9@sha256:2157d8b664937a8c3871c12e9a4ee90e7da1a3db2b240bdd320b5dc619b9b8d4" in 8.385s (8.385s including waiting). Image size: 662223062 bytes.

metallb-system

kubelet

frr-k8s-pp5nf

Pulled

Successfully pulled image "registry.redhat.io/openshift4/frr-rhel9@sha256:2157d8b664937a8c3871c12e9a4ee90e7da1a3db2b240bdd320b5dc619b9b8d4" in 8.385s (8.385s including waiting). Image size: 662223062 bytes.

metallb-system

kubelet

frr-k8s-webhook-server-bcc4b6f68-lrv7p

Pulled

Successfully pulled image "registry.redhat.io/openshift4/frr-rhel9@sha256:2157d8b664937a8c3871c12e9a4ee90e7da1a3db2b240bdd320b5dc619b9b8d4" in 8.658s (8.658s including waiting). Image size: 662223062 bytes.

openshift-nmstate

kubelet

nmstate-console-plugin-86f58fcf4-5wfxh

Pulled

Successfully pulled image "registry.redhat.io/openshift4/nmstate-console-plugin-rhel9@sha256:dcf6081eab6e9ce9595482d29ae143452dfc76682cc40354a9a64c8e3284c83a" in 5.736s (5.736s including waiting). Image size: 453916031 bytes.

openshift-nmstate

kubelet

nmstate-metrics-9b8c8685d-p28fn

Started

Started container kube-rbac-proxy

metallb-system

kubelet

frr-k8s-webhook-server-bcc4b6f68-lrv7p

Created

Created container: frr-k8s-webhook-server

metallb-system

kubelet

frr-k8s-webhook-server-bcc4b6f68-lrv7p

Started

Started container frr-k8s-webhook-server

openshift-nmstate

kubelet

nmstate-console-plugin-86f58fcf4-5wfxh

Created

Created container: nmstate-console-plugin

openshift-nmstate

kubelet

nmstate-console-plugin-86f58fcf4-5wfxh

Started

Started container nmstate-console-plugin

metallb-system

kubelet

frr-k8s-pp5nf

Started

Started container cp-reloader

metallb-system

kubelet

frr-k8s-pp5nf

Created

Created container: cp-reloader

metallb-system

kubelet

frr-k8s-pp5nf

Pulled

Container image "registry.redhat.io/openshift4/frr-rhel9@sha256:2157d8b664937a8c3871c12e9a4ee90e7da1a3db2b240bdd320b5dc619b9b8d4" already present on machine

metallb-system

kubelet

frr-k8s-pp5nf

Started

Started container cp-frr-files

metallb-system

kubelet

frr-k8s-pp5nf

Created

Created container: cp-frr-files

metallb-system

kubelet

frr-k8s-pp5nf

Created

Created container: cp-frr-files

metallb-system

kubelet

frr-k8s-pp5nf

Started

Started container cp-frr-files

metallb-system

kubelet

frr-k8s-pp5nf

Pulled

Container image "registry.redhat.io/openshift4/frr-rhel9@sha256:2157d8b664937a8c3871c12e9a4ee90e7da1a3db2b240bdd320b5dc619b9b8d4" already present on machine

metallb-system

kubelet

frr-k8s-pp5nf

Created

Created container: cp-reloader

openshift-nmstate

kubelet

nmstate-metrics-9b8c8685d-p28fn

Created

Created container: kube-rbac-proxy

openshift-nmstate

kubelet

nmstate-metrics-9b8c8685d-p28fn

Pulled

Container image "registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:4fdd6da66aba2523d2c21cef306b7650659926bbadb96dedd000d2b8c0229078" already present on machine

openshift-nmstate

kubelet

nmstate-metrics-9b8c8685d-p28fn

Started

Started container nmstate-metrics

metallb-system

kubelet

frr-k8s-pp5nf

Started

Started container cp-reloader

openshift-nmstate

kubelet

nmstate-handler-8ffsp

Started

Started container nmstate-handler

openshift-nmstate

kubelet

nmstate-handler-8ffsp

Created

Created container: nmstate-handler

openshift-nmstate

kubelet

nmstate-console-plugin-86f58fcf4-5wfxh

Started

Started container nmstate-console-plugin

openshift-nmstate

kubelet

nmstate-console-plugin-86f58fcf4-5wfxh

Created

Created container: nmstate-console-plugin

metallb-system

kubelet

frr-k8s-webhook-server-bcc4b6f68-lrv7p

Created

Created container: frr-k8s-webhook-server

metallb-system

kubelet

frr-k8s-webhook-server-bcc4b6f68-lrv7p

Started

Started container frr-k8s-webhook-server

openshift-nmstate

kubelet

nmstate-metrics-9b8c8685d-p28fn

Created

Created container: nmstate-metrics

openshift-nmstate

kubelet

nmstate-handler-8ffsp

Created

Created container: nmstate-handler

openshift-nmstate

kubelet

nmstate-webhook-5f558f5558-w8sq6

Created

Created container: nmstate-webhook

openshift-nmstate

kubelet

nmstate-webhook-5f558f5558-w8sq6

Started

Started container nmstate-webhook

openshift-nmstate

kubelet

nmstate-handler-8ffsp

Started

Started container nmstate-handler

openshift-nmstate

kubelet

nmstate-metrics-9b8c8685d-p28fn

Created

Created container: nmstate-metrics

openshift-nmstate

kubelet

nmstate-metrics-9b8c8685d-p28fn

Started

Started container nmstate-metrics

openshift-nmstate

kubelet

nmstate-webhook-5f558f5558-w8sq6

Created

Created container: nmstate-webhook

openshift-nmstate

kubelet

nmstate-webhook-5f558f5558-w8sq6

Started

Started container nmstate-webhook

openshift-nmstate

kubelet

nmstate-metrics-9b8c8685d-p28fn

Pulled

Container image "registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:4fdd6da66aba2523d2c21cef306b7650659926bbadb96dedd000d2b8c0229078" already present on machine

openshift-nmstate

kubelet

nmstate-metrics-9b8c8685d-p28fn

Created

Created container: kube-rbac-proxy

openshift-nmstate

kubelet

nmstate-metrics-9b8c8685d-p28fn

Started

Started container kube-rbac-proxy

metallb-system

kubelet

frr-k8s-pp5nf

Pulled

Container image "registry.redhat.io/openshift4/frr-rhel9@sha256:2157d8b664937a8c3871c12e9a4ee90e7da1a3db2b240bdd320b5dc619b9b8d4" already present on machine

metallb-system

kubelet

frr-k8s-pp5nf

Pulled

Container image "registry.redhat.io/openshift4/frr-rhel9@sha256:2157d8b664937a8c3871c12e9a4ee90e7da1a3db2b240bdd320b5dc619b9b8d4" already present on machine

metallb-system

kubelet

frr-k8s-pp5nf

Started

Started container cp-metrics

metallb-system

kubelet

frr-k8s-pp5nf

Created

Created container: cp-metrics

metallb-system

kubelet

frr-k8s-pp5nf

Started

Started container cp-metrics

metallb-system

kubelet

frr-k8s-pp5nf

Created

Created container: cp-metrics

metallb-system

kubelet

frr-k8s-pp5nf

Created

Created container: frr

metallb-system

kubelet

frr-k8s-pp5nf

Created

Created container: controller

metallb-system

kubelet

frr-k8s-pp5nf

Created

Created container: controller

metallb-system

kubelet

frr-k8s-pp5nf

Started

Started container controller

metallb-system

kubelet

frr-k8s-pp5nf

Pulled

Container image "registry.redhat.io/openshift4/frr-rhel9@sha256:2157d8b664937a8c3871c12e9a4ee90e7da1a3db2b240bdd320b5dc619b9b8d4" already present on machine

metallb-system

kubelet

frr-k8s-pp5nf

Created

Created container: frr

metallb-system

kubelet

frr-k8s-pp5nf

Started

Started container controller

metallb-system

kubelet

frr-k8s-pp5nf

Pulled

Container image "registry.redhat.io/openshift4/frr-rhel9@sha256:2157d8b664937a8c3871c12e9a4ee90e7da1a3db2b240bdd320b5dc619b9b8d4" already present on machine

metallb-system

kubelet

frr-k8s-pp5nf

Pulled

Container image "registry.redhat.io/openshift4/frr-rhel9@sha256:2157d8b664937a8c3871c12e9a4ee90e7da1a3db2b240bdd320b5dc619b9b8d4" already present on machine

metallb-system

kubelet

frr-k8s-pp5nf

Pulled

Container image "registry.redhat.io/openshift4/frr-rhel9@sha256:2157d8b664937a8c3871c12e9a4ee90e7da1a3db2b240bdd320b5dc619b9b8d4" already present on machine

metallb-system

kubelet

frr-k8s-pp5nf

Pulled

Container image "registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:4fdd6da66aba2523d2c21cef306b7650659926bbadb96dedd000d2b8c0229078" already present on machine

openshift-console-operator

console-operator-status-controller-statussyncer_console

console-operator

OperatorStatusChanged

Status for clusteroperator/console changed: Progressing message changed from "SyncLoopRefreshProgressing: working toward version 4.18.35, 1 replicas available" to "SyncLoopRefreshProgressing: working toward version 4.18.35, 2 replicas available"

openshift-console

deployment-controller

console

ScalingReplicaSet

Scaled down replica set console-d4699dbdf to 0 from 1

metallb-system

kubelet

frr-k8s-pp5nf

Created

Created container: kube-rbac-proxy

metallb-system

kubelet

frr-k8s-pp5nf

Started

Started container frr-metrics

metallb-system

kubelet

frr-k8s-pp5nf

Started

Started container frr

metallb-system

kubelet

frr-k8s-pp5nf

Pulled

Container image "registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:4fdd6da66aba2523d2c21cef306b7650659926bbadb96dedd000d2b8c0229078" already present on machine

metallb-system

kubelet

frr-k8s-pp5nf

Started

Started container frr-metrics

openshift-console

replicaset-controller

console-d4699dbdf

SuccessfulDelete

Deleted pod: console-d4699dbdf-6z95l

openshift-console

kubelet

console-d4699dbdf-6z95l

Killing

Stopping container console

metallb-system

kubelet

frr-k8s-pp5nf

Created

Created container: kube-rbac-proxy

metallb-system

kubelet

frr-k8s-pp5nf

Created

Created container: reloader

metallb-system

kubelet

frr-k8s-pp5nf

Created

Created container: frr-metrics

metallb-system

kubelet

frr-k8s-pp5nf

Pulled

Container image "registry.redhat.io/openshift4/frr-rhel9@sha256:2157d8b664937a8c3871c12e9a4ee90e7da1a3db2b240bdd320b5dc619b9b8d4" already present on machine

metallb-system

kubelet

frr-k8s-pp5nf

Pulled

Container image "registry.redhat.io/openshift4/frr-rhel9@sha256:2157d8b664937a8c3871c12e9a4ee90e7da1a3db2b240bdd320b5dc619b9b8d4" already present on machine

metallb-system

kubelet

frr-k8s-pp5nf

Started

Started container reloader
(x2)

openshift-console-operator

console-operator-status-controller-statussyncer_console

console-operator

OperatorStatusChanged

Status for clusteroperator/console changed: Progressing changed from True to False ("All is well")

metallb-system

kubelet

frr-k8s-pp5nf

Started

Started container reloader

metallb-system

kubelet

frr-k8s-pp5nf

Created

Created container: reloader

metallb-system

kubelet

frr-k8s-pp5nf

Pulled

Container image "registry.redhat.io/openshift4/frr-rhel9@sha256:2157d8b664937a8c3871c12e9a4ee90e7da1a3db2b240bdd320b5dc619b9b8d4" already present on machine

metallb-system

kubelet

frr-k8s-pp5nf

Pulled

Container image "registry.redhat.io/openshift4/frr-rhel9@sha256:2157d8b664937a8c3871c12e9a4ee90e7da1a3db2b240bdd320b5dc619b9b8d4" already present on machine

metallb-system

kubelet

frr-k8s-pp5nf

Started

Started container frr

metallb-system

kubelet

frr-k8s-pp5nf

Created

Created container: frr-metrics

openshift-storage

daemonset-controller

vg-manager

SuccessfulCreate

Created pod: vg-manager-rpqkf

openshift-storage

daemonset-controller

vg-manager

SuccessfulCreate

Created pod: vg-manager-rpqkf

openshift-storage

multus

vg-manager-rpqkf

AddedInterface

Add eth0 [10.128.0.138/23] from ovn-kubernetes

openshift-storage

multus

vg-manager-rpqkf

AddedInterface

Add eth0 [10.128.0.138/23] from ovn-kubernetes
(x2)

openshift-storage

kubelet

vg-manager-rpqkf

Pulled

Container image "registry.redhat.io/lvms4/lvms-rhel9-operator@sha256:be49109ec5be53a783b2de9bc6529db99930a69021577c05cfa0bf1312e06d69" already present on machine
(x2)

openshift-storage

kubelet

vg-manager-rpqkf

Pulled

Container image "registry.redhat.io/lvms4/lvms-rhel9-operator@sha256:be49109ec5be53a783b2de9bc6529db99930a69021577c05cfa0bf1312e06d69" already present on machine
(x2)

openshift-storage

kubelet

vg-manager-rpqkf

Created

Created container: vg-manager
(x2)

openshift-storage

kubelet

vg-manager-rpqkf

Started

Started container vg-manager
(x14)

openshift-storage

LVMClusterReconciler

lvmcluster

ResourceReconciliationIncomplete

LVMCluster's resources are not yet fully synchronized: csi node master-0 does not have driver topolvm.io
(x2)

openshift-storage

kubelet

vg-manager-rpqkf

Started

Started container vg-manager
(x14)

openshift-storage

LVMClusterReconciler

lvmcluster

ResourceReconciliationIncomplete

LVMCluster's resources are not yet fully synchronized: csi node master-0 does not have driver topolvm.io
(x2)

openshift-storage

kubelet

vg-manager-rpqkf

Created

Created container: vg-manager

openshift-kube-controller-manager

cluster-policy-controller-namespace-security-allocation-controller

kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openstack namespace

openshift-kube-controller-manager

cluster-policy-controller-namespace-security-allocation-controller

kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openstack-operators namespace

openstack-operators

multus

openstack-operator-index-5g65w

AddedInterface

Add eth0 [10.128.0.139/23] from ovn-kubernetes

openstack-operators

kubelet

openstack-operator-index-5g65w

Pulling

Pulling image "quay.io/openstack-k8s-operators/openstack-operator-index:latest"

openstack-operators

multus

openstack-operator-index-5g65w

AddedInterface

Add eth0 [10.128.0.139/23] from ovn-kubernetes

openstack-operators

kubelet

openstack-operator-index-5g65w

Pulling

Pulling image "quay.io/openstack-k8s-operators/openstack-operator-index:latest"

openstack-operators

kubelet

openstack-operator-index-5g65w

Started

Started container registry-server

openstack-operators

kubelet

openstack-operator-index-5g65w

Created

Created container: registry-server

openstack-operators

kubelet

openstack-operator-index-5g65w

Created

Created container: registry-server

openstack-operators

kubelet

openstack-operator-index-5g65w

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/openstack-operator-index:latest" in 972ms (972ms including waiting). Image size: 918642352 bytes.

openstack-operators

kubelet

openstack-operator-index-5g65w

Started

Started container registry-server
(x6)

default

operator-lifecycle-manager

openstack-operators

ResolutionFailed

error using catalogsource openstack-operators/openstack-operator-index: no registry client established for catalogsource openstack-operators/openstack-operator-index

openstack-operators

kubelet

openstack-operator-index-5g65w

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/openstack-operator-index:latest" in 972ms (972ms including waiting). Image size: 918642352 bytes.
(x4)

default

operator-lifecycle-manager

openstack-operators

ResolutionFailed

error using catalogsource openstack-operators/openstack-operator-index: failed to list bundles: rpc error: code = Unavailable desc = connection error: desc = "transport: Error while dialing: dial tcp 172.30.53.231:50051: connect: connection refused"

openstack-operators

job-controller

7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514cbbc26

SuccessfulCreate

Created pod: 7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514ckhsqj

openstack-operators

job-controller

7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514cbbc26

SuccessfulCreate

Created pod: 7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514ckhsqj

openstack-operators

kubelet

7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514ckhsqj

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6bba3f73c0066e42b24839e0d29f5dce2f36436f0a11f9f5e1029bccc5ed6578" already present on machine

openstack-operators

kubelet

7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514ckhsqj

Created

Created container: util

openstack-operators

multus

7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514ckhsqj

AddedInterface

Add eth0 [10.128.0.140/23] from ovn-kubernetes

openstack-operators

kubelet

7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514ckhsqj

Started

Started container util

openstack-operators

kubelet

7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514ckhsqj

Started

Started container util

openstack-operators

kubelet

7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514ckhsqj

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6bba3f73c0066e42b24839e0d29f5dce2f36436f0a11f9f5e1029bccc5ed6578" already present on machine

openstack-operators

kubelet

7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514ckhsqj

Created

Created container: util

openstack-operators

multus

7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514ckhsqj

AddedInterface

Add eth0 [10.128.0.140/23] from ovn-kubernetes

openstack-operators

kubelet

7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514ckhsqj

Pulling

Pulling image "quay.io/openstack-k8s-operators/openstack-operator-bundle:36856d22fbbd028e148ba6b5277b8d8be928cf7c"

openstack-operators

kubelet

7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514ckhsqj

Pulling

Pulling image "quay.io/openstack-k8s-operators/openstack-operator-bundle:36856d22fbbd028e148ba6b5277b8d8be928cf7c"

openstack-operators

kubelet

7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514ckhsqj

Started

Started container pull

openstack-operators

kubelet

7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514ckhsqj

Started

Started container pull

openstack-operators

kubelet

7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514ckhsqj

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/openstack-operator-bundle:36856d22fbbd028e148ba6b5277b8d8be928cf7c" in 1.092s (1.092s including waiting). Image size: 115773 bytes.

openstack-operators

kubelet

7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514ckhsqj

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/openstack-operator-bundle:36856d22fbbd028e148ba6b5277b8d8be928cf7c" in 1.092s (1.092s including waiting). Image size: 115773 bytes.

openstack-operators

kubelet

7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514ckhsqj

Created

Created container: pull

openstack-operators

kubelet

7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514ckhsqj

Created

Created container: pull

openstack-operators

kubelet

7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514ckhsqj

Started

Started container extract

openstack-operators

kubelet

7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514ckhsqj

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b00c42562d477ef44d51f35950253a26d7debc7de86e53270831aafef5795c1" already present on machine

openstack-operators

kubelet

7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514ckhsqj

Created

Created container: extract

openstack-operators

kubelet

7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514ckhsqj

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b00c42562d477ef44d51f35950253a26d7debc7de86e53270831aafef5795c1" already present on machine

openstack-operators

kubelet

7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514ckhsqj

Created

Created container: extract

openstack-operators

kubelet

7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514ckhsqj

Started

Started container extract

openstack-operators

job-controller

7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514cbbc26

Completed

Job completed

openstack-operators

job-controller

7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514cbbc26

Completed

Job completed

openstack-operators

operator-lifecycle-manager

openstack-operator.v0.6.0

RequirementsUnknown

requirements not yet checked

openstack-operators

operator-lifecycle-manager

openstack-operator.v0.6.0

RequirementsNotMet

one or more requirements couldn't be found

openstack-operators

operator-lifecycle-manager

openstack-operator.v0.6.0

RequirementsNotMet

one or more requirements couldn't be found

openstack-operators

operator-lifecycle-manager

openstack-operator.v0.6.0

RequirementsUnknown

requirements not yet checked

openstack-operators

replicaset-controller

openstack-operator-controller-init-b85c4d696

SuccessfulCreate

Created pod: openstack-operator-controller-init-b85c4d696-9zm7k

openstack-operators

operator-lifecycle-manager

openstack-operator.v0.6.0

InstallSucceeded

waiting for install components to report healthy

openstack-operators

deployment-controller

openstack-operator-controller-init

ScalingReplicaSet

Scaled up replica set openstack-operator-controller-init-b85c4d696 to 1

openstack-operators

replicaset-controller

openstack-operator-controller-init-b85c4d696

SuccessfulCreate

Created pod: openstack-operator-controller-init-b85c4d696-9zm7k

openstack-operators

deployment-controller

openstack-operator-controller-init

ScalingReplicaSet

Scaled up replica set openstack-operator-controller-init-b85c4d696 to 1

openstack-operators

operator-lifecycle-manager

openstack-operator.v0.6.0

InstallSucceeded

waiting for install components to report healthy

openstack-operators

operator-lifecycle-manager

openstack-operator.v0.6.0

AllRequirementsMet

all requirements found, attempting install

openstack-operators

operator-lifecycle-manager

openstack-operator.v0.6.0

AllRequirementsMet

all requirements found, attempting install

openstack-operators

multus

openstack-operator-controller-init-b85c4d696-9zm7k

AddedInterface

Add eth0 [10.128.0.141/23] from ovn-kubernetes

openstack-operators

operator-lifecycle-manager

openstack-operator.v0.6.0

InstallWaiting

installing: waiting for deployment openstack-operator-controller-init to become ready: deployment "openstack-operator-controller-init" not available: Deployment does not have minimum availability.

openstack-operators

kubelet

openstack-operator-controller-init-b85c4d696-9zm7k

Pulling

Pulling image "quay.io/openstack-k8s-operators/openstack-operator@sha256:5ab774ebf20af59fbaa509688a104dc74488b7cd8c4b0640d16924de8ead64fb"

openstack-operators

multus

openstack-operator-controller-init-b85c4d696-9zm7k

AddedInterface

Add eth0 [10.128.0.141/23] from ovn-kubernetes

openstack-operators

operator-lifecycle-manager

openstack-operator.v0.6.0

InstallWaiting

installing: waiting for deployment openstack-operator-controller-init to become ready: deployment "openstack-operator-controller-init" not available: Deployment does not have minimum availability.

openstack-operators

kubelet

openstack-operator-controller-init-b85c4d696-9zm7k

Pulling

Pulling image "quay.io/openstack-k8s-operators/openstack-operator@sha256:5ab774ebf20af59fbaa509688a104dc74488b7cd8c4b0640d16924de8ead64fb"

openstack-operators

kubelet

openstack-operator-controller-init-b85c4d696-9zm7k

Created

Created container: operator

openstack-operators

openstack-operator-controller-init-b85c4d696-9zm7k_6c005758-3cb7-49a8-ac1d-c69d9799f864

20ca801f.openstack.org

LeaderElection

openstack-operator-controller-init-b85c4d696-9zm7k_6c005758-3cb7-49a8-ac1d-c69d9799f864 became leader

openstack-operators

openstack-operator-controller-init-b85c4d696-9zm7k_6c005758-3cb7-49a8-ac1d-c69d9799f864

20ca801f.openstack.org

LeaderElection

openstack-operator-controller-init-b85c4d696-9zm7k_6c005758-3cb7-49a8-ac1d-c69d9799f864 became leader

openstack-operators

kubelet

openstack-operator-controller-init-b85c4d696-9zm7k

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/openstack-operator@sha256:5ab774ebf20af59fbaa509688a104dc74488b7cd8c4b0640d16924de8ead64fb" in 4.325s (4.325s including waiting). Image size: 293357387 bytes.

openstack-operators

kubelet

openstack-operator-controller-init-b85c4d696-9zm7k

Started

Started container operator

openstack-operators

kubelet

openstack-operator-controller-init-b85c4d696-9zm7k

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/openstack-operator@sha256:5ab774ebf20af59fbaa509688a104dc74488b7cd8c4b0640d16924de8ead64fb" in 4.325s (4.325s including waiting). Image size: 293357387 bytes.

openstack-operators

kubelet

openstack-operator-controller-init-b85c4d696-9zm7k

Started

Started container operator

openstack-operators

kubelet

openstack-operator-controller-init-b85c4d696-9zm7k

Created

Created container: operator

openstack-operators

operator-lifecycle-manager

openstack-operator.v0.6.0

InstallSucceeded

install strategy completed with no errors

openstack-operators

operator-lifecycle-manager

openstack-operator.v0.6.0

InstallSucceeded

install strategy completed with no errors

openstack-operators

cert-manager-certificates-key-manager

barbican-operator-metrics-certs

Generated

Stored new private key in temporary Secret resource "barbican-operator-metrics-certs-vxvbv"

openstack-operators

cert-manager-certificates-trigger

barbican-operator-metrics-certs

Issuing

Issuing certificate as Secret does not exist

openstack-operators

cert-manager-certificates-trigger

barbican-operator-metrics-certs

Issuing

Issuing certificate as Secret does not exist

openstack-operators

cert-manager-certificates-key-manager

cinder-operator-metrics-certs

Generated

Stored new private key in temporary Secret resource "cinder-operator-metrics-certs-ppqfw"

openstack-operators

cert-manager-certificates-trigger

cinder-operator-metrics-certs

Issuing

Issuing certificate as Secret does not exist

openstack-operators

cert-manager-certificates-key-manager

barbican-operator-metrics-certs

Generated

Stored new private key in temporary Secret resource "barbican-operator-metrics-certs-vxvbv"

openstack-operators

cert-manager-certificates-trigger

cinder-operator-metrics-certs

Issuing

Issuing certificate as Secret does not exist

openstack-operators

cert-manager-certificates-key-manager

cinder-operator-metrics-certs

Generated

Stored new private key in temporary Secret resource "cinder-operator-metrics-certs-ppqfw"

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

cinder-operator-metrics-certs-1

CertificateIssued

Certificate fetched from issuer successfully

openstack-operators

cert-manager-certificates-request-manager

barbican-operator-metrics-certs

Requested

Created new CertificateRequest resource "barbican-operator-metrics-certs-1"

openstack-operators

cert-manager-certificaterequests-issuer-ca

barbican-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-venafi

barbican-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

barbican-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-acme

barbican-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-venafi

barbican-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-vault

barbican-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-approver

barbican-operator-metrics-certs-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

barbican-operator-metrics-certs-1

CertificateIssued

Certificate fetched from issuer successfully

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

barbican-operator-metrics-certs-1

BadConfig

Certificate will be issued with an empty Issuer DN, which contravenes RFC 5280 and could break some strict clients

openstack-operators

cert-manager-certificaterequests-issuer-ca

barbican-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

barbican-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificates-request-manager

barbican-operator-metrics-certs

Requested

Created new CertificateRequest resource "barbican-operator-metrics-certs-1"

openstack-operators

cert-manager-certificates-issuing

barbican-operator-metrics-certs

Issuing

The certificate has been successfully issued

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

barbican-operator-metrics-certs-1

BadConfig

Certificate will be issued with an empty Issuer DN, which contravenes RFC 5280 and could break some strict clients

openstack-operators

cert-manager-certificates-trigger

glance-operator-metrics-certs

Issuing

Issuing certificate as Secret does not exist

openstack-operators

cert-manager-certificaterequests-issuer-acme

cinder-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificates-trigger

horizon-operator-metrics-certs

Issuing

Issuing certificate as Secret does not exist

openstack-operators

cert-manager-certificaterequests-approver

cinder-operator-metrics-certs-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack-operators

cert-manager-certificates-key-manager

designate-operator-metrics-certs

Generated

Stored new private key in temporary Secret resource "designate-operator-metrics-certs-djwbd"

openstack-operators

cert-manager-certificates-trigger

designate-operator-metrics-certs

Issuing

Issuing certificate as Secret does not exist

openstack-operators

cert-manager-certificaterequests-approver

barbican-operator-metrics-certs-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack-operators

cert-manager-certificates-key-manager

designate-operator-metrics-certs

Generated

Stored new private key in temporary Secret resource "designate-operator-metrics-certs-djwbd"

openstack-operators

cert-manager-certificaterequests-issuer-venafi

cinder-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificates-key-manager

glance-operator-metrics-certs

Generated

Stored new private key in temporary Secret resource "glance-operator-metrics-certs-t5278"

openstack-operators

cert-manager-certificaterequests-issuer-vault

barbican-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

barbican-operator-metrics-certs-1

CertificateIssued

Certificate fetched from issuer successfully

openstack-operators

cert-manager-certificates-issuing

cinder-operator-metrics-certs

Issuing

The certificate has been successfully issued

openstack-operators

cert-manager-certificaterequests-approver

cinder-operator-metrics-certs-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack-operators

cert-manager-certificates-trigger

heat-operator-metrics-certs

Issuing

Issuing certificate as Secret does not exist

openstack-operators

cert-manager-certificaterequests-issuer-vault

cinder-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificates-key-manager

glance-operator-metrics-certs

Generated

Stored new private key in temporary Secret resource "glance-operator-metrics-certs-t5278"

openstack-operators

cert-manager-certificates-trigger

glance-operator-metrics-certs

Issuing

Issuing certificate as Secret does not exist

openstack-operators

cert-manager-certificates-request-manager

cinder-operator-metrics-certs

Requested

Created new CertificateRequest resource "cinder-operator-metrics-certs-1"

openstack-operators

cert-manager-certificates-trigger

horizon-operator-metrics-certs

Issuing

Issuing certificate as Secret does not exist

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

cinder-operator-metrics-certs-1

CertificateIssued

Certificate fetched from issuer successfully

openstack-operators

cert-manager-certificaterequests-issuer-acme

barbican-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificates-trigger

designate-operator-metrics-certs

Issuing

Issuing certificate as Secret does not exist

openstack-operators

cert-manager-certificates-issuing

cinder-operator-metrics-certs

Issuing

The certificate has been successfully issued

openstack-operators

cert-manager-certificaterequests-issuer-ca

cinder-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificates-request-manager

cinder-operator-metrics-certs

Requested

Created new CertificateRequest resource "cinder-operator-metrics-certs-1"

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

cinder-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

cinder-operator-metrics-certs-1

BadConfig

Certificate will be issued with an empty Issuer DN, which contravenes RFC 5280 and could break some strict clients

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

cinder-operator-metrics-certs-1

BadConfig

Certificate will be issued with an empty Issuer DN, which contravenes RFC 5280 and could break some strict clients

openstack-operators

cert-manager-certificates-trigger

heat-operator-metrics-certs

Issuing

Issuing certificate as Secret does not exist

openstack-operators

cert-manager-certificates-issuing

barbican-operator-metrics-certs

Issuing

The certificate has been successfully issued

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

cinder-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-ca

cinder-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-vault

cinder-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-venafi

cinder-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-acme

cinder-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificates-trigger

manila-operator-metrics-certs

Issuing

Issuing certificate as Secret does not exist

openstack-operators

cert-manager-certificates-trigger

infra-operator-metrics-certs

Issuing

Issuing certificate as Secret does not exist

openstack-operators

cert-manager-certificates-trigger

infra-operator-metrics-certs

Issuing

Issuing certificate as Secret does not exist

openshift-machine-config-operator

machine-config-operator

machine-config-operator

ConfigMapUpdated

Updated ConfigMap/kube-rbac-proxy -n openshift-machine-config-operator: cause by changes in data.config-file.yaml

openstack-operators

cert-manager-certificates-trigger

ironic-operator-metrics-certs

Issuing

Issuing certificate as Secret does not exist

openstack-operators

cert-manager-certificates-trigger

ironic-operator-metrics-certs

Issuing

Issuing certificate as Secret does not exist

openstack-operators

cert-manager-certificates-trigger

manila-operator-metrics-certs

Issuing

Issuing certificate as Secret does not exist

openstack-operators

cert-manager-certificaterequests-issuer-ca

glance-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificates-trigger

octavia-operator-metrics-certs

Issuing

Issuing certificate as Secret does not exist

openstack-operators

cert-manager-certificates-key-manager

heat-operator-metrics-certs

Generated

Stored new private key in temporary Secret resource "heat-operator-metrics-certs-flvlv"

openstack-operators

cert-manager-certificaterequests-issuer-acme

glance-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-ca

designate-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

glance-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificates-trigger

octavia-operator-metrics-certs

Issuing

Issuing certificate as Secret does not exist

openstack-operators

cert-manager-certificaterequests-issuer-venafi

designate-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

glance-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-vault

glance-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificates-key-manager

horizon-operator-metrics-certs

Generated

Stored new private key in temporary Secret resource "horizon-operator-metrics-certs-rw482"

openstack-operators

cert-manager-certificaterequests-issuer-ca

glance-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificates-key-manager

heat-operator-metrics-certs

Generated

Stored new private key in temporary Secret resource "heat-operator-metrics-certs-flvlv"

openstack-operators

cert-manager-certificates-request-manager

designate-operator-metrics-certs

Requested

Created new CertificateRequest resource "designate-operator-metrics-certs-1"

openstack-operators

cert-manager-certificates-trigger

keystone-operator-metrics-certs

Issuing

Issuing certificate as Secret does not exist

openstack-operators

cert-manager-certificaterequests-issuer-vault

designate-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-venafi

designate-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-ca

designate-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

designate-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-acme

designate-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificates-trigger

keystone-operator-metrics-certs

Issuing

Issuing certificate as Secret does not exist

openstack-operators

cert-manager-certificates-request-manager

glance-operator-metrics-certs

Requested

Created new CertificateRequest resource "glance-operator-metrics-certs-1"

openstack-operators

cert-manager-certificates-request-manager

glance-operator-metrics-certs

Requested

Created new CertificateRequest resource "glance-operator-metrics-certs-1"

openstack-operators

cert-manager-certificaterequests-issuer-acme

glance-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificates-request-manager

designate-operator-metrics-certs

Requested

Created new CertificateRequest resource "designate-operator-metrics-certs-1"

openstack-operators

cert-manager-certificaterequests-issuer-vault

glance-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-venafi

glance-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-venafi

glance-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-vault

designate-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificates-key-manager

horizon-operator-metrics-certs

Generated

Stored new private key in temporary Secret resource "horizon-operator-metrics-certs-rw482"

openstack-operators

cert-manager-certificaterequests-issuer-acme

designate-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

designate-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificates-trigger

neutron-operator-metrics-certs

Issuing

Issuing certificate as Secret does not exist

openstack-operators

cert-manager-certificates-trigger

mariadb-operator-metrics-certs

Issuing

Issuing certificate as Secret does not exist

openstack-operators

cert-manager-certificates-key-manager

ironic-operator-metrics-certs

Generated

Stored new private key in temporary Secret resource "ironic-operator-metrics-certs-2p59c"

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

glance-operator-metrics-certs-1

CertificateIssued

Certificate fetched from issuer successfully

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

glance-operator-metrics-certs-1

BadConfig

Certificate will be issued with an empty Issuer DN, which contravenes RFC 5280 and could break some strict clients

openstack-operators

cert-manager-certificates-key-manager

ironic-operator-metrics-certs

Generated

Stored new private key in temporary Secret resource "ironic-operator-metrics-certs-2p59c"

openstack-operators

cert-manager-certificates-key-manager

manila-operator-metrics-certs

Generated

Stored new private key in temporary Secret resource "manila-operator-metrics-certs-2rlhk"

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

glance-operator-metrics-certs-1

CertificateIssued

Certificate fetched from issuer successfully

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

glance-operator-metrics-certs-1

BadConfig

Certificate will be issued with an empty Issuer DN, which contravenes RFC 5280 and could break some strict clients

openstack-operators

cert-manager-certificaterequests-approver

glance-operator-metrics-certs-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack-operators

cert-manager-certificates-trigger

mariadb-operator-metrics-certs

Issuing

Issuing certificate as Secret does not exist

openstack-operators

cert-manager-certificaterequests-approver

glance-operator-metrics-certs-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack-operators

cert-manager-certificates-key-manager

infra-operator-metrics-certs

Generated

Stored new private key in temporary Secret resource "infra-operator-metrics-certs-2zbtb"

openstack-operators

cert-manager-certificates-key-manager

infra-operator-metrics-certs

Generated

Stored new private key in temporary Secret resource "infra-operator-metrics-certs-2zbtb"

openstack-operators

cert-manager-certificates-key-manager

manila-operator-metrics-certs

Generated

Stored new private key in temporary Secret resource "manila-operator-metrics-certs-2rlhk"

openstack-operators

cert-manager-certificates-trigger

neutron-operator-metrics-certs

Issuing

Issuing certificate as Secret does not exist

openstack-operators

replicaset-controller

designate-operator-controller-manager-588d4d986b

SuccessfulCreate

Created pod: designate-operator-controller-manager-588d4d986b-bghwc

openstack-operators

replicaset-controller

openstack-baremetal-operator-controller-manager-74c4796899

SuccessfulCreate

Created pod: openstack-baremetal-operator-controller-manager-74c4796899s6vn4

openstack-operators

cert-manager-certificates-trigger

nova-operator-metrics-certs

Issuing

Issuing certificate as Secret does not exist

openstack-operators

replicaset-controller

openstack-baremetal-operator-controller-manager-74c4796899

SuccessfulCreate

Created pod: openstack-baremetal-operator-controller-manager-74c4796899s6vn4

openstack-operators

cert-manager-certificates-key-manager

octavia-operator-metrics-certs

Generated

Stored new private key in temporary Secret resource "octavia-operator-metrics-certs-hrr7h"

openstack-operators

replicaset-controller

ironic-operator-controller-manager-6f787dddc9

SuccessfulCreate

Created pod: ironic-operator-controller-manager-6f787dddc9-dq5f2

openstack-operators

deployment-controller

octavia-operator-controller-manager

ScalingReplicaSet

Scaled up replica set octavia-operator-controller-manager-5b9f45d989 to 1

openstack-operators

replicaset-controller

octavia-operator-controller-manager-5b9f45d989

SuccessfulCreate

Created pod: octavia-operator-controller-manager-5b9f45d989-fr295

openstack-operators

deployment-controller

ironic-operator-controller-manager

ScalingReplicaSet

Scaled up replica set ironic-operator-controller-manager-6f787dddc9 to 1

openstack-operators

deployment-controller

manila-operator-controller-manager

ScalingReplicaSet

Scaled up replica set manila-operator-controller-manager-55f864c847 to 1

openstack-operators

cert-manager-certificates-trigger

nova-operator-metrics-certs

Issuing

Issuing certificate as Secret does not exist

openstack-operators

replicaset-controller

manila-operator-controller-manager-55f864c847

SuccessfulCreate

Created pod: manila-operator-controller-manager-55f864c847-d49s9

openstack-operators

deployment-controller

nova-operator-controller-manager

ScalingReplicaSet

Scaled up replica set nova-operator-controller-manager-5d488d59fb to 1

openstack-operators

deployment-controller

placement-operator-controller-manager

ScalingReplicaSet

Scaled up replica set placement-operator-controller-manager-5784578c99 to 1

openstack-operators

replicaset-controller

cinder-operator-controller-manager-8d58dc466

SuccessfulCreate

Created pod: cinder-operator-controller-manager-8d58dc466-sfshp

openstack-operators

deployment-controller

infra-operator-controller-manager

ScalingReplicaSet

Scaled up replica set infra-operator-controller-manager-7dd6bb94c9 to 1

openstack-operators

replicaset-controller

nova-operator-controller-manager-5d488d59fb

SuccessfulCreate

Created pod: nova-operator-controller-manager-5d488d59fb-5rvds

openstack-operators

replicaset-controller

infra-operator-controller-manager-7dd6bb94c9

SuccessfulCreate

Created pod: infra-operator-controller-manager-7dd6bb94c9-7rp5f

openstack-operators

deployment-controller

ironic-operator-controller-manager

ScalingReplicaSet

Scaled up replica set ironic-operator-controller-manager-6f787dddc9 to 1

openstack-operators

replicaset-controller

ironic-operator-controller-manager-6f787dddc9

SuccessfulCreate

Created pod: ironic-operator-controller-manager-6f787dddc9-dq5f2

openstack-operators

replicaset-controller

keystone-operator-controller-manager-768b96df4c

SuccessfulCreate

Created pod: keystone-operator-controller-manager-768b96df4c-2zdk8

openstack-operators

deployment-controller

keystone-operator-controller-manager

ScalingReplicaSet

Scaled up replica set keystone-operator-controller-manager-768b96df4c to 1

openstack-operators

deployment-controller

cinder-operator-controller-manager

ScalingReplicaSet

Scaled up replica set cinder-operator-controller-manager-8d58dc466 to 1

openstack-operators

deployment-controller

ovn-operator-controller-manager

ScalingReplicaSet

Scaled up replica set ovn-operator-controller-manager-884679f54 to 1

openstack-operators

deployment-controller

nova-operator-controller-manager

ScalingReplicaSet

Scaled up replica set nova-operator-controller-manager-5d488d59fb to 1

openstack-operators

replicaset-controller

nova-operator-controller-manager-5d488d59fb

SuccessfulCreate

Created pod: nova-operator-controller-manager-5d488d59fb-5rvds

openstack-operators

replicaset-controller

ovn-operator-controller-manager-884679f54

SuccessfulCreate

Created pod: ovn-operator-controller-manager-884679f54-t875d

openstack-operators

replicaset-controller

barbican-operator-controller-manager-59bc569d95

SuccessfulCreate

Created pod: barbican-operator-controller-manager-59bc569d95-vds6m

openstack-operators

deployment-controller

barbican-operator-controller-manager

ScalingReplicaSet

Scaled up replica set barbican-operator-controller-manager-59bc569d95 to 1

openstack-operators

deployment-controller

infra-operator-controller-manager

ScalingReplicaSet

Scaled up replica set infra-operator-controller-manager-7dd6bb94c9 to 1

openstack-operators

replicaset-controller

infra-operator-controller-manager-7dd6bb94c9

SuccessfulCreate

Created pod: infra-operator-controller-manager-7dd6bb94c9-7rp5f

openstack-operators

deployment-controller

horizon-operator-controller-manager

ScalingReplicaSet

Scaled up replica set horizon-operator-controller-manager-8464cc45fb to 1

openstack-operators

replicaset-controller

horizon-operator-controller-manager-8464cc45fb

SuccessfulCreate

Created pod: horizon-operator-controller-manager-8464cc45fb-jbfpt

openstack-operators

replicaset-controller

barbican-operator-controller-manager-59bc569d95

SuccessfulCreate

Created pod: barbican-operator-controller-manager-59bc569d95-vds6m

openstack-operators

replicaset-controller

ovn-operator-controller-manager-884679f54

SuccessfulCreate

Created pod: ovn-operator-controller-manager-884679f54-t875d

openstack-operators

cert-manager-certificates-trigger

openstack-baremetal-operator-metrics-certs

Issuing

Issuing certificate as Secret does not exist

openstack-operators

replicaset-controller

cinder-operator-controller-manager-8d58dc466

SuccessfulCreate

Created pod: cinder-operator-controller-manager-8d58dc466-sfshp

openstack-operators

deployment-controller

cinder-operator-controller-manager

ScalingReplicaSet

Scaled up replica set cinder-operator-controller-manager-8d58dc466 to 1

openstack-operators

deployment-controller

heat-operator-controller-manager

ScalingReplicaSet

Scaled up replica set heat-operator-controller-manager-67dd5f86f5 to 1

openstack-operators

replicaset-controller

heat-operator-controller-manager-67dd5f86f5

SuccessfulCreate

Created pod: heat-operator-controller-manager-67dd5f86f5-5h2n9

openstack-operators

replicaset-controller

designate-operator-controller-manager-588d4d986b

SuccessfulCreate

Created pod: designate-operator-controller-manager-588d4d986b-bghwc

openstack-operators

deployment-controller

ovn-operator-controller-manager

ScalingReplicaSet

Scaled up replica set ovn-operator-controller-manager-884679f54 to 1

openstack-operators

deployment-controller

horizon-operator-controller-manager

ScalingReplicaSet

Scaled up replica set horizon-operator-controller-manager-8464cc45fb to 1

openstack-operators

deployment-controller

designate-operator-controller-manager

ScalingReplicaSet

Scaled up replica set designate-operator-controller-manager-588d4d986b to 1

openstack-operators

replicaset-controller

octavia-operator-controller-manager-5b9f45d989

SuccessfulCreate

Created pod: octavia-operator-controller-manager-5b9f45d989-fr295

openstack-operators

deployment-controller

octavia-operator-controller-manager

ScalingReplicaSet

Scaled up replica set octavia-operator-controller-manager-5b9f45d989 to 1

openstack-operators

deployment-controller

neutron-operator-controller-manager

ScalingReplicaSet

Scaled up replica set neutron-operator-controller-manager-767865f676 to 1

openstack-operators

replicaset-controller

neutron-operator-controller-manager-767865f676

SuccessfulCreate

Created pod: neutron-operator-controller-manager-767865f676-94tfj

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

designate-operator-metrics-certs-1

BadConfig

Certificate will be issued with an empty Issuer DN, which contravenes RFC 5280 and could break some strict clients

openstack-operators

cert-manager-certificaterequests-approver

designate-operator-metrics-certs-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

designate-operator-metrics-certs-1

CertificateIssued

Certificate fetched from issuer successfully

openstack-operators

deployment-controller

glance-operator-controller-manager

ScalingReplicaSet

Scaled up replica set glance-operator-controller-manager-79df6bcc97 to 1

openstack-operators

cert-manager-certificates-key-manager

octavia-operator-metrics-certs

Generated

Stored new private key in temporary Secret resource "octavia-operator-metrics-certs-hrr7h"

openstack-operators

deployment-controller

barbican-operator-controller-manager

ScalingReplicaSet

Scaled up replica set barbican-operator-controller-manager-59bc569d95 to 1

openstack-operators

replicaset-controller

glance-operator-controller-manager-79df6bcc97

SuccessfulCreate

Created pod: glance-operator-controller-manager-79df6bcc97-nt245

openstack-operators

deployment-controller

openstack-baremetal-operator-controller-manager

ScalingReplicaSet

Scaled up replica set openstack-baremetal-operator-controller-manager-74c4796899 to 1

openstack-operators

replicaset-controller

horizon-operator-controller-manager-8464cc45fb

SuccessfulCreate

Created pod: horizon-operator-controller-manager-8464cc45fb-jbfpt

openstack-operators

deployment-controller

mariadb-operator-controller-manager

ScalingReplicaSet

Scaled up replica set mariadb-operator-controller-manager-67ccfc9778 to 1

openstack-operators

deployment-controller

openstack-baremetal-operator-controller-manager

ScalingReplicaSet

Scaled up replica set openstack-baremetal-operator-controller-manager-74c4796899 to 1

openstack-operators

replicaset-controller

glance-operator-controller-manager-79df6bcc97

SuccessfulCreate

Created pod: glance-operator-controller-manager-79df6bcc97-nt245

openstack-operators

deployment-controller

glance-operator-controller-manager

ScalingReplicaSet

Scaled up replica set glance-operator-controller-manager-79df6bcc97 to 1

openstack-operators

deployment-controller

placement-operator-controller-manager

ScalingReplicaSet

Scaled up replica set placement-operator-controller-manager-5784578c99 to 1

openstack-operators

deployment-controller

keystone-operator-controller-manager

ScalingReplicaSet

Scaled up replica set keystone-operator-controller-manager-768b96df4c to 1

openstack-operators

replicaset-controller

mariadb-operator-controller-manager-67ccfc9778

SuccessfulCreate

Created pod: mariadb-operator-controller-manager-67ccfc9778-vxvcl

openstack-operators

replicaset-controller

keystone-operator-controller-manager-768b96df4c

SuccessfulCreate

Created pod: keystone-operator-controller-manager-768b96df4c-2zdk8

openstack-operators

deployment-controller

mariadb-operator-controller-manager

ScalingReplicaSet

Scaled up replica set mariadb-operator-controller-manager-67ccfc9778 to 1

openstack-operators

replicaset-controller

mariadb-operator-controller-manager-67ccfc9778

SuccessfulCreate

Created pod: mariadb-operator-controller-manager-67ccfc9778-vxvcl

openstack-operators

deployment-controller

neutron-operator-controller-manager

ScalingReplicaSet

Scaled up replica set neutron-operator-controller-manager-767865f676 to 1

openstack-operators

cert-manager-certificates-trigger

openstack-baremetal-operator-metrics-certs

Issuing

Issuing certificate as Secret does not exist

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

designate-operator-metrics-certs-1

CertificateIssued

Certificate fetched from issuer successfully

openstack-operators

cert-manager-certificaterequests-approver

designate-operator-metrics-certs-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

designate-operator-metrics-certs-1

BadConfig

Certificate will be issued with an empty Issuer DN, which contravenes RFC 5280 and could break some strict clients

openstack-operators

deployment-controller

designate-operator-controller-manager

ScalingReplicaSet

Scaled up replica set designate-operator-controller-manager-588d4d986b to 1

openstack-operators

replicaset-controller

neutron-operator-controller-manager-767865f676

SuccessfulCreate

Created pod: neutron-operator-controller-manager-767865f676-94tfj

openstack-operators

deployment-controller

manila-operator-controller-manager

ScalingReplicaSet

Scaled up replica set manila-operator-controller-manager-55f864c847 to 1

openstack-operators

replicaset-controller

manila-operator-controller-manager-55f864c847

SuccessfulCreate

Created pod: manila-operator-controller-manager-55f864c847-d49s9

openstack-operators

replicaset-controller

heat-operator-controller-manager-67dd5f86f5

SuccessfulCreate

Created pod: heat-operator-controller-manager-67dd5f86f5-5h2n9

openstack-operators

deployment-controller

heat-operator-controller-manager

ScalingReplicaSet

Scaled up replica set heat-operator-controller-manager-67dd5f86f5 to 1

openstack-operators

multus

cinder-operator-controller-manager-8d58dc466-sfshp

AddedInterface

Add eth0 [10.128.0.143/23] from ovn-kubernetes

openstack-operators

replicaset-controller

telemetry-operator-controller-manager-d6b694c5

SuccessfulCreate

Created pod: telemetry-operator-controller-manager-d6b694c5-lbvvg

openstack-operators

cert-manager-certificates-key-manager

keystone-operator-metrics-certs

Generated

Stored new private key in temporary Secret resource "keystone-operator-metrics-certs-lq9c7"

openstack-operators

deployment-controller

telemetry-operator-controller-manager

ScalingReplicaSet

Scaled up replica set telemetry-operator-controller-manager-d6b694c5 to 1

openstack-operators

deployment-controller

swift-operator-controller-manager

ScalingReplicaSet

Scaled up replica set swift-operator-controller-manager-c674c5965 to 1

openstack-operators

replicaset-controller

swift-operator-controller-manager-c674c5965

SuccessfulCreate

Created pod: swift-operator-controller-manager-c674c5965-srzmq

openstack-operators

replicaset-controller

placement-operator-controller-manager-5784578c99

SuccessfulCreate

Created pod: placement-operator-controller-manager-5784578c99-9wvcj

openstack-operators

deployment-controller

watcher-operator-controller-manager

ScalingReplicaSet

Scaled up replica set watcher-operator-controller-manager-6c4d75f7f9 to 1

openstack-operators

replicaset-controller

watcher-operator-controller-manager-6c4d75f7f9

SuccessfulCreate

Created pod: watcher-operator-controller-manager-6c4d75f7f9-kp7cg

openstack-operators

replicaset-controller

openstack-operator-controller-manager-86bd8996f6

SuccessfulCreate

Created pod: openstack-operator-controller-manager-86bd8996f6-cfkmp

openstack-operators

deployment-controller

openstack-operator-controller-manager

ScalingReplicaSet

Scaled up replica set openstack-operator-controller-manager-86bd8996f6 to 1

openstack-operators

deployment-controller

rabbitmq-cluster-operator-manager

ScalingReplicaSet

Scaled up replica set rabbitmq-cluster-operator-manager-668c99d594 to 1

openstack-operators

replicaset-controller

rabbitmq-cluster-operator-manager-668c99d594

SuccessfulCreate

Created pod: rabbitmq-cluster-operator-manager-668c99d594-s2dqt

openstack-operators

replicaset-controller

test-operator-controller-manager-5c5cb9c4d7

SuccessfulCreate

Created pod: test-operator-controller-manager-5c5cb9c4d7-nf9vn

openstack-operators

cert-manager-certificates-trigger

ovn-operator-metrics-certs

Issuing

Issuing certificate as Secret does not exist

openstack-operators

cert-manager-certificates-key-manager

keystone-operator-metrics-certs

Generated

Stored new private key in temporary Secret resource "keystone-operator-metrics-certs-lq9c7"

openstack-operators

deployment-controller

test-operator-controller-manager

ScalingReplicaSet

Scaled up replica set test-operator-controller-manager-5c5cb9c4d7 to 1

openstack-operators

multus

cinder-operator-controller-manager-8d58dc466-sfshp

AddedInterface

Add eth0 [10.128.0.143/23] from ovn-kubernetes

openstack-operators

deployment-controller

test-operator-controller-manager

ScalingReplicaSet

Scaled up replica set test-operator-controller-manager-5c5cb9c4d7 to 1

openstack-operators

replicaset-controller

openstack-operator-controller-manager-86bd8996f6

SuccessfulCreate

Created pod: openstack-operator-controller-manager-86bd8996f6-cfkmp

openstack-operators

deployment-controller

openstack-operator-controller-manager

ScalingReplicaSet

Scaled up replica set openstack-operator-controller-manager-86bd8996f6 to 1

openstack-operators

replicaset-controller

test-operator-controller-manager-5c5cb9c4d7

SuccessfulCreate

Created pod: test-operator-controller-manager-5c5cb9c4d7-nf9vn

openstack-operators

replicaset-controller

rabbitmq-cluster-operator-manager-668c99d594

SuccessfulCreate

Created pod: rabbitmq-cluster-operator-manager-668c99d594-s2dqt

openstack-operators

replicaset-controller

watcher-operator-controller-manager-6c4d75f7f9

SuccessfulCreate

Created pod: watcher-operator-controller-manager-6c4d75f7f9-kp7cg

openstack-operators

cert-manager-certificates-trigger

ovn-operator-metrics-certs

Issuing

Issuing certificate as Secret does not exist

openstack-operators

deployment-controller

watcher-operator-controller-manager

ScalingReplicaSet

Scaled up replica set watcher-operator-controller-manager-6c4d75f7f9 to 1

openstack-operators

deployment-controller

rabbitmq-cluster-operator-manager

ScalingReplicaSet

Scaled up replica set rabbitmq-cluster-operator-manager-668c99d594 to 1

openstack-operators

replicaset-controller

swift-operator-controller-manager-c674c5965

SuccessfulCreate

Created pod: swift-operator-controller-manager-c674c5965-srzmq

openstack-operators

deployment-controller

swift-operator-controller-manager

ScalingReplicaSet

Scaled up replica set swift-operator-controller-manager-c674c5965 to 1

openstack-operators

replicaset-controller

telemetry-operator-controller-manager-d6b694c5

SuccessfulCreate

Created pod: telemetry-operator-controller-manager-d6b694c5-lbvvg

openstack-operators

deployment-controller

telemetry-operator-controller-manager

ScalingReplicaSet

Scaled up replica set telemetry-operator-controller-manager-d6b694c5 to 1

openstack-operators

replicaset-controller

placement-operator-controller-manager-5784578c99

SuccessfulCreate

Created pod: placement-operator-controller-manager-5784578c99-9wvcj

openstack-operators

kubelet

designate-operator-controller-manager-588d4d986b-bghwc

Pulling

Pulling image "quay.io/openstack-k8s-operators/designate-operator@sha256:12841b27173f5f1beeb83112e057c8753f4cf411f583fba4f0610fac0f60b7ad"

openstack-operators

cert-manager-certificates-key-manager

neutron-operator-metrics-certs

Generated

Stored new private key in temporary Secret resource "neutron-operator-metrics-certs-jsrp2"

openstack-operators

kubelet

keystone-operator-controller-manager-768b96df4c-2zdk8

Pulling

Pulling image "quay.io/openstack-k8s-operators/keystone-operator@sha256:ec36a9083657587022f8471c9d5a71b87a7895398496e7fc546c73aa1eae4b56"

openstack-operators

kubelet

ironic-operator-controller-manager-6f787dddc9-dq5f2

Pulling

Pulling image "quay.io/openstack-k8s-operators/ironic-operator@sha256:9dd26bc51e7757d84736528d4988a1f980ad50ccb070aef6fc252e32c5c423a8"

openstack-operators

multus

designate-operator-controller-manager-588d4d986b-bghwc

AddedInterface

Add eth0 [10.128.0.144/23] from ovn-kubernetes

openstack-operators

kubelet

heat-operator-controller-manager-67dd5f86f5-5h2n9

Failed

Error: ErrImagePull

openstack-operators

multus

ironic-operator-controller-manager-6f787dddc9-dq5f2

AddedInterface

Add eth0 [10.128.0.149/23] from ovn-kubernetes

openstack-operators

kubelet

barbican-operator-controller-manager-59bc569d95-vds6m

Pulling

Pulling image "quay.io/openstack-k8s-operators/barbican-operator@sha256:7562d3e09bdac17f447f4523c5bd784c5f5ab5ca9cb2370a03b86126d6d7301d"

openstack-operators

multus

ironic-operator-controller-manager-6f787dddc9-dq5f2

AddedInterface

Add eth0 [10.128.0.149/23] from ovn-kubernetes

openstack-operators

kubelet

ironic-operator-controller-manager-6f787dddc9-dq5f2

Pulling

Pulling image "quay.io/openstack-k8s-operators/ironic-operator@sha256:9dd26bc51e7757d84736528d4988a1f980ad50ccb070aef6fc252e32c5c423a8"

openstack-operators

cert-manager-certificates-request-manager

infra-operator-metrics-certs

Requested

Created new CertificateRequest resource "infra-operator-metrics-certs-1"

openstack-operators

cert-manager-certificates-trigger

placement-operator-metrics-certs

Issuing

Issuing certificate as Secret does not exist

openstack-operators

cert-manager-certificates-request-manager

infra-operator-metrics-certs

Requested

Created new CertificateRequest resource "infra-operator-metrics-certs-1"

openstack-operators

kubelet

heat-operator-controller-manager-67dd5f86f5-5h2n9

Failed

Failed to pull image "quay.io/openstack-k8s-operators/heat-operator@sha256:c6ef5db244d874430a56c3cc9d27662e4bd57cdaa489e1f6059abcacf3aa0900": initializing source docker://quay.io/openstack-k8s-operators/heat-operator@sha256:c6ef5db244d874430a56c3cc9d27662e4bd57cdaa489e1f6059abcacf3aa0900: can't talk to a V1 container registry

openstack-operators

multus

barbican-operator-controller-manager-59bc569d95-vds6m

AddedInterface

Add eth0 [10.128.0.142/23] from ovn-kubernetes

openstack-operators

cert-manager-certificaterequests-issuer-acme

infra-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-venafi

infra-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-vault

infra-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

multus

heat-operator-controller-manager-67dd5f86f5-5h2n9

AddedInterface

Add eth0 [10.128.0.146/23] from ovn-kubernetes

openstack-operators

cert-manager-certificates-trigger

placement-operator-metrics-certs

Issuing

Issuing certificate as Secret does not exist

openstack-operators

cert-manager-certificaterequests-issuer-ca

infra-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

kubelet

cinder-operator-controller-manager-8d58dc466-sfshp

Pulling

Pulling image "quay.io/openstack-k8s-operators/cinder-operator@sha256:d8210bb21d4d298271a7b43f92fe58789393546e616aaaec1ce71bb2a754e777"

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

infra-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificates-key-manager

neutron-operator-metrics-certs

Generated

Stored new private key in temporary Secret resource "neutron-operator-metrics-certs-jsrp2"

openstack-operators

multus

glance-operator-controller-manager-79df6bcc97-nt245

AddedInterface

Add eth0 [10.128.0.145/23] from ovn-kubernetes

openstack-operators

kubelet

glance-operator-controller-manager-79df6bcc97-nt245

Pulling

Pulling image "quay.io/openstack-k8s-operators/glance-operator@sha256:76a1cde9f29fb39ed715b06be16adb803b9a2e24d68acb369911c0a88e33bc7d"

openstack-operators

cert-manager-certificates-key-manager

mariadb-operator-metrics-certs

Generated

Stored new private key in temporary Secret resource "mariadb-operator-metrics-certs-4cl8l"

openstack-operators

cert-manager-certificaterequests-issuer-acme

infra-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

multus

designate-operator-controller-manager-588d4d986b-bghwc

AddedInterface

Add eth0 [10.128.0.144/23] from ovn-kubernetes

openstack-operators

kubelet

designate-operator-controller-manager-588d4d986b-bghwc

Pulling

Pulling image "quay.io/openstack-k8s-operators/designate-operator@sha256:12841b27173f5f1beeb83112e057c8753f4cf411f583fba4f0610fac0f60b7ad"

openstack-operators

kubelet

glance-operator-controller-manager-79df6bcc97-nt245

Pulling

Pulling image "quay.io/openstack-k8s-operators/glance-operator@sha256:76a1cde9f29fb39ed715b06be16adb803b9a2e24d68acb369911c0a88e33bc7d"

openstack-operators

multus

glance-operator-controller-manager-79df6bcc97-nt245

AddedInterface

Add eth0 [10.128.0.145/23] from ovn-kubernetes

openstack-operators

multus

keystone-operator-controller-manager-768b96df4c-2zdk8

AddedInterface

Add eth0 [10.128.0.150/23] from ovn-kubernetes

openstack-operators

cert-manager-certificates-key-manager

nova-operator-metrics-certs

Generated

Stored new private key in temporary Secret resource "nova-operator-metrics-certs-x8bd5"

openstack-operators

cert-manager-certificates-key-manager

mariadb-operator-metrics-certs

Generated

Stored new private key in temporary Secret resource "mariadb-operator-metrics-certs-4cl8l"

openstack-operators

cert-manager-certificates-key-manager

nova-operator-metrics-certs

Generated

Stored new private key in temporary Secret resource "nova-operator-metrics-certs-x8bd5"

openstack-operators

cert-manager-certificaterequests-issuer-venafi

infra-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-vault

infra-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

kubelet

barbican-operator-controller-manager-59bc569d95-vds6m

Pulling

Pulling image "quay.io/openstack-k8s-operators/barbican-operator@sha256:7562d3e09bdac17f447f4523c5bd784c5f5ab5ca9cb2370a03b86126d6d7301d"

openstack-operators

cert-manager-certificaterequests-issuer-ca

infra-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

infra-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

kubelet

keystone-operator-controller-manager-768b96df4c-2zdk8

Pulling

Pulling image "quay.io/openstack-k8s-operators/keystone-operator@sha256:ec36a9083657587022f8471c9d5a71b87a7895398496e7fc546c73aa1eae4b56"

openstack-operators

multus

heat-operator-controller-manager-67dd5f86f5-5h2n9

AddedInterface

Add eth0 [10.128.0.146/23] from ovn-kubernetes

openstack-operators

kubelet

heat-operator-controller-manager-67dd5f86f5-5h2n9

Failed

Failed to pull image "quay.io/openstack-k8s-operators/heat-operator@sha256:c6ef5db244d874430a56c3cc9d27662e4bd57cdaa489e1f6059abcacf3aa0900": initializing source docker://quay.io/openstack-k8s-operators/heat-operator@sha256:c6ef5db244d874430a56c3cc9d27662e4bd57cdaa489e1f6059abcacf3aa0900: can't talk to a V1 container registry

openstack-operators

kubelet

heat-operator-controller-manager-67dd5f86f5-5h2n9

Failed

Error: ErrImagePull

openstack-operators

multus

keystone-operator-controller-manager-768b96df4c-2zdk8

AddedInterface

Add eth0 [10.128.0.150/23] from ovn-kubernetes

openstack-operators

kubelet

cinder-operator-controller-manager-8d58dc466-sfshp

Pulling

Pulling image "quay.io/openstack-k8s-operators/cinder-operator@sha256:d8210bb21d4d298271a7b43f92fe58789393546e616aaaec1ce71bb2a754e777"

openstack-operators

multus

barbican-operator-controller-manager-59bc569d95-vds6m

AddedInterface

Add eth0 [10.128.0.142/23] from ovn-kubernetes

openstack-operators

cert-manager-certificates-trigger

swift-operator-metrics-certs

Issuing

Issuing certificate as Secret does not exist

openstack-operators

multus

nova-operator-controller-manager-5d488d59fb-5rvds

AddedInterface

Add eth0 [10.128.0.154/23] from ovn-kubernetes

openstack-operators

kubelet

nova-operator-controller-manager-5d488d59fb-5rvds

Pulling

Pulling image "quay.io/openstack-k8s-operators/nova-operator@sha256:7398eb8fa5a4844d3326a5dff759d17199870c389b3ce3011a038b27bf95512a"

openstack-operators

kubelet

nova-operator-controller-manager-5d488d59fb-5rvds

Pulling

Pulling image "quay.io/openstack-k8s-operators/nova-operator@sha256:7398eb8fa5a4844d3326a5dff759d17199870c389b3ce3011a038b27bf95512a"

openstack-operators

kubelet

ovn-operator-controller-manager-884679f54-t875d

Pulling

Pulling image "quay.io/openstack-k8s-operators/ovn-operator@sha256:bef93f71d3b42a72d8b96c69bdb4db4b8bd797c5093a0a719443d7a5c9aaab55"

openstack-operators

kubelet

neutron-operator-controller-manager-767865f676-94tfj

Pulling

Pulling image "quay.io/openstack-k8s-operators/neutron-operator@sha256:526f9d4965431e1a5e4f8c3224bcee3f636a3108a5e0767296a994c2a517404a"

openstack-operators

kubelet

horizon-operator-controller-manager-8464cc45fb-jbfpt

Pulling

Pulling image "quay.io/openstack-k8s-operators/horizon-operator@sha256:703ad3a2b749bce100f1e2a445312b65dc3b8b45e8c8ba59f311d3f8f3368113"

openstack-operators

multus

horizon-operator-controller-manager-8464cc45fb-jbfpt

AddedInterface

Add eth0 [10.128.0.147/23] from ovn-kubernetes

openstack-operators

kubelet

octavia-operator-controller-manager-5b9f45d989-fr295

Pulling

Pulling image "quay.io/openstack-k8s-operators/octavia-operator@sha256:425fd66675becbe0ca2b2fe1a5a6694ac6e0b1cdce9a77a7a37f99785eadc74a"

openstack-operators

multus

swift-operator-controller-manager-c674c5965-srzmq

AddedInterface

Add eth0 [10.128.0.159/23] from ovn-kubernetes

openstack-operators

multus

manila-operator-controller-manager-55f864c847-d49s9

AddedInterface

Add eth0 [10.128.0.151/23] from ovn-kubernetes

openstack-operators

kubelet

manila-operator-controller-manager-55f864c847-d49s9

Pulling

Pulling image "quay.io/openstack-k8s-operators/manila-operator@sha256:f2e0b0fb34995b8acbbf1b0b60b5dbcf488b4f3899d1bb0763ae7dcee9bae6da"

openstack-operators

cert-manager-certificates-trigger

telemetry-operator-metrics-certs

Issuing

Issuing certificate as Secret does not exist

openstack-operators

multus

neutron-operator-controller-manager-767865f676-94tfj

AddedInterface

Add eth0 [10.128.0.153/23] from ovn-kubernetes

openstack-operators

kubelet

placement-operator-controller-manager-5784578c99-9wvcj

Pulling

Pulling image "quay.io/openstack-k8s-operators/placement-operator@sha256:c8743a6661d118b0e5ba3eb110643358a8a3237dc75984a8f9829880b55a1622"

openstack-operators

cert-manager-certificates-trigger

swift-operator-metrics-certs

Issuing

Issuing certificate as Secret does not exist

openstack-operators

multus

placement-operator-controller-manager-5784578c99-9wvcj

AddedInterface

Add eth0 [10.128.0.158/23] from ovn-kubernetes

openstack-operators

kubelet

mariadb-operator-controller-manager-67ccfc9778-vxvcl

Pulling

Pulling image "quay.io/openstack-k8s-operators/mariadb-operator@sha256:6e7552996253fc66667eaa3eb0e11b4e97145efa2ae577155ceabf8e9913ddc1"

openstack-operators

cert-manager-certificates-issuing

glance-operator-metrics-certs

Issuing

The certificate has been successfully issued

openstack-operators

kubelet

swift-operator-controller-manager-c674c5965-srzmq

Pulling

Pulling image "quay.io/openstack-k8s-operators/swift-operator@sha256:866844c5b88e1e0518ceb7490cac9d093da3fb8b2f27ba7bd9bd89f946b9ee6e"

openstack-operators

multus

mariadb-operator-controller-manager-67ccfc9778-vxvcl

AddedInterface

Add eth0 [10.128.0.152/23] from ovn-kubernetes

openstack-operators

multus

octavia-operator-controller-manager-5b9f45d989-fr295

AddedInterface

Add eth0 [10.128.0.155/23] from ovn-kubernetes

openstack-operators

cert-manager-certificates-issuing

designate-operator-metrics-certs

Issuing

The certificate has been successfully issued

openstack-operators

kubelet

watcher-operator-controller-manager-6c4d75f7f9-kp7cg

Pulling

Pulling image "quay.io/openstack-k8s-operators/watcher-operator@sha256:d9c55e8c6304a0e32289b5e8c69a87ea59b9968918a5c85b7c384633df82c807"

openstack-operators

multus

octavia-operator-controller-manager-5b9f45d989-fr295

AddedInterface

Add eth0 [10.128.0.155/23] from ovn-kubernetes

openstack-operators

multus

watcher-operator-controller-manager-6c4d75f7f9-kp7cg

AddedInterface

Add eth0 [10.128.0.162/23] from ovn-kubernetes

openstack-operators

multus

test-operator-controller-manager-5c5cb9c4d7-nf9vn

AddedInterface

Add eth0 [10.128.0.161/23] from ovn-kubernetes

openstack-operators

kubelet

test-operator-controller-manager-5c5cb9c4d7-nf9vn

Pulling

Pulling image "quay.io/openstack-k8s-operators/test-operator@sha256:43bd420bc05b4789243740bc75f61e10c7aac7883fc2f82b2d4d50085bc96c42"

openstack-operators

kubelet

swift-operator-controller-manager-c674c5965-srzmq

Pulling

Pulling image "quay.io/openstack-k8s-operators/swift-operator@sha256:866844c5b88e1e0518ceb7490cac9d093da3fb8b2f27ba7bd9bd89f946b9ee6e"

openstack-operators

multus

swift-operator-controller-manager-c674c5965-srzmq

AddedInterface

Add eth0 [10.128.0.159/23] from ovn-kubernetes

openstack-operators

kubelet

mariadb-operator-controller-manager-67ccfc9778-vxvcl

Pulling

Pulling image "quay.io/openstack-k8s-operators/mariadb-operator@sha256:6e7552996253fc66667eaa3eb0e11b4e97145efa2ae577155ceabf8e9913ddc1"

openstack-operators

multus

nova-operator-controller-manager-5d488d59fb-5rvds

AddedInterface

Add eth0 [10.128.0.154/23] from ovn-kubernetes

openstack-operators

multus

ovn-operator-controller-manager-884679f54-t875d

AddedInterface

Add eth0 [10.128.0.157/23] from ovn-kubernetes

openstack-operators

kubelet

horizon-operator-controller-manager-8464cc45fb-jbfpt

Pulling

Pulling image "quay.io/openstack-k8s-operators/horizon-operator@sha256:703ad3a2b749bce100f1e2a445312b65dc3b8b45e8c8ba59f311d3f8f3368113"

openstack-operators

multus

manila-operator-controller-manager-55f864c847-d49s9

AddedInterface

Add eth0 [10.128.0.151/23] from ovn-kubernetes

openstack-operators

kubelet

octavia-operator-controller-manager-5b9f45d989-fr295

Pulling

Pulling image "quay.io/openstack-k8s-operators/octavia-operator@sha256:425fd66675becbe0ca2b2fe1a5a6694ac6e0b1cdce9a77a7a37f99785eadc74a"

openstack-operators

cert-manager-certificaterequests-approver

infra-operator-metrics-certs-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack-operators

multus

ovn-operator-controller-manager-884679f54-t875d

AddedInterface

Add eth0 [10.128.0.157/23] from ovn-kubernetes

openstack-operators

multus

horizon-operator-controller-manager-8464cc45fb-jbfpt

AddedInterface

Add eth0 [10.128.0.147/23] from ovn-kubernetes

openstack-operators

kubelet

ovn-operator-controller-manager-884679f54-t875d

Pulling

Pulling image "quay.io/openstack-k8s-operators/ovn-operator@sha256:bef93f71d3b42a72d8b96c69bdb4db4b8bd797c5093a0a719443d7a5c9aaab55"

openstack-operators

cert-manager-certificates-issuing

designate-operator-metrics-certs

Issuing

The certificate has been successfully issued

openstack-operators

multus

mariadb-operator-controller-manager-67ccfc9778-vxvcl

AddedInterface

Add eth0 [10.128.0.152/23] from ovn-kubernetes

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

infra-operator-metrics-certs-1

BadConfig

Certificate will be issued with an empty Issuer DN, which contravenes RFC 5280 and could break some strict clients

openstack-operators

cert-manager-certificates-trigger

telemetry-operator-metrics-certs

Issuing

Issuing certificate as Secret does not exist

openstack-operators

multus

neutron-operator-controller-manager-767865f676-94tfj

AddedInterface

Add eth0 [10.128.0.153/23] from ovn-kubernetes

openstack-operators

kubelet

neutron-operator-controller-manager-767865f676-94tfj

Pulling

Pulling image "quay.io/openstack-k8s-operators/neutron-operator@sha256:526f9d4965431e1a5e4f8c3224bcee3f636a3108a5e0767296a994c2a517404a"

openstack-operators

multus

test-operator-controller-manager-5c5cb9c4d7-nf9vn

AddedInterface

Add eth0 [10.128.0.161/23] from ovn-kubernetes

openstack-operators

kubelet

test-operator-controller-manager-5c5cb9c4d7-nf9vn

Pulling

Pulling image "quay.io/openstack-k8s-operators/test-operator@sha256:43bd420bc05b4789243740bc75f61e10c7aac7883fc2f82b2d4d50085bc96c42"

openstack-operators

kubelet

manila-operator-controller-manager-55f864c847-d49s9

Pulling

Pulling image "quay.io/openstack-k8s-operators/manila-operator@sha256:f2e0b0fb34995b8acbbf1b0b60b5dbcf488b4f3899d1bb0763ae7dcee9bae6da"

openstack-operators

cert-manager-certificates-issuing

glance-operator-metrics-certs

Issuing

The certificate has been successfully issued

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

infra-operator-metrics-certs-1

CertificateIssued

Certificate fetched from issuer successfully

openstack-operators

cert-manager-certificaterequests-approver

infra-operator-metrics-certs-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

infra-operator-metrics-certs-1

BadConfig

Certificate will be issued with an empty Issuer DN, which contravenes RFC 5280 and could break some strict clients

openstack-operators

multus

watcher-operator-controller-manager-6c4d75f7f9-kp7cg

AddedInterface

Add eth0 [10.128.0.162/23] from ovn-kubernetes

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

infra-operator-metrics-certs-1

CertificateIssued

Certificate fetched from issuer successfully

openstack-operators

multus

placement-operator-controller-manager-5784578c99-9wvcj

AddedInterface

Add eth0 [10.128.0.158/23] from ovn-kubernetes

openstack-operators

kubelet

placement-operator-controller-manager-5784578c99-9wvcj

Pulling

Pulling image "quay.io/openstack-k8s-operators/placement-operator@sha256:c8743a6661d118b0e5ba3eb110643358a8a3237dc75984a8f9829880b55a1622"

openstack-operators

kubelet

watcher-operator-controller-manager-6c4d75f7f9-kp7cg

Pulling

Pulling image "quay.io/openstack-k8s-operators/watcher-operator@sha256:d9c55e8c6304a0e32289b5e8c69a87ea59b9968918a5c85b7c384633df82c807"

openstack-operators

cert-manager-certificaterequests-issuer-venafi

horizon-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

neutron-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-vault

neutron-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-acme

neutron-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-ca

neutron-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

multus

rabbitmq-cluster-operator-manager-668c99d594-s2dqt

AddedInterface

Add eth0 [10.128.0.164/23] from ovn-kubernetes

openstack-operators

cert-manager-certificaterequests-issuer-venafi

neutron-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

neutron-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-vault

neutron-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificates-trigger

openstack-baremetal-operator-serving-cert

Issuing

Issuing certificate as Secret does not exist

openstack-operators

cert-manager-certificates-key-manager

openstack-baremetal-operator-metrics-certs

Generated

Stored new private key in temporary Secret resource "openstack-baremetal-operator-metrics-certs-mcxqq"

openstack-operators

cert-manager-certificaterequests-issuer-acme

neutron-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificates-trigger

test-operator-metrics-certs

Issuing

Issuing certificate as Secret does not exist

openstack-operators

cert-manager-certificates-request-manager

mariadb-operator-metrics-certs

Requested

Created new CertificateRequest resource "mariadb-operator-metrics-certs-1"

openstack-operators

cert-manager-certificaterequests-issuer-ca

neutron-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-venafi

neutron-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-vault

mariadb-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificates-request-manager

neutron-operator-metrics-certs

Requested

Created new CertificateRequest resource "neutron-operator-metrics-certs-1"

openstack-operators

cert-manager-certificates-trigger

test-operator-metrics-certs

Issuing

Issuing certificate as Secret does not exist

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

mariadb-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-ca

horizon-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificates-request-manager

neutron-operator-metrics-certs

Requested

Created new CertificateRequest resource "neutron-operator-metrics-certs-1"

openstack-operators

cert-manager-certificaterequests-issuer-ca

mariadb-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-venafi

horizon-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-acme

mariadb-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-venafi

mariadb-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificates-request-manager

keystone-operator-metrics-certs

Requested

Created new CertificateRequest resource "keystone-operator-metrics-certs-1"

openstack-operators

cert-manager-certificates-key-manager

openstack-baremetal-operator-metrics-certs

Generated

Stored new private key in temporary Secret resource "openstack-baremetal-operator-metrics-certs-mcxqq"

openstack-operators

multus

telemetry-operator-controller-manager-d6b694c5-lbvvg

AddedInterface

Add eth0 [10.128.0.160/23] from ovn-kubernetes

openstack-operators

cert-manager-certificates-trigger

openstack-baremetal-operator-serving-cert

Issuing

Issuing certificate as Secret does not exist

openstack-operators

cert-manager-certificates-request-manager

keystone-operator-metrics-certs

Requested

Created new CertificateRequest resource "keystone-operator-metrics-certs-1"

openstack-operators

cert-manager-certificaterequests-issuer-acme

horizon-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-vault

keystone-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-ca

keystone-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-vault

horizon-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-acme

keystone-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-ca

nova-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-ca

horizon-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

keystone-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-venafi

keystone-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

horizon-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificates-request-manager

horizon-operator-metrics-certs

Requested

Created new CertificateRequest resource "horizon-operator-metrics-certs-1"

openstack-operators

cert-manager-certificaterequests-issuer-vault

nova-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-venafi

nova-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-acme

nova-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-acme

horizon-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificates-request-manager

mariadb-operator-metrics-certs

Requested

Created new CertificateRequest resource "mariadb-operator-metrics-certs-1"

openstack-operators

cert-manager-certificaterequests-issuer-vault

horizon-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

horizon-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-venafi

keystone-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

nova-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

keystone-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificates-trigger

watcher-operator-metrics-certs

Issuing

Issuing certificate as Secret does not exist

openstack-operators

cert-manager-certificaterequests-issuer-ca

nova-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

multus

rabbitmq-cluster-operator-manager-668c99d594-s2dqt

AddedInterface

Add eth0 [10.128.0.164/23] from ovn-kubernetes

openstack-operators

cert-manager-certificaterequests-issuer-vault

nova-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

multus

telemetry-operator-controller-manager-d6b694c5-lbvvg

AddedInterface

Add eth0 [10.128.0.160/23] from ovn-kubernetes

openstack-operators

cert-manager-certificates-request-manager

nova-operator-metrics-certs

Requested

Created new CertificateRequest resource "nova-operator-metrics-certs-1"

openstack-operators

cert-manager-certificates-request-manager

horizon-operator-metrics-certs

Requested

Created new CertificateRequest resource "horizon-operator-metrics-certs-1"

openstack-operators

cert-manager-certificaterequests-issuer-venafi

nova-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-acme

nova-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-acme

keystone-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-venafi

mariadb-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificates-trigger

watcher-operator-metrics-certs

Issuing

Issuing certificate as Secret does not exist

openstack-operators

cert-manager-certificaterequests-issuer-acme

mariadb-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-ca

keystone-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-ca

mariadb-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-vault

keystone-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

mariadb-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificates-request-manager

nova-operator-metrics-certs

Requested

Created new CertificateRequest resource "nova-operator-metrics-certs-1"

openstack-operators

cert-manager-certificates-trigger

infra-operator-serving-cert

Issuing

Issuing certificate as Secret does not exist

openstack-operators

cert-manager-certificates-trigger

infra-operator-serving-cert

Issuing

Issuing certificate as Secret does not exist

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

nova-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-vault

mariadb-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

keystone-operator-metrics-certs-1

BadConfig

Certificate will be issued with an empty Issuer DN, which contravenes RFC 5280 and could break some strict clients

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

manila-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-vault

manila-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

mariadb-operator-metrics-certs-1

BadConfig

Certificate will be issued with an empty Issuer DN, which contravenes RFC 5280 and could break some strict clients

openstack-operators

kubelet

rabbitmq-cluster-operator-manager-668c99d594-s2dqt

Pulling

Pulling image "quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2"

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

mariadb-operator-metrics-certs-1

CertificateIssued

Certificate fetched from issuer successfully

openstack-operators

cert-manager-certificaterequests-approver

horizon-operator-metrics-certs-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

horizon-operator-metrics-certs-1

BadConfig

Certificate will be issued with an empty Issuer DN, which contravenes RFC 5280 and could break some strict clients

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

horizon-operator-metrics-certs-1

CertificateIssued

Certificate fetched from issuer successfully

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

manila-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificates-trigger

openstack-operator-serving-cert

Issuing

Issuing certificate as Secret does not exist

openstack-operators

kubelet

telemetry-operator-controller-manager-d6b694c5-lbvvg

Pulling

Pulling image "quay.io/openstack-k8s-operators/telemetry-operator@sha256:c500fa7080b94105e85eeced772d8872e4168904e74ba02116e15ab66f522444"

openstack-operators

cert-manager-certificaterequests-issuer-venafi

manila-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificates-request-manager

heat-operator-metrics-certs

Requested

Created new CertificateRequest resource "heat-operator-metrics-certs-1"

openstack-operators

cert-manager-certificates-request-manager

manila-operator-metrics-certs

Requested

Created new CertificateRequest resource "manila-operator-metrics-certs-1"

openstack-operators

cert-manager-certificaterequests-issuer-vault

heat-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-acme

heat-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificates-request-manager

ironic-operator-metrics-certs

Requested

Created new CertificateRequest resource "ironic-operator-metrics-certs-1"

openstack-operators

cert-manager-certificaterequests-issuer-venafi

heat-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

heat-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-acme

manila-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-acme

heat-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-vault

heat-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-ca

heat-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

heat-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-venafi

heat-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved
(x2)

openstack-operators

kubelet

heat-operator-controller-manager-67dd5f86f5-5h2n9

Failed

Error: ImagePullBackOff
(x2)

openstack-operators

kubelet

heat-operator-controller-manager-67dd5f86f5-5h2n9

BackOff

Back-off pulling image "quay.io/openstack-k8s-operators/heat-operator@sha256:c6ef5db244d874430a56c3cc9d27662e4bd57cdaa489e1f6059abcacf3aa0900"

openstack-operators

cert-manager-certificaterequests-approver

mariadb-operator-metrics-certs-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack-operators

cert-manager-certificates-request-manager

heat-operator-metrics-certs

Requested

Created new CertificateRequest resource "heat-operator-metrics-certs-1"

openstack-operators

cert-manager-certificaterequests-issuer-ca

manila-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

kubelet

rabbitmq-cluster-operator-manager-668c99d594-s2dqt

Pulling

Pulling image "quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2"

openstack-operators

cert-manager-certificaterequests-issuer-ca

ironic-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

keystone-operator-metrics-certs-1

CertificateIssued

Certificate fetched from issuer successfully

openstack-operators

cert-manager-certificaterequests-approver

mariadb-operator-metrics-certs-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack-operators

cert-manager-certificaterequests-issuer-venafi

ironic-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-approver

keystone-operator-metrics-certs-1

cert-manager.io

Certificate request has been approved by cert-manager.io
(x2)

openstack-operators

kubelet

heat-operator-controller-manager-67dd5f86f5-5h2n9

Failed

Error: ImagePullBackOff

openstack-operators

cert-manager-certificaterequests-issuer-ca

ironic-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-vault

manila-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-vault

ironic-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

ironic-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-acme

ironic-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-venafi

ironic-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved
(x2)

openstack-operators

kubelet

heat-operator-controller-manager-67dd5f86f5-5h2n9

BackOff

Back-off pulling image "quay.io/openstack-k8s-operators/heat-operator@sha256:c6ef5db244d874430a56c3cc9d27662e4bd57cdaa489e1f6059abcacf3aa0900"

openstack-operators

cert-manager-certificates-trigger

openstack-operator-metrics-certs

Issuing

Issuing certificate as Secret does not exist

openstack-operators

cert-manager-certificaterequests-issuer-ca

heat-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-acme

ironic-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

ironic-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificates-key-manager

ovn-operator-metrics-certs

Generated

Stored new private key in temporary Secret resource "ovn-operator-metrics-certs-dswml"

openstack-operators

cert-manager-certificates-key-manager

ovn-operator-metrics-certs

Generated

Stored new private key in temporary Secret resource "ovn-operator-metrics-certs-dswml"

openstack-operators

cert-manager-certificaterequests-approver

keystone-operator-metrics-certs-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack-operators

cert-manager-certificaterequests-issuer-venafi

manila-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificates-request-manager

manila-operator-metrics-certs

Requested

Created new CertificateRequest resource "manila-operator-metrics-certs-1"

openstack-operators

cert-manager-certificaterequests-issuer-acme

manila-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-vault

ironic-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

keystone-operator-metrics-certs-1

BadConfig

Certificate will be issued with an empty Issuer DN, which contravenes RFC 5280 and could break some strict clients

openstack-operators

cert-manager-certificates-trigger

openstack-operator-metrics-certs

Issuing

Issuing certificate as Secret does not exist

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

keystone-operator-metrics-certs-1

CertificateIssued

Certificate fetched from issuer successfully

openstack-operators

cert-manager-certificaterequests-approver

horizon-operator-metrics-certs-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

horizon-operator-metrics-certs-1

BadConfig

Certificate will be issued with an empty Issuer DN, which contravenes RFC 5280 and could break some strict clients

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

mariadb-operator-metrics-certs-1

BadConfig

Certificate will be issued with an empty Issuer DN, which contravenes RFC 5280 and could break some strict clients

openstack-operators

cert-manager-certificates-trigger

openstack-operator-serving-cert

Issuing

Issuing certificate as Secret does not exist

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

mariadb-operator-metrics-certs-1

CertificateIssued

Certificate fetched from issuer successfully

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

horizon-operator-metrics-certs-1

CertificateIssued

Certificate fetched from issuer successfully

openstack-operators

kubelet

telemetry-operator-controller-manager-d6b694c5-lbvvg

Pulling

Pulling image "quay.io/openstack-k8s-operators/telemetry-operator@sha256:c500fa7080b94105e85eeced772d8872e4168904e74ba02116e15ab66f522444"

openstack-operators

cert-manager-certificaterequests-issuer-ca

manila-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificates-request-manager

ironic-operator-metrics-certs

Requested

Created new CertificateRequest resource "ironic-operator-metrics-certs-1"

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

neutron-operator-metrics-certs-1

BadConfig

Certificate will be issued with an empty Issuer DN, which contravenes RFC 5280 and could break some strict clients

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

neutron-operator-metrics-certs-1

CertificateIssued

Certificate fetched from issuer successfully

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

neutron-operator-metrics-certs-1

BadConfig

Certificate will be issued with an empty Issuer DN, which contravenes RFC 5280 and could break some strict clients

openstack-operators

cert-manager-certificaterequests-approver

neutron-operator-metrics-certs-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack-operators

cert-manager-certificaterequests-issuer-venafi

octavia-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-vault

octavia-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-ca

octavia-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-acme

octavia-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

nova-operator-metrics-certs-1

BadConfig

Certificate will be issued with an empty Issuer DN, which contravenes RFC 5280 and could break some strict clients

openstack-operators

cert-manager-certificaterequests-approver

nova-operator-metrics-certs-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack-operators

cert-manager-certificates-request-manager

octavia-operator-metrics-certs

Requested

Created new CertificateRequest resource "octavia-operator-metrics-certs-1"

openstack-operators

cert-manager-certificates-request-manager

octavia-operator-metrics-certs

Requested

Created new CertificateRequest resource "octavia-operator-metrics-certs-1"

openstack-operators

cert-manager-certificaterequests-approver

nova-operator-metrics-certs-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack-operators

cert-manager-certificates-key-manager

placement-operator-metrics-certs

Generated

Stored new private key in temporary Secret resource "placement-operator-metrics-certs-5ksw9"

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

neutron-operator-metrics-certs-1

CertificateIssued

Certificate fetched from issuer successfully

openstack-operators

cert-manager-certificaterequests-issuer-acme

octavia-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-ca

octavia-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificates-key-manager

placement-operator-metrics-certs

Generated

Stored new private key in temporary Secret resource "placement-operator-metrics-certs-5ksw9"

openstack-operators

cert-manager-certificaterequests-issuer-vault

octavia-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-venafi

octavia-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

nova-operator-metrics-certs-1

CertificateIssued

Certificate fetched from issuer successfully

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

nova-operator-metrics-certs-1

CertificateIssued

Certificate fetched from issuer successfully

openstack-operators

cert-manager-certificaterequests-approver

neutron-operator-metrics-certs-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

nova-operator-metrics-certs-1

BadConfig

Certificate will be issued with an empty Issuer DN, which contravenes RFC 5280 and could break some strict clients

openstack-operators

cert-manager-certificaterequests-approver

heat-operator-metrics-certs-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack-operators

cert-manager-certificaterequests-approver

manila-operator-metrics-certs-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

heat-operator-metrics-certs-1

CertificateIssued

Certificate fetched from issuer successfully

openstack-operators

cert-manager-certificaterequests-issuer-vault

openstack-baremetal-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificates-request-manager

openstack-baremetal-operator-metrics-certs

Requested

Created new CertificateRequest resource "openstack-baremetal-operator-metrics-certs-1"

openstack-operators

cert-manager-certificates-request-manager

openstack-baremetal-operator-metrics-certs

Requested

Created new CertificateRequest resource "openstack-baremetal-operator-metrics-certs-1"

openstack-operators

cert-manager-certificates-request-manager

ovn-operator-metrics-certs

Requested

Created new CertificateRequest resource "ovn-operator-metrics-certs-1"

openstack-operators

cert-manager-certificaterequests-issuer-ca

openstack-baremetal-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

openstack-baremetal-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-acme

openstack-baremetal-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-venafi

openstack-baremetal-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificates-key-manager

telemetry-operator-metrics-certs

Generated

Stored new private key in temporary Secret resource "telemetry-operator-metrics-certs-nnmsn"

openstack-operators

cert-manager-certificaterequests-approver

ironic-operator-metrics-certs-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack-operators

cert-manager-certificaterequests-approver

heat-operator-metrics-certs-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack-operators

cert-manager-certificaterequests-issuer-vault

openstack-baremetal-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-ca

openstack-baremetal-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificates-key-manager

telemetry-operator-metrics-certs

Generated

Stored new private key in temporary Secret resource "telemetry-operator-metrics-certs-nnmsn"

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

heat-operator-metrics-certs-1

CertificateIssued

Certificate fetched from issuer successfully

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

openstack-baremetal-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

ovn-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-acme

openstack-baremetal-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-venafi

openstack-baremetal-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-ca

ovn-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-acme

ovn-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificates-key-manager

swift-operator-metrics-certs

Generated

Stored new private key in temporary Secret resource "swift-operator-metrics-certs-qmf9f"

openstack-operators

cert-manager-certificaterequests-issuer-venafi

ovn-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

ironic-operator-metrics-certs-1

BadConfig

Certificate will be issued with an empty Issuer DN, which contravenes RFC 5280 and could break some strict clients

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

manila-operator-metrics-certs-1

CertificateIssued

Certificate fetched from issuer successfully

openstack-operators

cert-manager-certificaterequests-issuer-vault

ovn-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-vault

ovn-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificates-key-manager

swift-operator-metrics-certs

Generated

Stored new private key in temporary Secret resource "swift-operator-metrics-certs-qmf9f"

openstack-operators

cert-manager-certificaterequests-issuer-venafi

ovn-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-approver

octavia-operator-metrics-certs-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

octavia-operator-metrics-certs-1

CertificateIssued

Certificate fetched from issuer successfully

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

octavia-operator-metrics-certs-1

BadConfig

Certificate will be issued with an empty Issuer DN, which contravenes RFC 5280 and could break some strict clients

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

octavia-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-acme

ovn-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-ca

ovn-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

ovn-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

heat-operator-metrics-certs-1

BadConfig

Certificate will be issued with an empty Issuer DN, which contravenes RFC 5280 and could break some strict clients

openstack-operators

cert-manager-certificaterequests-approver

octavia-operator-metrics-certs-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

heat-operator-metrics-certs-1

BadConfig

Certificate will be issued with an empty Issuer DN, which contravenes RFC 5280 and could break some strict clients

openstack-operators

cert-manager-certificates-request-manager

ovn-operator-metrics-certs

Requested

Created new CertificateRequest resource "ovn-operator-metrics-certs-1"

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

octavia-operator-metrics-certs-1

CertificateIssued

Certificate fetched from issuer successfully

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

manila-operator-metrics-certs-1

CertificateIssued

Certificate fetched from issuer successfully

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

ironic-operator-metrics-certs-1

CertificateIssued

Certificate fetched from issuer successfully

openstack-operators

cert-manager-certificaterequests-approver

ironic-operator-metrics-certs-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

ironic-operator-metrics-certs-1

BadConfig

Certificate will be issued with an empty Issuer DN, which contravenes RFC 5280 and could break some strict clients

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

octavia-operator-metrics-certs-1

BadConfig

Certificate will be issued with an empty Issuer DN, which contravenes RFC 5280 and could break some strict clients

openstack-operators

cert-manager-certificaterequests-approver

manila-operator-metrics-certs-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

ironic-operator-metrics-certs-1

CertificateIssued

Certificate fetched from issuer successfully

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

manila-operator-metrics-certs-1

BadConfig

Certificate will be issued with an empty Issuer DN, which contravenes RFC 5280 and could break some strict clients

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

octavia-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

manila-operator-metrics-certs-1

BadConfig

Certificate will be issued with an empty Issuer DN, which contravenes RFC 5280 and could break some strict clients

openstack-operators

cert-manager-certificates-request-manager

placement-operator-metrics-certs

Requested

Created new CertificateRequest resource "placement-operator-metrics-certs-1"

openstack-operators

cert-manager-certificaterequests-issuer-venafi

placement-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-acme

placement-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

openstack-baremetal-operator-metrics-certs-1

BadConfig

Certificate will be issued with an empty Issuer DN, which contravenes RFC 5280 and could break some strict clients

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

ovn-operator-metrics-certs-1

CertificateIssued

Certificate fetched from issuer successfully

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

ovn-operator-metrics-certs-1

BadConfig

Certificate will be issued with an empty Issuer DN, which contravenes RFC 5280 and could break some strict clients

openstack-operators

cert-manager-certificaterequests-approver

ovn-operator-metrics-certs-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack-operators

cert-manager-certificaterequests-issuer-vault

placement-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

placement-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-ca

placement-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

placement-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificates-issuing

keystone-operator-metrics-certs

Issuing

The certificate has been successfully issued

openstack-operators

cert-manager-certificaterequests-issuer-ca

placement-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-venafi

placement-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-approver

ovn-operator-metrics-certs-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack-operators

cert-manager-certificaterequests-issuer-vault

placement-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

ovn-operator-metrics-certs-1

BadConfig

Certificate will be issued with an empty Issuer DN, which contravenes RFC 5280 and could break some strict clients

openstack-operators

cert-manager-certificaterequests-issuer-acme

placement-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-approver

openstack-baremetal-operator-metrics-certs-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack-operators

cert-manager-certificates-issuing

keystone-operator-metrics-certs

Issuing

The certificate has been successfully issued

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

openstack-baremetal-operator-metrics-certs-1

BadConfig

Certificate will be issued with an empty Issuer DN, which contravenes RFC 5280 and could break some strict clients

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

openstack-baremetal-operator-metrics-certs-1

CertificateIssued

Certificate fetched from issuer successfully

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

ovn-operator-metrics-certs-1

CertificateIssued

Certificate fetched from issuer successfully

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

openstack-baremetal-operator-metrics-certs-1

CertificateIssued

Certificate fetched from issuer successfully

openstack-operators

cert-manager-certificates-request-manager

placement-operator-metrics-certs

Requested

Created new CertificateRequest resource "placement-operator-metrics-certs-1"

openstack-operators

cert-manager-certificaterequests-approver

openstack-baremetal-operator-metrics-certs-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack-operators

cert-manager-certificaterequests-issuer-acme

swift-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-vault

swift-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-ca

telemetry-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-approver

placement-operator-metrics-certs-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack-operators

cert-manager-certificates-request-manager

telemetry-operator-metrics-certs

Requested

Created new CertificateRequest resource "telemetry-operator-metrics-certs-1"

openstack-operators

cert-manager-certificates-issuing

mariadb-operator-metrics-certs

Issuing

The certificate has been successfully issued

openstack-operators

cert-manager-certificaterequests-issuer-venafi

telemetry-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificates-key-manager

test-operator-metrics-certs

Generated

Stored new private key in temporary Secret resource "test-operator-metrics-certs-52s64"

openstack-operators

cert-manager-certificates-request-manager

swift-operator-metrics-certs

Requested

Created new CertificateRequest resource "swift-operator-metrics-certs-1"

openstack-operators

cert-manager-certificaterequests-issuer-vault

telemetry-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-acme

telemetry-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-ca

swift-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

swift-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-venafi

telemetry-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificates-request-manager

swift-operator-metrics-certs

Requested

Created new CertificateRequest resource "swift-operator-metrics-certs-1"

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

placement-operator-metrics-certs-1

BadConfig

Certificate will be issued with an empty Issuer DN, which contravenes RFC 5280 and could break some strict clients

openstack-operators

cert-manager-certificaterequests-issuer-ca

swift-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

swift-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-acme

telemetry-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-acme

swift-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-venafi

swift-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

placement-operator-metrics-certs-1

CertificateIssued

Certificate fetched from issuer successfully

openstack-operators

cert-manager-certificates-issuing

horizon-operator-metrics-certs

Issuing

The certificate has been successfully issued

openstack-operators

cert-manager-certificaterequests-issuer-venafi

swift-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-vault

swift-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificates-issuing

horizon-operator-metrics-certs

Issuing

The certificate has been successfully issued

openstack-operators

cert-manager-certificaterequests-issuer-vault

telemetry-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificates-key-manager

watcher-operator-metrics-certs

Generated

Stored new private key in temporary Secret resource "watcher-operator-metrics-certs-n6bg8"

openstack-operators

cert-manager-certificaterequests-issuer-ca

telemetry-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificates-request-manager

telemetry-operator-metrics-certs

Requested

Created new CertificateRequest resource "telemetry-operator-metrics-certs-1"

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

telemetry-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificates-key-manager

watcher-operator-metrics-certs

Generated

Stored new private key in temporary Secret resource "watcher-operator-metrics-certs-n6bg8"

openstack-operators

cert-manager-certificates-key-manager

infra-operator-serving-cert

Generated

Stored new private key in temporary Secret resource "infra-operator-serving-cert-8hddq"

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

telemetry-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificates-key-manager

test-operator-metrics-certs

Generated

Stored new private key in temporary Secret resource "test-operator-metrics-certs-52s64"

openstack-operators

cert-manager-certificates-issuing

mariadb-operator-metrics-certs

Issuing

The certificate has been successfully issued

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

placement-operator-metrics-certs-1

BadConfig

Certificate will be issued with an empty Issuer DN, which contravenes RFC 5280 and could break some strict clients

openstack-operators

cert-manager-certificaterequests-approver

placement-operator-metrics-certs-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack-operators

cert-manager-certificates-key-manager

infra-operator-serving-cert

Generated

Stored new private key in temporary Secret resource "infra-operator-serving-cert-8hddq"

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

placement-operator-metrics-certs-1

CertificateIssued

Certificate fetched from issuer successfully

openstack-operators

cert-manager-certificates-key-manager

openstack-baremetal-operator-serving-cert

Generated

Stored new private key in temporary Secret resource "openstack-baremetal-operator-serving-cert-xbf45"

openstack-operators

cert-manager-certificates-key-manager

openstack-baremetal-operator-serving-cert

Generated

Stored new private key in temporary Secret resource "openstack-baremetal-operator-serving-cert-xbf45"

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

telemetry-operator-metrics-certs-1

BadConfig

Certificate will be issued with an empty Issuer DN, which contravenes RFC 5280 and could break some strict clients

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

telemetry-operator-metrics-certs-1

CertificateIssued

Certificate fetched from issuer successfully

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

telemetry-operator-metrics-certs-1

CertificateIssued

Certificate fetched from issuer successfully

openstack-operators

cert-manager-certificates-key-manager

openstack-operator-metrics-certs

Generated

Stored new private key in temporary Secret resource "openstack-operator-metrics-certs-8ntw9"

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

telemetry-operator-metrics-certs-1

BadConfig

Certificate will be issued with an empty Issuer DN, which contravenes RFC 5280 and could break some strict clients

openstack-operators

cert-manager-certificaterequests-approver

telemetry-operator-metrics-certs-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack-operators

cert-manager-certificaterequests-approver

telemetry-operator-metrics-certs-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack-operators

cert-manager-certificates-key-manager

openstack-operator-metrics-certs

Generated

Stored new private key in temporary Secret resource "openstack-operator-metrics-certs-8ntw9"

openstack-operators

cert-manager-certificaterequests-issuer-venafi

test-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

swift-operator-metrics-certs-1

BadConfig

Certificate will be issued with an empty Issuer DN, which contravenes RFC 5280 and could break some strict clients

openstack-operators

cert-manager-certificaterequests-issuer-ca

watcher-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-venafi

watcher-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificates-issuing

nova-operator-metrics-certs

Issuing

The certificate has been successfully issued

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

infra-operator-serving-cert-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-acme

infra-operator-serving-cert-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-vault

test-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificates-request-manager

watcher-operator-metrics-certs

Requested

Created new CertificateRequest resource "watcher-operator-metrics-certs-1"

openstack-operators

cert-manager-certificates-issuing

nova-operator-metrics-certs

Issuing

The certificate has been successfully issued

openstack-operators

cert-manager-certificaterequests-issuer-vault

infra-operator-serving-cert-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificates-issuing

heat-operator-metrics-certs

Issuing

The certificate has been successfully issued

openstack-operators

cert-manager-certificaterequests-issuer-ca

infra-operator-serving-cert-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificates-issuing

heat-operator-metrics-certs

Issuing

The certificate has been successfully issued

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

infra-operator-serving-cert-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-venafi

infra-operator-serving-cert-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

swift-operator-metrics-certs-1

BadConfig

Certificate will be issued with an empty Issuer DN, which contravenes RFC 5280 and could break some strict clients

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

swift-operator-metrics-certs-1

CertificateIssued

Certificate fetched from issuer successfully

openstack-operators

cert-manager-certificates-request-manager

test-operator-metrics-certs

Requested

Created new CertificateRequest resource "test-operator-metrics-certs-1"

openstack-operators

cert-manager-certificates-request-manager

test-operator-metrics-certs

Requested

Created new CertificateRequest resource "test-operator-metrics-certs-1"

openstack-operators

cert-manager-certificaterequests-approver

swift-operator-metrics-certs-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

watcher-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-acme

infra-operator-serving-cert-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-ca

infra-operator-serving-cert-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

test-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-vault

infra-operator-serving-cert-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-ca

watcher-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-acme

watcher-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-ca

test-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-venafi

infra-operator-serving-cert-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-vault

test-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-venafi

watcher-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

swift-operator-metrics-certs-1

CertificateIssued

Certificate fetched from issuer successfully

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

test-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificates-request-manager

infra-operator-serving-cert

Requested

Created new CertificateRequest resource "infra-operator-serving-cert-1"

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

watcher-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-acme

test-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-acme

watcher-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-vault

watcher-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-vault

watcher-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-venafi

test-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificates-request-manager

infra-operator-serving-cert

Requested

Created new CertificateRequest resource "infra-operator-serving-cert-1"

openstack-operators

cert-manager-certificaterequests-approver

swift-operator-metrics-certs-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack-operators

cert-manager-certificates-request-manager

watcher-operator-metrics-certs

Requested

Created new CertificateRequest resource "watcher-operator-metrics-certs-1"

openstack-operators

cert-manager-certificaterequests-issuer-acme

test-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-ca

test-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-venafi

openstack-baremetal-operator-serving-cert-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-acme

openstack-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificates-key-manager

openstack-operator-serving-cert

Generated

Stored new private key in temporary Secret resource "openstack-operator-serving-cert-69bdz"

openstack-operators

cert-manager-certificaterequests-issuer-ca

openstack-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-vault

openstack-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-vault

openstack-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificates-issuing

ironic-operator-metrics-certs

Issuing

The certificate has been successfully issued

openstack-operators

cert-manager-certificaterequests-issuer-acme

openstack-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-ca

openstack-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificates-issuing

manila-operator-metrics-certs

Issuing

The certificate has been successfully issued

openstack-operators

cert-manager-certificaterequests-issuer-vault

openstack-baremetal-operator-serving-cert-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

openstack-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-venafi

openstack-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-vault

openstack-baremetal-operator-serving-cert-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificates-issuing

ironic-operator-metrics-certs

Issuing

The certificate has been successfully issued

openstack-operators

cert-manager-certificaterequests-issuer-venafi

openstack-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-acme

openstack-baremetal-operator-serving-cert-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

openstack-baremetal-operator-serving-cert-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificates-request-manager

openstack-operator-metrics-certs

Requested

Created new CertificateRequest resource "openstack-operator-metrics-certs-1"

openstack-operators

cert-manager-certificaterequests-issuer-ca

openstack-baremetal-operator-serving-cert-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificates-request-manager

openstack-baremetal-operator-serving-cert

Requested

Created new CertificateRequest resource "openstack-baremetal-operator-serving-cert-1"

openstack-operators

cert-manager-certificaterequests-issuer-venafi

openstack-baremetal-operator-serving-cert-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificates-request-manager

openstack-operator-metrics-certs

Requested

Created new CertificateRequest resource "openstack-operator-metrics-certs-1"

openstack-operators

cert-manager-certificates-request-manager

openstack-baremetal-operator-serving-cert

Requested

Created new CertificateRequest resource "openstack-baremetal-operator-serving-cert-1"

openstack-operators

cert-manager-certificaterequests-issuer-acme

openstack-baremetal-operator-serving-cert-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificates-key-manager

openstack-operator-serving-cert

Generated

Stored new private key in temporary Secret resource "openstack-operator-serving-cert-69bdz"

openstack-operators

cert-manager-certificaterequests-issuer-ca

openstack-baremetal-operator-serving-cert-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificates-issuing

manila-operator-metrics-certs

Issuing

The certificate has been successfully issued

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

openstack-baremetal-operator-serving-cert-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

openstack-operator-metrics-certs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

watcher-operator-metrics-certs-1

CertificateIssued

Certificate fetched from issuer successfully

openstack-operators

cert-manager-certificaterequests-approver

watcher-operator-metrics-certs-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack-operators

cert-manager-certificaterequests-approver

watcher-operator-metrics-certs-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

test-operator-metrics-certs-1

CertificateIssued

Certificate fetched from issuer successfully

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

watcher-operator-metrics-certs-1

BadConfig

Certificate will be issued with an empty Issuer DN, which contravenes RFC 5280 and could break some strict clients

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

test-operator-metrics-certs-1

BadConfig

Certificate will be issued with an empty Issuer DN, which contravenes RFC 5280 and could break some strict clients

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

test-operator-metrics-certs-1

CertificateIssued

Certificate fetched from issuer successfully

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

infra-operator-serving-cert-1

BadConfig

Certificate will be issued with an empty Issuer DN, which contravenes RFC 5280 and could break some strict clients

openstack-operators

cert-manager-certificaterequests-approver

test-operator-metrics-certs-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

watcher-operator-metrics-certs-1

BadConfig

Certificate will be issued with an empty Issuer DN, which contravenes RFC 5280 and could break some strict clients

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

infra-operator-serving-cert-1

CertificateIssued

Certificate fetched from issuer successfully

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

infra-operator-serving-cert-1

CertificateIssued

Certificate fetched from issuer successfully

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

test-operator-metrics-certs-1

BadConfig

Certificate will be issued with an empty Issuer DN, which contravenes RFC 5280 and could break some strict clients

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

infra-operator-serving-cert-1

BadConfig

Certificate will be issued with an empty Issuer DN, which contravenes RFC 5280 and could break some strict clients

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

watcher-operator-metrics-certs-1

CertificateIssued

Certificate fetched from issuer successfully

openstack-operators

cert-manager-certificaterequests-approver

test-operator-metrics-certs-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack-operators

cert-manager-certificaterequests-approver

infra-operator-serving-cert-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack-operators

cert-manager-certificaterequests-approver

infra-operator-serving-cert-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack-operators

cert-manager-certificates-issuing

neutron-operator-metrics-certs

Issuing

The certificate has been successfully issued

openstack-operators

cert-manager-certificaterequests-approver

openstack-operator-metrics-certs-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

openstack-operator-metrics-certs-1

BadConfig

Certificate will be issued with an empty Issuer DN, which contravenes RFC 5280 and could break some strict clients

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

openstack-operator-metrics-certs-1

CertificateIssued

Certificate fetched from issuer successfully

openstack-operators

cert-manager-certificaterequests-approver

openstack-operator-metrics-certs-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

openstack-operator-metrics-certs-1

BadConfig

Certificate will be issued with an empty Issuer DN, which contravenes RFC 5280 and could break some strict clients

openstack-operators

cert-manager-certificates-issuing

infra-operator-metrics-certs

Issuing

The certificate has been successfully issued

openstack-operators

cert-manager-certificates-issuing

infra-operator-metrics-certs

Issuing

The certificate has been successfully issued

openstack-operators

cert-manager-certificaterequests-approver

openstack-baremetal-operator-serving-cert-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

openstack-baremetal-operator-serving-cert-1

BadConfig

Certificate will be issued with an empty Issuer DN, which contravenes RFC 5280 and could break some strict clients

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

openstack-baremetal-operator-serving-cert-1

CertificateIssued

Certificate fetched from issuer successfully

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

openstack-operator-metrics-certs-1

CertificateIssued

Certificate fetched from issuer successfully

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

openstack-baremetal-operator-serving-cert-1

CertificateIssued

Certificate fetched from issuer successfully

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

openstack-baremetal-operator-serving-cert-1

BadConfig

Certificate will be issued with an empty Issuer DN, which contravenes RFC 5280 and could break some strict clients

openstack-operators

cert-manager-certificaterequests-approver

openstack-baremetal-operator-serving-cert-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack-operators

cert-manager-certificates-issuing

neutron-operator-metrics-certs

Issuing

The certificate has been successfully issued

openstack-operators

kubelet

barbican-operator-controller-manager-59bc569d95-vds6m

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/barbican-operator@sha256:7562d3e09bdac17f447f4523c5bd784c5f5ab5ca9cb2370a03b86126d6d7301d" in 14.132s (14.132s including waiting). Image size: 191122394 bytes.

openstack-operators

cert-manager-certificaterequests-issuer-ca

openstack-operator-serving-cert-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

openstack-operator-serving-cert-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificates-issuing

telemetry-operator-metrics-certs

Issuing

The certificate has been successfully issued

openstack-operators

kubelet

barbican-operator-controller-manager-59bc569d95-vds6m

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/barbican-operator@sha256:7562d3e09bdac17f447f4523c5bd784c5f5ab5ca9cb2370a03b86126d6d7301d" in 14.132s (14.132s including waiting). Image size: 191122394 bytes.

openstack-operators

kubelet

cinder-operator-controller-manager-8d58dc466-sfshp

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/cinder-operator@sha256:d8210bb21d4d298271a7b43f92fe58789393546e616aaaec1ce71bb2a754e777" in 14.322s (14.322s including waiting). Image size: 191447488 bytes.

openstack-operators

kubelet

keystone-operator-controller-manager-768b96df4c-2zdk8

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/keystone-operator@sha256:ec36a9083657587022f8471c9d5a71b87a7895398496e7fc546c73aa1eae4b56" in 13.545s (13.545s including waiting). Image size: 193037461 bytes.

openstack-operators

cert-manager-certificaterequests-issuer-venafi

openstack-operator-serving-cert-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-acme

openstack-operator-serving-cert-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-acme

openstack-operator-serving-cert-1

WaitingForApproval

Not signing CertificateRequest until it is Approved
(x6)

openstack-operators

kubelet

infra-operator-controller-manager-7dd6bb94c9-7rp5f

FailedMount

MountVolume.SetUp failed for volume "cert" : secret "infra-operator-webhook-server-cert" not found

openstack-operators

cert-manager-certificaterequests-issuer-vault

openstack-operator-serving-cert-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-ca

openstack-operator-serving-cert-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-venafi

openstack-operator-serving-cert-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

openstack-operator-serving-cert-1

WaitingForApproval

Not signing CertificateRequest until it is Approved
(x6)

openstack-operators

kubelet

infra-operator-controller-manager-7dd6bb94c9-7rp5f

FailedMount

MountVolume.SetUp failed for volume "cert" : secret "infra-operator-webhook-server-cert" not found

openstack-operators

cert-manager-certificates-issuing

ovn-operator-metrics-certs

Issuing

The certificate has been successfully issued

openstack-operators

cert-manager-certificates-issuing

ovn-operator-metrics-certs

Issuing

The certificate has been successfully issued

openstack-operators

cert-manager-certificates-issuing

telemetry-operator-metrics-certs

Issuing

The certificate has been successfully issued

openstack-operators

cert-manager-certificates-request-manager

openstack-operator-serving-cert

Requested

Created new CertificateRequest resource "openstack-operator-serving-cert-1"

openstack-operators

kubelet

keystone-operator-controller-manager-768b96df4c-2zdk8

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/keystone-operator@sha256:ec36a9083657587022f8471c9d5a71b87a7895398496e7fc546c73aa1eae4b56" in 13.545s (13.545s including waiting). Image size: 193037461 bytes.

openstack-operators

kubelet

cinder-operator-controller-manager-8d58dc466-sfshp

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/cinder-operator@sha256:d8210bb21d4d298271a7b43f92fe58789393546e616aaaec1ce71bb2a754e777" in 14.322s (14.322s including waiting). Image size: 191447488 bytes.

openstack-operators

cert-manager-certificates-request-manager

openstack-operator-serving-cert

Requested

Created new CertificateRequest resource "openstack-operator-serving-cert-1"

openstack-operators

cert-manager-certificaterequests-issuer-vault

openstack-operator-serving-cert-1

WaitingForApproval

Not signing CertificateRequest until it is Approved
(x6)

openstack-operators

kubelet

openstack-operator-controller-manager-86bd8996f6-cfkmp

FailedMount

MountVolume.SetUp failed for volume "webhook-certs" : secret "webhook-server-cert" not found

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

openstack-operator-serving-cert-1

CertificateIssued

Certificate fetched from issuer successfully
(x6)

openstack-operators

kubelet

openstack-operator-controller-manager-86bd8996f6-cfkmp

FailedMount

MountVolume.SetUp failed for volume "webhook-certs" : secret "webhook-server-cert" not found
(x6)

openstack-operators

kubelet

openstack-baremetal-operator-controller-manager-74c4796899s6vn4

FailedMount

MountVolume.SetUp failed for volume "cert" : secret "openstack-baremetal-operator-webhook-server-cert" not found

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

openstack-operator-serving-cert-1

BadConfig

Certificate will be issued with an empty Issuer DN, which contravenes RFC 5280 and could break some strict clients
(x6)

openstack-operators

kubelet

openstack-operator-controller-manager-86bd8996f6-cfkmp

FailedMount

MountVolume.SetUp failed for volume "metrics-certs" : secret "metrics-server-cert" not found
(x6)

openstack-operators

kubelet

openstack-operator-controller-manager-86bd8996f6-cfkmp

FailedMount

MountVolume.SetUp failed for volume "metrics-certs" : secret "metrics-server-cert" not found

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

openstack-operator-serving-cert-1

CertificateIssued

Certificate fetched from issuer successfully

openstack-operators

cert-manager-certificates-issuing

placement-operator-metrics-certs

Issuing

The certificate has been successfully issued
(x6)

openstack-operators

kubelet

openstack-baremetal-operator-controller-manager-74c4796899s6vn4

FailedMount

MountVolume.SetUp failed for volume "cert" : secret "openstack-baremetal-operator-webhook-server-cert" not found

openstack-operators

cert-manager-certificaterequests-approver

openstack-operator-serving-cert-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack-operators

cert-manager-certificates-issuing

placement-operator-metrics-certs

Issuing

The certificate has been successfully issued

openstack-operators

cert-manager-certificaterequests-approver

openstack-operator-serving-cert-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack-operators

cert-manager-certificaterequests-issuer-selfsigned

openstack-operator-serving-cert-1

BadConfig

Certificate will be issued with an empty Issuer DN, which contravenes RFC 5280 and could break some strict clients

openstack-operators

cert-manager-certificates-issuing

watcher-operator-metrics-certs

Issuing

The certificate has been successfully issued

openstack-operators

cert-manager-certificates-issuing

infra-operator-serving-cert

Issuing

The certificate has been successfully issued

openstack-operators

cert-manager-certificates-issuing

watcher-operator-metrics-certs

Issuing

The certificate has been successfully issued

openstack-operators

cert-manager-certificates-issuing

swift-operator-metrics-certs

Issuing

The certificate has been successfully issued
(x2)

openstack-operators

kubelet

heat-operator-controller-manager-67dd5f86f5-5h2n9

Pulling

Pulling image "quay.io/openstack-k8s-operators/heat-operator@sha256:c6ef5db244d874430a56c3cc9d27662e4bd57cdaa489e1f6059abcacf3aa0900"
(x2)

openstack-operators

kubelet

heat-operator-controller-manager-67dd5f86f5-5h2n9

Pulling

Pulling image "quay.io/openstack-k8s-operators/heat-operator@sha256:c6ef5db244d874430a56c3cc9d27662e4bd57cdaa489e1f6059abcacf3aa0900"

openstack-operators

cert-manager-certificates-issuing

swift-operator-metrics-certs

Issuing

The certificate has been successfully issued

openstack-operators

cert-manager-certificates-issuing

octavia-operator-metrics-certs

Issuing

The certificate has been successfully issued

openstack-operators

cert-manager-certificates-issuing

octavia-operator-metrics-certs

Issuing

The certificate has been successfully issued

openstack-operators

cert-manager-certificates-issuing

infra-operator-serving-cert

Issuing

The certificate has been successfully issued

openstack-operators

cert-manager-certificates-issuing

test-operator-metrics-certs

Issuing

The certificate has been successfully issued

openstack-operators

kubelet

neutron-operator-controller-manager-767865f676-94tfj

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/neutron-operator@sha256:526f9d4965431e1a5e4f8c3224bcee3f636a3108a5e0767296a994c2a517404a" in 16.7s (16.7s including waiting). Image size: 191045581 bytes.

openstack-operators

cert-manager-certificates-issuing

test-operator-metrics-certs

Issuing

The certificate has been successfully issued

openstack-operators

cert-manager-certificates-issuing

openstack-baremetal-operator-metrics-certs

Issuing

The certificate has been successfully issued

openstack-operators

cert-manager-certificates-issuing

openstack-baremetal-operator-metrics-certs

Issuing

The certificate has been successfully issued

openstack-operators

kubelet

horizon-operator-controller-manager-8464cc45fb-jbfpt

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/horizon-operator@sha256:703ad3a2b749bce100f1e2a445312b65dc3b8b45e8c8ba59f311d3f8f3368113" in 16.469s (16.469s including waiting). Image size: 190382026 bytes.

openstack-operators

cert-manager-certificates-issuing

openstack-baremetal-operator-serving-cert

Issuing

The certificate has been successfully issued

openstack-operators

kubelet

horizon-operator-controller-manager-8464cc45fb-jbfpt

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/horizon-operator@sha256:703ad3a2b749bce100f1e2a445312b65dc3b8b45e8c8ba59f311d3f8f3368113" in 16.469s (16.469s including waiting). Image size: 190382026 bytes.

openstack-operators

kubelet

ironic-operator-controller-manager-6f787dddc9-dq5f2

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/ironic-operator@sha256:9dd26bc51e7757d84736528d4988a1f980ad50ccb070aef6fc252e32c5c423a8" in 17.241s (17.241s including waiting). Image size: 191690176 bytes.

openstack-operators

cert-manager-certificates-issuing

openstack-baremetal-operator-serving-cert

Issuing

The certificate has been successfully issued

openstack-operators

kubelet

ironic-operator-controller-manager-6f787dddc9-dq5f2

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/ironic-operator@sha256:9dd26bc51e7757d84736528d4988a1f980ad50ccb070aef6fc252e32c5c423a8" in 17.241s (17.241s including waiting). Image size: 191690176 bytes.

openstack-operators

kubelet

neutron-operator-controller-manager-767865f676-94tfj

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/neutron-operator@sha256:526f9d4965431e1a5e4f8c3224bcee3f636a3108a5e0767296a994c2a517404a" in 16.7s (16.7s including waiting). Image size: 191045581 bytes.

openstack-operators

cert-manager-certificates-issuing

openstack-operator-metrics-certs

Issuing

The certificate has been successfully issued

openstack-operators

cert-manager-certificates-issuing

openstack-operator-metrics-certs

Issuing

The certificate has been successfully issued

openstack-operators

cert-manager-certificates-issuing

openstack-operator-serving-cert

Issuing

The certificate has been successfully issued

openstack-operators

cert-manager-certificates-issuing

openstack-operator-serving-cert

Issuing

The certificate has been successfully issued

openstack-operators

kubelet

manila-operator-controller-manager-55f864c847-d49s9

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/manila-operator@sha256:f2e0b0fb34995b8acbbf1b0b60b5dbcf488b4f3899d1bb0763ae7dcee9bae6da" in 18.714s (18.714s including waiting). Image size: 191263167 bytes.

openstack-operators

kubelet

manila-operator-controller-manager-55f864c847-d49s9

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/manila-operator@sha256:f2e0b0fb34995b8acbbf1b0b60b5dbcf488b4f3899d1bb0763ae7dcee9bae6da" in 18.714s (18.714s including waiting). Image size: 191263167 bytes.

openstack-operators

kubelet

telemetry-operator-controller-manager-d6b694c5-lbvvg

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/telemetry-operator@sha256:c500fa7080b94105e85eeced772d8872e4168904e74ba02116e15ab66f522444" in 17.594s (17.594s including waiting). Image size: 196297190 bytes.

openstack-operators

kubelet

glance-operator-controller-manager-79df6bcc97-nt245

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/glance-operator@sha256:76a1cde9f29fb39ed715b06be16adb803b9a2e24d68acb369911c0a88e33bc7d" in 20.971s (20.971s including waiting). Image size: 192008127 bytes.

openstack-operators

kubelet

telemetry-operator-controller-manager-d6b694c5-lbvvg

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/telemetry-operator@sha256:c500fa7080b94105e85eeced772d8872e4168904e74ba02116e15ab66f522444" in 17.594s (17.594s including waiting). Image size: 196297190 bytes.

openstack-operators

kubelet

ovn-operator-controller-manager-884679f54-t875d

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/ovn-operator@sha256:bef93f71d3b42a72d8b96c69bdb4db4b8bd797c5093a0a719443d7a5c9aaab55" in 19.779s (19.779s including waiting). Image size: 190114710 bytes.

openstack-operators

kubelet

placement-operator-controller-manager-5784578c99-9wvcj

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/placement-operator@sha256:c8743a6661d118b0e5ba3eb110643358a8a3237dc75984a8f9829880b55a1622" in 19.545s (19.545s including waiting). Image size: 190627813 bytes.

openstack-operators

kubelet

designate-operator-controller-manager-588d4d986b-bghwc

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/designate-operator@sha256:12841b27173f5f1beeb83112e057c8753f4cf411f583fba4f0610fac0f60b7ad" in 21.006s (21.006s including waiting). Image size: 195976677 bytes.

openstack-operators

kubelet

octavia-operator-controller-manager-5b9f45d989-fr295

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/octavia-operator@sha256:425fd66675becbe0ca2b2fe1a5a6694ac6e0b1cdce9a77a7a37f99785eadc74a" in 19.812s (19.812s including waiting). Image size: 193570760 bytes.

openstack-operators

kubelet

designate-operator-controller-manager-588d4d986b-bghwc

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/designate-operator@sha256:12841b27173f5f1beeb83112e057c8753f4cf411f583fba4f0610fac0f60b7ad" in 21.006s (21.006s including waiting). Image size: 195976677 bytes.

openstack-operators

kubelet

mariadb-operator-controller-manager-67ccfc9778-vxvcl

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/mariadb-operator@sha256:6e7552996253fc66667eaa3eb0e11b4e97145efa2ae577155ceabf8e9913ddc1" in 19.981s (19.981s including waiting). Image size: 189431506 bytes.

openstack-operators

kubelet

octavia-operator-controller-manager-5b9f45d989-fr295

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/octavia-operator@sha256:425fd66675becbe0ca2b2fe1a5a6694ac6e0b1cdce9a77a7a37f99785eadc74a" in 19.812s (19.812s including waiting). Image size: 193570760 bytes.

openstack-operators

kubelet

test-operator-controller-manager-5c5cb9c4d7-nf9vn

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/test-operator@sha256:43bd420bc05b4789243740bc75f61e10c7aac7883fc2f82b2d4d50085bc96c42" in 19.545s (19.545s including waiting). Image size: 188906426 bytes.

openstack-operators

kubelet

watcher-operator-controller-manager-6c4d75f7f9-kp7cg

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/watcher-operator@sha256:d9c55e8c6304a0e32289b5e8c69a87ea59b9968918a5c85b7c384633df82c807" in 19.574s (19.574s including waiting). Image size: 191011789 bytes.

openstack-operators

kubelet

watcher-operator-controller-manager-6c4d75f7f9-kp7cg

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/watcher-operator@sha256:d9c55e8c6304a0e32289b5e8c69a87ea59b9968918a5c85b7c384633df82c807" in 19.574s (19.574s including waiting). Image size: 191011789 bytes.

openstack-operators

kubelet

nova-operator-controller-manager-5d488d59fb-5rvds

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/nova-operator@sha256:7398eb8fa5a4844d3326a5dff759d17199870c389b3ce3011a038b27bf95512a" in 20.011s (20.011s including waiting). Image size: 193632103 bytes.

openstack-operators

kubelet

mariadb-operator-controller-manager-67ccfc9778-vxvcl

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/mariadb-operator@sha256:6e7552996253fc66667eaa3eb0e11b4e97145efa2ae577155ceabf8e9913ddc1" in 19.981s (19.981s including waiting). Image size: 189431506 bytes.

openstack-operators

kubelet

ovn-operator-controller-manager-884679f54-t875d

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/ovn-operator@sha256:bef93f71d3b42a72d8b96c69bdb4db4b8bd797c5093a0a719443d7a5c9aaab55" in 19.779s (19.779s including waiting). Image size: 190114710 bytes.

openstack-operators

kubelet

glance-operator-controller-manager-79df6bcc97-nt245

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/glance-operator@sha256:76a1cde9f29fb39ed715b06be16adb803b9a2e24d68acb369911c0a88e33bc7d" in 20.971s (20.971s including waiting). Image size: 192008127 bytes.

openstack-operators

kubelet

test-operator-controller-manager-5c5cb9c4d7-nf9vn

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/test-operator@sha256:43bd420bc05b4789243740bc75f61e10c7aac7883fc2f82b2d4d50085bc96c42" in 19.545s (19.545s including waiting). Image size: 188906426 bytes.

openstack-operators

kubelet

placement-operator-controller-manager-5784578c99-9wvcj

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/placement-operator@sha256:c8743a6661d118b0e5ba3eb110643358a8a3237dc75984a8f9829880b55a1622" in 19.545s (19.545s including waiting). Image size: 190627813 bytes.

openstack-operators

kubelet

nova-operator-controller-manager-5d488d59fb-5rvds

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/nova-operator@sha256:7398eb8fa5a4844d3326a5dff759d17199870c389b3ce3011a038b27bf95512a" in 20.011s (20.011s including waiting). Image size: 193632103 bytes.

openstack-operators

kubelet

swift-operator-controller-manager-c674c5965-srzmq

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/swift-operator@sha256:866844c5b88e1e0518ceb7490cac9d093da3fb8b2f27ba7bd9bd89f946b9ee6e" in 20.844s (20.844s including waiting). Image size: 192133556 bytes.

openstack-operators

kubelet

barbican-operator-controller-manager-59bc569d95-vds6m

Created

Created container: manager

openstack-operators

kubelet

swift-operator-controller-manager-c674c5965-srzmq

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/swift-operator@sha256:866844c5b88e1e0518ceb7490cac9d093da3fb8b2f27ba7bd9bd89f946b9ee6e" in 20.844s (20.844s including waiting). Image size: 192133556 bytes.

openstack-operators

kubelet

barbican-operator-controller-manager-59bc569d95-vds6m

Created

Created container: manager

openstack-operators

kubelet

cinder-operator-controller-manager-8d58dc466-sfshp

Created

Created container: manager

openstack-operators

kubelet

cinder-operator-controller-manager-8d58dc466-sfshp

Created

Created container: manager

openstack-operators

kubelet

keystone-operator-controller-manager-768b96df4c-2zdk8

Created

Created container: manager

openstack-operators

kubelet

rabbitmq-cluster-operator-manager-668c99d594-s2dqt

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" in 19.836s (19.836s including waiting). Image size: 176351298 bytes.

openstack-operators

kubelet

keystone-operator-controller-manager-768b96df4c-2zdk8

Started

Started container manager

openstack-operators

cinder-operator-controller-manager-8d58dc466-sfshp_5fcbc638-26c7-4fb8-b9be-bb1323b3986c

a6b6a260.openstack.org

LeaderElection

cinder-operator-controller-manager-8d58dc466-sfshp_5fcbc638-26c7-4fb8-b9be-bb1323b3986c became leader

openstack-operators

kubelet

barbican-operator-controller-manager-59bc569d95-vds6m

Started

Started container manager

openstack-operators

kubelet

cinder-operator-controller-manager-8d58dc466-sfshp

Started

Started container manager

openstack-operators

kubelet

rabbitmq-cluster-operator-manager-668c99d594-s2dqt

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" in 19.836s (19.836s including waiting). Image size: 176351298 bytes.

openstack-operators

barbican-operator-controller-manager-59bc569d95-vds6m_1ba9fd13-e385-4941-a366-05d1b925db97

8cc931b9.openstack.org

LeaderElection

barbican-operator-controller-manager-59bc569d95-vds6m_1ba9fd13-e385-4941-a366-05d1b925db97 became leader

openstack-operators

kubelet

cinder-operator-controller-manager-8d58dc466-sfshp

Started

Started container manager

openstack-operators

kubelet

keystone-operator-controller-manager-768b96df4c-2zdk8

Created

Created container: manager

openstack-operators

kubelet

heat-operator-controller-manager-67dd5f86f5-5h2n9

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/heat-operator@sha256:c6ef5db244d874430a56c3cc9d27662e4bd57cdaa489e1f6059abcacf3aa0900" in 7.694s (7.694s including waiting). Image size: 191633317 bytes.

openstack-operators

kubelet

manila-operator-controller-manager-55f864c847-d49s9

Started

Started container manager

openstack-operators

kubelet

manila-operator-controller-manager-55f864c847-d49s9

Started

Started container manager

openstack-operators

kubelet

manila-operator-controller-manager-55f864c847-d49s9

Created

Created container: manager

openstack-operators

barbican-operator-controller-manager-59bc569d95-vds6m_1ba9fd13-e385-4941-a366-05d1b925db97

8cc931b9.openstack.org

LeaderElection

barbican-operator-controller-manager-59bc569d95-vds6m_1ba9fd13-e385-4941-a366-05d1b925db97 became leader

openstack-operators

kubelet

manila-operator-controller-manager-55f864c847-d49s9

Created

Created container: manager

openstack-operators

cinder-operator-controller-manager-8d58dc466-sfshp_5fcbc638-26c7-4fb8-b9be-bb1323b3986c

a6b6a260.openstack.org

LeaderElection

cinder-operator-controller-manager-8d58dc466-sfshp_5fcbc638-26c7-4fb8-b9be-bb1323b3986c became leader

openstack-operators

kubelet

heat-operator-controller-manager-67dd5f86f5-5h2n9

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/heat-operator@sha256:c6ef5db244d874430a56c3cc9d27662e4bd57cdaa489e1f6059abcacf3aa0900" in 7.694s (7.694s including waiting). Image size: 191633317 bytes.

openstack-operators

kubelet

keystone-operator-controller-manager-768b96df4c-2zdk8

Started

Started container manager

openstack-operators

kubelet

barbican-operator-controller-manager-59bc569d95-vds6m

Started

Started container manager

openstack-operators

kubelet

neutron-operator-controller-manager-767865f676-94tfj

Started

Started container manager

openstack-operators

rabbitmq-cluster-operator-manager-668c99d594-s2dqt_894c896e-e62e-42f1-a996-b5fb9f021d17

rabbitmq-cluster-operator-leader-election

LeaderElection

rabbitmq-cluster-operator-manager-668c99d594-s2dqt_894c896e-e62e-42f1-a996-b5fb9f021d17 became leader

openstack-operators

kubelet

nova-operator-controller-manager-5d488d59fb-5rvds

Started

Started container manager

openstack-operators

kubelet

heat-operator-controller-manager-67dd5f86f5-5h2n9

Created

Created container: manager

openstack-operators

kubelet

nova-operator-controller-manager-5d488d59fb-5rvds

Created

Created container: manager

openstack-operators

ironic-operator-controller-manager-6f787dddc9-dq5f2_d276eab3-9df5-45f5-925e-16e3fb65ecd2

f92b5c2d.openstack.org

LeaderElection

ironic-operator-controller-manager-6f787dddc9-dq5f2_d276eab3-9df5-45f5-925e-16e3fb65ecd2 became leader

openstack-operators

kubelet

neutron-operator-controller-manager-767865f676-94tfj

Created

Created container: manager

openstack-operators

kubelet

placement-operator-controller-manager-5784578c99-9wvcj

Started

Started container manager

openstack-operators

neutron-operator-controller-manager-767865f676-94tfj_8054ae58-eaa8-4419-a273-2f36a45f6dea

972c7522.openstack.org

LeaderElection

neutron-operator-controller-manager-767865f676-94tfj_8054ae58-eaa8-4419-a273-2f36a45f6dea became leader

openstack-operators

kubelet

ovn-operator-controller-manager-884679f54-t875d

Created

Created container: manager

openstack-operators

kubelet

ovn-operator-controller-manager-884679f54-t875d

Started

Started container manager

openstack-operators

kubelet

watcher-operator-controller-manager-6c4d75f7f9-kp7cg

Created

Created container: manager

openstack-operators

kubelet

mariadb-operator-controller-manager-67ccfc9778-vxvcl

Started

Started container manager

openstack-operators

kubelet

designate-operator-controller-manager-588d4d986b-bghwc

Started

Started container manager

openstack-operators

kubelet

watcher-operator-controller-manager-6c4d75f7f9-kp7cg

Started

Started container manager

openstack-operators

kubelet

glance-operator-controller-manager-79df6bcc97-nt245

Started

Started container manager

openstack-operators

kubelet

glance-operator-controller-manager-79df6bcc97-nt245

Created

Created container: manager

openstack-operators

kubelet

watcher-operator-controller-manager-6c4d75f7f9-kp7cg

Started

Started container manager

openstack-operators

kubelet

glance-operator-controller-manager-79df6bcc97-nt245

Created

Created container: manager

openstack-operators

manila-operator-controller-manager-55f864c847-d49s9_2690a18b-0abc-4497-82a6-3ee1f1ef324d

858862a7.openstack.org

LeaderElection

manila-operator-controller-manager-55f864c847-d49s9_2690a18b-0abc-4497-82a6-3ee1f1ef324d became leader

openstack-operators

kubelet

designate-operator-controller-manager-588d4d986b-bghwc

Created

Created container: manager

openstack-operators

kubelet

mariadb-operator-controller-manager-67ccfc9778-vxvcl

Created

Created container: manager

openstack-operators

kubelet

swift-operator-controller-manager-c674c5965-srzmq

Started

Started container manager

openstack-operators

kubelet

telemetry-operator-controller-manager-d6b694c5-lbvvg

Started

Started container manager

openstack-operators

ironic-operator-controller-manager-6f787dddc9-dq5f2_d276eab3-9df5-45f5-925e-16e3fb65ecd2

f92b5c2d.openstack.org

LeaderElection

ironic-operator-controller-manager-6f787dddc9-dq5f2_d276eab3-9df5-45f5-925e-16e3fb65ecd2 became leader

openstack-operators

kubelet

swift-operator-controller-manager-c674c5965-srzmq

Created

Created container: manager

openstack-operators

kubelet

telemetry-operator-controller-manager-d6b694c5-lbvvg

Created

Created container: manager

openstack-operators

kubelet

glance-operator-controller-manager-79df6bcc97-nt245

Started

Started container manager

openstack-operators

kubelet

watcher-operator-controller-manager-6c4d75f7f9-kp7cg

Created

Created container: manager

openstack-operators

kubelet

designate-operator-controller-manager-588d4d986b-bghwc

Started

Started container manager

openstack-operators

kubelet

designate-operator-controller-manager-588d4d986b-bghwc

Created

Created container: manager

openstack-operators

kubelet

swift-operator-controller-manager-c674c5965-srzmq

Started

Started container manager

openstack-operators

kubelet

swift-operator-controller-manager-c674c5965-srzmq

Created

Created container: manager

openstack-operators

kubelet

rabbitmq-cluster-operator-manager-668c99d594-s2dqt

Started

Started container operator

openstack-operators

kubelet

heat-operator-controller-manager-67dd5f86f5-5h2n9

Created

Created container: manager

openstack-operators

kubelet

heat-operator-controller-manager-67dd5f86f5-5h2n9

Started

Started container manager

openstack-operators

kubelet

heat-operator-controller-manager-67dd5f86f5-5h2n9

Started

Started container manager

openstack-operators

kubelet

neutron-operator-controller-manager-767865f676-94tfj

Started

Started container manager

openstack-operators

kubelet

neutron-operator-controller-manager-767865f676-94tfj

Created

Created container: manager

openstack-operators

horizon-operator-controller-manager-8464cc45fb-jbfpt_180ca1c6-45c9-4c85-9006-d91be62ea844

5ad2eba0.openstack.org

LeaderElection

horizon-operator-controller-manager-8464cc45fb-jbfpt_180ca1c6-45c9-4c85-9006-d91be62ea844 became leader

openstack-operators

kubelet

placement-operator-controller-manager-5784578c99-9wvcj

Created

Created container: manager

openstack-operators

kubelet

placement-operator-controller-manager-5784578c99-9wvcj

Started

Started container manager

openstack-operators

keystone-operator-controller-manager-768b96df4c-2zdk8_48e684f3-d9ec-498d-823c-61dd04553534

6012128b.openstack.org

LeaderElection

keystone-operator-controller-manager-768b96df4c-2zdk8_48e684f3-d9ec-498d-823c-61dd04553534 became leader

openstack-operators

test-operator-controller-manager-5c5cb9c4d7-nf9vn_9f1b5ed3-a460-4e62-a59e-122382017ade

6cce095b.openstack.org

LeaderElection

test-operator-controller-manager-5c5cb9c4d7-nf9vn_9f1b5ed3-a460-4e62-a59e-122382017ade became leader

openstack-operators

placement-operator-controller-manager-5784578c99-9wvcj_ccc8f20e-527b-413f-b056-a219565c24e0

73d6b7ce.openstack.org

LeaderElection

placement-operator-controller-manager-5784578c99-9wvcj_ccc8f20e-527b-413f-b056-a219565c24e0 became leader

openstack-operators

kubelet

ovn-operator-controller-manager-884679f54-t875d

Created

Created container: manager

openstack-operators

mariadb-operator-controller-manager-67ccfc9778-vxvcl_250bc1e4-69f0-40ba-84fa-a3e3d7511ed1

7c2a6c6b.openstack.org

LeaderElection

mariadb-operator-controller-manager-67ccfc9778-vxvcl_250bc1e4-69f0-40ba-84fa-a3e3d7511ed1 became leader

openstack-operators

kubelet

ovn-operator-controller-manager-884679f54-t875d

Started

Started container manager

openstack-operators

kubelet

rabbitmq-cluster-operator-manager-668c99d594-s2dqt

Created

Created container: operator

openstack-operators

kubelet

nova-operator-controller-manager-5d488d59fb-5rvds

Started

Started container manager

openstack-operators

glance-operator-controller-manager-79df6bcc97-nt245_e792373f-b98b-488c-af34-037cb132d521

c569355b.openstack.org

LeaderElection

glance-operator-controller-manager-79df6bcc97-nt245_e792373f-b98b-488c-af34-037cb132d521 became leader

openstack-operators

glance-operator-controller-manager-79df6bcc97-nt245_e792373f-b98b-488c-af34-037cb132d521

c569355b.openstack.org

LeaderElection

glance-operator-controller-manager-79df6bcc97-nt245_e792373f-b98b-488c-af34-037cb132d521 became leader

openstack-operators

kubelet

horizon-operator-controller-manager-8464cc45fb-jbfpt

Created

Created container: manager

openstack-operators

rabbitmq-cluster-operator-manager-668c99d594-s2dqt_894c896e-e62e-42f1-a996-b5fb9f021d17

rabbitmq-cluster-operator-leader-election

LeaderElection

rabbitmq-cluster-operator-manager-668c99d594-s2dqt_894c896e-e62e-42f1-a996-b5fb9f021d17 became leader

openstack-operators

kubelet

horizon-operator-controller-manager-8464cc45fb-jbfpt

Started

Started container manager

openstack-operators

mariadb-operator-controller-manager-67ccfc9778-vxvcl_250bc1e4-69f0-40ba-84fa-a3e3d7511ed1

7c2a6c6b.openstack.org

LeaderElection

mariadb-operator-controller-manager-67ccfc9778-vxvcl_250bc1e4-69f0-40ba-84fa-a3e3d7511ed1 became leader

openstack-operators

placement-operator-controller-manager-5784578c99-9wvcj_ccc8f20e-527b-413f-b056-a219565c24e0

73d6b7ce.openstack.org

LeaderElection

placement-operator-controller-manager-5784578c99-9wvcj_ccc8f20e-527b-413f-b056-a219565c24e0 became leader

openstack-operators

test-operator-controller-manager-5c5cb9c4d7-nf9vn_9f1b5ed3-a460-4e62-a59e-122382017ade

6cce095b.openstack.org

LeaderElection

test-operator-controller-manager-5c5cb9c4d7-nf9vn_9f1b5ed3-a460-4e62-a59e-122382017ade became leader

openstack-operators

kubelet

mariadb-operator-controller-manager-67ccfc9778-vxvcl

Created

Created container: manager

openstack-operators

kubelet

mariadb-operator-controller-manager-67ccfc9778-vxvcl

Started

Started container manager

openstack-operators

keystone-operator-controller-manager-768b96df4c-2zdk8_48e684f3-d9ec-498d-823c-61dd04553534

6012128b.openstack.org

LeaderElection

keystone-operator-controller-manager-768b96df4c-2zdk8_48e684f3-d9ec-498d-823c-61dd04553534 became leader

openstack-operators

horizon-operator-controller-manager-8464cc45fb-jbfpt_180ca1c6-45c9-4c85-9006-d91be62ea844

5ad2eba0.openstack.org

LeaderElection

horizon-operator-controller-manager-8464cc45fb-jbfpt_180ca1c6-45c9-4c85-9006-d91be62ea844 became leader

openstack-operators

kubelet

telemetry-operator-controller-manager-d6b694c5-lbvvg

Created

Created container: manager

openstack-operators

kubelet

horizon-operator-controller-manager-8464cc45fb-jbfpt

Created

Created container: manager

openstack-operators

kubelet

telemetry-operator-controller-manager-d6b694c5-lbvvg

Started

Started container manager

openstack-operators

neutron-operator-controller-manager-767865f676-94tfj_8054ae58-eaa8-4419-a273-2f36a45f6dea

972c7522.openstack.org

LeaderElection

neutron-operator-controller-manager-767865f676-94tfj_8054ae58-eaa8-4419-a273-2f36a45f6dea became leader

openstack-operators

kubelet

horizon-operator-controller-manager-8464cc45fb-jbfpt

Started

Started container manager

openstack-operators

manila-operator-controller-manager-55f864c847-d49s9_2690a18b-0abc-4497-82a6-3ee1f1ef324d

858862a7.openstack.org

LeaderElection

manila-operator-controller-manager-55f864c847-d49s9_2690a18b-0abc-4497-82a6-3ee1f1ef324d became leader

openstack-operators

kubelet

nova-operator-controller-manager-5d488d59fb-5rvds

Created

Created container: manager

openstack-operators

kubelet

placement-operator-controller-manager-5784578c99-9wvcj

Created

Created container: manager

openstack-operators

kubelet

rabbitmq-cluster-operator-manager-668c99d594-s2dqt

Created

Created container: operator

openstack-operators

kubelet

rabbitmq-cluster-operator-manager-668c99d594-s2dqt

Started

Started container operator

openstack-operators

ovn-operator-controller-manager-884679f54-t875d_4a25d99f-11b8-4734-b435-680e810b5f4c

90840a60.openstack.org

LeaderElection

ovn-operator-controller-manager-884679f54-t875d_4a25d99f-11b8-4734-b435-680e810b5f4c became leader

openstack-operators

octavia-operator-controller-manager-5b9f45d989-fr295_979d1ec1-311e-4e0c-b6de-1a4a7ae30ddd

98809e87.openstack.org

LeaderElection

octavia-operator-controller-manager-5b9f45d989-fr295_979d1ec1-311e-4e0c-b6de-1a4a7ae30ddd became leader

openstack-operators

nova-operator-controller-manager-5d488d59fb-5rvds_9a05dbed-b8f3-403b-86f4-4619a7f625bf

f33036c1.openstack.org

LeaderElection

nova-operator-controller-manager-5d488d59fb-5rvds_9a05dbed-b8f3-403b-86f4-4619a7f625bf became leader

openstack-operators

heat-operator-controller-manager-67dd5f86f5-5h2n9_77487e03-2416-42a7-85aa-748a002d8628

c3c8b535.openstack.org

LeaderElection

heat-operator-controller-manager-67dd5f86f5-5h2n9_77487e03-2416-42a7-85aa-748a002d8628 became leader

openstack-operators

swift-operator-controller-manager-c674c5965-srzmq_2122b9f5-80e4-4969-aef2-c3cc13f30d67

83821f12.openstack.org

LeaderElection

swift-operator-controller-manager-c674c5965-srzmq_2122b9f5-80e4-4969-aef2-c3cc13f30d67 became leader

openstack-operators

designate-operator-controller-manager-588d4d986b-bghwc_925585aa-54a7-43ca-ae87-91ecd5454d62

f9497e05.openstack.org

LeaderElection

designate-operator-controller-manager-588d4d986b-bghwc_925585aa-54a7-43ca-ae87-91ecd5454d62 became leader

openstack-operators

watcher-operator-controller-manager-6c4d75f7f9-kp7cg_3f756e00-2ce2-47a1-a0f0-e6f92c4cbd32

5049980f.openstack.org

LeaderElection

watcher-operator-controller-manager-6c4d75f7f9-kp7cg_3f756e00-2ce2-47a1-a0f0-e6f92c4cbd32 became leader

openstack-operators

heat-operator-controller-manager-67dd5f86f5-5h2n9_77487e03-2416-42a7-85aa-748a002d8628

c3c8b535.openstack.org

LeaderElection

heat-operator-controller-manager-67dd5f86f5-5h2n9_77487e03-2416-42a7-85aa-748a002d8628 became leader

openstack-operators

nova-operator-controller-manager-5d488d59fb-5rvds_9a05dbed-b8f3-403b-86f4-4619a7f625bf

f33036c1.openstack.org

LeaderElection

nova-operator-controller-manager-5d488d59fb-5rvds_9a05dbed-b8f3-403b-86f4-4619a7f625bf became leader

openstack-operators

octavia-operator-controller-manager-5b9f45d989-fr295_979d1ec1-311e-4e0c-b6de-1a4a7ae30ddd

98809e87.openstack.org

LeaderElection

octavia-operator-controller-manager-5b9f45d989-fr295_979d1ec1-311e-4e0c-b6de-1a4a7ae30ddd became leader

openstack-operators

telemetry-operator-controller-manager-d6b694c5-lbvvg_b7452204-59e0-4892-8c96-4dfa45bbde94

fa1814a2.openstack.org

LeaderElection

telemetry-operator-controller-manager-d6b694c5-lbvvg_b7452204-59e0-4892-8c96-4dfa45bbde94 became leader

openstack-operators

ovn-operator-controller-manager-884679f54-t875d_4a25d99f-11b8-4734-b435-680e810b5f4c

90840a60.openstack.org

LeaderElection

ovn-operator-controller-manager-884679f54-t875d_4a25d99f-11b8-4734-b435-680e810b5f4c became leader

openstack-operators

telemetry-operator-controller-manager-d6b694c5-lbvvg_b7452204-59e0-4892-8c96-4dfa45bbde94

fa1814a2.openstack.org

LeaderElection

telemetry-operator-controller-manager-d6b694c5-lbvvg_b7452204-59e0-4892-8c96-4dfa45bbde94 became leader

openstack-operators

swift-operator-controller-manager-c674c5965-srzmq_2122b9f5-80e4-4969-aef2-c3cc13f30d67

83821f12.openstack.org

LeaderElection

swift-operator-controller-manager-c674c5965-srzmq_2122b9f5-80e4-4969-aef2-c3cc13f30d67 became leader

openstack-operators

watcher-operator-controller-manager-6c4d75f7f9-kp7cg_3f756e00-2ce2-47a1-a0f0-e6f92c4cbd32

5049980f.openstack.org

LeaderElection

watcher-operator-controller-manager-6c4d75f7f9-kp7cg_3f756e00-2ce2-47a1-a0f0-e6f92c4cbd32 became leader

openstack-operators

designate-operator-controller-manager-588d4d986b-bghwc_925585aa-54a7-43ca-ae87-91ecd5454d62

f9497e05.openstack.org

LeaderElection

designate-operator-controller-manager-588d4d986b-bghwc_925585aa-54a7-43ca-ae87-91ecd5454d62 became leader

openstack-operators

kubelet

infra-operator-controller-manager-7dd6bb94c9-7rp5f

Pulling

Pulling image "quay.io/openstack-k8s-operators/infra-operator@sha256:a4cb438fef247332815b032c8a248bc65b873274aaac92478a22aa2f915798db"

openstack-operators

multus

infra-operator-controller-manager-7dd6bb94c9-7rp5f

AddedInterface

Add eth0 [10.128.0.148/23] from ovn-kubernetes

openstack-operators

kubelet

infra-operator-controller-manager-7dd6bb94c9-7rp5f

Pulling

Pulling image "quay.io/openstack-k8s-operators/infra-operator@sha256:a4cb438fef247332815b032c8a248bc65b873274aaac92478a22aa2f915798db"

openstack-operators

multus

infra-operator-controller-manager-7dd6bb94c9-7rp5f

AddedInterface

Add eth0 [10.128.0.148/23] from ovn-kubernetes

openstack-operators

kubelet

openstack-operator-controller-manager-86bd8996f6-cfkmp

Created

Created container: manager

openstack-operators

kubelet

openstack-operator-controller-manager-86bd8996f6-cfkmp

Pulled

Container image "quay.io/openstack-k8s-operators/openstack-operator@sha256:5ab774ebf20af59fbaa509688a104dc74488b7cd8c4b0640d16924de8ead64fb" already present on machine

openstack-operators

multus

openstack-baremetal-operator-controller-manager-74c4796899s6vn4

AddedInterface

Add eth0 [10.128.0.156/23] from ovn-kubernetes

openstack-operators

multus

openstack-baremetal-operator-controller-manager-74c4796899s6vn4

AddedInterface

Add eth0 [10.128.0.156/23] from ovn-kubernetes

openstack-operators

kubelet

openstack-baremetal-operator-controller-manager-74c4796899s6vn4

Pulling

Pulling image "quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:bf7cdbfb125c4327b35870f8640cbed9ddc32d6f07fedd117c6fd59f16463329"

openstack-operators

kubelet

openstack-baremetal-operator-controller-manager-74c4796899s6vn4

Pulling

Pulling image "quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:bf7cdbfb125c4327b35870f8640cbed9ddc32d6f07fedd117c6fd59f16463329"

openstack-operators

openstack-operator-controller-manager-86bd8996f6-cfkmp_e76856f9-6fb1-4550-b7c4-346e92be1537

40ba705e.openstack.org

LeaderElection

openstack-operator-controller-manager-86bd8996f6-cfkmp_e76856f9-6fb1-4550-b7c4-346e92be1537 became leader

openstack-operators

kubelet

openstack-operator-controller-manager-86bd8996f6-cfkmp

Started

Started container manager

openstack-operators

multus

openstack-operator-controller-manager-86bd8996f6-cfkmp

AddedInterface

Add eth0 [10.128.0.163/23] from ovn-kubernetes

openstack-operators

kubelet

openstack-operator-controller-manager-86bd8996f6-cfkmp

Started

Started container manager

openstack-operators

multus

openstack-operator-controller-manager-86bd8996f6-cfkmp

AddedInterface

Add eth0 [10.128.0.163/23] from ovn-kubernetes

openstack-operators

kubelet

openstack-operator-controller-manager-86bd8996f6-cfkmp

Pulled

Container image "quay.io/openstack-k8s-operators/openstack-operator@sha256:5ab774ebf20af59fbaa509688a104dc74488b7cd8c4b0640d16924de8ead64fb" already present on machine

openstack-operators

kubelet

openstack-operator-controller-manager-86bd8996f6-cfkmp

Created

Created container: manager

openstack-operators

openstack-operator-controller-manager-86bd8996f6-cfkmp_e76856f9-6fb1-4550-b7c4-346e92be1537

40ba705e.openstack.org

LeaderElection

openstack-operator-controller-manager-86bd8996f6-cfkmp_e76856f9-6fb1-4550-b7c4-346e92be1537 became leader

openstack-operators

kubelet

openstack-baremetal-operator-controller-manager-74c4796899s6vn4

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:bf7cdbfb125c4327b35870f8640cbed9ddc32d6f07fedd117c6fd59f16463329" in 2.633s (2.633s including waiting). Image size: 190544999 bytes.

openstack-operators

kubelet

infra-operator-controller-manager-7dd6bb94c9-7rp5f

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/infra-operator@sha256:a4cb438fef247332815b032c8a248bc65b873274aaac92478a22aa2f915798db" in 3.017s (3.017s including waiting). Image size: 192852400 bytes.

openstack-operators

kubelet

openstack-baremetal-operator-controller-manager-74c4796899s6vn4

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:bf7cdbfb125c4327b35870f8640cbed9ddc32d6f07fedd117c6fd59f16463329" in 2.633s (2.633s including waiting). Image size: 190544999 bytes.

openstack-operators

kubelet

infra-operator-controller-manager-7dd6bb94c9-7rp5f

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/infra-operator@sha256:a4cb438fef247332815b032c8a248bc65b873274aaac92478a22aa2f915798db" in 3.017s (3.017s including waiting). Image size: 192852400 bytes.

openstack-operators

infra-operator-controller-manager-7dd6bb94c9-7rp5f_3cac61a6-9595-4685-91d4-6c9614882f1a

c8c223a1.openstack.org

LeaderElection

infra-operator-controller-manager-7dd6bb94c9-7rp5f_3cac61a6-9595-4685-91d4-6c9614882f1a became leader

openstack-operators

openstack-baremetal-operator-controller-manager-74c4796899s6vn4_c80638bc-2e20-4c3e-a65d-fe3fbd8a726c

dedc2245.openstack.org

LeaderElection

openstack-baremetal-operator-controller-manager-74c4796899s6vn4_c80638bc-2e20-4c3e-a65d-fe3fbd8a726c became leader

openstack-operators

kubelet

infra-operator-controller-manager-7dd6bb94c9-7rp5f

Started

Started container manager

openstack-operators

kubelet

infra-operator-controller-manager-7dd6bb94c9-7rp5f

Created

Created container: manager

openstack-operators

kubelet

infra-operator-controller-manager-7dd6bb94c9-7rp5f

Created

Created container: manager

openstack-operators

kubelet

infra-operator-controller-manager-7dd6bb94c9-7rp5f

Started

Started container manager

openstack-operators

infra-operator-controller-manager-7dd6bb94c9-7rp5f_3cac61a6-9595-4685-91d4-6c9614882f1a

c8c223a1.openstack.org

LeaderElection

infra-operator-controller-manager-7dd6bb94c9-7rp5f_3cac61a6-9595-4685-91d4-6c9614882f1a became leader

openstack-operators

openstack-baremetal-operator-controller-manager-74c4796899s6vn4_c80638bc-2e20-4c3e-a65d-fe3fbd8a726c

dedc2245.openstack.org

LeaderElection

openstack-baremetal-operator-controller-manager-74c4796899s6vn4_c80638bc-2e20-4c3e-a65d-fe3fbd8a726c became leader

openstack

cert-manager-certificates-trigger

rootca-public

Issuing

Issuing certificate as Secret does not exist

openstack

cert-manager-certificates-key-manager

rootca-public

Generated

Stored new private key in temporary Secret resource "rootca-public-55fsf"
(x2)

openstack

cert-manager-issuers

rootca-internal

ErrGetKeyPair

Error getting keypair for CA issuer: secrets "rootca-internal" not found

openstack

cert-manager-certificaterequests-issuer-ca

rootca-public-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificates-issuing

rootca-public

Issuing

The certificate has been successfully issued

openstack

cert-manager-certificates-request-manager

rootca-public

Requested

Created new CertificateRequest resource "rootca-public-1"

openstack

cert-manager-certificaterequests-issuer-selfsigned

rootca-public-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificates-trigger

rootca-internal

Issuing

Issuing certificate as Secret does not exist
(x3)

openstack

cert-manager-issuers

rootca-public

ErrInitIssuer

Error initializing issuer: secrets "rootca-public" not found
(x3)

openstack

cert-manager-issuers

rootca-public

ErrGetKeyPair

Error getting keypair for CA issuer: secrets "rootca-public" not found
(x2)

openstack

cert-manager-issuers

rootca-internal

ErrInitIssuer

Error initializing issuer: secrets "rootca-internal" not found

openstack

cert-manager-certificaterequests-issuer-selfsigned

rootca-public-1

CertificateIssued

Certificate fetched from issuer successfully

openstack

cert-manager-certificaterequests-approver

rootca-public-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack

cert-manager-certificaterequests-issuer-acme

rootca-public-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-vault

rootca-public-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-venafi

rootca-public-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-ca

rootca-internal-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-approver

rootca-internal-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack

cert-manager-certificaterequests-issuer-venafi

rootca-internal-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificates-key-manager

rootca-internal

Generated

Stored new private key in temporary Secret resource "rootca-internal-tbvp2"

openstack

cert-manager-certificates-request-manager

rootca-internal

Requested

Created new CertificateRequest resource "rootca-internal-1"

openstack

cert-manager-certificaterequests-issuer-acme

rootca-internal-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-selfsigned

rootca-internal-1

CertificateIssued

Certificate fetched from issuer successfully

openstack

cert-manager-certificaterequests-issuer-selfsigned

rootca-internal-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificates-issuing

rootca-internal

Issuing

The certificate has been successfully issued

openstack

cert-manager-certificates-trigger

rootca-libvirt

Issuing

Issuing certificate as Secret does not exist
(x2)

openstack

cert-manager-issuers

rootca-libvirt

ErrInitIssuer

Error initializing issuer: secrets "rootca-libvirt" not found

openstack

cert-manager-certificaterequests-issuer-vault

rootca-internal-1

WaitingForApproval

Not signing CertificateRequest until it is Approved
(x2)

openstack

cert-manager-issuers

rootca-libvirt

ErrGetKeyPair

Error getting keypair for CA issuer: secrets "rootca-libvirt" not found

openstack

cert-manager-certificates-key-manager

rootca-libvirt

Generated

Stored new private key in temporary Secret resource "rootca-libvirt-7rrzb"

openstack

cert-manager-certificaterequests-issuer-selfsigned

rootca-libvirt-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-ca

rootca-libvirt-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-vault

rootca-libvirt-1

WaitingForApproval

Not signing CertificateRequest until it is Approved
(x2)

openstack

cert-manager-issuers

rootca-ovn

ErrGetKeyPair

Error getting keypair for CA issuer: secrets "rootca-ovn" not found

openstack

cert-manager-certificaterequests-issuer-acme

rootca-libvirt-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificates-trigger

rootca-ovn

Issuing

Issuing certificate as Secret does not exist

openstack

cert-manager-certificates-issuing

rootca-libvirt

Issuing

The certificate has been successfully issued

openstack

cert-manager-certificaterequests-issuer-venafi

rootca-libvirt-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificates-request-manager

rootca-libvirt

Requested

Created new CertificateRequest resource "rootca-libvirt-1"

openstack

cert-manager-certificaterequests-issuer-selfsigned

rootca-libvirt-1

CertificateIssued

Certificate fetched from issuer successfully

openstack

cert-manager-certificaterequests-approver

rootca-libvirt-1

cert-manager.io

Certificate request has been approved by cert-manager.io
(x2)

openstack

cert-manager-issuers

rootca-ovn

ErrInitIssuer

Error initializing issuer: secrets "rootca-ovn" not found

openstack

cert-manager-certificaterequests-issuer-venafi

rootca-ovn-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-ca

rootca-ovn-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificates-request-manager

rootca-ovn

Requested

Created new CertificateRequest resource "rootca-ovn-1"

openstack

cert-manager-certificaterequests-issuer-vault

rootca-ovn-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-acme

rootca-ovn-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-selfsigned

rootca-ovn-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-approver

rootca-ovn-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack

cert-manager-certificates-key-manager

rootca-ovn

Generated

Stored new private key in temporary Secret resource "rootca-ovn-8qz5k"

openstack

cert-manager-certificaterequests-issuer-selfsigned

rootca-ovn-1

CertificateIssued

Certificate fetched from issuer successfully

openstack

replicaset-controller

dnsmasq-dns-8476fd89bc

SuccessfulCreate

Created pod: dnsmasq-dns-8476fd89bc-24d8n

openstack

cert-manager-certificates-issuing

rootca-ovn

Issuing

The certificate has been successfully issued

openstack

deployment-controller

dnsmasq-dns

ScalingReplicaSet

Scaled up replica set dnsmasq-dns-8476fd89bc to 1

openstack

deployment-controller

dnsmasq-dns

ScalingReplicaSet

Scaled up replica set dnsmasq-dns-685c76cf85 to 1

openstack

replicaset-controller

dnsmasq-dns-685c76cf85

SuccessfulCreate

Created pod: dnsmasq-dns-685c76cf85-qwz94
(x2)

openstack

metallb-controller

dnsmasq-dns

deprecatedAnnotation

Service uses deprecated annotation metallb.universe.tf/allow-shared-ip
(x2)

openstack

metallb-controller

dnsmasq-dns

deprecatedAnnotation

Service uses deprecated annotation metallb.universe.tf/address-pool
(x2)

openstack

metallb-controller

dnsmasq-dns

deprecatedAnnotation

Service uses deprecated annotation metallb.universe.tf/loadBalancerIPs

openstack

metallb-controller

dnsmasq-dns

IPAllocated

Assigned IP ["192.168.122.80"]
(x3)

openstack

cert-manager-issuers

rootca-internal

KeyPairVerified

Signing CA verified

openstack

multus

dnsmasq-dns-685c76cf85-qwz94

AddedInterface

Add eth0 [10.128.0.165/23] from ovn-kubernetes

openstack

cert-manager-certificaterequests-approver

rabbitmq-svc-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack

cert-manager-certificaterequests-issuer-selfsigned

rabbitmq-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

multus

dnsmasq-dns-8476fd89bc-24d8n

AddedInterface

Add eth0 [10.128.0.166/23] from ovn-kubernetes

openstack

kubelet

dnsmasq-dns-8476fd89bc-24d8n

Pulling

Pulling image "quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:fbb5be29e9e4fa11f0743e7f74f2e80dcc7445d24770709ea0e038147f752c51"

openstack

cert-manager-certificates-trigger

rabbitmq-cell1-svc

Issuing

Issuing certificate as Secret does not exist

openstack

cert-manager-certificaterequests-issuer-ca

rabbitmq-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificates-trigger

rabbitmq-svc

Issuing

Issuing certificate as Secret does not exist

openstack

cert-manager-certificates-key-manager

rabbitmq-svc

Generated

Stored new private key in temporary Secret resource "rabbitmq-svc-vq54f"

openstack

cert-manager-certificates-request-manager

rabbitmq-svc

Requested

Created new CertificateRequest resource "rabbitmq-svc-1"

openstack

cert-manager-certificates-key-manager

rabbitmq-cell1-svc

Generated

Stored new private key in temporary Secret resource "rabbitmq-cell1-svc-jsczn"

openstack

kubelet

dnsmasq-dns-685c76cf85-qwz94

Pulling

Pulling image "quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:fbb5be29e9e4fa11f0743e7f74f2e80dcc7445d24770709ea0e038147f752c51"

openstack

cert-manager-certificaterequests-issuer-ca

rabbitmq-svc-1

CertificateIssued

Certificate fetched from issuer successfully

openstack

cert-manager-certificaterequests-issuer-acme

rabbitmq-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-venafi

rabbitmq-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-vault

rabbitmq-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificates-request-manager

rabbitmq-cell1-svc

Requested

Created new CertificateRequest resource "rabbitmq-cell1-svc-1"

openstack

cert-manager-certificates-issuing

rabbitmq-svc

Issuing

The certificate has been successfully issued
(x3)

openstack

cert-manager-issuers

rootca-libvirt

KeyPairVerified

Signing CA verified

openstack

cert-manager-certificaterequests-issuer-ca

rabbitmq-cell1-svc-1

CertificateIssued

Certificate fetched from issuer successfully

openstack

cert-manager-certificaterequests-approver

rabbitmq-cell1-svc-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack

cert-manager-certificaterequests-issuer-selfsigned

rabbitmq-cell1-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-vault

rabbitmq-cell1-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-acme

rabbitmq-cell1-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-ca

rabbitmq-cell1-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-venafi

rabbitmq-cell1-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

rabbitmqcluster-controller

rabbitmq-cell1

SuccessfulCreate

created resource rabbitmq-cell1-erlang-cookie of Type *v1.Secret

openstack

rabbitmqcluster-controller

rabbitmq-cell1

SuccessfulCreate

created resource rabbitmq-cell1-nodes of Type *v1.Service

openstack

rabbitmqcluster-controller

rabbitmq-cell1

SuccessfulCreate

created resource rabbitmq-cell1-plugins-conf of Type *v1.ConfigMap

openstack

rabbitmqcluster-controller

rabbitmq

SuccessfulCreate

created resource rabbitmq-erlang-cookie of Type *v1.Secret
(x2)

openstack

metallb-controller

rabbitmq

deprecatedAnnotation

Service uses deprecated annotation metallb.universe.tf/loadBalancerIPs
(x2)

openstack

metallb-controller

rabbitmq

deprecatedAnnotation

Service uses deprecated annotation metallb.universe.tf/address-pool

openstack

rabbitmqcluster-controller

rabbitmq-cell1

SuccessfulCreate

created resource rabbitmq-cell1-default-user of Type *v1.Secret

openstack

rabbitmqcluster-controller

rabbitmq

SuccessfulCreate

created resource rabbitmq-default-user of Type *v1.Secret

openstack

rabbitmqcluster-controller

rabbitmq

SuccessfulCreate

(combined from similar events): created resource rabbitmq-server of Type *v1.StatefulSet

openstack

rabbitmqcluster-controller

rabbitmq

SuccessfulCreate

created resource rabbitmq-server of Type *v1.RoleBinding
(x2)

openstack

metallb-controller

rabbitmq-cell1

deprecatedAnnotation

Service uses deprecated annotation metallb.universe.tf/loadBalancerIPs
(x2)

openstack

metallb-controller

rabbitmq-cell1

deprecatedAnnotation

Service uses deprecated annotation metallb.universe.tf/address-pool

openstack

metallb-controller

rabbitmq

IPAllocated

Assigned IP ["172.17.0.85"]

openstack

rabbitmqcluster-controller

rabbitmq

SuccessfulCreate

created resource rabbitmq-plugins-conf of Type *v1.ConfigMap

openstack

rabbitmqcluster-controller

rabbitmq

SuccessfulCreate

created resource rabbitmq of Type *v1.Service

openstack

rabbitmqcluster-controller

rabbitmq

SuccessfulCreate

created resource rabbitmq-nodes of Type *v1.Service

openstack

replicaset-controller

dnsmasq-dns-685c76cf85

SuccessfulDelete

Deleted pod: dnsmasq-dns-685c76cf85-qwz94

openstack

metallb-controller

rabbitmq-cell1

IPAllocated

Assigned IP ["172.17.0.86"]

openstack

rabbitmqcluster-controller

rabbitmq-cell1

SuccessfulCreate

created resource rabbitmq-cell1 of Type *v1.Service

openstack

deployment-controller

dnsmasq-dns

ScalingReplicaSet

Scaled up replica set dnsmasq-dns-586dbdbb8c to 1 from 0

openstack

cert-manager-certificates-trigger

galera-openstack-svc

Issuing

Issuing certificate as Secret does not exist

openstack

replicaset-controller

dnsmasq-dns-586dbdbb8c

SuccessfulCreate

Created pod: dnsmasq-dns-586dbdbb8c-28nq4

openstack

cert-manager-certificates-issuing

rabbitmq-cell1-svc

Issuing

The certificate has been successfully issued

openstack

rabbitmqcluster-controller

rabbitmq-cell1

SuccessfulCreate

created resource rabbitmq-cell1-server of Type *v1.ServiceAccount

openstack

persistentvolume-controller

persistence-rabbitmq-server-0

WaitForFirstConsumer

waiting for first consumer to be created before binding

openstack

replicaset-controller

dnsmasq-dns-6ff8fd9d5c

SuccessfulCreate

Created pod: dnsmasq-dns-6ff8fd9d5c-47t6r

openstack

rabbitmqcluster-controller

rabbitmq

SuccessfulCreate

created resource rabbitmq-server-conf of Type *v1.ConfigMap
(x2)

openstack

persistentvolume-controller

persistence-rabbitmq-server-0

ExternalProvisioning

Waiting for a volume to be created either by the external provisioner 'topolvm.io' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered.

openstack

topolvm.io_lvms-operator-dd6b59969-zzdqx_ce63b60a-9fe5-4592-9fa1-06008d67137b

persistence-rabbitmq-server-0

Provisioning

External provisioner is provisioning volume for claim "openstack/persistence-rabbitmq-server-0"

openstack

rabbitmqcluster-controller

rabbitmq-cell1

SuccessfulCreate

created resource rabbitmq-cell1-peer-discovery of Type *v1.Role

openstack

rabbitmqcluster-controller

rabbitmq-cell1

SuccessfulCreate

created resource rabbitmq-cell1-server-conf of Type *v1.ConfigMap

openstack

deployment-controller

dnsmasq-dns

ScalingReplicaSet

Scaled down replica set dnsmasq-dns-8476fd89bc to 0 from 1

openstack

replicaset-controller

dnsmasq-dns-8476fd89bc

SuccessfulDelete

Deleted pod: dnsmasq-dns-8476fd89bc-24d8n

openstack

deployment-controller

dnsmasq-dns

ScalingReplicaSet

Scaled up replica set dnsmasq-dns-6ff8fd9d5c to 1 from 0

openstack

statefulset-controller

rabbitmq-server

SuccessfulCreate

create Claim persistence-rabbitmq-server-0 Pod rabbitmq-server-0 in StatefulSet rabbitmq-server success

openstack

statefulset-controller

rabbitmq-server

SuccessfulCreate

create Pod rabbitmq-server-0 in StatefulSet rabbitmq-server successful

openstack

rabbitmqcluster-controller

rabbitmq

SuccessfulCreate

created resource rabbitmq-peer-discovery of Type *v1.Role

openstack

rabbitmqcluster-controller

rabbitmq

SuccessfulCreate

created resource rabbitmq-server of Type *v1.ServiceAccount

openstack

deployment-controller

dnsmasq-dns

ScalingReplicaSet

Scaled down replica set dnsmasq-dns-685c76cf85 to 0 from 1
(x3)

openstack

cert-manager-issuers

rootca-ovn

KeyPairVerified

Signing CA verified

openstack

multus

dnsmasq-dns-586dbdbb8c-28nq4

AddedInterface

Add eth0 [10.128.0.167/23] from ovn-kubernetes

openstack

persistentvolume-controller

persistence-rabbitmq-cell1-server-0

ExternalProvisioning

Waiting for a volume to be created either by the external provisioner 'topolvm.io' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered.

openstack

statefulset-controller

rabbitmq-cell1-server

SuccessfulCreate

create Claim persistence-rabbitmq-cell1-server-0 Pod rabbitmq-cell1-server-0 in StatefulSet rabbitmq-cell1-server success

openstack

cert-manager-certificaterequests-approver

galera-openstack-svc-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack

cert-manager-certificaterequests-issuer-venafi

galera-openstack-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-ca

galera-openstack-svc-1

CertificateIssued

Certificate fetched from issuer successfully

openstack

rabbitmqcluster-controller

rabbitmq-cell1

SuccessfulCreate

(combined from similar events): created resource rabbitmq-cell1-server of Type *v1.StatefulSet

openstack

cert-manager-certificaterequests-issuer-vault

galera-openstack-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-ca

galera-openstack-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

rabbitmqcluster-controller

rabbitmq-cell1

SuccessfulCreate

created resource rabbitmq-cell1-server of Type *v1.RoleBinding

openstack

cert-manager-certificaterequests-issuer-acme

galera-openstack-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-selfsigned

galera-openstack-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

multus

dnsmasq-dns-6ff8fd9d5c-47t6r

AddedInterface

Add eth0 [10.128.0.168/23] from ovn-kubernetes

openstack

cert-manager-certificates-key-manager

galera-openstack-svc

Generated

Stored new private key in temporary Secret resource "galera-openstack-svc-bnfg8"

openstack

cert-manager-certificates-request-manager

galera-openstack-svc

Requested

Created new CertificateRequest resource "galera-openstack-svc-1"

openstack

persistentvolume-controller

persistence-rabbitmq-cell1-server-0

WaitForFirstConsumer

waiting for first consumer to be created before binding

openstack

statefulset-controller

rabbitmq-cell1-server

SuccessfulCreate

create Pod rabbitmq-cell1-server-0 in StatefulSet rabbitmq-cell1-server successful

openstack

cert-manager-certificates-issuing

galera-openstack-svc

Issuing

The certificate has been successfully issued

openstack

kubelet

dnsmasq-dns-6ff8fd9d5c-47t6r

Pulling

Pulling image "quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:fbb5be29e9e4fa11f0743e7f74f2e80dcc7445d24770709ea0e038147f752c51"

openstack

kubelet

dnsmasq-dns-586dbdbb8c-28nq4

Pulling

Pulling image "quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:fbb5be29e9e4fa11f0743e7f74f2e80dcc7445d24770709ea0e038147f752c51"

openstack

cert-manager-certificates-trigger

galera-openstack-cell1-svc

Issuing

Issuing certificate as Secret does not exist

openstack

cert-manager-certificaterequests-issuer-selfsigned

galera-openstack-cell1-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

persistentvolume-controller

mysql-db-openstack-cell1-galera-0

WaitForFirstConsumer

waiting for first consumer to be created before binding

openstack

statefulset-controller

openstack-galera

SuccessfulCreate

create Claim mysql-db-openstack-galera-0 Pod openstack-galera-0 in StatefulSet openstack-galera success

openstack

statefulset-controller

openstack-galera

SuccessfulCreate

create Pod openstack-galera-0 in StatefulSet openstack-galera successful

openstack

cert-manager-certificaterequests-issuer-vault

galera-openstack-cell1-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-ca

galera-openstack-cell1-svc-1

CertificateIssued

Certificate fetched from issuer successfully

openstack

cert-manager-certificaterequests-issuer-ca

galera-openstack-cell1-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-venafi

galera-openstack-cell1-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

statefulset-controller

openstack-cell1-galera

SuccessfulCreate

create Claim mysql-db-openstack-cell1-galera-0 Pod openstack-cell1-galera-0 in StatefulSet openstack-cell1-galera success

openstack

cert-manager-certificaterequests-issuer-acme

galera-openstack-cell1-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-approver

galera-openstack-cell1-svc-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack

topolvm.io_lvms-operator-dd6b59969-zzdqx_ce63b60a-9fe5-4592-9fa1-06008d67137b

persistence-rabbitmq-server-0

ProvisioningSucceeded

Successfully provisioned volume pvc-3a9e7590-481d-455b-97e9-a4c82858f1df
(x2)

openstack

persistentvolume-controller

mysql-db-openstack-galera-0

ExternalProvisioning

Waiting for a volume to be created either by the external provisioner 'topolvm.io' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered.

openstack

persistentvolume-controller

mysql-db-openstack-galera-0

WaitForFirstConsumer

waiting for first consumer to be created before binding

openstack

persistentvolume-controller

mysql-db-openstack-cell1-galera-0

ExternalProvisioning

Waiting for a volume to be created either by the external provisioner 'topolvm.io' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered.

openstack

statefulset-controller

openstack-cell1-galera

SuccessfulCreate

create Pod openstack-cell1-galera-0 in StatefulSet openstack-cell1-galera successful

openstack

cert-manager-certificates-key-manager

galera-openstack-cell1-svc

Generated

Stored new private key in temporary Secret resource "galera-openstack-cell1-svc-nswxv"

openstack

cert-manager-certificates-request-manager

galera-openstack-cell1-svc

Requested

Created new CertificateRequest resource "galera-openstack-cell1-svc-1"

openstack

cert-manager-certificates-issuing

galera-openstack-cell1-svc

Issuing

The certificate has been successfully issued

openstack

cert-manager-certificates-trigger

memcached-svc

Issuing

Issuing certificate as Secret does not exist

openstack

topolvm.io_lvms-operator-dd6b59969-zzdqx_ce63b60a-9fe5-4592-9fa1-06008d67137b

persistence-rabbitmq-cell1-server-0

Provisioning

External provisioner is provisioning volume for claim "openstack/persistence-rabbitmq-cell1-server-0"

openstack

cert-manager-certificaterequests-approver

memcached-svc-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack

cert-manager-certificaterequests-issuer-selfsigned

memcached-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-vault

memcached-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-ca

memcached-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-acme

memcached-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-venafi

memcached-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-ca

memcached-svc-1

CertificateIssued

Certificate fetched from issuer successfully

openstack

cert-manager-certificates-key-manager

memcached-svc

Generated

Stored new private key in temporary Secret resource "memcached-svc-mgjrc"

openstack

cert-manager-certificates-request-manager

memcached-svc

Requested

Created new CertificateRequest resource "memcached-svc-1"

openstack

cert-manager-certificates-issuing

memcached-svc

Issuing

The certificate has been successfully issued

openstack

topolvm.io_lvms-operator-dd6b59969-zzdqx_ce63b60a-9fe5-4592-9fa1-06008d67137b

mysql-db-openstack-galera-0

Provisioning

External provisioner is provisioning volume for claim "openstack/mysql-db-openstack-galera-0"

openstack

cert-manager-certificates-trigger

ovn-metrics

Issuing

Issuing certificate as Secret does not exist

openstack

cert-manager-certificaterequests-issuer-ca

ovn-metrics-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-vault

ovn-metrics-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-selfsigned

ovn-metrics-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-venafi

ovn-metrics-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-acme

ovn-metrics-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-approver

ovn-metrics-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack

cert-manager-certificaterequests-issuer-ca

ovn-metrics-1

CertificateIssued

Certificate fetched from issuer successfully

openstack

topolvm.io_lvms-operator-dd6b59969-zzdqx_ce63b60a-9fe5-4592-9fa1-06008d67137b

persistence-rabbitmq-cell1-server-0

ProvisioningSucceeded

Successfully provisioned volume pvc-83923b42-a350-4692-9612-54b29c1c60d2

openstack

cert-manager-certificates-key-manager

ovn-metrics

Generated

Stored new private key in temporary Secret resource "ovn-metrics-6q855"

openstack

cert-manager-certificates-request-manager

ovn-metrics

Requested

Created new CertificateRequest resource "ovn-metrics-1"

openstack

statefulset-controller

memcached

SuccessfulCreate

create Pod memcached-0 in StatefulSet memcached successful

openstack

cert-manager-certificates-trigger

ovndbcluster-nb-ovndbs

Issuing

Issuing certificate as Secret does not exist

openstack

cert-manager-certificates-trigger

ovnnorthd-ovndbs

Issuing

Issuing certificate as Secret does not exist

openstack

cert-manager-certificates-issuing

ovn-metrics

Issuing

The certificate has been successfully issued

openstack

topolvm.io_lvms-operator-dd6b59969-zzdqx_ce63b60a-9fe5-4592-9fa1-06008d67137b

mysql-db-openstack-cell1-galera-0

Provisioning

External provisioner is provisioning volume for claim "openstack/mysql-db-openstack-cell1-galera-0"

openstack

topolvm.io_lvms-operator-dd6b59969-zzdqx_ce63b60a-9fe5-4592-9fa1-06008d67137b

mysql-db-openstack-galera-0

ProvisioningSucceeded

Successfully provisioned volume pvc-7f5b2eea-768b-474c-90fe-c7bb50fdf552

openstack

cert-manager-certificates-key-manager

ovnnorthd-ovndbs

Generated

Stored new private key in temporary Secret resource "ovnnorthd-ovndbs-flmvh"

openstack

cert-manager-certificates-trigger

ovncontroller-ovndbs

Issuing

Issuing certificate as Secret does not exist

openstack

cert-manager-certificaterequests-issuer-ca

ovnnorthd-ovndbs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificates-key-manager

ovndbcluster-nb-ovndbs

Generated

Stored new private key in temporary Secret resource "ovndbcluster-nb-ovndbs-xnnk6"

openstack

topolvm.io_lvms-operator-dd6b59969-zzdqx_ce63b60a-9fe5-4592-9fa1-06008d67137b

mysql-db-openstack-cell1-galera-0

ProvisioningSucceeded

Successfully provisioned volume pvc-759b0178-3d51-437c-811b-ccaeafcbe2c4

openstack

cert-manager-certificates-trigger

neutron-ovndbs

Issuing

Issuing certificate as Secret does not exist

openstack

cert-manager-certificates-key-manager

ovncontroller-ovndbs

Generated

Stored new private key in temporary Secret resource "ovncontroller-ovndbs-mp4pl"

openstack

cert-manager-certificaterequests-issuer-vault

ovnnorthd-ovndbs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-venafi

ovnnorthd-ovndbs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-acme

ovnnorthd-ovndbs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-selfsigned

ovnnorthd-ovndbs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificates-request-manager

ovnnorthd-ovndbs

Requested

Created new CertificateRequest resource "ovnnorthd-ovndbs-1"

openstack

cert-manager-certificaterequests-issuer-ca

ovncontroller-ovndbs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-vault

ovndbcluster-nb-ovndbs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificates-key-manager

neutron-ovndbs

Generated

Stored new private key in temporary Secret resource "neutron-ovndbs-d8wpm"

openstack

cert-manager-certificaterequests-approver

ovndbcluster-nb-ovndbs-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack

cert-manager-certificaterequests-issuer-selfsigned

ovncontroller-ovndbs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-acme

ovncontroller-ovndbs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-venafi

ovndbcluster-nb-ovndbs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificates-request-manager

ovncontroller-ovndbs

Requested

Created new CertificateRequest resource "ovncontroller-ovndbs-1"

openstack

cert-manager-certificaterequests-issuer-ca

ovndbcluster-nb-ovndbs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-ca

ovndbcluster-nb-ovndbs-1

CertificateIssued

Certificate fetched from issuer successfully

openstack

cert-manager-certificaterequests-issuer-selfsigned

ovndbcluster-nb-ovndbs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-ca

ovnnorthd-ovndbs-1

CertificateIssued

Certificate fetched from issuer successfully

openstack

cert-manager-certificaterequests-issuer-acme

ovndbcluster-nb-ovndbs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-vault

ovncontroller-ovndbs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificates-request-manager

ovndbcluster-nb-ovndbs

Requested

Created new CertificateRequest resource "ovndbcluster-nb-ovndbs-1"

openstack

cert-manager-certificaterequests-approver

ovnnorthd-ovndbs-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack

cert-manager-certificaterequests-issuer-venafi

ovncontroller-ovndbs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificates-issuing

ovnnorthd-ovndbs

Issuing

The certificate has been successfully issued

openstack

cert-manager-certificaterequests-issuer-ca

ovncontroller-ovndbs-1

CertificateIssued

Certificate fetched from issuer successfully

openstack

cert-manager-certificaterequests-approver

ovncontroller-ovndbs-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack

cert-manager-certificates-request-manager

neutron-ovndbs

Requested

Created new CertificateRequest resource "neutron-ovndbs-1"

openstack

daemonset-controller

ovn-controller

SuccessfulCreate

Created pod: ovn-controller-q5t48

openstack

statefulset-controller

ovsdbserver-nb

SuccessfulCreate

create Claim ovndbcluster-nb-etc-ovn-ovsdbserver-nb-0 Pod ovsdbserver-nb-0 in StatefulSet ovsdbserver-nb success

openstack

cert-manager-certificates-issuing

ovndbcluster-nb-ovndbs

Issuing

The certificate has been successfully issued

openstack

topolvm.io_lvms-operator-dd6b59969-zzdqx_ce63b60a-9fe5-4592-9fa1-06008d67137b

ovndbcluster-nb-etc-ovn-ovsdbserver-nb-0

Provisioning

External provisioner is provisioning volume for claim "openstack/ovndbcluster-nb-etc-ovn-ovsdbserver-nb-0"
(x2)

openstack

persistentvolume-controller

ovndbcluster-nb-etc-ovn-ovsdbserver-nb-0

ExternalProvisioning

Waiting for a volume to be created either by the external provisioner 'topolvm.io' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered.

openstack

persistentvolume-controller

ovndbcluster-nb-etc-ovn-ovsdbserver-nb-0

WaitForFirstConsumer

waiting for first consumer to be created before binding

openstack

daemonset-controller

ovn-controller-ovs

SuccessfulCreate

Created pod: ovn-controller-ovs-hz4j4

openstack

cert-manager-certificates-issuing

ovncontroller-ovndbs

Issuing

The certificate has been successfully issued

openstack

statefulset-controller

ovsdbserver-nb

SuccessfulCreate

create Pod ovsdbserver-nb-0 in StatefulSet ovsdbserver-nb successful

openstack

cert-manager-certificaterequests-issuer-ca

neutron-ovndbs-1

CertificateIssued

Certificate fetched from issuer successfully

openstack

cert-manager-certificaterequests-approver

neutron-ovndbs-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack

cert-manager-certificaterequests-issuer-vault

neutron-ovndbs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificates-trigger

ovndbcluster-sb-ovndbs

Issuing

Issuing certificate as Secret does not exist

openstack

cert-manager-certificaterequests-issuer-ca

neutron-ovndbs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-acme

neutron-ovndbs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-venafi

neutron-ovndbs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-selfsigned

neutron-ovndbs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-vault

ovndbcluster-sb-ovndbs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-ca

ovndbcluster-sb-ovndbs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificates-request-manager

ovndbcluster-sb-ovndbs

Requested

Created new CertificateRequest resource "ovndbcluster-sb-ovndbs-1"

openstack

cert-manager-certificates-key-manager

ovndbcluster-sb-ovndbs

Generated

Stored new private key in temporary Secret resource "ovndbcluster-sb-ovndbs-fmkvs"

openstack

cert-manager-certificaterequests-issuer-selfsigned

ovndbcluster-sb-ovndbs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-acme

ovndbcluster-sb-ovndbs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificates-issuing

neutron-ovndbs

Issuing

The certificate has been successfully issued

openstack

cert-manager-certificaterequests-issuer-venafi

ovndbcluster-sb-ovndbs-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

topolvm.io_lvms-operator-dd6b59969-zzdqx_ce63b60a-9fe5-4592-9fa1-06008d67137b

ovndbcluster-nb-etc-ovn-ovsdbserver-nb-0

ProvisioningSucceeded

Successfully provisioned volume pvc-c9f7f5f2-b30e-4ea6-89f6-a25b5651e9cf

openstack

cert-manager-certificaterequests-issuer-ca

ovndbcluster-sb-ovndbs-1

CertificateIssued

Certificate fetched from issuer successfully

openstack

cert-manager-certificates-issuing

ovndbcluster-sb-ovndbs

Issuing

The certificate has been successfully issued

openstack

cert-manager-certificaterequests-approver

ovndbcluster-sb-ovndbs-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack

persistentvolume-controller

ovndbcluster-sb-etc-ovn-ovsdbserver-sb-0

WaitForFirstConsumer

waiting for first consumer to be created before binding
(x2)

openstack

persistentvolume-controller

ovndbcluster-sb-etc-ovn-ovsdbserver-sb-0

ExternalProvisioning

Waiting for a volume to be created either by the external provisioner 'topolvm.io' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered.
(x3)

openstack

cert-manager-issuers

rootca-public

KeyPairVerified

Signing CA verified

openstack

topolvm.io_lvms-operator-dd6b59969-zzdqx_ce63b60a-9fe5-4592-9fa1-06008d67137b

ovndbcluster-sb-etc-ovn-ovsdbserver-sb-0

Provisioning

External provisioner is provisioning volume for claim "openstack/ovndbcluster-sb-etc-ovn-ovsdbserver-sb-0"

openstack

statefulset-controller

ovsdbserver-sb

SuccessfulCreate

create Pod ovsdbserver-sb-0 in StatefulSet ovsdbserver-sb successful

openstack

statefulset-controller

ovsdbserver-sb

SuccessfulCreate

create Claim ovndbcluster-sb-etc-ovn-ovsdbserver-sb-0 Pod ovsdbserver-sb-0 in StatefulSet ovsdbserver-sb success

openstack

topolvm.io_lvms-operator-dd6b59969-zzdqx_ce63b60a-9fe5-4592-9fa1-06008d67137b

ovndbcluster-sb-etc-ovn-ovsdbserver-sb-0

ProvisioningSucceeded

Successfully provisioned volume pvc-5b494783-e9ed-438b-9ae2-64b525749c17

openstack

kubelet

dnsmasq-dns-8476fd89bc-24d8n

Pulled

Successfully pulled image "quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:fbb5be29e9e4fa11f0743e7f74f2e80dcc7445d24770709ea0e038147f752c51" in 19.852s (19.852s including waiting). Image size: 679076174 bytes.

openstack

kubelet

dnsmasq-dns-685c76cf85-qwz94

Pulled

Successfully pulled image "quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:fbb5be29e9e4fa11f0743e7f74f2e80dcc7445d24770709ea0e038147f752c51" in 19.789s (19.789s including waiting). Image size: 679076174 bytes.

openstack

kubelet

dnsmasq-dns-6ff8fd9d5c-47t6r

Pulled

Successfully pulled image "quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:fbb5be29e9e4fa11f0743e7f74f2e80dcc7445d24770709ea0e038147f752c51" in 16.354s (16.354s including waiting). Image size: 679076174 bytes.

openstack

kubelet

dnsmasq-dns-586dbdbb8c-28nq4

Pulled

Successfully pulled image "quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:fbb5be29e9e4fa11f0743e7f74f2e80dcc7445d24770709ea0e038147f752c51" in 16.357s (16.357s including waiting). Image size: 679076174 bytes.

openstack

multus

ovn-controller-ovs-hz4j4

AddedInterface

Add eth0 [10.128.0.175/23] from ovn-kubernetes

openstack

kubelet

ovsdbserver-nb-0

Pulling

Pulling image "quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server@sha256:ac71d8f4475d08f0a40a993cf5f083aead99232c2d5d8cd9514d63a345d0c128"

openstack

multus

openstack-cell1-galera-0

AddedInterface

Add eth0 [10.128.0.173/23] from ovn-kubernetes

openstack

kubelet

dnsmasq-dns-8476fd89bc-24d8n

Created

Created container: init

openstack

kubelet

dnsmasq-dns-586dbdbb8c-28nq4

Started

Started container init

openstack

kubelet

dnsmasq-dns-8476fd89bc-24d8n

Started

Started container init

openstack

multus

ovsdbserver-nb-0

AddedInterface

Add internalapi [172.17.0.30/24] from openstack/internalapi

openstack

multus

memcached-0

AddedInterface

Add eth0 [10.128.0.170/23] from ovn-kubernetes

openstack

kubelet

openstack-galera-0

Pulling

Pulling image "quay.io/podified-antelope-centos9/openstack-mariadb@sha256:4caef2b55e01b9a7ee88a22bc69db1893521a91d95c7ad4c8e593f14f17a5f95"

openstack

kubelet

rabbitmq-server-0

Pulling

Pulling image "quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:2087a09e7ea9f1dbadd433366bb46cc93dd5460ac9606b65f430460f4c2ee18d"

openstack

multus

ovn-controller-q5t48

AddedInterface

Add eth0 [10.128.0.174/23] from ovn-kubernetes

openstack

multus

ovsdbserver-nb-0

AddedInterface

Add eth0 [10.128.0.176/23] from ovn-kubernetes

openstack

kubelet

dnsmasq-dns-685c76cf85-qwz94

Created

Created container: init

openstack

kubelet

dnsmasq-dns-685c76cf85-qwz94

Started

Started container init

openstack

multus

rabbitmq-server-0

AddedInterface

Add eth0 [10.128.0.169/23] from ovn-kubernetes

openstack

kubelet

dnsmasq-dns-6ff8fd9d5c-47t6r

Created

Created container: init

openstack

kubelet

dnsmasq-dns-6ff8fd9d5c-47t6r

Started

Started container init

openstack

kubelet

dnsmasq-dns-586dbdbb8c-28nq4

Created

Created container: init

openstack

multus

openstack-galera-0

AddedInterface

Add eth0 [10.128.0.172/23] from ovn-kubernetes

openstack

kubelet

rabbitmq-cell1-server-0

Pulling

Pulling image "quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:2087a09e7ea9f1dbadd433366bb46cc93dd5460ac9606b65f430460f4c2ee18d"

openstack

multus

rabbitmq-cell1-server-0

AddedInterface

Add eth0 [10.128.0.171/23] from ovn-kubernetes

openstack

kubelet

openstack-cell1-galera-0

Pulling

Pulling image "quay.io/podified-antelope-centos9/openstack-mariadb@sha256:4caef2b55e01b9a7ee88a22bc69db1893521a91d95c7ad4c8e593f14f17a5f95"

openstack

kubelet

dnsmasq-dns-6ff8fd9d5c-47t6r

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:fbb5be29e9e4fa11f0743e7f74f2e80dcc7445d24770709ea0e038147f752c51" already present on machine

openstack

kubelet

ovn-controller-ovs-hz4j4

Pulling

Pulling image "quay.io/podified-antelope-centos9/openstack-ovn-base@sha256:943db2d546cbc78f663edcd102c478b71d755a66f99d24fea1b4e628c4125104"

openstack

multus

ovn-controller-ovs-hz4j4

AddedInterface

Add tenant [172.19.0.30/24] from openstack/tenant

openstack

kubelet

ovn-controller-q5t48

Pulling

Pulling image "quay.io/podified-antelope-centos9/openstack-ovn-controller@sha256:79213c923a25c7aa65998a66c3c2c2fbd8973f837cfb94f867e567cd71614af0"

openstack

multus

ovn-controller-ovs-hz4j4

AddedInterface

Add datacentre [] from openstack/datacentre

openstack

multus

ovsdbserver-sb-0

AddedInterface

Add internalapi [172.17.0.31/24] from openstack/internalapi

openstack

kubelet

dnsmasq-dns-586dbdbb8c-28nq4

Failed

Error: container create failed: mount `/var/lib/kubelet/pods/cbb818c8-8429-4748-a924-2b7f77d812da/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory

openstack

kubelet

ovsdbserver-sb-0

Pulling

Pulling image "quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server@sha256:bba31d7d170c92451c1d62346da1057e9c0e941a074a32cc54219cb79a4ea24a"

openstack

kubelet

memcached-0

Pulling

Pulling image "quay.io/podified-antelope-centos9/openstack-memcached@sha256:0d759b31e4da88b3fa1b823ab634d982fd713e81ed648626de1d8ec40ae7cad4"

openstack

kubelet

dnsmasq-dns-6ff8fd9d5c-47t6r

Started

Started container dnsmasq-dns

openstack

kubelet

dnsmasq-dns-6ff8fd9d5c-47t6r

Created

Created container: dnsmasq-dns

openstack

multus

ovsdbserver-sb-0

AddedInterface

Add eth0 [10.128.0.177/23] from ovn-kubernetes

openstack

kubelet

dnsmasq-dns-586dbdbb8c-28nq4

Started

Started container dnsmasq-dns

openstack

kubelet

dnsmasq-dns-586dbdbb8c-28nq4

Created

Created container: dnsmasq-dns
(x2)

openstack

kubelet

dnsmasq-dns-586dbdbb8c-28nq4

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:fbb5be29e9e4fa11f0743e7f74f2e80dcc7445d24770709ea0e038147f752c51" already present on machine

openstack

kubelet

dnsmasq-dns-586dbdbb8c-28nq4

Killing

Stopping container dnsmasq-dns

openstack

replicaset-controller

dnsmasq-dns-586dbdbb8c

SuccessfulDelete

Deleted pod: dnsmasq-dns-586dbdbb8c-28nq4

openstack

deployment-controller

dnsmasq-dns

ScalingReplicaSet

Scaled down replica set dnsmasq-dns-586dbdbb8c to 0 from 1

openstack

kubelet

memcached-0

Pulled

Successfully pulled image "quay.io/podified-antelope-centos9/openstack-memcached@sha256:0d759b31e4da88b3fa1b823ab634d982fd713e81ed648626de1d8ec40ae7cad4" in 10.16s (10.16s including waiting). Image size: 277692612 bytes.

openstack

kubelet

openstack-galera-0

Created

Created container: mysql-bootstrap

openstack

kubelet

memcached-0

Created

Created container: memcached

openstack

kubelet

rabbitmq-cell1-server-0

Pulled

Successfully pulled image "quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:2087a09e7ea9f1dbadd433366bb46cc93dd5460ac9606b65f430460f4c2ee18d" in 12.544s (12.544s including waiting). Image size: 304732739 bytes.

openstack

kubelet

openstack-cell1-galera-0

Pulled

Successfully pulled image "quay.io/podified-antelope-centos9/openstack-mariadb@sha256:4caef2b55e01b9a7ee88a22bc69db1893521a91d95c7ad4c8e593f14f17a5f95" in 12.352s (12.352s including waiting). Image size: 429677374 bytes.

openstack

kubelet

ovsdbserver-nb-0

Pulling

Pulling image "quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:156f95f51d0a91422548c574e96ee37f07a200c948e173b22523982f24f1e79c"

openstack

kubelet

ovsdbserver-nb-0

Started

Started container ovsdbserver-nb

openstack

kubelet

ovsdbserver-nb-0

Created

Created container: ovsdbserver-nb

openstack

kubelet

ovsdbserver-sb-0

Pulled

Successfully pulled image "quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server@sha256:bba31d7d170c92451c1d62346da1057e9c0e941a074a32cc54219cb79a4ea24a" in 12.073s (12.073s including waiting). Image size: 346963744 bytes.

openstack

kubelet

ovsdbserver-sb-0

Created

Created container: ovsdbserver-sb

openstack

kubelet

ovsdbserver-nb-0

Pulled

Successfully pulled image "quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server@sha256:ac71d8f4475d08f0a40a993cf5f083aead99232c2d5d8cd9514d63a345d0c128" in 12.347s (12.348s including waiting). Image size: 346963744 bytes.

openstack

kubelet

ovn-controller-ovs-hz4j4

Started

Started container ovsdb-server-init

openstack

kubelet

ovn-controller-ovs-hz4j4

Pulled

Successfully pulled image "quay.io/podified-antelope-centos9/openstack-ovn-base@sha256:943db2d546cbc78f663edcd102c478b71d755a66f99d24fea1b4e628c4125104" in 12.175s (12.175s including waiting). Image size: 324510219 bytes.

openstack

kubelet

memcached-0

Started

Started container memcached

openstack

kubelet

openstack-cell1-galera-0

Created

Created container: mysql-bootstrap

openstack

kubelet

openstack-cell1-galera-0

Started

Started container mysql-bootstrap

openstack

kubelet

rabbitmq-server-0

Pulled

Successfully pulled image "quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:2087a09e7ea9f1dbadd433366bb46cc93dd5460ac9606b65f430460f4c2ee18d" in 12.769s (12.769s including waiting). Image size: 304732739 bytes.

openstack

kubelet

ovsdbserver-sb-0

Started

Started container ovsdbserver-sb

openstack

kubelet

ovsdbserver-sb-0

Pulling

Pulling image "quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:156f95f51d0a91422548c574e96ee37f07a200c948e173b22523982f24f1e79c"

openstack

kubelet

openstack-galera-0

Pulled

Successfully pulled image "quay.io/podified-antelope-centos9/openstack-mariadb@sha256:4caef2b55e01b9a7ee88a22bc69db1893521a91d95c7ad4c8e593f14f17a5f95" in 12.506s (12.506s including waiting). Image size: 429677374 bytes.

openstack

kubelet

ovn-controller-ovs-hz4j4

Created

Created container: ovsdb-server-init

openstack

kubelet

openstack-galera-0

Started

Started container mysql-bootstrap

openstack

kubelet

ovn-controller-q5t48

Started

Started container ovn-controller

openstack

kubelet

ovn-controller-q5t48

Created

Created container: ovn-controller

openstack

kubelet

ovn-controller-q5t48

Pulled

Successfully pulled image "quay.io/podified-antelope-centos9/openstack-ovn-controller@sha256:79213c923a25c7aa65998a66c3c2c2fbd8973f837cfb94f867e567cd71614af0" in 12.272s (12.272s including waiting). Image size: 346792900 bytes.

openstack

kubelet

rabbitmq-cell1-server-0

Started

Started container setup-container

openstack

kubelet

rabbitmq-cell1-server-0

Created

Created container: setup-container

openstack

kubelet

ovn-controller-ovs-hz4j4

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-ovn-base@sha256:943db2d546cbc78f663edcd102c478b71d755a66f99d24fea1b4e628c4125104" already present on machine

openstack

kubelet

rabbitmq-server-0

Started

Started container setup-container

openstack

kubelet

rabbitmq-server-0

Created

Created container: setup-container

openstack

kubelet

ovn-controller-ovs-hz4j4

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-ovn-base@sha256:943db2d546cbc78f663edcd102c478b71d755a66f99d24fea1b4e628c4125104" already present on machine

openstack

kubelet

ovn-controller-ovs-hz4j4

Created

Created container: ovs-vswitchd

openstack

kubelet

ovn-controller-ovs-hz4j4

Started

Started container ovs-vswitchd

openstack

kubelet

ovn-controller-ovs-hz4j4

Started

Started container ovsdb-server

openstack

kubelet

ovn-controller-ovs-hz4j4

Created

Created container: ovsdb-server

openstack

kubelet

ovsdbserver-sb-0

Started

Started container openstack-network-exporter

openstack

kubelet

openstack-galera-0

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-mariadb@sha256:4caef2b55e01b9a7ee88a22bc69db1893521a91d95c7ad4c8e593f14f17a5f95" already present on machine

openstack

kubelet

ovsdbserver-sb-0

Created

Created container: openstack-network-exporter

openstack

kubelet

ovsdbserver-sb-0

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:156f95f51d0a91422548c574e96ee37f07a200c948e173b22523982f24f1e79c" in 5.86s (5.86s including waiting). Image size: 165206333 bytes.

openstack

kubelet

openstack-cell1-galera-0

Started

Started container galera

openstack

kubelet

openstack-cell1-galera-0

Created

Created container: galera

openstack

kubelet

openstack-cell1-galera-0

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-mariadb@sha256:4caef2b55e01b9a7ee88a22bc69db1893521a91d95c7ad4c8e593f14f17a5f95" already present on machine

openstack

kubelet

openstack-galera-0

Created

Created container: galera

openstack

kubelet

openstack-galera-0

Started

Started container galera

openstack

kubelet

ovsdbserver-nb-0

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:156f95f51d0a91422548c574e96ee37f07a200c948e173b22523982f24f1e79c" in 5.899s (5.899s including waiting). Image size: 165206333 bytes.

openstack

kubelet

ovsdbserver-nb-0

Started

Started container openstack-network-exporter

openstack

kubelet

ovsdbserver-nb-0

Created

Created container: openstack-network-exporter

openstack

deployment-controller

dnsmasq-dns

ScalingReplicaSet

Scaled down replica set dnsmasq-dns-65db7fd8ff to 0 from 1

openstack

deployment-controller

dnsmasq-dns

ScalingReplicaSet

Scaled up replica set dnsmasq-dns-65db7fd8ff to 1

openstack

statefulset-controller

ovn-northd

SuccessfulCreate

create Pod ovn-northd-0 in StatefulSet ovn-northd successful

openstack

replicaset-controller

dnsmasq-dns-76f498f559

SuccessfulCreate

Created pod: dnsmasq-dns-76f498f559-52t76

openstack

daemonset-controller

ovn-controller-metrics

SuccessfulCreate

Created pod: ovn-controller-metrics-p4sw4

openstack

replicaset-controller

dnsmasq-dns-65db7fd8ff

SuccessfulDelete

Deleted pod: dnsmasq-dns-65db7fd8ff-r6h25

openstack

replicaset-controller

dnsmasq-dns-65db7fd8ff

SuccessfulCreate

Created pod: dnsmasq-dns-65db7fd8ff-r6h25

openstack

multus

dnsmasq-dns-65db7fd8ff-r6h25

AddedInterface

Add eth0 [10.128.0.178/23] from ovn-kubernetes

openstack

multus

ovn-controller-metrics-p4sw4

AddedInterface

Add eth0 [10.128.0.179/23] from ovn-kubernetes

openstack

kubelet

dnsmasq-dns-76f498f559-52t76

Started

Started container init

openstack

kubelet

dnsmasq-dns-76f498f559-52t76

Created

Created container: init

openstack

kubelet

dnsmasq-dns-65db7fd8ff-r6h25

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:fbb5be29e9e4fa11f0743e7f74f2e80dcc7445d24770709ea0e038147f752c51" already present on machine

openstack

kubelet

dnsmasq-dns-65db7fd8ff-r6h25

Created

Created container: init

openstack

metal3-baremetal-controller

bmh1

Registered

Registered new host

openstack

metal3-baremetal-controller

bmh0

Registered

Registered new host

openstack

kubelet

dnsmasq-dns-65db7fd8ff-r6h25

Started

Started container init

openstack

kubelet

ovn-northd-0

Pulling

Pulling image "quay.io/podified-antelope-centos9/openstack-ovn-northd@sha256:3d4aa78bc0932fd39a377beb5a649e47832c0de33a62c413776de2f9de31763e"

openstack

multus

dnsmasq-dns-76f498f559-52t76

AddedInterface

Add eth0 [10.128.0.180/23] from ovn-kubernetes

openstack

multus

ovn-northd-0

AddedInterface

Add eth0 [10.128.0.181/23] from ovn-kubernetes

openstack

kubelet

dnsmasq-dns-76f498f559-52t76

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:fbb5be29e9e4fa11f0743e7f74f2e80dcc7445d24770709ea0e038147f752c51" already present on machine

openstack

kubelet

ovn-controller-metrics-p4sw4

Pulled

Container image "quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:156f95f51d0a91422548c574e96ee37f07a200c948e173b22523982f24f1e79c" already present on machine

openstack

kubelet

ovn-controller-metrics-p4sw4

Created

Created container: openstack-network-exporter

openstack

kubelet

ovn-controller-metrics-p4sw4

Started

Started container openstack-network-exporter

openstack

kubelet

ovn-northd-0

Pulled

Successfully pulled image "quay.io/podified-antelope-centos9/openstack-ovn-northd@sha256:3d4aa78bc0932fd39a377beb5a649e47832c0de33a62c413776de2f9de31763e" in 1.26s (1.26s including waiting). Image size: 346960837 bytes.

openstack

kubelet

dnsmasq-dns-76f498f559-52t76

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:fbb5be29e9e4fa11f0743e7f74f2e80dcc7445d24770709ea0e038147f752c51" already present on machine

openstack

kubelet

dnsmasq-dns-76f498f559-52t76

Started

Started container dnsmasq-dns

openstack

kubelet

ovn-northd-0

Started

Started container openstack-network-exporter

openstack

kubelet

ovn-northd-0

Created

Created container: ovn-northd

openstack

kubelet

dnsmasq-dns-76f498f559-52t76

Created

Created container: dnsmasq-dns

openstack

kubelet

ovn-northd-0

Created

Created container: openstack-network-exporter

openstack

kubelet

ovn-northd-0

Pulled

Container image "quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:156f95f51d0a91422548c574e96ee37f07a200c948e173b22523982f24f1e79c" already present on machine

openstack

kubelet

ovn-northd-0

Started

Started container ovn-northd
(x5)

openstack

rabbitmqcluster-controller

rabbitmq

SuccessfulUpdate

updated resource rabbitmq of Type *v1.Service
(x5)

openstack

rabbitmqcluster-controller

rabbitmq

SuccessfulUpdate

updated resource rabbitmq-server of Type *v1.StatefulSet

openstack

job-controller

keystone-ec41-account-create-update

SuccessfulCreate

Created pod: keystone-ec41-account-create-update-wlbg9

openstack

job-controller

placement-db-create

SuccessfulCreate

Created pod: placement-db-create-dqlc9

openstack

job-controller

placement-ffa9-account-create-update

SuccessfulCreate

Created pod: placement-ffa9-account-create-update-22wx5
(x5)

openstack

rabbitmqcluster-controller

rabbitmq-cell1

SuccessfulUpdate

updated resource rabbitmq-cell1 of Type *v1.Service
(x5)

openstack

rabbitmqcluster-controller

rabbitmq-cell1

SuccessfulUpdate

updated resource rabbitmq-cell1-server of Type *v1.StatefulSet

openstack

job-controller

keystone-db-create

SuccessfulCreate

Created pod: keystone-db-create-cslxs

openstack

multus

placement-ffa9-account-create-update-22wx5

AddedInterface

Add eth0 [10.128.0.185/23] from ovn-kubernetes

openstack

topolvm.io_lvms-operator-dd6b59969-zzdqx_ce63b60a-9fe5-4592-9fa1-06008d67137b

swift-swift-storage-0

Provisioning

External provisioner is provisioning volume for claim "openstack/swift-swift-storage-0"

openstack

kubelet

keystone-ec41-account-create-update-wlbg9

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-mariadb@sha256:4caef2b55e01b9a7ee88a22bc69db1893521a91d95c7ad4c8e593f14f17a5f95" already present on machine

openstack

kubelet

dnsmasq-dns-76f498f559-52t76

Killing

Stopping container dnsmasq-dns

openstack

statefulset-controller

swift-storage

SuccessfulCreate

create Claim swift-swift-storage-0 Pod swift-storage-0 in StatefulSet swift-storage success

openstack

multus

keystone-ec41-account-create-update-wlbg9

AddedInterface

Add eth0 [10.128.0.182/23] from ovn-kubernetes

openstack

statefulset-controller

swift-storage

SuccessfulCreate

create Pod swift-storage-0 in StatefulSet swift-storage successful

openstack

replicaset-controller

dnsmasq-dns-76f498f559

SuccessfulDelete

Deleted pod: dnsmasq-dns-76f498f559-52t76

openstack

persistentvolume-controller

swift-swift-storage-0

WaitForFirstConsumer

waiting for first consumer to be created before binding

openstack

persistentvolume-controller

swift-swift-storage-0

ExternalProvisioning

Waiting for a volume to be created either by the external provisioner 'topolvm.io' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered.

openstack

kubelet

keystone-db-create-cslxs

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-mariadb@sha256:4caef2b55e01b9a7ee88a22bc69db1893521a91d95c7ad4c8e593f14f17a5f95" already present on machine

openstack

multus

keystone-db-create-cslxs

AddedInterface

Add eth0 [10.128.0.183/23] from ovn-kubernetes
(x2)

openstack

metallb-controller

swift-internal

deprecatedAnnotation

Service uses deprecated annotation metallb.universe.tf/loadBalancerIPs

openstack

cert-manager-certificates-trigger

swift-internal-svc

Issuing

Issuing certificate as Secret does not exist

openstack

metallb-controller

swift-internal

IPAllocated

Assigned IP ["172.17.0.80"]

openstack

multus

placement-db-create-dqlc9

AddedInterface

Add eth0 [10.128.0.184/23] from ovn-kubernetes
(x2)

openstack

metallb-controller

swift-internal

deprecatedAnnotation

Service uses deprecated annotation metallb.universe.tf/address-pool

openstack

replicaset-controller

dnsmasq-dns-5bf8b865dc

SuccessfulCreate

Created pod: dnsmasq-dns-5bf8b865dc-dhd96
(x2)

openstack

metallb-controller

swift-internal

deprecatedAnnotation

Service uses deprecated annotation metallb.universe.tf/allow-shared-ip

openstack

cert-manager-certificaterequests-issuer-selfsigned

swift-internal-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

kubelet

keystone-ec41-account-create-update-wlbg9

Created

Created container: mariadb-account-create-update

openstack

kubelet

placement-db-create-dqlc9

Started

Started container mariadb-database-create

openstack

cert-manager-certificates-trigger

swift-public-svc

Issuing

Issuing certificate as Secret does not exist

openstack

kubelet

keystone-db-create-cslxs

Started

Started container mariadb-database-create

openstack

kubelet

placement-db-create-dqlc9

Created

Created container: mariadb-database-create

openstack

cert-manager-certificaterequests-approver

swift-internal-svc-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack

kubelet

placement-db-create-dqlc9

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-mariadb@sha256:4caef2b55e01b9a7ee88a22bc69db1893521a91d95c7ad4c8e593f14f17a5f95" already present on machine

openstack

cert-manager-certificaterequests-issuer-vault

swift-internal-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificates-issuing

swift-internal-svc

Issuing

The certificate has been successfully issued

openstack

cert-manager-certificates-request-manager

swift-internal-svc

Requested

Created new CertificateRequest resource "swift-internal-svc-1"

openstack

cert-manager-certificaterequests-issuer-ca

swift-internal-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

kubelet

placement-ffa9-account-create-update-22wx5

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-mariadb@sha256:4caef2b55e01b9a7ee88a22bc69db1893521a91d95c7ad4c8e593f14f17a5f95" already present on machine

openstack

kubelet

placement-ffa9-account-create-update-22wx5

Created

Created container: mariadb-account-create-update

openstack

kubelet

keystone-ec41-account-create-update-wlbg9

Started

Started container mariadb-account-create-update

openstack

cert-manager-certificaterequests-issuer-ca

swift-internal-svc-1

CertificateIssued

Certificate fetched from issuer successfully

openstack

cert-manager-certificaterequests-issuer-venafi

swift-internal-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-acme

swift-internal-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificates-key-manager

swift-internal-svc

Generated

Stored new private key in temporary Secret resource "swift-internal-svc-58s65"

openstack

kubelet

placement-ffa9-account-create-update-22wx5

Started

Started container mariadb-account-create-update

openstack

kubelet

keystone-db-create-cslxs

Created

Created container: mariadb-database-create

openstack

cert-manager-certificaterequests-issuer-ca

swift-public-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-ca

swift-public-svc-1

CertificateIssued

Certificate fetched from issuer successfully

openstack

cert-manager-certificates-trigger

swift-public-route

Issuing

Issuing certificate as Secret does not exist

openstack

cert-manager-certificaterequests-approver

swift-public-svc-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack

kubelet

dnsmasq-dns-5bf8b865dc-dhd96

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:fbb5be29e9e4fa11f0743e7f74f2e80dcc7445d24770709ea0e038147f752c51" already present on machine

openstack

cert-manager-certificaterequests-issuer-venafi

swift-public-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-selfsigned

swift-public-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

kubelet

dnsmasq-dns-5bf8b865dc-dhd96

Started

Started container init

openstack

cert-manager-certificaterequests-issuer-vault

swift-public-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-acme

swift-public-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

kubelet

dnsmasq-dns-5bf8b865dc-dhd96

Created

Created container: init

openstack

cert-manager-certificates-key-manager

swift-public-svc

Generated

Stored new private key in temporary Secret resource "swift-public-svc-vbm22"

openstack

cert-manager-certificates-request-manager

swift-public-svc

Requested

Created new CertificateRequest resource "swift-public-svc-1"

openstack

cert-manager-certificates-issuing

swift-public-svc

Issuing

The certificate has been successfully issued

openstack

topolvm.io_lvms-operator-dd6b59969-zzdqx_ce63b60a-9fe5-4592-9fa1-06008d67137b

swift-swift-storage-0

ProvisioningSucceeded

Successfully provisioned volume pvc-5b0423b7-8d71-4226-b20d-e3df8d969082

openstack

multus

dnsmasq-dns-5bf8b865dc-dhd96

AddedInterface

Add eth0 [10.128.0.186/23] from ovn-kubernetes

openstack

kubelet

dnsmasq-dns-5bf8b865dc-dhd96

Started

Started container dnsmasq-dns

openstack

cert-manager-certificaterequests-issuer-acme

swift-public-route-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-selfsigned

swift-public-route-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-approver

swift-public-route-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack

cert-manager-certificaterequests-issuer-vault

swift-public-route-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-ca

swift-public-route-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-venafi

swift-public-route-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificates-key-manager

swift-public-route

Generated

Stored new private key in temporary Secret resource "swift-public-route-4d9zr"

openstack

cert-manager-certificates-request-manager

swift-public-route

Requested

Created new CertificateRequest resource "swift-public-route-1"

openstack

kubelet

dnsmasq-dns-5bf8b865dc-dhd96

Created

Created container: dnsmasq-dns

openstack

kubelet

dnsmasq-dns-5bf8b865dc-dhd96

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:fbb5be29e9e4fa11f0743e7f74f2e80dcc7445d24770709ea0e038147f752c51" already present on machine

openstack

cert-manager-certificaterequests-issuer-ca

swift-public-route-1

CertificateIssued

Certificate fetched from issuer successfully

openstack

cert-manager-certificates-issuing

swift-public-route

Issuing

The certificate has been successfully issued

openstack

job-controller

glance-db-create

SuccessfulCreate

Created pod: glance-db-create-l87pv

openstack

job-controller

glance-fb00-account-create-update

SuccessfulCreate

Created pod: glance-fb00-account-create-update-9vcp7

openstack

kubelet

glance-fb00-account-create-update-9vcp7

Created

Created container: mariadb-account-create-update

openstack

kubelet

glance-db-create-l87pv

Started

Started container mariadb-database-create

openstack

kubelet

glance-fb00-account-create-update-9vcp7

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-mariadb@sha256:4caef2b55e01b9a7ee88a22bc69db1893521a91d95c7ad4c8e593f14f17a5f95" already present on machine

openstack

multus

glance-fb00-account-create-update-9vcp7

AddedInterface

Add eth0 [10.128.0.189/23] from ovn-kubernetes

openstack

kubelet

glance-fb00-account-create-update-9vcp7

Started

Started container mariadb-account-create-update

openstack

ovnk-controlplane

swift-ring-rebalance-2zgwx

ErrorUpdatingResource

addLogicalPort failed for openstack/swift-ring-rebalance-2zgwx: failed to update pod openstack/swift-ring-rebalance-2zgwx: Operation cannot be fulfilled on pods "swift-ring-rebalance-2zgwx": the object has been modified; please apply your changes to the latest version and try again

openstack

job-controller

keystone-db-create

Completed

Job completed

openstack

multus

glance-db-create-l87pv

AddedInterface

Add eth0 [10.128.0.188/23] from ovn-kubernetes

openstack

kubelet

glance-db-create-l87pv

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-mariadb@sha256:4caef2b55e01b9a7ee88a22bc69db1893521a91d95c7ad4c8e593f14f17a5f95" already present on machine

openstack

job-controller

swift-ring-rebalance

SuccessfulCreate

Created pod: swift-ring-rebalance-2zgwx

openstack

job-controller

swift-ring-rebalance

SuccessfulCreate

Created pod: swift-ring-rebalance-p9cff

openstack

job-controller

placement-ffa9-account-create-update

Completed

Job completed

openstack

kubelet

glance-db-create-l87pv

Created

Created container: mariadb-database-create

openstack

job-controller

root-account-create-update

SuccessfulCreate

Created pod: root-account-create-update-nc4r5

openstack

kubelet

swift-ring-rebalance-p9cff

Pulling

Pulling image "quay.io/podified-antelope-centos9/openstack-swift-proxy-server@sha256:6c3eb966650a7a98feb4ddb31e1bdba1095b0c62e349196aca6a423681d7e5fb"

openstack

multus

root-account-create-update-nc4r5

AddedInterface

Add eth0 [10.128.0.193/23] from ovn-kubernetes

openstack

job-controller

placement-db-create

Completed

Job completed

openstack

multus

swift-ring-rebalance-p9cff

AddedInterface

Add eth0 [10.128.0.192/23] from ovn-kubernetes

openstack

job-controller

keystone-ec41-account-create-update

Completed

Job completed

openstack

kubelet

root-account-create-update-nc4r5

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-mariadb@sha256:4caef2b55e01b9a7ee88a22bc69db1893521a91d95c7ad4c8e593f14f17a5f95" already present on machine

openstack

kubelet

root-account-create-update-nc4r5

Started

Started container mariadb-account-create-update

openstack

kubelet

root-account-create-update-nc4r5

Created

Created container: mariadb-account-create-update

openstack

job-controller

glance-fb00-account-create-update

Completed

Job completed

openstack

job-controller

glance-db-create

Completed

Job completed

openstack

job-controller

glance-db-sync

SuccessfulCreate

Created pod: glance-db-sync-662t6

openstack

kubelet

swift-ring-rebalance-p9cff

Created

Created container: swift-ring-rebalance

openstack

kubelet

swift-ring-rebalance-p9cff

Started

Started container swift-ring-rebalance

openstack

kubelet

swift-ring-rebalance-p9cff

Pulled

Successfully pulled image "quay.io/podified-antelope-centos9/openstack-swift-proxy-server@sha256:6c3eb966650a7a98feb4ddb31e1bdba1095b0c62e349196aca6a423681d7e5fb" in 4.006s (4.006s including waiting). Image size: 500200203 bytes.
(x5)

openstack

kubelet

swift-storage-0

FailedMount

MountVolume.SetUp failed for volume "etc-swift" : configmap "swift-ring-files" not found

openstack

job-controller

root-account-create-update

Completed

Job completed

openstack

multus

glance-db-sync-662t6

AddedInterface

Add eth0 [10.128.0.194/23] from ovn-kubernetes

openstack

multus

glance-db-sync-662t6

AddedInterface

Add storage [172.18.0.30/24] from openstack/storage

openstack

replicaset-controller

dnsmasq-dns-6ff8fd9d5c

SuccessfulDelete

Deleted pod: dnsmasq-dns-6ff8fd9d5c-47t6r

openstack

kubelet

dnsmasq-dns-6ff8fd9d5c-47t6r

Killing

Stopping container dnsmasq-dns

openstack

kubelet

glance-db-sync-662t6

Pulling

Pulling image "quay.io/podified-antelope-centos9/openstack-glance-api@sha256:dae5e39780d5a15eed030c7009f8e5317139d447558ac83f038497be594be120"

openstack

metal3-baremetal-controller

bmh1

BMCAccessValidated

Verified access to BMC

openstack

metal3-baremetal-controller

bmh0

BMCAccessValidated

Verified access to BMC

openstack

metal3-baremetal-controller

bmh1

ProfileSet

Hardware profile set: unknown

openstack

metal3-baremetal-controller

bmh0

ProfileSet

Hardware profile set: unknown

openstack

metal3-baremetal-controller

bmh1

PowerOff

Host soft powered off

openstack

job-controller

root-account-create-update

SuccessfulCreate

Created pod: root-account-create-update-54pxn

openstack

metal3-baremetal-controller

bmh0

PowerOff

Host soft powered off

openstack

kubelet

root-account-create-update-54pxn

Created

Created container: mariadb-account-create-update

openstack

multus

root-account-create-update-54pxn

AddedInterface

Add eth0 [10.128.0.195/23] from ovn-kubernetes

openstack

kubelet

root-account-create-update-54pxn

Started

Started container mariadb-account-create-update

openstack

kubelet

root-account-create-update-54pxn

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-mariadb@sha256:4caef2b55e01b9a7ee88a22bc69db1893521a91d95c7ad4c8e593f14f17a5f95" already present on machine

openstack

job-controller

ovn-controller-q5t48-config

SuccessfulCreate

Created pod: ovn-controller-q5t48-config-bfd5h

openstack

kubelet

glance-db-sync-662t6

Started

Started container glance-db-sync

openstack

kubelet

rabbitmq-server-0

Started

Started container rabbitmq

openstack

kubelet

rabbitmq-server-0

Created

Created container: rabbitmq

openstack

kubelet

rabbitmq-server-0

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:2087a09e7ea9f1dbadd433366bb46cc93dd5460ac9606b65f430460f4c2ee18d" already present on machine

openstack

kubelet

rabbitmq-cell1-server-0

Created

Created container: rabbitmq
(x2)

openstack

kubelet

ovn-controller-q5t48

Unhealthy

Readiness probe failed: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status

openstack

kubelet

rabbitmq-cell1-server-0

Started

Started container rabbitmq

openstack

kubelet

glance-db-sync-662t6

Created

Created container: glance-db-sync

openstack

kubelet

glance-db-sync-662t6

Pulled

Successfully pulled image "quay.io/podified-antelope-centos9/openstack-glance-api@sha256:dae5e39780d5a15eed030c7009f8e5317139d447558ac83f038497be594be120" in 13.2s (13.2s including waiting). Image size: 983055212 bytes.

openstack

multus

swift-storage-0

AddedInterface

Add eth0 [10.128.0.187/23] from ovn-kubernetes

openstack

kubelet

swift-storage-0

Pulling

Pulling image "quay.io/podified-antelope-centos9/openstack-swift-account@sha256:f919023e9754d0d94b3fa3e7f571e6d22330ad3cdbb17b20d6143d2581b49ef1"

openstack

kubelet

rabbitmq-cell1-server-0

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:2087a09e7ea9f1dbadd433366bb46cc93dd5460ac9606b65f430460f4c2ee18d" already present on machine

openstack

multus

ovn-controller-q5t48-config-bfd5h

AddedInterface

Add eth0 [10.128.0.196/23] from ovn-kubernetes

openstack

kubelet

ovn-controller-q5t48-config-bfd5h

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-ovn-controller@sha256:79213c923a25c7aa65998a66c3c2c2fbd8973f837cfb94f867e567cd71614af0" already present on machine

openstack

kubelet

ovn-controller-q5t48-config-bfd5h

Created

Created container: ovn-config

openstack

kubelet

ovn-controller-q5t48-config-bfd5h

Started

Started container ovn-config

openstack

job-controller

root-account-create-update

Completed

Job completed

openstack

kubelet

swift-storage-0

Created

Created container: account-replicator

openstack

kubelet

swift-storage-0

Started

Started container account-replicator

openstack

job-controller

swift-ring-rebalance

Completed

Job completed

openstack

kubelet

swift-storage-0

Started

Started container account-server

openstack

kubelet

swift-storage-0

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-swift-account@sha256:f919023e9754d0d94b3fa3e7f571e6d22330ad3cdbb17b20d6143d2581b49ef1" already present on machine

openstack

kubelet

swift-storage-0

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-swift-account@sha256:f919023e9754d0d94b3fa3e7f571e6d22330ad3cdbb17b20d6143d2581b49ef1" already present on machine

openstack

kubelet

swift-storage-0

Started

Started container account-auditor

openstack

kubelet

swift-storage-0

Pulled

Successfully pulled image "quay.io/podified-antelope-centos9/openstack-swift-account@sha256:f919023e9754d0d94b3fa3e7f571e6d22330ad3cdbb17b20d6143d2581b49ef1" in 1.495s (1.495s including waiting). Image size: 445167617 bytes.

openstack

kubelet

swift-storage-0

Created

Created container: account-auditor

openstack

kubelet

swift-storage-0

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-swift-account@sha256:f919023e9754d0d94b3fa3e7f571e6d22330ad3cdbb17b20d6143d2581b49ef1" already present on machine

openstack

kubelet

swift-storage-0

Created

Created container: account-server

openstack

kubelet

swift-storage-0

Pulling

Pulling image "quay.io/podified-antelope-centos9/openstack-swift-container@sha256:37da219a7d5254e5fa6cac571f99d8ca7c600d3243b68ffb282a6c70ff8b3ff2"

openstack

kubelet

swift-storage-0

Created

Created container: account-reaper

openstack

kubelet

swift-storage-0

Started

Started container account-reaper

openstack

kubelet

swift-storage-0

Pulled

Successfully pulled image "quay.io/podified-antelope-centos9/openstack-swift-container@sha256:37da219a7d5254e5fa6cac571f99d8ca7c600d3243b68ffb282a6c70ff8b3ff2" in 1.334s (1.334s including waiting). Image size: 445183491 bytes.

openstack

kubelet

swift-storage-0

Created

Created container: container-server

openstack

kubelet

swift-storage-0

Started

Started container container-server

openstack

kubelet

swift-storage-0

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-swift-container@sha256:37da219a7d5254e5fa6cac571f99d8ca7c600d3243b68ffb282a6c70ff8b3ff2" already present on machine

openstack

job-controller

ovn-controller-q5t48-config

SuccessfulCreate

Created pod: ovn-controller-q5t48-config-944s7

openstack

job-controller

ovn-controller-q5t48-config

Completed

Job completed

openstack

kubelet

swift-storage-0

Created

Created container: container-replicator

openstack

kubelet

swift-storage-0

Started

Started container container-replicator

openstack

job-controller

root-account-create-update

SuccessfulCreate

Created pod: root-account-create-update-ndh4m

openstack

multus

ovn-controller-q5t48-config-944s7

AddedInterface

Add eth0 [10.128.0.197/23] from ovn-kubernetes

openstack

kubelet

ovn-controller-q5t48-config-944s7

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-ovn-controller@sha256:79213c923a25c7aa65998a66c3c2c2fbd8973f837cfb94f867e567cd71614af0" already present on machine

openstack

replicaset-controller

dnsmasq-dns-6c8447fc9

SuccessfulCreate

Created pod: dnsmasq-dns-6c8447fc9-wqvmw

openstack

kubelet

ovn-controller-q5t48-config-944s7

Created

Created container: ovn-config

openstack

kubelet

ovn-controller-q5t48-config-944s7

Started

Started container ovn-config

openstack

kubelet

root-account-create-update-ndh4m

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-mariadb@sha256:4caef2b55e01b9a7ee88a22bc69db1893521a91d95c7ad4c8e593f14f17a5f95" already present on machine

openstack

multus

root-account-create-update-ndh4m

AddedInterface

Add eth0 [10.128.0.199/23] from ovn-kubernetes

openstack

kubelet

root-account-create-update-ndh4m

Created

Created container: mariadb-account-create-update

openstack

kubelet

root-account-create-update-ndh4m

Started

Started container mariadb-account-create-update

openstack

multus

dnsmasq-dns-6c8447fc9-wqvmw

AddedInterface

Add eth0 [10.128.0.198/23] from ovn-kubernetes

openstack

kubelet

dnsmasq-dns-6c8447fc9-wqvmw

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:fbb5be29e9e4fa11f0743e7f74f2e80dcc7445d24770709ea0e038147f752c51" already present on machine

openstack

kubelet

dnsmasq-dns-6c8447fc9-wqvmw

Created

Created container: init

openstack

kubelet

dnsmasq-dns-6c8447fc9-wqvmw

Started

Started container init

openstack

kubelet

dnsmasq-dns-6c8447fc9-wqvmw

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:fbb5be29e9e4fa11f0743e7f74f2e80dcc7445d24770709ea0e038147f752c51" already present on machine

openstack

kubelet

dnsmasq-dns-6c8447fc9-wqvmw

Created

Created container: dnsmasq-dns

openstack

kubelet

dnsmasq-dns-6c8447fc9-wqvmw

Started

Started container dnsmasq-dns

openstack

replicaset-controller

dnsmasq-dns-6c8447fc9

SuccessfulDelete

Deleted pod: dnsmasq-dns-6c8447fc9-wqvmw

openstack

replicaset-controller

dnsmasq-dns-6974f6bb85

SuccessfulCreate

Created pod: dnsmasq-dns-6974f6bb85-z8znc

openstack

rabbitmq-server-0/rabbitmq_peer_discovery

pod/rabbitmq-server-0

Created

Node rabbit@rabbitmq-server-0.rabbitmq-nodes.openstack is registered

openstack

rabbitmq-cell1-server-0/rabbitmq_peer_discovery

pod/rabbitmq-cell1-server-0

Created

Node rabbit@rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack is registered

openstack

kubelet

dnsmasq-dns-6974f6bb85-z8znc

Created

Created container: dnsmasq-dns

openstack

kubelet

dnsmasq-dns-6974f6bb85-z8znc

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:fbb5be29e9e4fa11f0743e7f74f2e80dcc7445d24770709ea0e038147f752c51" already present on machine

openstack

kubelet

dnsmasq-dns-6c8447fc9-wqvmw

Killing

Stopping container dnsmasq-dns

openstack

multus

dnsmasq-dns-6974f6bb85-z8znc

AddedInterface

Add eth0 [10.128.0.200/23] from ovn-kubernetes

openstack

kubelet

dnsmasq-dns-6974f6bb85-z8znc

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:fbb5be29e9e4fa11f0743e7f74f2e80dcc7445d24770709ea0e038147f752c51" already present on machine

openstack

job-controller

ovn-controller-q5t48-config

Completed

Job completed

openstack

kubelet

dnsmasq-dns-6974f6bb85-z8znc

Started

Started container init

openstack

kubelet

dnsmasq-dns-6974f6bb85-z8znc

Started

Started container dnsmasq-dns

openstack

kubelet

dnsmasq-dns-6974f6bb85-z8znc

Created

Created container: init

openstack

job-controller

root-account-create-update

Completed

Job completed

openstack

job-controller

glance-db-sync

Completed

Job completed

openstack

metallb-controller

glance-default-internal

IPAllocated

Assigned IP ["172.17.0.80"]
(x2)

openstack

metallb-controller

glance-default-internal

deprecatedAnnotation

Service uses deprecated annotation metallb.universe.tf/allow-shared-ip
(x2)

openstack

metallb-controller

glance-default-internal

deprecatedAnnotation

Service uses deprecated annotation metallb.universe.tf/loadBalancerIPs
(x2)

openstack

metallb-controller

glance-default-internal

deprecatedAnnotation

Service uses deprecated annotation metallb.universe.tf/address-pool

openstack

replicaset-controller

dnsmasq-dns-7d58447bc7

SuccessfulDelete

Deleted pod: dnsmasq-dns-7d58447bc7-xn5wl

openstack

job-controller

cinder-db-create

SuccessfulCreate

Created pod: cinder-db-create-vtvmv

openstack

deployment-controller

edpm-a-provisionserver-openstackprovisionserver

ScalingReplicaSet

Scaled up replica set edpm-a-provisionserver-openstackprovisionserver-78fd94b97 to 1

openstack

cert-manager-certificates-issuing

glance-default-internal-svc

Issuing

The certificate has been successfully issued

openstack

cert-manager-certificates-request-manager

glance-default-internal-svc

Requested

Created new CertificateRequest resource "glance-default-internal-svc-1"

openstack

cert-manager-certificates-key-manager

glance-default-internal-svc

Generated

Stored new private key in temporary Secret resource "glance-default-internal-svc-flzvr"

openstack

cert-manager-certificates-trigger

glance-default-internal-svc

Issuing

Issuing certificate as Secret does not exist

openstack

metallb-speaker

rabbitmq

nodeAssigned

announcing from node "master-0" with protocol "layer2"

openstack

replicaset-controller

dnsmasq-dns-5d96bd4f7c

SuccessfulDelete

Deleted pod: dnsmasq-dns-5d96bd4f7c-7qqmd

openstack

cert-manager-certificaterequests-approver

glance-default-internal-svc-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack

cert-manager-certificaterequests-issuer-ca

glance-default-internal-svc-1

CertificateIssued

Certificate fetched from issuer successfully

openstack

cert-manager-certificaterequests-issuer-acme

glance-default-internal-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-venafi

glance-default-internal-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-ca

glance-default-internal-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-selfsigned

glance-default-internal-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-vault

glance-default-internal-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

job-controller

cinder-3574-account-create-update

SuccessfulCreate

Created pod: cinder-3574-account-create-update-mw5s6

openstack

replicaset-controller

dnsmasq-dns-5d96bd4f7c

SuccessfulCreate

Created pod: dnsmasq-dns-5d96bd4f7c-7qqmd

openstack

replicaset-controller

dnsmasq-dns-7d58447bc7

SuccessfulCreate

Created pod: dnsmasq-dns-7d58447bc7-xn5wl

openstack

replicaset-controller

dnsmasq-dns-b5b845b79

SuccessfulCreate

Created pod: dnsmasq-dns-b5b845b79-xn624

openstack

replicaset-controller

dnsmasq-dns-6974f6bb85

SuccessfulDelete

Deleted pod: dnsmasq-dns-6974f6bb85-z8znc

openstack

kubelet

dnsmasq-dns-6974f6bb85-z8znc

Killing

Stopping container dnsmasq-dns

openstack

replicaset-controller

edpm-a-provisionserver-openstackprovisionserver-78fd94b97

SuccessfulCreate

Created pod: edpm-a-provisionserver-openstackprovisionserver-78fd94b97-tkt7j

openstack

job-controller

neutron-3e52-account-create-update

SuccessfulCreate

Created pod: neutron-3e52-account-create-update-qlzvf

openstack

job-controller

keystone-db-sync

SuccessfulCreate

Created pod: keystone-db-sync-jg5hc

openstack

multus

dnsmasq-dns-b5b845b79-xn624

AddedInterface

Add eth0 [10.128.0.203/23] from ovn-kubernetes

openstack

kubelet

dnsmasq-dns-b5b845b79-xn624

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:fbb5be29e9e4fa11f0743e7f74f2e80dcc7445d24770709ea0e038147f752c51" already present on machine

openstack

multus

cinder-db-create-vtvmv

AddedInterface

Add eth0 [10.128.0.204/23] from ovn-kubernetes

openstack

job-controller

neutron-db-create

SuccessfulCreate

Created pod: neutron-db-create-bbvbz

openstack

multus

cinder-3574-account-create-update-mw5s6

AddedInterface

Add eth0 [10.128.0.205/23] from ovn-kubernetes

openstack

kubelet

cinder-db-create-vtvmv

Started

Started container mariadb-database-create

openstack

replicaset-controller

edpm-b-provisionserver-openstackprovisionserver-6bf75656cf

SuccessfulCreate

Created pod: edpm-b-provisionserver-openstackprovisionserver-6bf75656cftnmfr

openstack

kubelet

neutron-db-create-bbvbz

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-mariadb@sha256:4caef2b55e01b9a7ee88a22bc69db1893521a91d95c7ad4c8e593f14f17a5f95" already present on machine

openstack

kubelet

dnsmasq-dns-b5b845b79-xn624

Started

Started container init

openstack

multus

neutron-db-create-bbvbz

AddedInterface

Add eth0 [10.128.0.207/23] from ovn-kubernetes

openstack

kubelet

dnsmasq-dns-b5b845b79-xn624

Created

Created container: init

openstack

kubelet

keystone-db-sync-jg5hc

Pulling

Pulling image "quay.io/podified-antelope-centos9/openstack-keystone@sha256:3a148c87899d4cbbb6bbe1203ad6c237fb295b5f42abda425dc0329305723414"

openstack

multus

keystone-db-sync-jg5hc

AddedInterface

Add eth0 [10.128.0.206/23] from ovn-kubernetes

openstack

metallb-speaker

rabbitmq-cell1

nodeAssigned

announcing from node "master-0" with protocol "layer2"

openstack

kubelet

cinder-db-create-vtvmv

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-mariadb@sha256:4caef2b55e01b9a7ee88a22bc69db1893521a91d95c7ad4c8e593f14f17a5f95" already present on machine

openstack

kubelet

cinder-3574-account-create-update-mw5s6

Started

Started container mariadb-account-create-update

openstack

deployment-controller

edpm-b-provisionserver-openstackprovisionserver

ScalingReplicaSet

Scaled up replica set edpm-b-provisionserver-openstackprovisionserver-6bf75656cf to 1

openstack

kubelet

cinder-3574-account-create-update-mw5s6

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-mariadb@sha256:4caef2b55e01b9a7ee88a22bc69db1893521a91d95c7ad4c8e593f14f17a5f95" already present on machine

openstack

kubelet

cinder-db-create-vtvmv

Created

Created container: mariadb-database-create

openstack

kubelet

edpm-a-provisionserver-openstackprovisionserver-78fd94b97-tkt7j

Pulling

Pulling image "quay.io/podified-antelope-centos9/edpm-hardened-uefi@sha256:894ce79b38510973ca610423cc34a7383b7761b6ceb47d18637daffaa93336f7"

openstack

kubelet

cinder-3574-account-create-update-mw5s6

Created

Created container: mariadb-account-create-update

openstack

cert-manager-certificates-trigger

glance-default-public-svc

Issuing

Issuing certificate as Secret does not exist

openstack

cert-manager-certificates-key-manager

glance-default-public-route

Generated

Stored new private key in temporary Secret resource "glance-default-public-route-vvjfq"

openstack

cert-manager-certificaterequests-issuer-venafi

glance-default-public-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

kubelet

neutron-3e52-account-create-update-qlzvf

Started

Started container mariadb-account-create-update

openstack

cert-manager-certificaterequests-issuer-selfsigned

glance-default-public-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

kubelet

edpm-b-provisionserver-openstackprovisionserver-6bf75656cftnmfr

Pulling

Pulling image "quay.io/podified-antelope-centos9/edpm-hardened-uefi@sha256:894ce79b38510973ca610423cc34a7383b7761b6ceb47d18637daffaa93336f7"

openstack

cert-manager-certificaterequests-approver

glance-default-public-svc-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack

multus

neutron-3e52-account-create-update-qlzvf

AddedInterface

Add eth0 [10.128.0.208/23] from ovn-kubernetes

openstack

cert-manager-certificates-issuing

glance-default-public-svc

Issuing

The certificate has been successfully issued

openstack

cert-manager-certificaterequests-issuer-ca

glance-default-public-svc-1

CertificateIssued

Certificate fetched from issuer successfully

openstack

cert-manager-certificates-request-manager

glance-default-public-svc

Requested

Created new CertificateRequest resource "glance-default-public-svc-1"

openstack

cert-manager-certificaterequests-issuer-ca

glance-default-public-route-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-vault

glance-default-public-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-ca

glance-default-public-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificates-issuing

glance-default-public-route

Issuing

The certificate has been successfully issued

openstack

cert-manager-certificates-key-manager

glance-default-public-svc

Generated

Stored new private key in temporary Secret resource "glance-default-public-svc-dchbv"

openstack

cert-manager-certificates-request-manager

glance-default-public-route

Requested

Created new CertificateRequest resource "glance-default-public-route-1"

openstack

cert-manager-certificaterequests-issuer-acme

glance-default-public-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

kubelet

neutron-db-create-bbvbz

Started

Started container mariadb-database-create

openstack

cert-manager-certificates-trigger

glance-default-public-route

Issuing

Issuing certificate as Secret does not exist

openstack

cert-manager-certificaterequests-issuer-ca

glance-default-public-route-1

CertificateIssued

Certificate fetched from issuer successfully

openstack

cert-manager-certificaterequests-approver

glance-default-public-route-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack

kubelet

neutron-3e52-account-create-update-qlzvf

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-mariadb@sha256:4caef2b55e01b9a7ee88a22bc69db1893521a91d95c7ad4c8e593f14f17a5f95" already present on machine

openstack

cert-manager-certificaterequests-issuer-venafi

glance-default-public-route-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

kubelet

dnsmasq-dns-b5b845b79-xn624

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:fbb5be29e9e4fa11f0743e7f74f2e80dcc7445d24770709ea0e038147f752c51" already present on machine

openstack

kubelet

neutron-3e52-account-create-update-qlzvf

Created

Created container: mariadb-account-create-update

openstack

cert-manager-certificaterequests-issuer-selfsigned

glance-default-public-route-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

kubelet

neutron-db-create-bbvbz

Created

Created container: mariadb-database-create

openstack

cert-manager-certificaterequests-issuer-acme

glance-default-public-route-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-vault

glance-default-public-route-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

kubelet

dnsmasq-dns-b5b845b79-xn624

Created

Created container: dnsmasq-dns

openstack

kubelet

dnsmasq-dns-b5b845b79-xn624

Started

Started container dnsmasq-dns

openstack

job-controller

neutron-3e52-account-create-update

Completed

Job completed

openstack

job-controller

cinder-3574-account-create-update

Completed

Job completed

openstack

job-controller

cinder-db-create

Completed

Job completed

openstack

job-controller

neutron-db-create

Completed

Job completed

openstack

replicaset-controller

dnsmasq-dns-5bf8b865dc

SuccessfulDelete

Deleted pod: dnsmasq-dns-5bf8b865dc-dhd96

openstack

kubelet

keystone-db-sync-jg5hc

Started

Started container keystone-db-sync

openstack

kubelet

keystone-db-sync-jg5hc

Created

Created container: keystone-db-sync

openstack

kubelet

dnsmasq-dns-5bf8b865dc-dhd96

Unhealthy

Readiness probe failed: dial tcp 10.128.0.186:5353: connect: connection refused

openstack

kubelet

keystone-db-sync-jg5hc

Pulled

Successfully pulled image "quay.io/podified-antelope-centos9/openstack-keystone@sha256:3a148c87899d4cbbb6bbe1203ad6c237fb295b5f42abda425dc0329305723414" in 8.268s (8.268s including waiting). Image size: 520208900 bytes.

openstack

kubelet

dnsmasq-dns-5bf8b865dc-dhd96

Killing

Stopping container dnsmasq-dns

openstack

job-controller

keystone-db-sync

Completed

Job completed

openstack

statefulset-controller

glance-27086-default-internal-api

SuccessfulCreate

create Claim glance-glance-27086-default-internal-api-0 Pod glance-27086-default-internal-api-0 in StatefulSet glance-27086-default-internal-api success

openstack

persistentvolume-controller

glance-glance-27086-default-external-api-0

WaitForFirstConsumer

waiting for first consumer to be created before binding
(x2)

openstack

metallb-controller

placement-internal

deprecatedAnnotation

Service uses deprecated annotation metallb.universe.tf/allow-shared-ip
(x2)

openstack

metallb-controller

placement-internal

deprecatedAnnotation

Service uses deprecated annotation metallb.universe.tf/loadBalancerIPs
(x2)

openstack

persistentvolume-controller

glance-glance-27086-default-internal-api-0

ExternalProvisioning

Waiting for a volume to be created either by the external provisioner 'topolvm.io' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered.

openstack

replicaset-controller

dnsmasq-dns-5bd59cc7c7

SuccessfulCreate

Created pod: dnsmasq-dns-5bd59cc7c7-jcbnm

openstack

persistentvolume-controller

glance-glance-27086-default-internal-api-0

WaitForFirstConsumer

waiting for first consumer to be created before binding

openstack

topolvm.io_lvms-operator-dd6b59969-zzdqx_ce63b60a-9fe5-4592-9fa1-06008d67137b

glance-glance-27086-default-external-api-0

Provisioning

External provisioner is provisioning volume for claim "openstack/glance-glance-27086-default-external-api-0"

openstack

persistentvolume-controller

glance-glance-27086-default-external-api-0

ExternalProvisioning

Waiting for a volume to be created either by the external provisioner 'topolvm.io' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered.
(x2)

openstack

metallb-controller

placement-internal

deprecatedAnnotation

Service uses deprecated annotation metallb.universe.tf/address-pool
(x2)

openstack

metallb-controller

keystone-internal

deprecatedAnnotation

Service uses deprecated annotation metallb.universe.tf/allow-shared-ip
(x2)

openstack

metallb-controller

keystone-internal

deprecatedAnnotation

Service uses deprecated annotation metallb.universe.tf/address-pool

openstack

metallb-controller

placement-internal

IPAllocated

Assigned IP ["172.17.0.80"]

openstack

statefulset-controller

glance-27086-default-external-api

SuccessfulCreate

create Claim glance-glance-27086-default-external-api-0 Pod glance-27086-default-external-api-0 in StatefulSet glance-27086-default-external-api success

openstack

job-controller

cinder-c920a-db-sync

SuccessfulCreate

Created pod: cinder-c920a-db-sync-5sx9b
(x2)

openstack

metallb-controller

keystone-internal

deprecatedAnnotation

Service uses deprecated annotation metallb.universe.tf/loadBalancerIPs

openstack

metallb-controller

keystone-internal

IPAllocated

Assigned IP ["172.17.0.80"]

openstack

job-controller

neutron-db-sync

SuccessfulCreate

Created pod: neutron-db-sync-q6j55

openstack

job-controller

keystone-bootstrap

SuccessfulCreate

Created pod: keystone-bootstrap-xz2l7

openstack

cert-manager-certificates-trigger

keystone-internal-svc

Issuing

Issuing certificate as Secret does not exist

openstack

multus

dnsmasq-dns-5bd59cc7c7-jcbnm

AddedInterface

Add eth0 [10.128.0.210/23] from ovn-kubernetes

openstack

kubelet

dnsmasq-dns-5bd59cc7c7-jcbnm

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:fbb5be29e9e4fa11f0743e7f74f2e80dcc7445d24770709ea0e038147f752c51" already present on machine

openstack

replicaset-controller

dnsmasq-dns-74db7c7d5f

SuccessfulCreate

Created pod: dnsmasq-dns-74db7c7d5f-t4vkw

openstack

job-controller

placement-db-sync

SuccessfulCreate

Created pod: placement-db-sync-8jqss

openstack

cert-manager-certificaterequests-issuer-selfsigned

keystone-internal-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-venafi

keystone-internal-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-acme

keystone-internal-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-ca

keystone-internal-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-vault

keystone-internal-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-approver

keystone-internal-svc-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack

cert-manager-certificaterequests-issuer-ca

keystone-internal-svc-1

CertificateIssued

Certificate fetched from issuer successfully

openstack

cert-manager-certificates-key-manager

keystone-internal-svc

Generated

Stored new private key in temporary Secret resource "keystone-internal-svc-nrppb"

openstack

cert-manager-certificates-request-manager

keystone-internal-svc

Requested

Created new CertificateRequest resource "keystone-internal-svc-1"

openstack

cert-manager-certificates-issuing

keystone-internal-svc

Issuing

The certificate has been successfully issued

openstack

cert-manager-certificaterequests-issuer-venafi

keystone-public-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-acme

keystone-public-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-vault

keystone-public-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-ca

keystone-public-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-selfsigned

keystone-public-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-approver

keystone-public-svc-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack

cert-manager-certificaterequests-issuer-ca

keystone-public-svc-1

CertificateIssued

Certificate fetched from issuer successfully

openstack

cert-manager-certificates-trigger

keystone-public-svc

Issuing

Issuing certificate as Secret does not exist

openstack

cert-manager-certificates-key-manager

keystone-public-svc

Generated

Stored new private key in temporary Secret resource "keystone-public-svc-zwkrq"

openstack

cert-manager-certificates-request-manager

keystone-public-svc

Requested

Created new CertificateRequest resource "keystone-public-svc-1"

openstack

topolvm.io_lvms-operator-dd6b59969-zzdqx_ce63b60a-9fe5-4592-9fa1-06008d67137b

glance-glance-27086-default-internal-api-0

Provisioning

External provisioner is provisioning volume for claim "openstack/glance-glance-27086-default-internal-api-0"
(x16)

openstack

deployment-controller

dnsmasq-dns

ScalingReplicaSet

(combined from similar events): Scaled down replica set dnsmasq-dns-5bd59cc7c7 to 0 from 1

openstack

replicaset-controller

dnsmasq-dns-5bd59cc7c7

SuccessfulDelete

Deleted pod: dnsmasq-dns-5bd59cc7c7-jcbnm

openstack

topolvm.io_lvms-operator-dd6b59969-zzdqx_ce63b60a-9fe5-4592-9fa1-06008d67137b

glance-glance-27086-default-external-api-0

ProvisioningSucceeded

Successfully provisioned volume pvc-c52c5b10-3247-40dd-a4c1-8cd8b0dd6647

openstack

kubelet

neutron-db-sync-q6j55

Created

Created container: neutron-db-sync

openstack

topolvm.io_lvms-operator-dd6b59969-zzdqx_ce63b60a-9fe5-4592-9fa1-06008d67137b

glance-glance-27086-default-internal-api-0

ProvisioningSucceeded

Successfully provisioned volume pvc-ed7fc861-4795-473d-84e6-66068cd18122

openstack

kubelet

keystone-bootstrap-xz2l7

Created

Created container: keystone-bootstrap

openstack

kubelet

keystone-bootstrap-xz2l7

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-keystone@sha256:3a148c87899d4cbbb6bbe1203ad6c237fb295b5f42abda425dc0329305723414" already present on machine

openstack

multus

keystone-bootstrap-xz2l7

AddedInterface

Add eth0 [10.128.0.209/23] from ovn-kubernetes

openstack

kubelet

placement-db-sync-8jqss

Pulling

Pulling image "quay.io/podified-antelope-centos9/openstack-placement-api@sha256:b8a5d052890fb9cefa333baf10b607add227ed5d79aa108b576a97b21e89327a"

openstack

kubelet

dnsmasq-dns-74db7c7d5f-t4vkw

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:fbb5be29e9e4fa11f0743e7f74f2e80dcc7445d24770709ea0e038147f752c51" already present on machine

openstack

multus

dnsmasq-dns-74db7c7d5f-t4vkw

AddedInterface

Add eth0 [10.128.0.214/23] from ovn-kubernetes

openstack

kubelet

neutron-db-sync-q6j55

Started

Started container neutron-db-sync

openstack

kubelet

cinder-c920a-db-sync-5sx9b

Pulling

Pulling image "quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:574a17f0877c175128a764f2b37fc02456649c8514689125718ce6ca974bfb6b"

openstack

multus

cinder-c920a-db-sync-5sx9b

AddedInterface

Add eth0 [10.128.0.211/23] from ovn-kubernetes

openstack

multus

neutron-db-sync-q6j55

AddedInterface

Add eth0 [10.128.0.212/23] from ovn-kubernetes

openstack

kubelet

keystone-bootstrap-xz2l7

Started

Started container keystone-bootstrap

openstack

multus

placement-db-sync-8jqss

AddedInterface

Add eth0 [10.128.0.213/23] from ovn-kubernetes

openstack

kubelet

dnsmasq-dns-5bd59cc7c7-jcbnm

Created

Created container: init

openstack

cert-manager-certificaterequests-issuer-venafi

keystone-public-route-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

kubelet

dnsmasq-dns-5bd59cc7c7-jcbnm

Started

Started container init

openstack

cert-manager-certificaterequests-issuer-ca

keystone-public-route-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

kubelet

neutron-db-sync-q6j55

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:fbb5be29e9e4fa11f0743e7f74f2e80dcc7445d24770709ea0e038147f752c51" already present on machine

openstack

cert-manager-certificaterequests-issuer-vault

keystone-public-route-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-acme

keystone-public-route-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificates-issuing

keystone-public-svc

Issuing

The certificate has been successfully issued

openstack

cert-manager-certificaterequests-issuer-selfsigned

keystone-public-route-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-approver

keystone-public-route-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack

cert-manager-certificaterequests-issuer-ca

keystone-public-route-1

CertificateIssued

Certificate fetched from issuer successfully

openstack

cert-manager-certificates-trigger

keystone-public-route

Issuing

Issuing certificate as Secret does not exist

openstack

cert-manager-certificates-key-manager

keystone-public-route

Generated

Stored new private key in temporary Secret resource "keystone-public-route-5hz9w"

openstack

cert-manager-certificates-request-manager

keystone-public-route

Requested

Created new CertificateRequest resource "keystone-public-route-1"

openstack

cert-manager-certificates-issuing

keystone-public-route

Issuing

The certificate has been successfully issued

openstack

cert-manager-certificates-trigger

placement-internal-svc

Issuing

Issuing certificate as Secret does not exist

openstack

cert-manager-certificaterequests-issuer-ca

placement-internal-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificates-key-manager

placement-internal-svc

Generated

Stored new private key in temporary Secret resource "placement-internal-svc-2snd4"

openstack

cert-manager-certificaterequests-issuer-ca

placement-internal-svc-1

CertificateIssued

Certificate fetched from issuer successfully

openstack

cert-manager-certificaterequests-approver

placement-internal-svc-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack

kubelet

dnsmasq-dns-74db7c7d5f-t4vkw

Created

Created container: dnsmasq-dns

openstack

kubelet

dnsmasq-dns-74db7c7d5f-t4vkw

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:fbb5be29e9e4fa11f0743e7f74f2e80dcc7445d24770709ea0e038147f752c51" already present on machine

openstack

kubelet

dnsmasq-dns-74db7c7d5f-t4vkw

Started

Started container init

openstack

kubelet

dnsmasq-dns-74db7c7d5f-t4vkw

Created

Created container: init

openstack

cert-manager-certificates-request-manager

placement-internal-svc

Requested

Created new CertificateRequest resource "placement-internal-svc-1"

openstack

cert-manager-certificaterequests-issuer-vault

placement-internal-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-selfsigned

placement-internal-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-acme

placement-internal-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-venafi

placement-internal-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

kubelet

dnsmasq-dns-74db7c7d5f-t4vkw

Started

Started container dnsmasq-dns

openstack

cert-manager-certificates-issuing

placement-internal-svc

Issuing

The certificate has been successfully issued

openstack

cert-manager-certificates-trigger

placement-public-svc

Issuing

Issuing certificate as Secret does not exist

openstack

cert-manager-certificates-key-manager

placement-public-svc

Generated

Stored new private key in temporary Secret resource "placement-public-svc-6rg5z"

openshift-machine-config-operator

machine-config-operator

machine-config-operator

ConfigMapUpdated

Updated ConfigMap/kube-rbac-proxy -n openshift-machine-config-operator: cause by changes in data.config-file.yaml

openstack

cert-manager-certificates-trigger

placement-public-route

Issuing

Issuing certificate as Secret does not exist

openstack

cert-manager-certificaterequests-issuer-vault

placement-public-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-approver

placement-public-svc-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack

cert-manager-certificaterequests-issuer-venafi

placement-public-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificates-request-manager

placement-public-svc

Requested

Created new CertificateRequest resource "placement-public-svc-1"

openstack

cert-manager-certificaterequests-issuer-selfsigned

placement-public-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-acme

placement-public-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-ca

placement-public-svc-1

CertificateIssued

Certificate fetched from issuer successfully

openstack

cert-manager-certificates-issuing

placement-public-svc

Issuing

The certificate has been successfully issued

openstack

cert-manager-certificaterequests-issuer-ca

placement-public-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-selfsigned

placement-public-route-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificates-request-manager

placement-public-route

Requested

Created new CertificateRequest resource "placement-public-route-1"

openstack

cert-manager-certificaterequests-issuer-ca

placement-public-route-1

CertificateIssued

Certificate fetched from issuer successfully

openstack

cert-manager-certificaterequests-approver

placement-public-route-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack

cert-manager-certificaterequests-issuer-acme

placement-public-route-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-vault

placement-public-route-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-venafi

placement-public-route-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-ca

placement-public-route-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificates-issuing

placement-public-route

Issuing

The certificate has been successfully issued

openstack

cert-manager-certificates-key-manager

placement-public-route

Generated

Stored new private key in temporary Secret resource "placement-public-route-7vlc5"

openstack

replicaset-controller

dnsmasq-dns-b5b845b79

SuccessfulDelete

Deleted pod: dnsmasq-dns-b5b845b79-xn624

openstack

kubelet

dnsmasq-dns-b5b845b79-xn624

Killing

Stopping container dnsmasq-dns
(x2)

openstack

kubelet

dnsmasq-dns-b5b845b79-xn624

Unhealthy

Readiness probe failed: dial tcp 10.128.0.203:5353: connect: connection refused

openstack

job-controller

keystone-bootstrap

SuccessfulCreate

Created pod: keystone-bootstrap-x7bjf

openstack

kubelet

cinder-c920a-db-sync-5sx9b

Pulled

Successfully pulled image "quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:574a17f0877c175128a764f2b37fc02456649c8514689125718ce6ca974bfb6b" in 20.61s (20.61s including waiting). Image size: 1161166113 bytes.

openstack

job-controller

keystone-bootstrap

Completed

Job completed

openstack

kubelet

edpm-a-provisionserver-openstackprovisionserver-78fd94b97-tkt7j

Pulled

Successfully pulled image "quay.io/podified-antelope-centos9/edpm-hardened-uefi@sha256:894ce79b38510973ca610423cc34a7383b7761b6ceb47d18637daffaa93336f7" in 39.661s (39.661s including waiting). Image size: 964857828 bytes.

openstack

kubelet

edpm-b-provisionserver-openstackprovisionserver-6bf75656cftnmfr

Pulled

Successfully pulled image "quay.io/podified-antelope-centos9/edpm-hardened-uefi@sha256:894ce79b38510973ca610423cc34a7383b7761b6ceb47d18637daffaa93336f7" in 38.289s (38.289s including waiting). Image size: 964857828 bytes.

openstack

kubelet

placement-db-sync-8jqss

Pulled

Successfully pulled image "quay.io/podified-antelope-centos9/openstack-placement-api@sha256:b8a5d052890fb9cefa333baf10b607add227ed5d79aa108b576a97b21e89327a" in 20.048s (20.048s including waiting). Image size: 472783568 bytes.

openstack

kubelet

keystone-bootstrap-x7bjf

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-keystone@sha256:3a148c87899d4cbbb6bbe1203ad6c237fb295b5f42abda425dc0329305723414" already present on machine

openstack

multus

keystone-bootstrap-x7bjf

AddedInterface

Add eth0 [10.128.0.219/23] from ovn-kubernetes

openstack

multus

glance-27086-default-external-api-0

AddedInterface

Add eth0 [10.128.0.217/23] from ovn-kubernetes

openstack

multus

glance-27086-default-external-api-0

AddedInterface

Add storage [172.18.0.30/24] from openstack/storage

openstack

kubelet

edpm-b-provisionserver-openstackprovisionserver-6bf75656cftnmfr

Started

Started container init

openstack

multus

glance-27086-default-internal-api-0

AddedInterface

Add eth0 [10.128.0.218/23] from ovn-kubernetes

openstack

multus

glance-27086-default-internal-api-0

AddedInterface

Add storage [172.18.0.31/24] from openstack/storage

openstack

kubelet

edpm-a-provisionserver-openstackprovisionserver-78fd94b97-tkt7j

Created

Created container: init

openstack

kubelet

edpm-b-provisionserver-openstackprovisionserver-6bf75656cftnmfr

Created

Created container: init

openstack

kubelet

glance-27086-default-external-api-0

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-glance-api@sha256:dae5e39780d5a15eed030c7009f8e5317139d447558ac83f038497be594be120" already present on machine

openstack

kubelet

glance-27086-default-internal-api-0

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-glance-api@sha256:dae5e39780d5a15eed030c7009f8e5317139d447558ac83f038497be594be120" already present on machine

openstack

kubelet

placement-db-sync-8jqss

Started

Started container placement-db-sync

openstack

kubelet

edpm-a-provisionserver-openstackprovisionserver-78fd94b97-tkt7j

Started

Started container init

openstack

kubelet

placement-db-sync-8jqss

Created

Created container: placement-db-sync

openstack

kubelet

glance-27086-default-internal-api-0

Created

Created container: glance-log

openstack

kubelet

cinder-c920a-db-sync-5sx9b

Started

Started container cinder-c920a-db-sync

openstack

kubelet

keystone-bootstrap-x7bjf

Created

Created container: keystone-bootstrap

openstack

kubelet

glance-27086-default-external-api-0

Started

Started container glance-log

openstack

kubelet

glance-27086-default-external-api-0

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-glance-api@sha256:dae5e39780d5a15eed030c7009f8e5317139d447558ac83f038497be594be120" already present on machine

openstack

kubelet

glance-27086-default-internal-api-0

Started

Started container glance-log

openstack

kubelet

glance-27086-default-internal-api-0

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-glance-api@sha256:dae5e39780d5a15eed030c7009f8e5317139d447558ac83f038497be594be120" already present on machine

openstack

kubelet

glance-27086-default-external-api-0

Created

Created container: glance-log

openstack

kubelet

cinder-c920a-db-sync-5sx9b

Created

Created container: cinder-c920a-db-sync

openstack

kubelet

keystone-bootstrap-x7bjf

Started

Started container keystone-bootstrap

openstack

kubelet

glance-27086-default-external-api-0

Created

Created container: glance-httpd

openstack

kubelet

glance-27086-default-external-api-0

Started

Started container glance-httpd

openstack

kubelet

glance-27086-default-internal-api-0

Created

Created container: glance-httpd

openstack

kubelet

glance-27086-default-internal-api-0

Started

Started container glance-httpd

openstack

kubelet

dnsmasq-dns-b5b845b79-xn624

Unhealthy

Readiness probe failed: dial tcp 10.128.0.203:5353: i/o timeout

openstack

kubelet

edpm-a-provisionserver-openstackprovisionserver-78fd94b97-tkt7j

Pulling

Pulling image "registry.redhat.io/ubi9/httpd-24@sha256:e0697f36760789183fefc807dafa3bfeb4098725f923eb9a8f034725a01fbf9f"

openstack

kubelet

edpm-b-provisionserver-openstackprovisionserver-6bf75656cftnmfr

Pulling

Pulling image "registry.redhat.io/ubi9/httpd-24@sha256:e0697f36760789183fefc807dafa3bfeb4098725f923eb9a8f034725a01fbf9f"

openstack

deployment-controller

placement

ScalingReplicaSet

Scaled up replica set placement-8b47bf878 to 1

openstack

replicaset-controller

placement-8b47bf878

SuccessfulCreate

Created pod: placement-8b47bf878-2ftwn

openstack

job-controller

placement-db-sync

Completed

Job completed

openstack

kubelet

placement-8b47bf878-2ftwn

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-placement-api@sha256:b8a5d052890fb9cefa333baf10b607add227ed5d79aa108b576a97b21e89327a" already present on machine

openstack

multus

placement-8b47bf878-2ftwn

AddedInterface

Add eth0 [10.128.0.220/23] from ovn-kubernetes

openstack

kubelet

placement-8b47bf878-2ftwn

Started

Started container placement-api

openstack

kubelet

placement-8b47bf878-2ftwn

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-placement-api@sha256:b8a5d052890fb9cefa333baf10b607add227ed5d79aa108b576a97b21e89327a" already present on machine

openstack

kubelet

placement-8b47bf878-2ftwn

Created

Created container: placement-log

openstack

kubelet

placement-8b47bf878-2ftwn

Created

Created container: placement-api

openstack

kubelet

placement-8b47bf878-2ftwn

Started

Started container placement-log

openstack

kubelet

edpm-a-provisionserver-openstackprovisionserver-78fd94b97-tkt7j

Created

Created container: osp-httpd

openstack

kubelet

edpm-b-provisionserver-openstackprovisionserver-6bf75656cftnmfr

Pulled

Successfully pulled image "registry.redhat.io/ubi9/httpd-24@sha256:e0697f36760789183fefc807dafa3bfeb4098725f923eb9a8f034725a01fbf9f" in 11.291s (11.291s including waiting). Image size: 313866450 bytes.

openstack

kubelet

edpm-b-provisionserver-openstackprovisionserver-6bf75656cftnmfr

Created

Created container: osp-httpd

openstack

kubelet

edpm-b-provisionserver-openstackprovisionserver-6bf75656cftnmfr

Started

Started container osp-httpd

openstack

kubelet

edpm-a-provisionserver-openstackprovisionserver-78fd94b97-tkt7j

Pulled

Successfully pulled image "registry.redhat.io/ubi9/httpd-24@sha256:e0697f36760789183fefc807dafa3bfeb4098725f923eb9a8f034725a01fbf9f" in 11.243s (11.243s including waiting). Image size: 313866450 bytes.

openstack

kubelet

edpm-a-provisionserver-openstackprovisionserver-78fd94b97-tkt7j

Started

Started container osp-httpd

openstack

deployment-controller

keystone

ScalingReplicaSet

Scaled up replica set keystone-5987d8f9fd to 1

openstack

replicaset-controller

keystone-5987d8f9fd

SuccessfulCreate

Created pod: keystone-5987d8f9fd-q4m4b

openstack

job-controller

keystone-bootstrap

Completed

Job completed

openstack

job-controller

neutron-db-sync

Completed

Job completed
(x2)

openstack

metallb-controller

neutron-internal

deprecatedAnnotation

Service uses deprecated annotation metallb.universe.tf/allow-shared-ip

openstack

metallb-controller

neutron-internal

IPAllocated

Assigned IP ["172.17.0.80"]

openstack

multus

keystone-5987d8f9fd-q4m4b

AddedInterface

Add eth0 [10.128.0.221/23] from ovn-kubernetes
(x2)

openstack

metallb-controller

neutron-internal

deprecatedAnnotation

Service uses deprecated annotation metallb.universe.tf/address-pool

openstack

replicaset-controller

dnsmasq-dns-7bcb7479f5

SuccessfulCreate

Created pod: dnsmasq-dns-7bcb7479f5-z74kg
(x2)

openstack

metallb-controller

neutron-internal

deprecatedAnnotation

Service uses deprecated annotation metallb.universe.tf/loadBalancerIPs

openstack

job-controller

cinder-c920a-db-sync

Completed

Job completed

openstack

kubelet

keystone-5987d8f9fd-q4m4b

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-keystone@sha256:3a148c87899d4cbbb6bbe1203ad6c237fb295b5f42abda425dc0329305723414" already present on machine

openstack

replicaset-controller

dnsmasq-dns-7cd5f68b57

SuccessfulCreate

Created pod: dnsmasq-dns-7cd5f68b57-cbttm

openstack

replicaset-controller

dnsmasq-dns-7bcb7479f5

SuccessfulDelete

Deleted pod: dnsmasq-dns-7bcb7479f5-z74kg

openstack

deployment-controller

neutron

ScalingReplicaSet

Scaled up replica set neutron-555d58bf7b to 1

openstack

metallb-controller

cinder-internal

IPAllocated

Assigned IP ["172.17.0.80"]

openstack

replicaset-controller

neutron-555d58bf7b

SuccessfulCreate

Created pod: neutron-555d58bf7b-nplsv

openstack

kubelet

keystone-5987d8f9fd-q4m4b

Created

Created container: keystone-api
(x2)

openstack

metallb-controller

cinder-internal

deprecatedAnnotation

Service uses deprecated annotation metallb.universe.tf/address-pool
(x2)

openstack

metallb-controller

cinder-internal

deprecatedAnnotation

Service uses deprecated annotation metallb.universe.tf/loadBalancerIPs
(x2)

openstack

metallb-controller

cinder-internal

deprecatedAnnotation

Service uses deprecated annotation metallb.universe.tf/allow-shared-ip

openstack

kubelet

keystone-5987d8f9fd-q4m4b

Started

Started container keystone-api

openstack

cert-manager-certificaterequests-issuer-acme

cinder-internal-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

kubelet

cinder-c920a-scheduler-0

Pulling

Pulling image "quay.io/podified-antelope-centos9/openstack-cinder-scheduler@sha256:9e7c747eeeefb391dc6dedaaac57fa694c4d08b991c54bb99aa6de77451e792f"

openstack

cert-manager-certificaterequests-issuer-venafi

cinder-internal-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

kubelet

dnsmasq-dns-7bcb7479f5-z74kg

Created

Created container: init

openstack

cert-manager-certificates-request-manager

cinder-internal-svc

Requested

Created new CertificateRequest resource "cinder-internal-svc-1"

openstack

cert-manager-certificates-trigger

cinder-public-svc

Issuing

Issuing certificate as Secret does not exist

openstack

cert-manager-certificaterequests-issuer-vault

cinder-internal-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificates-issuing

cinder-internal-svc

Issuing

The certificate has been successfully issued

openstack

cert-manager-certificaterequests-issuer-ca

cinder-internal-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

kubelet

cinder-c920a-volume-lvm-iscsi-0

Pulling

Pulling image "quay.io/podified-antelope-centos9/openstack-cinder-volume@sha256:7423f71c91f5a1d0aec9dcf0993db6e2495b520b5e5bbcf1615b9ac9759c0a58"

openstack

cert-manager-certificaterequests-issuer-selfsigned

cinder-internal-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

multus

cinder-c920a-volume-lvm-iscsi-0

AddedInterface

Add eth0 [10.128.0.223/23] from ovn-kubernetes

openstack

kubelet

dnsmasq-dns-7bcb7479f5-z74kg

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:fbb5be29e9e4fa11f0743e7f74f2e80dcc7445d24770709ea0e038147f752c51" already present on machine

openstack

kubelet

dnsmasq-dns-7bcb7479f5-z74kg

Started

Started container init

openstack

multus

cinder-c920a-api-0

AddedInterface

Add eth0 [10.128.0.228/23] from ovn-kubernetes

openstack

kubelet

neutron-555d58bf7b-nplsv

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:fbb5be29e9e4fa11f0743e7f74f2e80dcc7445d24770709ea0e038147f752c51" already present on machine

openstack

kubelet

neutron-555d58bf7b-nplsv

Created

Created container: neutron-api

openstack

multus

cinder-c920a-scheduler-0

AddedInterface

Add eth0 [10.128.0.224/23] from ovn-kubernetes

openstack

cert-manager-certificaterequests-approver

cinder-internal-svc-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack

kubelet

cinder-c920a-api-0

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:574a17f0877c175128a764f2b37fc02456649c8514689125718ce6ca974bfb6b" already present on machine

openstack

multus

neutron-555d58bf7b-nplsv

AddedInterface

Add internalapi [172.17.0.32/24] from openstack/internalapi

openstack

cert-manager-certificaterequests-issuer-ca

cinder-internal-svc-1

CertificateIssued

Certificate fetched from issuer successfully

openstack

cert-manager-certificates-trigger

cinder-internal-svc

Issuing

Issuing certificate as Secret does not exist

openstack

cert-manager-certificates-key-manager

cinder-internal-svc

Generated

Stored new private key in temporary Secret resource "cinder-internal-svc-7swmm"

openstack

multus

neutron-555d58bf7b-nplsv

AddedInterface

Add eth0 [10.128.0.226/23] from ovn-kubernetes

openstack

multus

dnsmasq-dns-7bcb7479f5-z74kg

AddedInterface

Add eth0 [10.128.0.222/23] from ovn-kubernetes

openstack

cert-manager-certificaterequests-issuer-ca

cinder-public-svc-1

CertificateIssued

Certificate fetched from issuer successfully

openstack

cert-manager-certificates-request-manager

cinder-public-svc

Requested

Created new CertificateRequest resource "cinder-public-svc-1"

openstack

kubelet

neutron-555d58bf7b-nplsv

Started

Started container neutron-httpd

openstack

kubelet

neutron-555d58bf7b-nplsv

Created

Created container: neutron-httpd

openstack

kubelet

cinder-c920a-api-0

Created

Created container: cinder-c920a-api-log

openstack

kubelet

cinder-c920a-api-0

Started

Started container cinder-c920a-api-log

openstack

kubelet

neutron-555d58bf7b-nplsv

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:fbb5be29e9e4fa11f0743e7f74f2e80dcc7445d24770709ea0e038147f752c51" already present on machine

openstack

kubelet

cinder-c920a-api-0

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:574a17f0877c175128a764f2b37fc02456649c8514689125718ce6ca974bfb6b" already present on machine

openstack

kubelet

cinder-c920a-volume-lvm-iscsi-0

Pulled

Successfully pulled image "quay.io/podified-antelope-centos9/openstack-cinder-volume@sha256:7423f71c91f5a1d0aec9dcf0993db6e2495b520b5e5bbcf1615b9ac9759c0a58" in 953ms (953ms including waiting). Image size: 1084059606 bytes.

openstack

kubelet

cinder-c920a-volume-lvm-iscsi-0

Created

Created container: cinder-volume

openstack

kubelet

cinder-c920a-volume-lvm-iscsi-0

Started

Started container cinder-volume

openstack

cert-manager-certificaterequests-issuer-selfsigned

cinder-public-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-ca

cinder-public-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-venafi

cinder-public-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

kubelet

neutron-555d58bf7b-nplsv

Started

Started container neutron-api

openstack

multus

cinder-c920a-backup-0

AddedInterface

Add eth0 [10.128.0.225/23] from ovn-kubernetes

openstack

multus

cinder-c920a-backup-0

AddedInterface

Add storage [172.18.0.32/24] from openstack/storage

openstack

kubelet

cinder-c920a-backup-0

Pulling

Pulling image "quay.io/podified-antelope-centos9/openstack-cinder-backup@sha256:2781f3bed351ce4c77a235e2381576637203459384fd93e05584a0013b3fe93e"

openstack

kubelet

cinder-c920a-volume-lvm-iscsi-0

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-cinder-volume@sha256:7423f71c91f5a1d0aec9dcf0993db6e2495b520b5e5bbcf1615b9ac9759c0a58" already present on machine

openstack

kubelet

dnsmasq-dns-7cd5f68b57-cbttm

Created

Created container: dnsmasq-dns

openstack

kubelet

dnsmasq-dns-7cd5f68b57-cbttm

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:fbb5be29e9e4fa11f0743e7f74f2e80dcc7445d24770709ea0e038147f752c51" already present on machine

openstack

cert-manager-certificaterequests-issuer-acme

cinder-public-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-approver

cinder-public-svc-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack

kubelet

dnsmasq-dns-7cd5f68b57-cbttm

Started

Started container init

openstack

kubelet

dnsmasq-dns-7cd5f68b57-cbttm

Created

Created container: init

openstack

kubelet

cinder-c920a-scheduler-0

Pulled

Successfully pulled image "quay.io/podified-antelope-centos9/openstack-cinder-scheduler@sha256:9e7c747eeeefb391dc6dedaaac57fa694c4d08b991c54bb99aa6de77451e792f" in 1.038s (1.038s including waiting). Image size: 1083121304 bytes.

openstack

kubelet

dnsmasq-dns-7cd5f68b57-cbttm

Started

Started container dnsmasq-dns

openstack

kubelet

dnsmasq-dns-7cd5f68b57-cbttm

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:fbb5be29e9e4fa11f0743e7f74f2e80dcc7445d24770709ea0e038147f752c51" already present on machine

openstack

cert-manager-certificates-issuing

cinder-public-svc

Issuing

The certificate has been successfully issued

openstack

multus

dnsmasq-dns-7cd5f68b57-cbttm

AddedInterface

Add eth0 [10.128.0.227/23] from ovn-kubernetes

openstack

cert-manager-certificates-trigger

cinder-public-route

Issuing

Issuing certificate as Secret does not exist

openstack

cert-manager-certificates-key-manager

cinder-public-svc

Generated

Stored new private key in temporary Secret resource "cinder-public-svc-kfqzr"

openstack

cert-manager-certificaterequests-issuer-vault

cinder-public-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

kubelet

cinder-c920a-api-0

Created

Created container: cinder-api

openstack

kubelet

cinder-c920a-backup-0

Created

Created container: cinder-backup

openstack

cert-manager-certificaterequests-issuer-acme

cinder-public-route-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-vault

cinder-public-route-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-ca

cinder-public-route-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

kubelet

cinder-c920a-backup-0

Started

Started container cinder-backup

openstack

cert-manager-certificaterequests-approver

cinder-public-route-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack

cert-manager-certificaterequests-issuer-ca

cinder-public-route-1

CertificateIssued

Certificate fetched from issuer successfully

openstack

cert-manager-certificaterequests-issuer-selfsigned

cinder-public-route-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

kubelet

cinder-c920a-volume-lvm-iscsi-0

Created

Created container: probe

openstack

cert-manager-certificates-key-manager

cinder-public-route

Generated

Stored new private key in temporary Secret resource "cinder-public-route-vnnlb"

openstack

kubelet

cinder-c920a-volume-lvm-iscsi-0

Started

Started container probe

openstack

kubelet

cinder-c920a-scheduler-0

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-cinder-scheduler@sha256:9e7c747eeeefb391dc6dedaaac57fa694c4d08b991c54bb99aa6de77451e792f" already present on machine

openstack

cert-manager-certificates-request-manager

cinder-public-route

Requested

Created new CertificateRequest resource "cinder-public-route-1"

openstack

cert-manager-certificates-issuing

cinder-public-route

Issuing

The certificate has been successfully issued

openstack

kubelet

cinder-c920a-scheduler-0

Started

Started container cinder-scheduler

openstack

kubelet

cinder-c920a-api-0

Started

Started container cinder-api

openstack

kubelet

cinder-c920a-backup-0

Pulled

Successfully pulled image "quay.io/podified-antelope-centos9/openstack-cinder-backup@sha256:2781f3bed351ce4c77a235e2381576637203459384fd93e05584a0013b3fe93e" in 984ms (984ms including waiting). Image size: 1083126547 bytes.

openstack

kubelet

cinder-c920a-backup-0

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-cinder-backup@sha256:2781f3bed351ce4c77a235e2381576637203459384fd93e05584a0013b3fe93e" already present on machine

openstack

cert-manager-certificaterequests-issuer-venafi

cinder-public-route-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

kubelet

cinder-c920a-backup-0

Created

Created container: probe

openstack

kubelet

cinder-c920a-backup-0

Started

Started container probe

openstack

kubelet

cinder-c920a-scheduler-0

Created

Created container: cinder-scheduler

openstack

statefulset-controller

cinder-c920a-api

SuccessfulDelete

delete Pod cinder-c920a-api-0 in StatefulSet cinder-c920a-api successful

openstack

job-controller

edpm-b-provisionserver-checksum-discovery

SuccessfulCreate

Created pod: edpm-b-provisionserver-checksum-discovery-mf2jz

openstack

kubelet

cinder-c920a-scheduler-0

Started

Started container probe

openstack

kubelet

cinder-c920a-scheduler-0

Created

Created container: probe

openstack

cert-manager-certificates-trigger

neutron-internal-svc

Issuing

Issuing certificate as Secret does not exist

openstack

cert-manager-certificaterequests-issuer-selfsigned

neutron-internal-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

kubelet

cinder-c920a-api-0

Killing

Stopping container cinder-api

openstack

cert-manager-certificates-key-manager

neutron-internal-svc

Generated

Stored new private key in temporary Secret resource "neutron-internal-svc-kzkp9"

openstack

kubelet

edpm-b-provisionserver-checksum-discovery-mf2jz

Pulled

Container image "quay.io/podified-antelope-centos9/edpm-hardened-uefi@sha256:894ce79b38510973ca610423cc34a7383b7761b6ceb47d18637daffaa93336f7" already present on machine

openstack

cert-manager-certificaterequests-approver

neutron-internal-svc-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack

cert-manager-certificaterequests-issuer-vault

neutron-internal-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-acme

neutron-internal-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-ca

neutron-internal-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificates-issuing

neutron-internal-svc

Issuing

The certificate has been successfully issued

openstack

multus

edpm-b-provisionserver-checksum-discovery-mf2jz

AddedInterface

Add eth0 [10.128.0.229/23] from ovn-kubernetes

openstack

cert-manager-certificaterequests-issuer-venafi

neutron-internal-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-ca

neutron-internal-svc-1

CertificateIssued

Certificate fetched from issuer successfully

openstack

kubelet

cinder-c920a-api-0

Killing

Stopping container cinder-c920a-api-log

openstack

cert-manager-certificates-request-manager

neutron-internal-svc

Requested

Created new CertificateRequest resource "neutron-internal-svc-1"

openstack

cert-manager-certificates-trigger

neutron-public-svc

Issuing

Issuing certificate as Secret does not exist

openstack

cert-manager-certificaterequests-issuer-ca

neutron-public-svc-1

CertificateIssued

Certificate fetched from issuer successfully

openstack

kubelet

edpm-b-provisionserver-checksum-discovery-mf2jz

Created

Created container: init

openstack

kubelet

edpm-b-provisionserver-checksum-discovery-mf2jz

Started

Started container init

openstack

cert-manager-certificates-request-manager

neutron-public-svc

Requested

Created new CertificateRequest resource "neutron-public-svc-1"

openstack

cert-manager-certificates-key-manager

neutron-public-svc

Generated

Stored new private key in temporary Secret resource "neutron-public-svc-nhfqv"

openstack

cert-manager-certificaterequests-approver

neutron-public-svc-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack

cert-manager-certificaterequests-issuer-vault

neutron-public-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-venafi

neutron-public-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-selfsigned

neutron-public-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-acme

neutron-public-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-ca

neutron-public-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificates-issuing

neutron-public-svc

Issuing

The certificate has been successfully issued
(x2)

openstack

statefulset-controller

cinder-c920a-api

SuccessfulCreate

create Pod cinder-c920a-api-0 in StatefulSet cinder-c920a-api successful

openstack

cert-manager-certificaterequests-issuer-ca

neutron-public-route-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-vault

neutron-public-route-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-venafi

neutron-public-route-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-selfsigned

neutron-public-route-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-acme

neutron-public-route-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificates-trigger

neutron-public-route

Issuing

Issuing certificate as Secret does not exist

openstack

cert-manager-certificaterequests-approver

neutron-public-route-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack

cert-manager-certificates-issuing

neutron-public-route

Issuing

The certificate has been successfully issued

openstack

cert-manager-certificates-request-manager

neutron-public-route

Requested

Created new CertificateRequest resource "neutron-public-route-1"

openstack

cert-manager-certificates-key-manager

neutron-public-route

Generated

Stored new private key in temporary Secret resource "neutron-public-route-ftfdt"

openstack

cert-manager-certificaterequests-issuer-ca

neutron-public-route-1

CertificateIssued

Certificate fetched from issuer successfully

openstack

job-controller

edpm-a-provisionserver-checksum-discovery

SuccessfulCreate

Created pod: edpm-a-provisionserver-checksum-discovery-49xmz

openstack

kubelet

cinder-c920a-volume-lvm-iscsi-0

Unhealthy

Startup probe failed: HTTP probe failed with statuscode: 500

openstack

kubelet

dnsmasq-dns-74db7c7d5f-t4vkw

Killing

Stopping container dnsmasq-dns

openstack

kubelet

edpm-a-provisionserver-checksum-discovery-49xmz

Pulled

Container image "quay.io/podified-antelope-centos9/edpm-hardened-uefi@sha256:894ce79b38510973ca610423cc34a7383b7761b6ceb47d18637daffaa93336f7" already present on machine

openstack

kubelet

edpm-a-provisionserver-checksum-discovery-49xmz

Started

Started container init

openstack

statefulset-controller

cinder-c920a-scheduler

SuccessfulDelete

delete Pod cinder-c920a-scheduler-0 in StatefulSet cinder-c920a-scheduler successful

openstack

kubelet

cinder-c920a-scheduler-0

Killing

Stopping container cinder-scheduler

openstack

kubelet

cinder-c920a-api-0

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:574a17f0877c175128a764f2b37fc02456649c8514689125718ce6ca974bfb6b" already present on machine

openstack

deployment-controller

neutron

ScalingReplicaSet

Scaled up replica set neutron-6bc775ffc9 to 1

openstack

kubelet

cinder-c920a-backup-0

Killing

Stopping container probe

openstack

kubelet

cinder-c920a-scheduler-0

Killing

Stopping container probe

openstack

replicaset-controller

neutron-6bc775ffc9

SuccessfulCreate

Created pod: neutron-6bc775ffc9-jcrc5

openstack

kubelet

edpm-a-provisionserver-checksum-discovery-49xmz

Created

Created container: init

openstack

multus

cinder-c920a-api-0

AddedInterface

Add eth0 [10.128.0.230/23] from ovn-kubernetes

openstack

multus

edpm-a-provisionserver-checksum-discovery-49xmz

AddedInterface

Add eth0 [10.128.0.231/23] from ovn-kubernetes

openstack

replicaset-controller

dnsmasq-dns-74db7c7d5f

SuccessfulDelete

Deleted pod: dnsmasq-dns-74db7c7d5f-t4vkw

openstack

statefulset-controller

cinder-c920a-backup

SuccessfulDelete

delete Pod cinder-c920a-backup-0 in StatefulSet cinder-c920a-backup successful

openstack

kubelet

cinder-c920a-backup-0

Killing

Stopping container cinder-backup

openstack

kubelet

cinder-c920a-api-0

Started

Started container cinder-c920a-api-log

openstack

kubelet

cinder-c920a-api-0

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:574a17f0877c175128a764f2b37fc02456649c8514689125718ce6ca974bfb6b" already present on machine

openstack

multus

neutron-6bc775ffc9-jcrc5

AddedInterface

Add eth0 [10.128.0.232/23] from ovn-kubernetes

openstack

kubelet

neutron-6bc775ffc9-jcrc5

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:fbb5be29e9e4fa11f0743e7f74f2e80dcc7445d24770709ea0e038147f752c51" already present on machine

openstack

kubelet

cinder-c920a-api-0

Created

Created container: cinder-c920a-api-log

openstack

multus

neutron-6bc775ffc9-jcrc5

AddedInterface

Add internalapi [172.17.0.33/24] from openstack/internalapi

openstack

kubelet

neutron-6bc775ffc9-jcrc5

Started

Started container neutron-api

openstack

kubelet

neutron-6bc775ffc9-jcrc5

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:fbb5be29e9e4fa11f0743e7f74f2e80dcc7445d24770709ea0e038147f752c51" already present on machine

openstack

kubelet

neutron-6bc775ffc9-jcrc5

Created

Created container: neutron-httpd

openstack

kubelet

neutron-6bc775ffc9-jcrc5

Created

Created container: neutron-api

openstack

kubelet

neutron-6bc775ffc9-jcrc5

Started

Started container neutron-httpd

openstack

kubelet

cinder-c920a-api-0

Started

Started container cinder-api

openstack

kubelet

cinder-c920a-api-0

Created

Created container: cinder-api
(x2)

openstack

statefulset-controller

cinder-c920a-backup

SuccessfulCreate

create Pod cinder-c920a-backup-0 in StatefulSet cinder-c920a-backup successful

openstack

statefulset-controller

cinder-c920a-volume-lvm-iscsi

SuccessfulDelete

delete Pod cinder-c920a-volume-lvm-iscsi-0 in StatefulSet cinder-c920a-volume-lvm-iscsi successful

openstack

kubelet

edpm-b-provisionserver-checksum-discovery-mf2jz

Pulling

Pulling image "quay.io/openstack-k8s-operators/openstack-baremetal-operator-agent@sha256:55a1ea09e9211b1719a2377854012765e767931e1d437fc1b5d4722863dcb4fb"
(x2)

openstack

statefulset-controller

cinder-c920a-scheduler

SuccessfulCreate

create Pod cinder-c920a-scheduler-0 in StatefulSet cinder-c920a-scheduler successful

openstack

kubelet

cinder-c920a-volume-lvm-iscsi-0

Killing

Stopping container cinder-volume

openstack

kubelet

cinder-c920a-backup-0

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-cinder-backup@sha256:2781f3bed351ce4c77a235e2381576637203459384fd93e05584a0013b3fe93e" already present on machine

openstack

multus

cinder-c920a-backup-0

AddedInterface

Add storage [172.18.0.32/24] from openstack/storage

openstack

kubelet

dnsmasq-dns-74db7c7d5f-t4vkw

Unhealthy

Readiness probe failed: dial tcp 10.128.0.214:5353: i/o timeout

openstack

kubelet

cinder-c920a-backup-0

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-cinder-backup@sha256:2781f3bed351ce4c77a235e2381576637203459384fd93e05584a0013b3fe93e" already present on machine

openstack

kubelet

cinder-c920a-backup-0

Created

Created container: cinder-backup

openstack

kubelet

cinder-c920a-backup-0

Started

Started container cinder-backup

openstack

multus

cinder-c920a-backup-0

AddedInterface

Add eth0 [10.128.0.233/23] from ovn-kubernetes

openstack

multus

cinder-c920a-scheduler-0

AddedInterface

Add eth0 [10.128.0.234/23] from ovn-kubernetes

openstack

kubelet

cinder-c920a-scheduler-0

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-cinder-scheduler@sha256:9e7c747eeeefb391dc6dedaaac57fa694c4d08b991c54bb99aa6de77451e792f" already present on machine

openstack

kubelet

cinder-c920a-volume-lvm-iscsi-0

Killing

Stopping container probe

openstack

kubelet

cinder-c920a-scheduler-0

Started

Started container cinder-scheduler

openstack

kubelet

cinder-c920a-scheduler-0

Created

Created container: cinder-scheduler

openstack

kubelet

cinder-c920a-backup-0

Started

Started container probe

openstack

kubelet

cinder-c920a-backup-0

Created

Created container: probe

openstack

kubelet

edpm-b-provisionserver-checksum-discovery-mf2jz

Created

Created container: edpm-b-provisionserver-checksum-discovery

openstack

kubelet

cinder-c920a-scheduler-0

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-cinder-scheduler@sha256:9e7c747eeeefb391dc6dedaaac57fa694c4d08b991c54bb99aa6de77451e792f" already present on machine

openstack

kubelet

edpm-b-provisionserver-checksum-discovery-mf2jz

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/openstack-baremetal-operator-agent@sha256:55a1ea09e9211b1719a2377854012765e767931e1d437fc1b5d4722863dcb4fb" in 3.245s (3.245s including waiting). Image size: 159969701 bytes.

openstack

kubelet

cinder-c920a-scheduler-0

Created

Created container: probe

openstack

kubelet

cinder-c920a-scheduler-0

Started

Started container probe

openstack

kubelet

edpm-b-provisionserver-checksum-discovery-mf2jz

Started

Started container edpm-b-provisionserver-checksum-discovery
(x2)

openstack

statefulset-controller

cinder-c920a-volume-lvm-iscsi

SuccessfulCreate

create Pod cinder-c920a-volume-lvm-iscsi-0 in StatefulSet cinder-c920a-volume-lvm-iscsi successful

openstack

kubelet

edpm-a-provisionserver-checksum-discovery-49xmz

Pulled

Container image "quay.io/openstack-k8s-operators/openstack-baremetal-operator-agent@sha256:55a1ea09e9211b1719a2377854012765e767931e1d437fc1b5d4722863dcb4fb" already present on machine

openstack

kubelet

edpm-a-provisionserver-checksum-discovery-49xmz

Created

Created container: edpm-a-provisionserver-checksum-discovery

openstack

kubelet

edpm-a-provisionserver-checksum-discovery-49xmz

Started

Started container edpm-a-provisionserver-checksum-discovery

openstack

multus

cinder-c920a-volume-lvm-iscsi-0

AddedInterface

Add eth0 [10.128.0.235/23] from ovn-kubernetes

openstack

kubelet

cinder-c920a-volume-lvm-iscsi-0

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-cinder-volume@sha256:7423f71c91f5a1d0aec9dcf0993db6e2495b520b5e5bbcf1615b9ac9759c0a58" already present on machine

openstack

kubelet

cinder-c920a-volume-lvm-iscsi-0

Created

Created container: probe

openstack

replicaset-controller

placement-7754cdbcb6

SuccessfulCreate

Created pod: placement-7754cdbcb6-lblgz

openstack

kubelet

cinder-c920a-volume-lvm-iscsi-0

Started

Started container probe

openstack

kubelet

cinder-c920a-volume-lvm-iscsi-0

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-cinder-volume@sha256:7423f71c91f5a1d0aec9dcf0993db6e2495b520b5e5bbcf1615b9ac9759c0a58" already present on machine

openstack

deployment-controller

placement

ScalingReplicaSet

Scaled up replica set placement-7754cdbcb6 to 1

openstack

kubelet

cinder-c920a-volume-lvm-iscsi-0

Started

Started container cinder-volume

openstack

kubelet

cinder-c920a-volume-lvm-iscsi-0

Created

Created container: cinder-volume

openstack

job-controller

edpm-b-provisionserver-checksum-discovery

Completed

Job completed

openstack

multus

placement-7754cdbcb6-lblgz

AddedInterface

Add eth0 [10.128.0.236/23] from ovn-kubernetes

openstack

kubelet

placement-7754cdbcb6-lblgz

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-placement-api@sha256:b8a5d052890fb9cefa333baf10b607add227ed5d79aa108b576a97b21e89327a" already present on machine

openstack

kubelet

placement-7754cdbcb6-lblgz

Created

Created container: placement-log

openstack

kubelet

placement-7754cdbcb6-lblgz

Started

Started container placement-log

openstack

kubelet

placement-7754cdbcb6-lblgz

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-placement-api@sha256:b8a5d052890fb9cefa333baf10b607add227ed5d79aa108b576a97b21e89327a" already present on machine

openstack

metallb-speaker

cinder-internal

nodeAssigned

announcing from node "master-0" with protocol "layer2"

openstack

kubelet

placement-7754cdbcb6-lblgz

Created

Created container: placement-api

openstack

kubelet

placement-7754cdbcb6-lblgz

Started

Started container placement-api

openstack

job-controller

edpm-a-provisionserver-checksum-discovery

Completed

Job completed

openstack

metallb-speaker

keystone-internal

nodeAssigned

announcing from node "master-0" with protocol "layer2"

openstack

multus

openstackclient

AddedInterface

Add eth0 [10.128.0.237/23] from ovn-kubernetes

openstack

kubelet

openstackclient

Pulling

Pulling image "quay.io/podified-antelope-centos9/openstack-openstackclient@sha256:d765b589a5f7bc8573b3b192ed265654699012e6342cc4829bd8ea65a7b239a5"

openstack

deployment-controller

swift-proxy

ScalingReplicaSet

Scaled up replica set swift-proxy-84fcf8cbb to 1

openstack

multus

swift-proxy-84fcf8cbb-q9wtd

AddedInterface

Add eth0 [10.128.0.238/23] from ovn-kubernetes

openstack

replicaset-controller

swift-proxy-84fcf8cbb

SuccessfulCreate

Created pod: swift-proxy-84fcf8cbb-q9wtd

openstack

kubelet

neutron-555d58bf7b-nplsv

Killing

Stopping container neutron-api

openstack

kubelet

neutron-555d58bf7b-nplsv

Killing

Stopping container neutron-httpd

openstack

replicaset-controller

neutron-555d58bf7b

SuccessfulDelete

Deleted pod: neutron-555d58bf7b-nplsv

openstack

deployment-controller

neutron

ScalingReplicaSet

Scaled down replica set neutron-555d58bf7b to 0 from 1

openstack

kubelet

swift-proxy-84fcf8cbb-q9wtd

Started

Started container proxy-server

openstack

kubelet

swift-proxy-84fcf8cbb-q9wtd

Created

Created container: proxy-server

openstack

kubelet

swift-proxy-84fcf8cbb-q9wtd

Killing

Container proxy-httpd failed liveness probe, will be restarted
(x3)

openstack

kubelet

swift-proxy-84fcf8cbb-q9wtd

Unhealthy

Liveness probe failed: HTTP probe failed with statuscode: 503
(x2)

openstack

kubelet

swift-proxy-84fcf8cbb-q9wtd

Unhealthy

Readiness probe failed: HTTP probe failed with statuscode: 503
(x2)

openstack

kubelet

swift-proxy-84fcf8cbb-q9wtd

Unhealthy

Readiness probe failed: HTTP probe failed with statuscode: 503
(x2)

openstack

kubelet

swift-proxy-84fcf8cbb-q9wtd

Unhealthy

Liveness probe failed: HTTP probe failed with statuscode: 503
(x2)

openstack

kubelet

swift-proxy-84fcf8cbb-q9wtd

Started

Started container proxy-httpd

openstack

kubelet

glance-27086-default-external-api-0

Killing

Stopping container glance-log

openstack

kubelet

openstackclient

Pulled

Successfully pulled image "quay.io/podified-antelope-centos9/openstack-openstackclient@sha256:d765b589a5f7bc8573b3b192ed265654699012e6342cc4829bd8ea65a7b239a5" in 16.557s (16.557s including waiting). Image size: 594372457 bytes.
(x2)

openstack

kubelet

swift-proxy-84fcf8cbb-q9wtd

Created

Created container: proxy-httpd

openstack

kubelet

openstackclient

Created

Created container: openstackclient

openstack

kubelet

openstackclient

Started

Started container openstackclient
(x2)

openstack

statefulset-controller

glance-27086-default-external-api

SuccessfulDelete

delete Pod glance-27086-default-external-api-0 in StatefulSet glance-27086-default-external-api successful

openstack

kubelet

glance-27086-default-external-api-0

Killing

Stopping container glance-httpd
(x2)

openstack

kubelet

swift-proxy-84fcf8cbb-q9wtd

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-swift-proxy-server@sha256:6c3eb966650a7a98feb4ddb31e1bdba1095b0c62e349196aca6a423681d7e5fb" already present on machine

openstack

kubelet

swift-proxy-84fcf8cbb-q9wtd

Unhealthy

Liveness probe failed: Get "https://10.128.0.238:8080/healthcheck": dial tcp 10.128.0.238:8080: connect: connection refused
(x2)

openstack

kubelet

swift-proxy-84fcf8cbb-q9wtd

Unhealthy

Readiness probe failed: Get "https://10.128.0.238:8080/healthcheck": dial tcp 10.128.0.238:8080: connect: connection refused
(x4)

openstack

metallb-speaker

neutron-internal

nodeAssigned

announcing from node "master-0" with protocol "layer2"

openstack

kubelet

swift-proxy-84fcf8cbb-q9wtd

Unhealthy

Readiness probe failed: Get "https://10.128.0.238:8080/healthcheck": dial tcp 10.128.0.238:8080: connect: connection refused

openstack

kubelet

swift-proxy-84fcf8cbb-q9wtd

Killing

Container proxy-server failed liveness probe, will be restarted
(x2)

openstack

statefulset-controller

glance-27086-default-internal-api

SuccessfulDelete

delete Pod glance-27086-default-internal-api-0 in StatefulSet glance-27086-default-internal-api successful

openstack

kubelet

glance-27086-default-internal-api-0

Killing

Stopping container glance-httpd

openstack

kubelet

glance-27086-default-internal-api-0

Killing

Stopping container glance-log
(x2)

openstack

kubelet

swift-proxy-84fcf8cbb-q9wtd

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-swift-proxy-server@sha256:6c3eb966650a7a98feb4ddb31e1bdba1095b0c62e349196aca6a423681d7e5fb" already present on machine

openstack

deployment-controller

placement

ScalingReplicaSet

Scaled down replica set placement-8b47bf878 to 0 from 1

openstack

job-controller

nova-cell1-db-create

SuccessfulCreate

Created pod: nova-cell1-db-create-fh7jr

openstack

job-controller

nova-cell1-3995-account-create-update

SuccessfulCreate

Created pod: nova-cell1-3995-account-create-update-7pfxc

openstack

job-controller

nova-api-db-create

SuccessfulCreate

Created pod: nova-api-db-create-hdkgz

openstack

kubelet

placement-8b47bf878-2ftwn

Killing

Stopping container placement-log

openstack

job-controller

nova-cell0-db-create

SuccessfulCreate

Created pod: nova-cell0-db-create-th2rs

openstack

job-controller

nova-api-af59-account-create-update

SuccessfulCreate

Created pod: nova-api-af59-account-create-update-tk2tl

openstack

kubelet

placement-8b47bf878-2ftwn

Killing

Stopping container placement-api

openstack

job-controller

nova-cell0-11a2-account-create-update

SuccessfulCreate

Created pod: nova-cell0-11a2-account-create-update-gkhvg

openstack

replicaset-controller

placement-8b47bf878

SuccessfulDelete

Deleted pod: placement-8b47bf878-2ftwn

openstack

multus

nova-api-db-create-hdkgz

AddedInterface

Add eth0 [10.128.0.239/23] from ovn-kubernetes

openstack

multus

nova-api-af59-account-create-update-tk2tl

AddedInterface

Add eth0 [10.128.0.241/23] from ovn-kubernetes

openstack

multus

nova-cell0-db-create-th2rs

AddedInterface

Add eth0 [10.128.0.240/23] from ovn-kubernetes

openstack

multus

nova-cell0-11a2-account-create-update-gkhvg

AddedInterface

Add eth0 [10.128.0.243/23] from ovn-kubernetes

openstack

multus

nova-cell1-db-create-fh7jr

AddedInterface

Add eth0 [10.128.0.242/23] from ovn-kubernetes

openstack

kubelet

nova-api-db-create-hdkgz

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-mariadb@sha256:4caef2b55e01b9a7ee88a22bc69db1893521a91d95c7ad4c8e593f14f17a5f95" already present on machine

openstack

kubelet

nova-cell0-db-create-th2rs

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-mariadb@sha256:4caef2b55e01b9a7ee88a22bc69db1893521a91d95c7ad4c8e593f14f17a5f95" already present on machine

openstack

kubelet

nova-api-af59-account-create-update-tk2tl

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-mariadb@sha256:4caef2b55e01b9a7ee88a22bc69db1893521a91d95c7ad4c8e593f14f17a5f95" already present on machine

openstack

multus

nova-cell1-3995-account-create-update-7pfxc

AddedInterface

Add eth0 [10.128.0.244/23] from ovn-kubernetes

openstack

kubelet

nova-cell0-11a2-account-create-update-gkhvg

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-mariadb@sha256:4caef2b55e01b9a7ee88a22bc69db1893521a91d95c7ad4c8e593f14f17a5f95" already present on machine

openstack

kubelet

nova-cell1-db-create-fh7jr

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-mariadb@sha256:4caef2b55e01b9a7ee88a22bc69db1893521a91d95c7ad4c8e593f14f17a5f95" already present on machine

openstack

kubelet

nova-cell1-3995-account-create-update-7pfxc

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-mariadb@sha256:4caef2b55e01b9a7ee88a22bc69db1893521a91d95c7ad4c8e593f14f17a5f95" already present on machine

openstack

kubelet

nova-api-db-create-hdkgz

Created

Created container: mariadb-database-create

openstack

kubelet

nova-api-db-create-hdkgz

Started

Started container mariadb-database-create

openstack

kubelet

nova-cell1-db-create-fh7jr

Created

Created container: mariadb-database-create

openstack

kubelet

nova-api-af59-account-create-update-tk2tl

Started

Started container mariadb-account-create-update

openstack

kubelet

nova-api-af59-account-create-update-tk2tl

Created

Created container: mariadb-account-create-update

openstack

kubelet

nova-cell1-3995-account-create-update-7pfxc

Started

Started container mariadb-account-create-update

openstack

kubelet

nova-cell1-3995-account-create-update-7pfxc

Created

Created container: mariadb-account-create-update

openstack

kubelet

nova-cell0-11a2-account-create-update-gkhvg

Created

Created container: mariadb-account-create-update

openstack

kubelet

nova-cell0-11a2-account-create-update-gkhvg

Started

Started container mariadb-account-create-update

openstack

kubelet

nova-cell1-db-create-fh7jr

Started

Started container mariadb-database-create

openstack

kubelet

nova-cell0-db-create-th2rs

Started

Started container mariadb-database-create

openstack

kubelet

nova-cell0-db-create-th2rs

Created

Created container: mariadb-database-create
(x3)

openstack

statefulset-controller

glance-27086-default-external-api

SuccessfulCreate

create Pod glance-27086-default-external-api-0 in StatefulSet glance-27086-default-external-api successful
(x3)

openstack

statefulset-controller

glance-27086-default-internal-api

SuccessfulCreate

create Pod glance-27086-default-internal-api-0 in StatefulSet glance-27086-default-internal-api successful

openstack

multus

glance-27086-default-external-api-0

AddedInterface

Add eth0 [10.128.0.245/23] from ovn-kubernetes
(x5)

openstack

metallb-speaker

placement-internal

nodeAssigned

announcing from node "master-0" with protocol "layer2"

openstack

multus

glance-27086-default-external-api-0

AddedInterface

Add storage [172.18.0.30/24] from openstack/storage

openstack

kubelet

glance-27086-default-external-api-0

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-glance-api@sha256:dae5e39780d5a15eed030c7009f8e5317139d447558ac83f038497be594be120" already present on machine

openstack

job-controller

nova-cell0-11a2-account-create-update

Completed

Job completed

openstack

multus

glance-27086-default-internal-api-0

AddedInterface

Add eth0 [10.128.0.246/23] from ovn-kubernetes

openstack

job-controller

nova-cell1-db-create

Completed

Job completed

openstack

job-controller

nova-api-db-create

Completed

Job completed

openstack

kubelet

glance-27086-default-external-api-0

Created

Created container: glance-log

openstack

kubelet

glance-27086-default-external-api-0

Started

Started container glance-log

openstack

job-controller

nova-api-af59-account-create-update

Completed

Job completed

openstack

kubelet

glance-27086-default-external-api-0

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-glance-api@sha256:dae5e39780d5a15eed030c7009f8e5317139d447558ac83f038497be594be120" already present on machine

openstack

multus

glance-27086-default-internal-api-0

AddedInterface

Add storage [172.18.0.31/24] from openstack/storage

openstack

kubelet

glance-27086-default-internal-api-0

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-glance-api@sha256:dae5e39780d5a15eed030c7009f8e5317139d447558ac83f038497be594be120" already present on machine

openstack

metallb-speaker

swift-internal

nodeAssigned

announcing from node "master-0" with protocol "layer2"

openstack

job-controller

nova-cell0-db-create

Completed

Job completed

openstack

job-controller

nova-cell1-3995-account-create-update

Completed

Job completed

openstack

kubelet

glance-27086-default-internal-api-0

Created

Created container: glance-log

openstack

kubelet

glance-27086-default-external-api-0

Created

Created container: glance-httpd

openstack

kubelet

glance-27086-default-external-api-0

Started

Started container glance-httpd

openstack

kubelet

glance-27086-default-internal-api-0

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-glance-api@sha256:dae5e39780d5a15eed030c7009f8e5317139d447558ac83f038497be594be120" already present on machine

openstack

kubelet

glance-27086-default-internal-api-0

Started

Started container glance-log

openstack

kubelet

glance-27086-default-internal-api-0

Created

Created container: glance-httpd

openstack

kubelet

glance-27086-default-internal-api-0

Started

Started container glance-httpd

openstack

job-controller

nova-cell0-conductor-db-sync

SuccessfulCreate

Created pod: nova-cell0-conductor-db-sync-zzsdf

openstack

multus

nova-cell0-conductor-db-sync-zzsdf

AddedInterface

Add eth0 [10.128.0.247/23] from ovn-kubernetes

openstack

kubelet

nova-cell0-conductor-db-sync-zzsdf

Pulling

Pulling image "quay.io/podified-antelope-centos9/openstack-nova-conductor@sha256:d347d48e9a8ae4136dd99c5222480ceccb2819beaf80b11048644d4acf0a4305"

openstack

kubelet

nova-cell0-conductor-db-sync-zzsdf

Pulled

Successfully pulled image "quay.io/podified-antelope-centos9/openstack-nova-conductor@sha256:d347d48e9a8ae4136dd99c5222480ceccb2819beaf80b11048644d4acf0a4305" in 11.378s (11.378s including waiting). Image size: 667925477 bytes.

openstack

kubelet

nova-cell0-conductor-db-sync-zzsdf

Created

Created container: nova-cell0-conductor-db-sync

openstack

kubelet

nova-cell0-conductor-db-sync-zzsdf

Started

Started container nova-cell0-conductor-db-sync
(x3)

openstack

metallb-speaker

glance-default-internal

nodeAssigned

announcing from node "master-0" with protocol "layer2"

openstack

job-controller

nova-cell0-conductor-db-sync

Completed

Job completed

openstack

statefulset-controller

nova-cell0-conductor

SuccessfulCreate

create Pod nova-cell0-conductor-0 in StatefulSet nova-cell0-conductor successful

openstack

kubelet

nova-cell0-conductor-0

Started

Started container nova-cell0-conductor-conductor

openstack

kubelet

nova-cell0-conductor-0

Created

Created container: nova-cell0-conductor-conductor

openstack

kubelet

nova-cell0-conductor-0

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-nova-conductor@sha256:d347d48e9a8ae4136dd99c5222480ceccb2819beaf80b11048644d4acf0a4305" already present on machine

openstack

multus

nova-cell0-conductor-0

AddedInterface

Add eth0 [10.128.0.248/23] from ovn-kubernetes

openstack

job-controller

nova-cell0-cell-mapping

SuccessfulCreate

Created pod: nova-cell0-cell-mapping-czjjp

openstack

multus

nova-cell0-cell-mapping-czjjp

AddedInterface

Add eth0 [10.128.0.249/23] from ovn-kubernetes

openstack

statefulset-controller

nova-metadata

SuccessfulCreate

create Pod nova-metadata-0 in StatefulSet nova-metadata successful

openstack

kubelet

nova-cell0-cell-mapping-czjjp

Started

Started container nova-manage

openstack

metallb-controller

nova-metadata-internal

IPAllocated

Assigned IP ["172.17.0.80"]

openstack

statefulset-controller

nova-cell1-novncproxy

SuccessfulCreate

create Pod nova-cell1-novncproxy-0 in StatefulSet nova-cell1-novncproxy successful

openstack

statefulset-controller

nova-api

SuccessfulCreate

create Pod nova-api-0 in StatefulSet nova-api successful

openstack

job-controller

nova-cell1-conductor-db-sync

SuccessfulCreate

Created pod: nova-cell1-conductor-db-sync-bcpbl

openstack

kubelet

nova-cell0-cell-mapping-czjjp

Created

Created container: nova-manage

openstack

statefulset-controller

nova-scheduler

SuccessfulCreate

create Pod nova-scheduler-0 in StatefulSet nova-scheduler successful

openstack

kubelet

nova-cell0-cell-mapping-czjjp

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-nova-conductor@sha256:d347d48e9a8ae4136dd99c5222480ceccb2819beaf80b11048644d4acf0a4305" already present on machine

openstack

replicaset-controller

dnsmasq-dns-75597d7cbf

SuccessfulCreate

Created pod: dnsmasq-dns-75597d7cbf-d9w8g

openstack

multus

nova-cell1-conductor-db-sync-bcpbl

AddedInterface

Add eth0 [10.128.0.250/23] from ovn-kubernetes

openstack

kubelet

nova-cell1-conductor-db-sync-bcpbl

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-nova-conductor@sha256:d347d48e9a8ae4136dd99c5222480ceccb2819beaf80b11048644d4acf0a4305" already present on machine

openstack

multus

nova-cell1-novncproxy-0

AddedInterface

Add eth0 [10.128.0.252/23] from ovn-kubernetes
(x2)

openstack

metallb-controller

nova-metadata-internal

deprecatedAnnotation

Service uses deprecated annotation metallb.universe.tf/allow-shared-ip

openstack

multus

nova-api-0

AddedInterface

Add eth0 [10.128.0.251/23] from ovn-kubernetes

openstack

kubelet

nova-api-0

Pulling

Pulling image "quay.io/podified-antelope-centos9/openstack-nova-api@sha256:faf711e1e5fa2ad74a73d3dfffd88f6312fb045cb69e9b7e6331558784163d16"

openstack

multus

nova-metadata-0

AddedInterface

Add eth0 [10.128.0.254/23] from ovn-kubernetes
(x2)

openstack

metallb-controller

nova-metadata-internal

deprecatedAnnotation

Service uses deprecated annotation metallb.universe.tf/loadBalancerIPs
(x2)

openstack

metallb-controller

nova-metadata-internal

deprecatedAnnotation

Service uses deprecated annotation metallb.universe.tf/address-pool

openstack

kubelet

nova-cell1-conductor-db-sync-bcpbl

Started

Started container nova-cell1-conductor-db-sync

openstack

cert-manager-certificates-trigger

nova-metadata-internal-svc

Issuing

Issuing certificate as Secret does not exist

openstack

kubelet

nova-metadata-0

Pulling

Pulling image "quay.io/podified-antelope-centos9/openstack-nova-api@sha256:faf711e1e5fa2ad74a73d3dfffd88f6312fb045cb69e9b7e6331558784163d16"

openstack

kubelet

nova-cell1-novncproxy-0

Pulling

Pulling image "quay.io/podified-antelope-centos9/openstack-nova-novncproxy@sha256:228b3c59ea6527048a4b3d1e340c15f22dcf9f9ba8f302d6263f2e4ef79463ff"

openstack

kubelet

nova-cell1-conductor-db-sync-bcpbl

Created

Created container: nova-cell1-conductor-db-sync

openstack

kubelet

dnsmasq-dns-75597d7cbf-d9w8g

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:fbb5be29e9e4fa11f0743e7f74f2e80dcc7445d24770709ea0e038147f752c51" already present on machine

openstack

kubelet

nova-scheduler-0

Pulling

Pulling image "quay.io/podified-antelope-centos9/openstack-nova-scheduler@sha256:d89e44b4641e8bd60abf1b674253975596fafc490022169681555069174a414e"

openstack

multus

nova-scheduler-0

AddedInterface

Add eth0 [10.128.0.253/23] from ovn-kubernetes

openstack

multus

dnsmasq-dns-75597d7cbf-d9w8g

AddedInterface

Add eth0 [10.128.0.255/23] from ovn-kubernetes

openstack

kubelet

dnsmasq-dns-75597d7cbf-d9w8g

Created

Created container: init

openstack

kubelet

dnsmasq-dns-75597d7cbf-d9w8g

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:fbb5be29e9e4fa11f0743e7f74f2e80dcc7445d24770709ea0e038147f752c51" already present on machine

openstack

kubelet

dnsmasq-dns-75597d7cbf-d9w8g

Started

Started container init

openstack

cert-manager-certificates-key-manager

nova-novncproxy-cell1-public-svc

Generated

Stored new private key in temporary Secret resource "nova-novncproxy-cell1-public-svc-b7hp5"

openstack

cert-manager-certificates-trigger

nova-novncproxy-cell1-public-svc

Issuing

Issuing certificate as Secret does not exist

openstack

kubelet

dnsmasq-dns-75597d7cbf-d9w8g

Created

Created container: dnsmasq-dns

openstack

kubelet

dnsmasq-dns-75597d7cbf-d9w8g

Started

Started container dnsmasq-dns

openstack

cert-manager-certificates-key-manager

nova-metadata-internal-svc

Generated

Stored new private key in temporary Secret resource "nova-metadata-internal-svc-jtl8x"

openstack

kubelet

nova-api-0

Pulled

Successfully pulled image "quay.io/podified-antelope-centos9/openstack-nova-api@sha256:faf711e1e5fa2ad74a73d3dfffd88f6312fb045cb69e9b7e6331558784163d16" in 6.11s (6.11s including waiting). Image size: 684724449 bytes.

openstack

kubelet

nova-api-0

Started

Started container nova-api-api

openstack

kubelet

nova-cell1-novncproxy-0

Created

Created container: nova-cell1-novncproxy-novncproxy

openstack

kubelet

nova-metadata-0

Created

Created container: nova-metadata-log

openstack

kubelet

nova-metadata-0

Pulled

Successfully pulled image "quay.io/podified-antelope-centos9/openstack-nova-api@sha256:faf711e1e5fa2ad74a73d3dfffd88f6312fb045cb69e9b7e6331558784163d16" in 6.153s (6.153s including waiting). Image size: 684724449 bytes.

openstack

kubelet

nova-api-0

Created

Created container: nova-api-log

openstack

kubelet

nova-api-0

Started

Started container nova-api-log

openstack

kubelet

nova-metadata-0

Started

Started container nova-metadata-log

openstack

kubelet

nova-api-0

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-nova-api@sha256:faf711e1e5fa2ad74a73d3dfffd88f6312fb045cb69e9b7e6331558784163d16" already present on machine

openstack

kubelet

nova-metadata-0

Created

Created container: nova-metadata-metadata

openstack

kubelet

nova-metadata-0

Started

Started container nova-metadata-metadata

openstack

kubelet

nova-api-0

Created

Created container: nova-api-api

openstack

kubelet

nova-metadata-0

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-nova-api@sha256:faf711e1e5fa2ad74a73d3dfffd88f6312fb045cb69e9b7e6331558784163d16" already present on machine

openstack

kubelet

nova-scheduler-0

Pulled

Successfully pulled image "quay.io/podified-antelope-centos9/openstack-nova-scheduler@sha256:d89e44b4641e8bd60abf1b674253975596fafc490022169681555069174a414e" in 5.873s (5.873s including waiting). Image size: 667925989 bytes.

openstack

kubelet

nova-scheduler-0

Created

Created container: nova-scheduler-scheduler

openstack

kubelet

nova-scheduler-0

Started

Started container nova-scheduler-scheduler

openstack

kubelet

nova-cell1-novncproxy-0

Pulled

Successfully pulled image "quay.io/podified-antelope-centos9/openstack-nova-novncproxy@sha256:228b3c59ea6527048a4b3d1e340c15f22dcf9f9ba8f302d6263f2e4ef79463ff" in 6.048s (6.048s including waiting). Image size: 670289900 bytes.

openstack

kubelet

nova-cell1-novncproxy-0

Started

Started container nova-cell1-novncproxy-novncproxy

openstack

kubelet

nova-api-0

Unhealthy

Startup probe failed: Get "http://10.128.0.251:8774/": context deadline exceeded (Client.Timeout exceeded while awaiting headers)

openstack

kubelet

nova-api-0

Unhealthy

Startup probe failed: Get "http://10.128.0.251:8774/": context deadline exceeded (Client.Timeout exceeded while awaiting headers)

openstack

kubelet

nova-metadata-0

Unhealthy

Startup probe failed: Get "http://10.128.0.254:8775/": context deadline exceeded (Client.Timeout exceeded while awaiting headers)

openstack

kubelet

nova-metadata-0

Unhealthy

Startup probe failed: Get "http://10.128.0.254:8775/": context deadline exceeded (Client.Timeout exceeded while awaiting headers)

metallb-system

kubelet

metallb-operator-controller-manager-64ff997757-xgs48

Pulled

Container image "registry.redhat.io/openshift4/metallb-rhel9-operator@sha256:9d74242d31d5f83bb8207d71e2a766ce9ababf218795d5c6fbb50450af5c29e8" already present on machine

metallb-system

kubelet

metallb-operator-controller-manager-64ff997757-xgs48

Pulled

Container image "registry.redhat.io/openshift4/metallb-rhel9-operator@sha256:9d74242d31d5f83bb8207d71e2a766ce9ababf218795d5c6fbb50450af5c29e8" already present on machine

openstack-operators

kubelet

octavia-operator-controller-manager-5b9f45d989-fr295

Unhealthy

Readiness probe failed: Get "http://10.128.0.155:8081/readyz": dial tcp 10.128.0.155:8081: connect: connection refused
(x2)

metallb-system

kubelet

metallb-operator-controller-manager-64ff997757-xgs48

Created

Created container: manager

openstack-operators

kubelet

octavia-operator-controller-manager-5b9f45d989-fr295

Unhealthy

Liveness probe failed: Get "http://10.128.0.155:8081/healthz": dial tcp 10.128.0.155:8081: connect: connection refused

openstack-operators

kubelet

octavia-operator-controller-manager-5b9f45d989-fr295

Unhealthy

Liveness probe failed: Get "http://10.128.0.155:8081/healthz": dial tcp 10.128.0.155:8081: connect: connection refused

openstack-operators

kubelet

octavia-operator-controller-manager-5b9f45d989-fr295

Unhealthy

Readiness probe failed: Get "http://10.128.0.155:8081/readyz": dial tcp 10.128.0.155:8081: connect: connection refused
(x2)

metallb-system

kubelet

metallb-operator-controller-manager-64ff997757-xgs48

Created

Created container: manager

openstack-operators

kubelet

test-operator-controller-manager-5c5cb9c4d7-nf9vn

Unhealthy

Readiness probe failed: Get "http://10.128.0.161:8081/readyz": dial tcp 10.128.0.161:8081: connect: connection refused
(x2)

metallb-system

kubelet

metallb-operator-controller-manager-64ff997757-xgs48

Started

Started container manager

openstack-operators

kubelet

test-operator-controller-manager-5c5cb9c4d7-nf9vn

Unhealthy

Readiness probe failed: Get "http://10.128.0.161:8081/readyz": dial tcp 10.128.0.161:8081: connect: connection refused

openstack-operators

kubelet

test-operator-controller-manager-5c5cb9c4d7-nf9vn

Unhealthy

Liveness probe failed: Get "http://10.128.0.161:8081/healthz": dial tcp 10.128.0.161:8081: connect: connection refused

openstack-operators

kubelet

test-operator-controller-manager-5c5cb9c4d7-nf9vn

Unhealthy

Liveness probe failed: Get "http://10.128.0.161:8081/healthz": dial tcp 10.128.0.161:8081: connect: connection refused
(x2)

openshift-kube-controller-manager

kubelet

kube-controller-manager-master-0

Pulled

Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b23c544d3894e5b31f66a18c554f03b0d29f92c2000c46b57b1c96da7ec25db9" already present on machine
(x2)

metallb-system

kubelet

metallb-operator-controller-manager-64ff997757-xgs48

Started

Started container manager
(x2)

openstack-operators

kubelet

openstack-baremetal-operator-controller-manager-74c4796899s6vn4

Started

Started container manager
(x2)

openstack-operators

kubelet

ironic-operator-controller-manager-6f787dddc9-dq5f2

Created

Created container: manager

openstack-operators

kubelet

octavia-operator-controller-manager-5b9f45d989-fr295

Pulled

Container image "quay.io/openstack-k8s-operators/octavia-operator@sha256:425fd66675becbe0ca2b2fe1a5a6694ac6e0b1cdce9a77a7a37f99785eadc74a" already present on machine

openstack-operators

kubelet

ironic-operator-controller-manager-6f787dddc9-dq5f2

Pulled

Container image "quay.io/openstack-k8s-operators/ironic-operator@sha256:9dd26bc51e7757d84736528d4988a1f980ad50ccb070aef6fc252e32c5c423a8" already present on machine
(x2)

openstack-operators

kubelet

test-operator-controller-manager-5c5cb9c4d7-nf9vn

Created

Created container: manager
(x2)

openstack-operators

kubelet

octavia-operator-controller-manager-5b9f45d989-fr295

Started

Started container manager
(x2)

openstack-operators

kubelet

openstack-baremetal-operator-controller-manager-74c4796899s6vn4

Started

Started container manager
(x2)

openstack-operators

kubelet

test-operator-controller-manager-5c5cb9c4d7-nf9vn

Started

Started container manager
(x2)

openstack-operators

kubelet

test-operator-controller-manager-5c5cb9c4d7-nf9vn

Created

Created container: manager

openstack-operators

kubelet

openstack-baremetal-operator-controller-manager-74c4796899s6vn4

Pulled

Container image "quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:bf7cdbfb125c4327b35870f8640cbed9ddc32d6f07fedd117c6fd59f16463329" already present on machine

openstack-operators

kubelet

openstack-baremetal-operator-controller-manager-74c4796899s6vn4

Pulled

Container image "quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:bf7cdbfb125c4327b35870f8640cbed9ddc32d6f07fedd117c6fd59f16463329" already present on machine

openstack-operators

kubelet

ironic-operator-controller-manager-6f787dddc9-dq5f2

Pulled

Container image "quay.io/openstack-k8s-operators/ironic-operator@sha256:9dd26bc51e7757d84736528d4988a1f980ad50ccb070aef6fc252e32c5c423a8" already present on machine
(x2)

openstack-operators

kubelet

ironic-operator-controller-manager-6f787dddc9-dq5f2

Started

Started container manager
(x2)

openstack-operators

kubelet

openstack-baremetal-operator-controller-manager-74c4796899s6vn4

Created

Created container: manager
(x2)

openstack-operators

kubelet

openstack-baremetal-operator-controller-manager-74c4796899s6vn4

Created

Created container: manager
(x2)

openstack-operators

kubelet

octavia-operator-controller-manager-5b9f45d989-fr295

Created

Created container: manager

openstack-operators

kubelet

octavia-operator-controller-manager-5b9f45d989-fr295

Pulled

Container image "quay.io/openstack-k8s-operators/octavia-operator@sha256:425fd66675becbe0ca2b2fe1a5a6694ac6e0b1cdce9a77a7a37f99785eadc74a" already present on machine
(x2)

openstack-operators

kubelet

ironic-operator-controller-manager-6f787dddc9-dq5f2

Created

Created container: manager
(x2)

openstack-operators

kubelet

octavia-operator-controller-manager-5b9f45d989-fr295

Started

Started container manager
(x2)

openstack-operators

kubelet

octavia-operator-controller-manager-5b9f45d989-fr295

Created

Created container: manager

openstack-operators

kubelet

test-operator-controller-manager-5c5cb9c4d7-nf9vn

Pulled

Container image "quay.io/openstack-k8s-operators/test-operator@sha256:43bd420bc05b4789243740bc75f61e10c7aac7883fc2f82b2d4d50085bc96c42" already present on machine

openstack-operators

kubelet

test-operator-controller-manager-5c5cb9c4d7-nf9vn

Pulled

Container image "quay.io/openstack-k8s-operators/test-operator@sha256:43bd420bc05b4789243740bc75f61e10c7aac7883fc2f82b2d4d50085bc96c42" already present on machine
(x2)

openstack-operators

kubelet

ironic-operator-controller-manager-6f787dddc9-dq5f2

Started

Started container manager
(x2)

openstack-operators

kubelet

test-operator-controller-manager-5c5cb9c4d7-nf9vn

Started

Started container manager
(x2)

openshift-kube-controller-manager

kubelet

kube-controller-manager-master-0

Created

Created container: kube-controller-manager
(x2)

openshift-kube-controller-manager

kubelet

kube-controller-manager-master-0

Started

Started container kube-controller-manager

openstack

cert-manager-certificaterequests-approver

nova-novncproxy-cell1-public-svc-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack

cert-manager-certificaterequests-issuer-selfsigned

nova-metadata-internal-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificates-request-manager

nova-novncproxy-cell1-public-svc

Requested

Created new CertificateRequest resource "nova-novncproxy-cell1-public-svc-1"

openstack

cert-manager-certificaterequests-issuer-ca

nova-novncproxy-cell1-public-svc-1

CertificateIssued

Certificate fetched from issuer successfully

openstack

cert-manager-certificaterequests-issuer-venafi

nova-novncproxy-cell1-public-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-ca

nova-novncproxy-cell1-public-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-selfsigned

nova-novncproxy-cell1-public-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-vault

nova-novncproxy-cell1-public-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-acme

nova-novncproxy-cell1-public-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-vault

nova-metadata-internal-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificates-request-manager

nova-metadata-internal-svc

Requested

Created new CertificateRequest resource "nova-metadata-internal-svc-1"

openstack

cert-manager-certificaterequests-issuer-venafi

nova-metadata-internal-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-acme

nova-metadata-internal-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-ca

nova-metadata-internal-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificates-issuing

nova-novncproxy-cell1-public-svc

Issuing

The certificate has been successfully issued

openstack

cert-manager-certificaterequests-issuer-ca

nova-metadata-internal-svc-1

CertificateIssued

Certificate fetched from issuer successfully

openstack

cert-manager-certificaterequests-approver

nova-metadata-internal-svc-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack

cert-manager-certificates-issuing

nova-metadata-internal-svc

Issuing

The certificate has been successfully issued

openshift-kube-controller-manager-operator

kube-controller-manager-operator-status-controller-statussyncer_kube-controller-manager

kube-controller-manager-operator

OperatorStatusChanged

Status for clusteroperator/kube-controller-manager changed: Degraded message changed from "NodeControllerDegraded: All master nodes are ready" to "NodeControllerDegraded: All master nodes are ready\nStaticPodsDegraded: pod/kube-controller-manager-master-0 container \"kube-controller-manager\" is terminated: Error: ing\" logger=\"replicaset-controller\" kind=\"ReplicaSet\" key=\"openstack/dnsmasq-dns-75597d7cbf\" duration=\"139.074µs\"\nStaticPodsDegraded: I0320 09:08:57.711218 1 replica_set.go:679] \"Finished syncing\" logger=\"replicaset-controller\" kind=\"ReplicaSet\" key=\"openstack/dnsmasq-dns-75597d7cbf\" duration=\"194.066µs\"\nStaticPodsDegraded: I0320 09:08:57.857644 1 replica_set.go:679] \"Finished syncing\" logger=\"replicaset-controller\" kind=\"ReplicaSet\" key=\"openstack/dnsmasq-dns-75597d7cbf\" duration=\"143.604µs\"\nStaticPodsDegraded: I0320 09:08:58.694209 1 job_controller.go:573] \"enqueueing job\" logger=\"job-controller\" key=\"openstack/nova-cell1-conductor-db-sync\" delay=\"1s\"\nStaticPodsDegraded: I0320 09:08:58.773423 1 job_controller.go:573] \"enqueueing job\" logger=\"job-controller\" key=\"openstack/nova-cell0-cell-mapping\" delay=\"1s\"\nStaticPodsDegraded: I0320 09:08:58.902725 1 replica_set.go:679] \"Finished syncing\" logger=\"replicaset-controller\" kind=\"ReplicaSet\" key=\"openstack/dnsmasq-dns-75597d7cbf\" duration=\"100.582µs\"\nStaticPodsDegraded: I0320 09:09:00.034719 1 job_controller.go:573] \"enqueueing job\" logger=\"job-controller\" key=\"openstack/nova-cell1-conductor-db-sync\" delay=\"1s\"\nStaticPodsDegraded: W0320 09:09:06.617243 1 transport.go:356] Unable to cancel request for *app.rejectIfNotReadyHeaderRT\nStaticPodsDegraded: E0320 09:09:06.617410 1 leaderelection.go:429] Failed to update lock optimitically: Put \"https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers), falling back to slow path\nStaticPodsDegraded: E0320 09:09:12.617444 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get \"https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s\": context deadline exceeded\nStaticPodsDegraded: I0320 09:09:12.617514 1 leaderelection.go:297] failed to renew lease kube-system/kube-controller-manager: timed out waiting for the condition\nStaticPodsDegraded: E0320 09:09:12.617751 1 controllermanager.go:373] \"leaderelection lost\"\nStaticPodsDegraded: "

openstack

cert-manager-certificates-key-manager

nova-novncproxy-cell1-vencrypt

Generated

Stored new private key in temporary Secret resource "nova-novncproxy-cell1-vencrypt-pg4bj"

openstack

cert-manager-certificaterequests-approver

nova-novncproxy-cell1-vencrypt-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack

cert-manager-certificates-issuing

nova-novncproxy-cell1-public-route

Issuing

The certificate has been successfully issued

openstack

cert-manager-certificaterequests-issuer-ca

nova-novncproxy-cell1-public-route-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-ca

nova-novncproxy-cell1-public-route-1

CertificateIssued

Certificate fetched from issuer successfully

openstack

cert-manager-certificates-issuing

nova-novncproxy-cell1-vencrypt

Issuing

The certificate has been successfully issued

openstack

cert-manager-certificaterequests-approver

nova-novncproxy-cell1-public-route-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack

cert-manager-certificaterequests-issuer-selfsigned

nova-novncproxy-cell1-public-route-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-vault

nova-novncproxy-cell1-public-route-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificates-request-manager

nova-novncproxy-cell1-vencrypt

Requested

Created new CertificateRequest resource "nova-novncproxy-cell1-vencrypt-1"

openstack

cert-manager-certificaterequests-issuer-acme

nova-novncproxy-cell1-public-route-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificates-trigger

nova-novncproxy-cell1-vencrypt

Issuing

Issuing certificate as Secret does not exist

openstack

cert-manager-certificaterequests-issuer-ca

nova-novncproxy-cell1-vencrypt-1

CertificateIssued

Certificate fetched from issuer successfully

openstack

cert-manager-certificates-request-manager

nova-novncproxy-cell1-public-route

Requested

Created new CertificateRequest resource "nova-novncproxy-cell1-public-route-1"

openstack

cert-manager-certificaterequests-issuer-selfsigned

nova-novncproxy-cell1-vencrypt-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificates-trigger

nova-novncproxy-cell1-public-route

Issuing

Issuing certificate as Secret does not exist

openstack

cert-manager-certificates-key-manager

nova-novncproxy-cell1-public-route

Generated

Stored new private key in temporary Secret resource "nova-novncproxy-cell1-public-route-tkp7p"

openshift-kube-controller-manager-operator

kube-controller-manager-operator-status-controller-statussyncer_kube-controller-manager

kube-controller-manager-operator

OperatorStatusChanged

Status for clusteroperator/kube-controller-manager changed: Degraded message changed from "NodeControllerDegraded: All master nodes are ready\nStaticPodsDegraded: pod/kube-controller-manager-master-0 container \"kube-controller-manager\" is terminated: Error: ing\" logger=\"replicaset-controller\" kind=\"ReplicaSet\" key=\"openstack/dnsmasq-dns-75597d7cbf\" duration=\"139.074µs\"\nStaticPodsDegraded: I0320 09:08:57.711218 1 replica_set.go:679] \"Finished syncing\" logger=\"replicaset-controller\" kind=\"ReplicaSet\" key=\"openstack/dnsmasq-dns-75597d7cbf\" duration=\"194.066µs\"\nStaticPodsDegraded: I0320 09:08:57.857644 1 replica_set.go:679] \"Finished syncing\" logger=\"replicaset-controller\" kind=\"ReplicaSet\" key=\"openstack/dnsmasq-dns-75597d7cbf\" duration=\"143.604µs\"\nStaticPodsDegraded: I0320 09:08:58.694209 1 job_controller.go:573] \"enqueueing job\" logger=\"job-controller\" key=\"openstack/nova-cell1-conductor-db-sync\" delay=\"1s\"\nStaticPodsDegraded: I0320 09:08:58.773423 1 job_controller.go:573] \"enqueueing job\" logger=\"job-controller\" key=\"openstack/nova-cell0-cell-mapping\" delay=\"1s\"\nStaticPodsDegraded: I0320 09:08:58.902725 1 replica_set.go:679] \"Finished syncing\" logger=\"replicaset-controller\" kind=\"ReplicaSet\" key=\"openstack/dnsmasq-dns-75597d7cbf\" duration=\"100.582µs\"\nStaticPodsDegraded: I0320 09:09:00.034719 1 job_controller.go:573] \"enqueueing job\" logger=\"job-controller\" key=\"openstack/nova-cell1-conductor-db-sync\" delay=\"1s\"\nStaticPodsDegraded: W0320 09:09:06.617243 1 transport.go:356] Unable to cancel request for *app.rejectIfNotReadyHeaderRT\nStaticPodsDegraded: E0320 09:09:06.617410 1 leaderelection.go:429] Failed to update lock optimitically: Put \"https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers), falling back to slow path\nStaticPodsDegraded: E0320 09:09:12.617444 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get \"https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s\": context deadline exceeded\nStaticPodsDegraded: I0320 09:09:12.617514 1 leaderelection.go:297] failed to renew lease kube-system/kube-controller-manager: timed out waiting for the condition\nStaticPodsDegraded: E0320 09:09:12.617751 1 controllermanager.go:373] \"leaderelection lost\"\nStaticPodsDegraded: " to "NodeControllerDegraded: All master nodes are ready"

openstack

cert-manager-certificaterequests-issuer-venafi

nova-novncproxy-cell1-vencrypt-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-vault

nova-novncproxy-cell1-vencrypt-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-acme

nova-novncproxy-cell1-vencrypt-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-venafi

nova-novncproxy-cell1-public-route-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-ca

nova-novncproxy-cell1-vencrypt-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack-operators

openstack-baremetal-operator-controller-manager-74c4796899s6vn4_ad688592-a215-4a6f-b0de-3fa6d09b1a4f

dedc2245.openstack.org

LeaderElection

openstack-baremetal-operator-controller-manager-74c4796899s6vn4_ad688592-a215-4a6f-b0de-3fa6d09b1a4f became leader

openstack-operators

octavia-operator-controller-manager-5b9f45d989-fr295_291cd54b-4999-4325-9daf-3b0fab874122

98809e87.openstack.org

LeaderElection

octavia-operator-controller-manager-5b9f45d989-fr295_291cd54b-4999-4325-9daf-3b0fab874122 became leader

openstack-operators

openstack-baremetal-operator-controller-manager-74c4796899s6vn4_ad688592-a215-4a6f-b0de-3fa6d09b1a4f

dedc2245.openstack.org

LeaderElection

openstack-baremetal-operator-controller-manager-74c4796899s6vn4_ad688592-a215-4a6f-b0de-3fa6d09b1a4f became leader

openstack-operators

octavia-operator-controller-manager-5b9f45d989-fr295_291cd54b-4999-4325-9daf-3b0fab874122

98809e87.openstack.org

LeaderElection

octavia-operator-controller-manager-5b9f45d989-fr295_291cd54b-4999-4325-9daf-3b0fab874122 became leader
(x2)

openstack

metallb-controller

nova-internal

deprecatedAnnotation

Service uses deprecated annotation metallb.universe.tf/loadBalancerIPs

default

garbage-collector-controller

frr-k8s-validating-webhook-configuration

OwnerRefInvalidNamespace

ownerRef [metallb.io/v1beta1/MetalLB, namespace: , name: metallb, uid: 1c3f6435-b6df-436c-b702-585794b6aeb6] does not exist in namespace ""

openstack

kubelet

nova-metadata-0

Killing

Stopping container nova-metadata-log

openstack

job-controller

nova-cell1-conductor-db-sync

Completed

Job completed

openstack

kubelet

nova-metadata-0

Killing

Stopping container nova-metadata-metadata

openstack

statefulset-controller

nova-cell1-conductor

SuccessfulCreate

create Pod nova-cell1-conductor-0 in StatefulSet nova-cell1-conductor successful

metallb-system

metallb-operator-controller-manager-64ff997757-xgs48_f797bfc2-b728-48ea-88e2-3461d7edaed4

metallb.io.metallboperator

LeaderElection

metallb-operator-controller-manager-64ff997757-xgs48_f797bfc2-b728-48ea-88e2-3461d7edaed4 became leader

openstack

deployment-controller

dnsmasq-dns

ScalingReplicaSet

Scaled down replica set dnsmasq-dns-7cd5f68b57 to 0 from 1

openstack-operators

test-operator-controller-manager-5c5cb9c4d7-nf9vn_8fc114ba-10d2-4af0-91af-5fbe89bde1d4

6cce095b.openstack.org

LeaderElection

test-operator-controller-manager-5c5cb9c4d7-nf9vn_8fc114ba-10d2-4af0-91af-5fbe89bde1d4 became leader

openstack-operators

test-operator-controller-manager-5c5cb9c4d7-nf9vn_8fc114ba-10d2-4af0-91af-5fbe89bde1d4

6cce095b.openstack.org

LeaderElection

test-operator-controller-manager-5c5cb9c4d7-nf9vn_8fc114ba-10d2-4af0-91af-5fbe89bde1d4 became leader

metallb-system

metallb-operator-controller-manager-64ff997757-xgs48_f797bfc2-b728-48ea-88e2-3461d7edaed4

metallb.io.metallboperator

LeaderElection

metallb-operator-controller-manager-64ff997757-xgs48_f797bfc2-b728-48ea-88e2-3461d7edaed4 became leader
(x2)

openstack

metallb-controller

nova-internal

deprecatedAnnotation

Service uses deprecated annotation metallb.universe.tf/allow-shared-ip
(x2)

openstack

metallb-controller

nova-internal

deprecatedAnnotation

Service uses deprecated annotation metallb.universe.tf/address-pool

openstack

metallb-controller

nova-internal

IPAllocated

Assigned IP ["172.17.0.80"]

default

node-controller

master-0

RegisteredNode

Node master-0 event: Registered Node master-0 in Controller

kube-system

kube-controller-manager

kube-controller-manager

LeaderElection

master-0_91d2d57a-44ae-47bf-910a-0ac4097eb8be became leader

openstack

kubelet

nova-cell1-novncproxy-0

Killing

Stopping container nova-cell1-novncproxy-novncproxy

openstack

replicaset-controller

dnsmasq-dns-7cd5f68b57

SuccessfulDelete

Deleted pod: dnsmasq-dns-7cd5f68b57-cbttm

openstack

job-controller

nova-cell0-cell-mapping

Completed

Job completed

openstack

statefulset-controller

nova-cell1-novncproxy

SuccessfulDelete

delete Pod nova-cell1-novncproxy-0 in StatefulSet nova-cell1-novncproxy successful

openstack

replicaset-controller

dnsmasq-dns-5cc86b957c

SuccessfulCreate

Created pod: dnsmasq-dns-5cc86b957c-nqjxf

openstack-operators

ironic-operator-controller-manager-6f787dddc9-dq5f2_bdd41352-86a9-4683-be16-fb41e4fca75c

f92b5c2d.openstack.org

LeaderElection

ironic-operator-controller-manager-6f787dddc9-dq5f2_bdd41352-86a9-4683-be16-fb41e4fca75c became leader

openstack

deployment-controller

dnsmasq-dns

ScalingReplicaSet

Scaled up replica set dnsmasq-dns-5cc86b957c to 1

openstack

kubelet

nova-api-0

Killing

Stopping container nova-api-log

openstack

kubelet

nova-api-0

Killing

Stopping container nova-api-api

openstack

kubelet

nova-cell1-novncproxy-0

Unhealthy

Readiness probe failed: Get "http://10.128.0.252:6080/vnc_lite.html": dial tcp 10.128.0.252:6080: connect: connection refused

openstack-operators

ironic-operator-controller-manager-6f787dddc9-dq5f2_bdd41352-86a9-4683-be16-fb41e4fca75c

f92b5c2d.openstack.org

LeaderElection

ironic-operator-controller-manager-6f787dddc9-dq5f2_bdd41352-86a9-4683-be16-fb41e4fca75c became leader

openstack

kubelet

dnsmasq-dns-7cd5f68b57-cbttm

Killing

Stopping container dnsmasq-dns

metallb-system

operator-lifecycle-manager

metallb-operator.v4.18.0-202603040208

ComponentUnhealthy

installing: waiting for deployment metallb-operator-controller-manager to become ready: deployment "metallb-operator-controller-manager" not available: Deployment does not have minimum availability.

metallb-system

operator-lifecycle-manager

metallb-operator.v4.18.0-202603040208

ComponentUnhealthy

installing: waiting for deployment metallb-operator-controller-manager to become ready: deployment "metallb-operator-controller-manager" not available: Deployment does not have minimum availability.

openstack

kubelet

nova-cell1-conductor-0

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-nova-conductor@sha256:d347d48e9a8ae4136dd99c5222480ceccb2819beaf80b11048644d4acf0a4305" already present on machine

openstack

cert-manager-certificates-trigger

nova-internal-svc

Issuing

Issuing certificate as Secret does not exist

openstack

multus

nova-cell1-conductor-0

AddedInterface

Add eth0 [10.128.1.0/23] from ovn-kubernetes

openstack

multus

dnsmasq-dns-5cc86b957c-nqjxf

AddedInterface

Add eth0 [10.128.1.1/23] from ovn-kubernetes

openstack

cert-manager-certificaterequests-issuer-venafi

nova-internal-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

kubelet

dnsmasq-dns-5cc86b957c-nqjxf

Started

Started container init

openstack

cert-manager-certificates-issuing

nova-internal-svc

Issuing

The certificate has been successfully issued
(x11)

openstack

rabbitmqcluster-controller

rabbitmq-cell1

SuccessfulUpdate

updated resource rabbitmq-cell1-nodes of Type *v1.Service

openstack

cert-manager-certificaterequests-issuer-vault

nova-internal-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificates-key-manager

nova-internal-svc

Generated

Stored new private key in temporary Secret resource "nova-internal-svc-2kfkx"

openstack

kubelet

dnsmasq-dns-5cc86b957c-nqjxf

Created

Created container: init

openstack

cert-manager-certificaterequests-issuer-ca

nova-internal-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-acme

nova-internal-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificates-request-manager

nova-internal-svc

Requested

Created new CertificateRequest resource "nova-internal-svc-1"

openstack

kubelet

nova-scheduler-0

Killing

Stopping container nova-scheduler-scheduler

openstack

cert-manager-certificaterequests-approver

nova-internal-svc-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack

kubelet

dnsmasq-dns-5cc86b957c-nqjxf

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:fbb5be29e9e4fa11f0743e7f74f2e80dcc7445d24770709ea0e038147f752c51" already present on machine

openstack

kubelet

nova-cell1-conductor-0

Started

Started container nova-cell1-conductor-conductor

openstack

cert-manager-certificaterequests-issuer-selfsigned

nova-internal-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-ca

nova-internal-svc-1

CertificateIssued

Certificate fetched from issuer successfully

openstack

kubelet

nova-cell1-conductor-0

Created

Created container: nova-cell1-conductor-conductor

openstack

kubelet

nova-metadata-0

Unhealthy

Readiness probe failed: Get "http://10.128.0.254:8775/": read tcp 10.128.0.2:43876->10.128.0.254:8775: read: connection reset by peer

openstack

kubelet

nova-metadata-0

Unhealthy

Readiness probe failed: Get "http://10.128.0.254:8775/": read tcp 10.128.0.2:43860->10.128.0.254:8775: read: connection reset by peer

openstack

kubelet

dnsmasq-dns-5cc86b957c-nqjxf

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:fbb5be29e9e4fa11f0743e7f74f2e80dcc7445d24770709ea0e038147f752c51" already present on machine
(x12)

openstack

rabbitmqcluster-controller

rabbitmq

SuccessfulUpdate

updated resource rabbitmq-nodes of Type *v1.Service

openstack

kubelet

nova-api-0

Unhealthy

Readiness probe failed: Get "http://10.128.0.251:8774/": read tcp 10.128.0.2:41056->10.128.0.251:8774: read: connection reset by peer

openstack

kubelet

nova-api-0

Unhealthy

Readiness probe failed: Get "http://10.128.0.251:8774/": read tcp 10.128.0.2:41068->10.128.0.251:8774: read: connection reset by peer

openstack

kubelet

nova-scheduler-0

Unhealthy

Readiness probe errored: rpc error: code = NotFound desc = container is not created or running: checking if PID of c0a20a321758b89a7197046f590fbf478d8bb8ff29a8be0633a162d93309aede is running failed: container process not found

openstack

statefulset-controller

nova-cell1-novncproxy

SuccessfulCreate

create Pod nova-cell1-novncproxy-0 in StatefulSet nova-cell1-novncproxy successful

openstack

kubelet

dnsmasq-dns-5cc86b957c-nqjxf

Created

Created container: dnsmasq-dns

openstack

kubelet

dnsmasq-dns-5cc86b957c-nqjxf

Started

Started container dnsmasq-dns

openstack

kubelet

dnsmasq-dns-7cd5f68b57-cbttm

Unhealthy

Readiness probe failed: dial tcp 10.128.0.227:5353: i/o timeout

openstack

multus

nova-cell1-novncproxy-0

AddedInterface

Add eth0 [10.128.1.2/23] from ovn-kubernetes

openstack

kubelet

nova-cell1-novncproxy-0

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-nova-novncproxy@sha256:228b3c59ea6527048a4b3d1e340c15f22dcf9f9ba8f302d6263f2e4ef79463ff" already present on machine

openstack

cert-manager-certificates-trigger

nova-public-svc

Issuing

Issuing certificate as Secret does not exist

openstack

kubelet

nova-cell1-novncproxy-0

Created

Created container: nova-cell1-novncproxy-novncproxy

openstack

cert-manager-certificates-key-manager

nova-public-svc

Generated

Stored new private key in temporary Secret resource "nova-public-svc-7nzmq"

openstack

kubelet

nova-cell1-novncproxy-0

Started

Started container nova-cell1-novncproxy-novncproxy

openstack

cert-manager-certificaterequests-issuer-acme

nova-public-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-selfsigned

nova-public-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-ca

nova-public-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-vault

nova-public-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-venafi

nova-public-svc-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificates-request-manager

nova-public-svc

Requested

Created new CertificateRequest resource "nova-public-svc-1"

openstack

cert-manager-certificaterequests-issuer-ca

nova-public-svc-1

CertificateIssued

Certificate fetched from issuer successfully

openstack

cert-manager-certificaterequests-approver

nova-public-svc-1

cert-manager.io

Certificate request has been approved by cert-manager.io

metallb-system

operator-lifecycle-manager

metallb-operator.v4.18.0-202603040208

NeedsReinstall

installing: waiting for deployment metallb-operator-controller-manager to become ready: deployment "metallb-operator-controller-manager" not available: Deployment does not have minimum availability.

metallb-system

operator-lifecycle-manager

metallb-operator.v4.18.0-202603040208

NeedsReinstall

installing: waiting for deployment metallb-operator-controller-manager to become ready: deployment "metallb-operator-controller-manager" not available: Deployment does not have minimum availability.

openstack

multus

nova-metadata-0

AddedInterface

Add eth0 [10.128.1.3/23] from ovn-kubernetes

openstack

cert-manager-certificates-issuing

nova-public-svc

Issuing

The certificate has been successfully issued

openstack

kubelet

nova-metadata-0

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-nova-api@sha256:faf711e1e5fa2ad74a73d3dfffd88f6312fb045cb69e9b7e6331558784163d16" already present on machine

openstack

kubelet

nova-metadata-0

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-nova-api@sha256:faf711e1e5fa2ad74a73d3dfffd88f6312fb045cb69e9b7e6331558784163d16" already present on machine

openstack

kubelet

nova-metadata-0

Created

Created container: nova-metadata-log

openstack

kubelet

nova-metadata-0

Started

Started container nova-metadata-log

openstack

replicaset-controller

dnsmasq-dns-75597d7cbf

SuccessfulDelete

Deleted pod: dnsmasq-dns-75597d7cbf-d9w8g

openstack

deployment-controller

dnsmasq-dns

ScalingReplicaSet

Scaled down replica set dnsmasq-dns-75597d7cbf to 0 from 1
(x38)

openstack

metallb-speaker

dnsmasq-dns

nodeAssigned

announcing from node "master-0" with protocol "layer2"

openstack

kubelet

nova-metadata-0

Created

Created container: nova-metadata-metadata

openstack

kubelet

nova-metadata-0

Started

Started container nova-metadata-metadata

openstack

cert-manager-certificates-key-manager

nova-public-route

Generated

Stored new private key in temporary Secret resource "nova-public-route-2rdsh"
(x3)

metallb-system

operator-lifecycle-manager

metallb-operator.v4.18.0-202603040208

AllRequirementsMet

all requirements found, attempting install

openstack

kubelet

dnsmasq-dns-75597d7cbf-d9w8g

Killing

Stopping container dnsmasq-dns

openstack

multus

nova-api-0

AddedInterface

Add eth0 [10.128.1.4/23] from ovn-kubernetes

openshift-machine-config-operator

machine-config-operator

machine-config-operator

ConfigMapUpdated

Updated ConfigMap/kube-rbac-proxy -n openshift-machine-config-operator: cause by changes in data.config-file.yaml

openstack

cert-manager-certificates-trigger

nova-public-route

Issuing

Issuing certificate as Secret does not exist
(x3)

metallb-system

operator-lifecycle-manager

metallb-operator.v4.18.0-202603040208

AllRequirementsMet

all requirements found, attempting install

openstack

multus

nova-scheduler-0

AddedInterface

Add eth0 [10.128.1.5/23] from ovn-kubernetes

openstack

job-controller

nova-cell1-cell-mapping

SuccessfulCreate

Created pod: nova-cell1-cell-mapping-4h8z9

openstack

cert-manager-certificaterequests-issuer-acme

nova-public-route-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

kubelet

nova-api-0

Created

Created container: nova-api-log

openstack

kubelet

nova-api-0

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-nova-api@sha256:faf711e1e5fa2ad74a73d3dfffd88f6312fb045cb69e9b7e6331558784163d16" already present on machine

openstack

kubelet

nova-api-0

Started

Started container nova-api-log

openstack

kubelet

nova-api-0

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-nova-api@sha256:faf711e1e5fa2ad74a73d3dfffd88f6312fb045cb69e9b7e6331558784163d16" already present on machine

openstack

kubelet

nova-cell1-cell-mapping-4h8z9

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-nova-conductor@sha256:d347d48e9a8ae4136dd99c5222480ceccb2819beaf80b11048644d4acf0a4305" already present on machine

openstack

cert-manager-certificaterequests-issuer-selfsigned

nova-public-route-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

multus

nova-cell1-cell-mapping-4h8z9

AddedInterface

Add eth0 [10.128.1.6/23] from ovn-kubernetes

openstack

cert-manager-certificaterequests-approver

nova-public-route-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack

cert-manager-certificates-issuing

nova-public-route

Issuing

The certificate has been successfully issued

openstack

kubelet

nova-scheduler-0

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-nova-scheduler@sha256:d89e44b4641e8bd60abf1b674253975596fafc490022169681555069174a414e" already present on machine

openstack

cert-manager-certificaterequests-issuer-ca

nova-public-route-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-ca

nova-public-route-1

CertificateIssued

Certificate fetched from issuer successfully

openstack

cert-manager-certificaterequests-issuer-vault

nova-public-route-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-venafi

nova-public-route-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

kubelet

nova-scheduler-0

Started

Started container nova-scheduler-scheduler

openstack

kubelet

nova-scheduler-0

Created

Created container: nova-scheduler-scheduler

openstack

cert-manager-certificates-request-manager

nova-public-route

Requested

Created new CertificateRequest resource "nova-public-route-1"

openstack

kubelet

nova-api-0

Started

Started container nova-api-api

openstack

kubelet

nova-api-0

Created

Created container: nova-api-api

openstack

kubelet

nova-cell1-cell-mapping-4h8z9

Created

Created container: nova-manage

openstack

kubelet

nova-cell1-cell-mapping-4h8z9

Started

Started container nova-manage

openstack

kubelet

nova-api-0

Killing

Stopping container nova-api-api

openstack

kubelet

nova-api-0

Killing

Stopping container nova-api-log
(x3)

metallb-system

operator-lifecycle-manager

metallb-operator.v4.18.0-202603040208

InstallSucceeded

waiting for install components to report healthy
(x3)

metallb-system

operator-lifecycle-manager

metallb-operator.v4.18.0-202603040208

InstallSucceeded

waiting for install components to report healthy
(x2)

openstack

kubelet

nova-metadata-0

Unhealthy

Startup probe failed: Get "https://10.128.1.3:8775/": net/http: request canceled (Client.Timeout exceeded while awaiting headers)
(x2)

openstack

kubelet

nova-metadata-0

Unhealthy

Startup probe failed: Get "https://10.128.1.3:8775/": net/http: request canceled (Client.Timeout exceeded while awaiting headers)

openstack

multus

nova-api-0

AddedInterface

Add eth0 [10.128.1.7/23] from ovn-kubernetes

openstack

kubelet

nova-api-0

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-nova-api@sha256:faf711e1e5fa2ad74a73d3dfffd88f6312fb045cb69e9b7e6331558784163d16" already present on machine

openstack

kubelet

nova-api-0

Created

Created container: nova-api-log

openstack

kubelet

nova-api-0

Started

Started container nova-api-log

openstack

kubelet

nova-api-0

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-nova-api@sha256:faf711e1e5fa2ad74a73d3dfffd88f6312fb045cb69e9b7e6331558784163d16" already present on machine

openstack

kubelet

nova-api-0

Created

Created container: nova-api-api

openstack

kubelet

nova-api-0

Started

Started container nova-api-api

openstack

kubelet

nova-api-0

Unhealthy

Startup probe failed: Get "https://10.128.1.7:8774/": net/http: request canceled (Client.Timeout exceeded while awaiting headers)

openstack

kubelet

nova-api-0

Unhealthy

Startup probe failed: Get "https://10.128.1.7:8774/": net/http: request canceled (Client.Timeout exceeded while awaiting headers)

openstack

job-controller

nova-cell1-cell-mapping

Completed

Job completed
(x2)

openstack

statefulset-controller

nova-scheduler

SuccessfulDelete

delete Pod nova-scheduler-0 in StatefulSet nova-scheduler successful
(x3)

openstack

statefulset-controller

nova-api

SuccessfulDelete

delete Pod nova-api-0 in StatefulSet nova-api successful

openstack

kubelet

nova-api-0

Killing

Stopping container nova-api-api

openstack

kubelet

nova-scheduler-0

Killing

Stopping container nova-scheduler-scheduler

openstack

kubelet

nova-api-0

Killing

Stopping container nova-api-log
(x2)

metallb-system

operator-lifecycle-manager

metallb-operator.v4.18.0-202603040208

InstallSucceeded

install strategy completed with no errors
(x2)

metallb-system

operator-lifecycle-manager

metallb-operator.v4.18.0-202603040208

InstallSucceeded

install strategy completed with no errors

openstack

kubelet

nova-metadata-0

Killing

Stopping container nova-metadata-metadata

openstack

kubelet

nova-metadata-0

Killing

Stopping container nova-metadata-log
(x2)

openstack

statefulset-controller

nova-metadata

SuccessfulDelete

delete Pod nova-metadata-0 in StatefulSet nova-metadata successful
(x2)

openstack

statefulset-controller

nova-scheduler

SuccessfulCreate

create Pod nova-scheduler-0 in StatefulSet nova-scheduler successful
(x2)

openstack

statefulset-controller

nova-metadata

SuccessfulCreate

create Pod nova-metadata-0 in StatefulSet nova-metadata successful

openstack

kubelet

nova-scheduler-0

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-nova-scheduler@sha256:d89e44b4641e8bd60abf1b674253975596fafc490022169681555069174a414e" already present on machine

openstack

multus

nova-scheduler-0

AddedInterface

Add eth0 [10.128.1.8/23] from ovn-kubernetes

openstack

kubelet

nova-scheduler-0

Created

Created container: nova-scheduler-scheduler

openstack

multus

nova-metadata-0

AddedInterface

Add eth0 [10.128.1.9/23] from ovn-kubernetes

openstack

kubelet

nova-scheduler-0

Started

Started container nova-scheduler-scheduler

openstack

kubelet

nova-metadata-0

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-nova-api@sha256:faf711e1e5fa2ad74a73d3dfffd88f6312fb045cb69e9b7e6331558784163d16" already present on machine

openstack

kubelet

nova-metadata-0

Created

Created container: nova-metadata-log

openstack

kubelet

nova-metadata-0

Started

Started container nova-metadata-log

openstack

kubelet

nova-metadata-0

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-nova-api@sha256:faf711e1e5fa2ad74a73d3dfffd88f6312fb045cb69e9b7e6331558784163d16" already present on machine

openstack

kubelet

nova-metadata-0

Created

Created container: nova-metadata-metadata
(x3)

openstack

statefulset-controller

nova-api

SuccessfulCreate

create Pod nova-api-0 in StatefulSet nova-api successful

openstack

kubelet

nova-metadata-0

Started

Started container nova-metadata-metadata

openstack

kubelet

nova-api-0

Created

Created container: nova-api-log

openstack

kubelet

nova-api-0

Started

Started container nova-api-api

openstack

multus

nova-api-0

AddedInterface

Add eth0 [10.128.1.10/23] from ovn-kubernetes

openstack

kubelet

nova-api-0

Created

Created container: nova-api-api

openstack

kubelet

nova-api-0

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-nova-api@sha256:faf711e1e5fa2ad74a73d3dfffd88f6312fb045cb69e9b7e6331558784163d16" already present on machine

openstack

kubelet

nova-api-0

Started

Started container nova-api-log

openstack

kubelet

nova-api-0

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-nova-api@sha256:faf711e1e5fa2ad74a73d3dfffd88f6312fb045cb69e9b7e6331558784163d16" already present on machine

openstack

kubelet

nova-metadata-0

Unhealthy

Startup probe failed: Get "https://10.128.1.9:8775/": net/http: request canceled (Client.Timeout exceeded while awaiting headers)

openstack

kubelet

nova-metadata-0

Unhealthy

Startup probe failed: Get "https://10.128.1.9:8775/": net/http: request canceled (Client.Timeout exceeded while awaiting headers)

openstack

kubelet

nova-api-0

Unhealthy

Startup probe failed: Get "https://10.128.1.10:8774/": net/http: request canceled (Client.Timeout exceeded while awaiting headers)

openstack

kubelet

nova-api-0

Unhealthy

Startup probe failed: Get "https://10.128.1.10:8774/": net/http: request canceled (Client.Timeout exceeded while awaiting headers)
(x5)

openstack

metallb-speaker

nova-metadata-internal

nodeAssigned

announcing from node "master-0" with protocol "layer2"
(x3)

openstack

metallb-speaker

nova-internal

nodeAssigned

announcing from node "master-0" with protocol "layer2"

openshift-machine-config-operator

machine-config-operator

machine-config-operator

ConfigMapUpdated

Updated ConfigMap/kube-rbac-proxy -n openshift-machine-config-operator: cause by changes in data.config-file.yaml

openstack

metal3-baremetal-controller

bmh0

ProvisioningStarted

Image provisioning started for http://192.168.32.10:6190/edpm-hardened-uefi.qcow2

openstack

metal3-baremetal-controller

bmh1

ProvisioningStarted

Image provisioning started for http://192.168.32.10:6191/edpm-hardened-uefi.qcow2

openshift-machine-config-operator

machine-config-operator

machine-config-operator

ConfigMapUpdated

Updated ConfigMap/kube-rbac-proxy -n openshift-machine-config-operator: cause by changes in data.config-file.yaml

openstack

metal3-baremetal-controller

bmh0

ProvisioningComplete

Image provisioning completed for http://192.168.32.10:6190/edpm-hardened-uefi.qcow2

openstack

metal3-baremetal-controller

bmh1

ProvisioningComplete

Image provisioning completed for http://192.168.32.10:6191/edpm-hardened-uefi.qcow2

openstack

job-controller

bootstrap-dataplane-step-1-edpm-b

SuccessfulCreate

Created pod: bootstrap-dataplane-step-1-edpm-b-k2lgn

openstack

job-controller

bootstrap-dataplane-step-1-edpm-a

SuccessfulCreate

Created pod: bootstrap-dataplane-step-1-edpm-a-54cs7

openstack

multus

bootstrap-dataplane-step-1-edpm-b-k2lgn

AddedInterface

Add eth0 [10.128.1.12/23] from ovn-kubernetes

openstack

kubelet

bootstrap-dataplane-step-1-edpm-a-54cs7

Pulling

Pulling image "quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:a8da62d734643927e28a259fa0b64b81a85cf108a29154c4ddaf9cf271c05610"

openstack

multus

bootstrap-dataplane-step-1-edpm-a-54cs7

AddedInterface

Add ctlplane [192.168.122.30/24] from openstack/ctlplane

openstack

multus

bootstrap-dataplane-step-1-edpm-a-54cs7

AddedInterface

Add eth0 [10.128.1.11/23] from ovn-kubernetes

openstack

kubelet

bootstrap-dataplane-step-1-edpm-b-k2lgn

Pulling

Pulling image "quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:a8da62d734643927e28a259fa0b64b81a85cf108a29154c4ddaf9cf271c05610"

openstack

multus

bootstrap-dataplane-step-1-edpm-b-k2lgn

AddedInterface

Add ctlplane [192.168.122.31/24] from openstack/ctlplane

openstack

kubelet

bootstrap-dataplane-step-1-edpm-b-k2lgn

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:a8da62d734643927e28a259fa0b64b81a85cf108a29154c4ddaf9cf271c05610" in 11.396s (11.396s including waiting). Image size: 487820744 bytes.

openstack

kubelet

bootstrap-dataplane-step-1-edpm-a-54cs7

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:a8da62d734643927e28a259fa0b64b81a85cf108a29154c4ddaf9cf271c05610" in 11.284s (11.286s including waiting). Image size: 487820744 bytes.

openstack

kubelet

bootstrap-dataplane-step-1-edpm-a-54cs7

Created

Created container: bootstrap-dataplane-step-1-edpm-a

openstack

kubelet

bootstrap-dataplane-step-1-edpm-a-54cs7

Started

Started container bootstrap-dataplane-step-1-edpm-a

openstack

kubelet

bootstrap-dataplane-step-1-edpm-b-k2lgn

Created

Created container: bootstrap-dataplane-step-1-edpm-b

openstack

kubelet

bootstrap-dataplane-step-1-edpm-b-k2lgn

Started

Started container bootstrap-dataplane-step-1-edpm-b

openstack

job-controller

configure-network-dataplane-step-1-edpm-a

SuccessfulCreate

Created pod: configure-network-dataplane-step-1-edpm-a-4hslh

openstack

job-controller

bootstrap-dataplane-step-1-edpm-a

Completed

Job completed

openstack

multus

configure-network-dataplane-step-1-edpm-a-4hslh

AddedInterface

Add ctlplane [192.168.122.30/24] from openstack/ctlplane

openstack

multus

configure-network-dataplane-step-1-edpm-a-4hslh

AddedInterface

Add eth0 [10.128.1.13/23] from ovn-kubernetes

openstack

kubelet

configure-network-dataplane-step-1-edpm-a-4hslh

Pulling

Pulling image "quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:a8da62d734643927e28a259fa0b64b81a85cf108a29154c4ddaf9cf271c05610"

openstack

kubelet

configure-network-dataplane-step-1-edpm-a-4hslh

Started

Started container configure-network-dataplane-step-1-edpm-a

openstack

kubelet

configure-network-dataplane-step-1-edpm-a-4hslh

Created

Created container: configure-network-dataplane-step-1-edpm-a

openstack

kubelet

configure-network-dataplane-step-1-edpm-a-4hslh

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:a8da62d734643927e28a259fa0b64b81a85cf108a29154c4ddaf9cf271c05610" in 529ms (529ms including waiting). Image size: 487820744 bytes.

openstack

job-controller

bootstrap-dataplane-step-1-edpm-b

Completed

Job completed

openstack

job-controller

configure-network-dataplane-step-1-edpm-b

SuccessfulCreate

Created pod: configure-network-dataplane-step-1-edpm-b-5xnfd

openstack

multus

configure-network-dataplane-step-1-edpm-b-5xnfd

AddedInterface

Add eth0 [10.128.1.14/23] from ovn-kubernetes

openstack

multus

configure-network-dataplane-step-1-edpm-b-5xnfd

AddedInterface

Add ctlplane [192.168.122.31/24] from openstack/ctlplane

openstack

kubelet

configure-network-dataplane-step-1-edpm-b-5xnfd

Pulling

Pulling image "quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:a8da62d734643927e28a259fa0b64b81a85cf108a29154c4ddaf9cf271c05610"

openstack

kubelet

configure-network-dataplane-step-1-edpm-b-5xnfd

Created

Created container: configure-network-dataplane-step-1-edpm-b

openstack

kubelet

configure-network-dataplane-step-1-edpm-b-5xnfd

Started

Started container configure-network-dataplane-step-1-edpm-b

openstack

kubelet

configure-network-dataplane-step-1-edpm-b-5xnfd

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:a8da62d734643927e28a259fa0b64b81a85cf108a29154c4ddaf9cf271c05610" in 1.217s (1.217s including waiting). Image size: 487820744 bytes.

openshift-machine-config-operator

machine-config-operator

machine-config-operator

ConfigMapUpdated

Updated ConfigMap/kube-rbac-proxy -n openshift-machine-config-operator: cause by changes in data.config-file.yaml

openstack

job-controller

validate-network-dataplane-step-1-edpm-a

SuccessfulCreate

Created pod: validate-network-dataplane-step-1-edpm-a-b7h26

openstack

job-controller

configure-network-dataplane-step-1-edpm-a

Completed

Job completed

openstack

kubelet

validate-network-dataplane-step-1-edpm-a-b7h26

Pulling

Pulling image "quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:a8da62d734643927e28a259fa0b64b81a85cf108a29154c4ddaf9cf271c05610"

openstack

multus

validate-network-dataplane-step-1-edpm-a-b7h26

AddedInterface

Add eth0 [10.128.1.15/23] from ovn-kubernetes

openstack

multus

validate-network-dataplane-step-1-edpm-a-b7h26

AddedInterface

Add ctlplane [192.168.122.30/24] from openstack/ctlplane

openstack

kubelet

validate-network-dataplane-step-1-edpm-a-b7h26

Created

Created container: validate-network-dataplane-step-1-edpm-a

openstack

kubelet

validate-network-dataplane-step-1-edpm-a-b7h26

Started

Started container validate-network-dataplane-step-1-edpm-a

openstack

kubelet

validate-network-dataplane-step-1-edpm-a-b7h26

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:a8da62d734643927e28a259fa0b64b81a85cf108a29154c4ddaf9cf271c05610" in 454ms (454ms including waiting). Image size: 487820744 bytes.

openstack

multus

validate-network-dataplane-step-1-edpm-b-x6qvv

AddedInterface

Add eth0 [10.128.1.16/23] from ovn-kubernetes

openstack

job-controller

validate-network-dataplane-step-1-edpm-b

SuccessfulCreate

Created pod: validate-network-dataplane-step-1-edpm-b-x6qvv

openstack

job-controller

configure-network-dataplane-step-1-edpm-b

Completed

Job completed

openstack

multus

validate-network-dataplane-step-1-edpm-b-x6qvv

AddedInterface

Add ctlplane [192.168.122.31/24] from openstack/ctlplane

openstack

kubelet

validate-network-dataplane-step-1-edpm-b-x6qvv

Started

Started container validate-network-dataplane-step-1-edpm-b

openstack

kubelet

validate-network-dataplane-step-1-edpm-b-x6qvv

Created

Created container: validate-network-dataplane-step-1-edpm-b

openstack

kubelet

validate-network-dataplane-step-1-edpm-b-x6qvv

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:a8da62d734643927e28a259fa0b64b81a85cf108a29154c4ddaf9cf271c05610" in 499ms (499ms including waiting). Image size: 487820744 bytes.

openstack

kubelet

validate-network-dataplane-step-1-edpm-b-x6qvv

Pulling

Pulling image "quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:a8da62d734643927e28a259fa0b64b81a85cf108a29154c4ddaf9cf271c05610"

openstack

job-controller

validate-network-dataplane-step-1-edpm-a

Completed

Job completed

openstack

job-controller

install-os-dataplane-step-1-edpm-a

SuccessfulCreate

Created pod: install-os-dataplane-step-1-edpm-a-4zvvk

openstack

kubelet

install-os-dataplane-step-1-edpm-a-4zvvk

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:a8da62d734643927e28a259fa0b64b81a85cf108a29154c4ddaf9cf271c05610" in 456ms (456ms including waiting). Image size: 487820744 bytes.

openstack

kubelet

install-os-dataplane-step-1-edpm-a-4zvvk

Pulling

Pulling image "quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:a8da62d734643927e28a259fa0b64b81a85cf108a29154c4ddaf9cf271c05610"

openstack

multus

install-os-dataplane-step-1-edpm-a-4zvvk

AddedInterface

Add ctlplane [192.168.122.30/24] from openstack/ctlplane

openstack

multus

install-os-dataplane-step-1-edpm-a-4zvvk

AddedInterface

Add eth0 [10.128.1.17/23] from ovn-kubernetes

openstack

kubelet

install-os-dataplane-step-1-edpm-a-4zvvk

Started

Started container install-os-dataplane-step-1-edpm-a

openstack

kubelet

install-os-dataplane-step-1-edpm-a-4zvvk

Created

Created container: install-os-dataplane-step-1-edpm-a

openstack

job-controller

install-os-dataplane-step-1-edpm-b

SuccessfulCreate

Created pod: install-os-dataplane-step-1-edpm-b-6gmw5

openstack

job-controller

validate-network-dataplane-step-1-edpm-b

Completed

Job completed

openstack

kubelet

install-os-dataplane-step-1-edpm-b-6gmw5

Pulling

Pulling image "quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:a8da62d734643927e28a259fa0b64b81a85cf108a29154c4ddaf9cf271c05610"

openstack

multus

install-os-dataplane-step-1-edpm-b-6gmw5

AddedInterface

Add ctlplane [192.168.122.31/24] from openstack/ctlplane

openstack

multus

install-os-dataplane-step-1-edpm-b-6gmw5

AddedInterface

Add eth0 [10.128.1.18/23] from ovn-kubernetes

openstack

kubelet

install-os-dataplane-step-1-edpm-b-6gmw5

Started

Started container install-os-dataplane-step-1-edpm-b

openstack

kubelet

install-os-dataplane-step-1-edpm-b-6gmw5

Created

Created container: install-os-dataplane-step-1-edpm-b

openstack

kubelet

install-os-dataplane-step-1-edpm-b-6gmw5

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:a8da62d734643927e28a259fa0b64b81a85cf108a29154c4ddaf9cf271c05610" in 575ms (575ms including waiting). Image size: 487820744 bytes.

openstack

job-controller

configure-os-dataplane-step-1-edpm-a

SuccessfulCreate

Created pod: configure-os-dataplane-step-1-edpm-a-sgpm2

openstack

job-controller

install-os-dataplane-step-1-edpm-a

Completed

Job completed

openstack

multus

configure-os-dataplane-step-1-edpm-a-sgpm2

AddedInterface

Add eth0 [10.128.1.19/23] from ovn-kubernetes

openstack

multus

configure-os-dataplane-step-1-edpm-a-sgpm2

AddedInterface

Add ctlplane [192.168.122.30/24] from openstack/ctlplane

openstack

kubelet

configure-os-dataplane-step-1-edpm-a-sgpm2

Pulling

Pulling image "quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:a8da62d734643927e28a259fa0b64b81a85cf108a29154c4ddaf9cf271c05610"

openstack

kubelet

configure-os-dataplane-step-1-edpm-a-sgpm2

Started

Started container configure-os-dataplane-step-1-edpm-a

openstack

kubelet

configure-os-dataplane-step-1-edpm-a-sgpm2

Created

Created container: configure-os-dataplane-step-1-edpm-a

openstack

kubelet

configure-os-dataplane-step-1-edpm-a-sgpm2

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:a8da62d734643927e28a259fa0b64b81a85cf108a29154c4ddaf9cf271c05610" in 560ms (560ms including waiting). Image size: 487820744 bytes.

openstack

job-controller

configure-os-dataplane-step-1-edpm-b

SuccessfulCreate

Created pod: configure-os-dataplane-step-1-edpm-b-jdmpl

openstack

job-controller

install-os-dataplane-step-1-edpm-b

Completed

Job completed

openstack

kubelet

configure-os-dataplane-step-1-edpm-b-jdmpl

Pulling

Pulling image "quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:a8da62d734643927e28a259fa0b64b81a85cf108a29154c4ddaf9cf271c05610"

openstack

multus

configure-os-dataplane-step-1-edpm-b-jdmpl

AddedInterface

Add ctlplane [192.168.122.31/24] from openstack/ctlplane

openstack

multus

configure-os-dataplane-step-1-edpm-b-jdmpl

AddedInterface

Add eth0 [10.128.1.20/23] from ovn-kubernetes

openstack

kubelet

configure-os-dataplane-step-1-edpm-b-jdmpl

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:a8da62d734643927e28a259fa0b64b81a85cf108a29154c4ddaf9cf271c05610" in 923ms (923ms including waiting). Image size: 487820744 bytes.

openstack

kubelet

configure-os-dataplane-step-1-edpm-b-jdmpl

Created

Created container: configure-os-dataplane-step-1-edpm-b

openstack

kubelet

configure-os-dataplane-step-1-edpm-b-jdmpl

Started

Started container configure-os-dataplane-step-1-edpm-b

openstack

job-controller

configure-os-dataplane-step-1-edpm-a

Completed

Job completed

openstack

job-controller

ssh-known-hosts-dataplane-step-1

SuccessfulCreate

Created pod: ssh-known-hosts-dataplane-step-1-kt96m

openstack

kubelet

ssh-known-hosts-dataplane-step-1-kt96m

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:a8da62d734643927e28a259fa0b64b81a85cf108a29154c4ddaf9cf271c05610" in 652ms (652ms including waiting). Image size: 487820744 bytes.

openstack

multus

ssh-known-hosts-dataplane-step-1-kt96m

AddedInterface

Add ctlplane [192.168.122.30/24] from openstack/ctlplane

openstack

kubelet

ssh-known-hosts-dataplane-step-1-kt96m

Pulling

Pulling image "quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:a8da62d734643927e28a259fa0b64b81a85cf108a29154c4ddaf9cf271c05610"

openstack

multus

ssh-known-hosts-dataplane-step-1-kt96m

AddedInterface

Add eth0 [10.128.1.21/23] from ovn-kubernetes

openstack

kubelet

ssh-known-hosts-dataplane-step-1-kt96m

Created

Created container: ssh-known-hosts-dataplane-step-1

openstack

kubelet

ssh-known-hosts-dataplane-step-1-kt96m

Started

Started container ssh-known-hosts-dataplane-step-1

openstack

job-controller

run-os-dataplane-step-1-edpm-b

SuccessfulCreate

Created pod: run-os-dataplane-step-1-edpm-b-4xvvh

openstack

job-controller

configure-os-dataplane-step-1-edpm-b

Completed

Job completed

openstack

multus

run-os-dataplane-step-1-edpm-b-4xvvh

AddedInterface

Add eth0 [10.128.1.22/23] from ovn-kubernetes

openstack

kubelet

run-os-dataplane-step-1-edpm-b-4xvvh

Created

Created container: run-os-dataplane-step-1-edpm-b

openstack

kubelet

run-os-dataplane-step-1-edpm-b-4xvvh

Pulling

Pulling image "quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:a8da62d734643927e28a259fa0b64b81a85cf108a29154c4ddaf9cf271c05610"

openstack

kubelet

run-os-dataplane-step-1-edpm-b-4xvvh

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:a8da62d734643927e28a259fa0b64b81a85cf108a29154c4ddaf9cf271c05610" in 401ms (401ms including waiting). Image size: 487820744 bytes.

openstack

kubelet

run-os-dataplane-step-1-edpm-b-4xvvh

Started

Started container run-os-dataplane-step-1-edpm-b

openstack

multus

run-os-dataplane-step-1-edpm-b-4xvvh

AddedInterface

Add ctlplane [192.168.122.31/24] from openstack/ctlplane

openstack

job-controller

ssh-known-hosts-dataplane-step-1

Completed

Job completed

openstack

job-controller

run-os-dataplane-step-1-edpm-a

SuccessfulCreate

Created pod: run-os-dataplane-step-1-edpm-a-mtbhq

openstack

kubelet

run-os-dataplane-step-1-edpm-a-mtbhq

Pulling

Pulling image "quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:a8da62d734643927e28a259fa0b64b81a85cf108a29154c4ddaf9cf271c05610"

openstack

multus

run-os-dataplane-step-1-edpm-a-mtbhq

AddedInterface

Add eth0 [10.128.1.23/23] from ovn-kubernetes

openstack

kubelet

run-os-dataplane-step-1-edpm-a-mtbhq

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:a8da62d734643927e28a259fa0b64b81a85cf108a29154c4ddaf9cf271c05610" in 544ms (544ms including waiting). Image size: 487820744 bytes.

openstack

multus

run-os-dataplane-step-1-edpm-a-mtbhq

AddedInterface

Add ctlplane [192.168.122.30/24] from openstack/ctlplane

openstack

kubelet

run-os-dataplane-step-1-edpm-a-mtbhq

Started

Started container run-os-dataplane-step-1-edpm-a

openstack

kubelet

run-os-dataplane-step-1-edpm-a-mtbhq

Created

Created container: run-os-dataplane-step-1-edpm-a

openstack

job-controller

run-os-dataplane-step-1-edpm-b

Completed

Job completed

openstack

job-controller

reboot-os-dataplane-step-1-edpm-b

SuccessfulCreate

Created pod: reboot-os-dataplane-step-1-edpm-b-9dw2p

openstack

kubelet

reboot-os-dataplane-step-1-edpm-b-9dw2p

Pulling

Pulling image "quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:a8da62d734643927e28a259fa0b64b81a85cf108a29154c4ddaf9cf271c05610"

openstack

multus

reboot-os-dataplane-step-1-edpm-b-9dw2p

AddedInterface

Add ctlplane [192.168.122.31/24] from openstack/ctlplane

openstack

multus

reboot-os-dataplane-step-1-edpm-b-9dw2p

AddedInterface

Add eth0 [10.128.1.24/23] from ovn-kubernetes

openstack

kubelet

reboot-os-dataplane-step-1-edpm-b-9dw2p

Started

Started container reboot-os-dataplane-step-1-edpm-b

openstack

kubelet

reboot-os-dataplane-step-1-edpm-b-9dw2p

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:a8da62d734643927e28a259fa0b64b81a85cf108a29154c4ddaf9cf271c05610" in 511ms (511ms including waiting). Image size: 487820744 bytes.

openstack

kubelet

reboot-os-dataplane-step-1-edpm-b-9dw2p

Created

Created container: reboot-os-dataplane-step-1-edpm-b

openstack

job-controller

reboot-os-dataplane-step-1-edpm-a

SuccessfulCreate

Created pod: reboot-os-dataplane-step-1-edpm-a-rwc7j

openstack

job-controller

run-os-dataplane-step-1-edpm-a

Completed

Job completed

openstack

multus

reboot-os-dataplane-step-1-edpm-a-rwc7j

AddedInterface

Add ctlplane [192.168.122.30/24] from openstack/ctlplane

openstack

kubelet

reboot-os-dataplane-step-1-edpm-a-rwc7j

Pulling

Pulling image "quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:a8da62d734643927e28a259fa0b64b81a85cf108a29154c4ddaf9cf271c05610"

openstack

multus

reboot-os-dataplane-step-1-edpm-a-rwc7j

AddedInterface

Add eth0 [10.128.1.25/23] from ovn-kubernetes

openstack

kubelet

reboot-os-dataplane-step-1-edpm-a-rwc7j

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:a8da62d734643927e28a259fa0b64b81a85cf108a29154c4ddaf9cf271c05610" in 424ms (424ms including waiting). Image size: 487820744 bytes.

openstack

kubelet

reboot-os-dataplane-step-1-edpm-a-rwc7j

Created

Created container: reboot-os-dataplane-step-1-edpm-a

openstack

kubelet

reboot-os-dataplane-step-1-edpm-a-rwc7j

Started

Started container reboot-os-dataplane-step-1-edpm-a

openstack

job-controller

reboot-os-dataplane-step-1-edpm-b

Completed

Job completed

openstack

job-controller

reboot-os-dataplane-step-1-edpm-a

Completed

Job completed

openstack

cert-manager-certificates-key-manager

ovn-default-edpm-compute-a-0

Generated

Stored new private key in temporary Secret resource "ovn-default-edpm-compute-a-0-xk7z5"

openstack

cert-manager-certificaterequests-issuer-ca

neutron-metadata-default-edpm-compute-a-0-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-ca

ovn-default-edpm-compute-a-0-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-selfsigned

ovn-default-edpm-compute-a-0-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificates-request-manager

neutron-metadata-default-edpm-compute-a-0

Requested

Created new CertificateRequest resource "neutron-metadata-default-edpm-compute-a-0-1"

openstack

cert-manager-certificates-key-manager

neutron-metadata-default-edpm-compute-a-0

Generated

Stored new private key in temporary Secret resource "neutron-metadata-default-edpm-compute-a-0-mz6z9"

openstack

cert-manager-certificates-trigger

neutron-metadata-default-edpm-compute-a-0

Issuing

Issuing certificate as Secret does not exist

openstack

cert-manager-certificates-trigger

libvirt-default-edpm-compute-a-0

Issuing

Issuing certificate as Secret does not exist

openstack

cert-manager-certificaterequests-issuer-ca

neutron-metadata-default-edpm-compute-a-0-1

CertificateIssued

Certificate fetched from issuer successfully

openstack

cert-manager-certificaterequests-approver

neutron-metadata-default-edpm-compute-a-0-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack

cert-manager-certificaterequests-issuer-selfsigned

neutron-metadata-default-edpm-compute-a-0-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-vault

ovn-default-edpm-compute-a-0-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-venafi

neutron-metadata-default-edpm-compute-a-0-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificates-issuing

neutron-metadata-default-edpm-compute-a-0

Issuing

The certificate has been successfully issued

openstack

cert-manager-certificaterequests-issuer-vault

neutron-metadata-default-edpm-compute-a-0-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-acme

neutron-metadata-default-edpm-compute-a-0-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-acme

ovn-default-edpm-compute-a-0-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-venafi

ovn-default-edpm-compute-a-0-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-approver

ovn-default-edpm-compute-a-0-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack

cert-manager-certificaterequests-issuer-ca

ovn-default-edpm-compute-a-0-1

CertificateIssued

Certificate fetched from issuer successfully

openstack

cert-manager-certificates-trigger

ovn-default-edpm-compute-a-0

Issuing

Issuing certificate as Secret does not exist

openstack

cert-manager-certificates-issuing

ovn-default-edpm-compute-a-0

Issuing

The certificate has been successfully issued

openstack

cert-manager-certificates-request-manager

ovn-default-edpm-compute-a-0

Requested

Created new CertificateRequest resource "ovn-default-edpm-compute-a-0-1"

openstack

cert-manager-certificaterequests-issuer-ca

libvirt-default-edpm-compute-a-0-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-acme

libvirt-default-edpm-compute-a-0-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-selfsigned

libvirt-default-edpm-compute-a-0-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificates-issuing

libvirt-default-edpm-compute-a-0

Issuing

The certificate has been successfully issued

openstack

cert-manager-certificates-request-manager

libvirt-default-edpm-compute-a-0

Requested

Created new CertificateRequest resource "libvirt-default-edpm-compute-a-0-1"

openstack

cert-manager-certificates-key-manager

libvirt-default-edpm-compute-a-0

Generated

Stored new private key in temporary Secret resource "libvirt-default-edpm-compute-a-0-ds6rj"

openstack

cert-manager-certificaterequests-issuer-venafi

libvirt-default-edpm-compute-a-0-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-ca

libvirt-default-edpm-compute-a-0-1

CertificateIssued

Certificate fetched from issuer successfully

openstack

cert-manager-certificaterequests-approver

libvirt-default-edpm-compute-a-0-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack

cert-manager-certificaterequests-issuer-vault

libvirt-default-edpm-compute-a-0-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-approver

ovn-default-edpm-compute-b-0-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack

cert-manager-certificaterequests-issuer-acme

ovn-default-edpm-compute-b-0-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-ca

ovn-default-edpm-compute-b-0-1

CertificateIssued

Certificate fetched from issuer successfully

openstack

cert-manager-certificates-trigger

ovn-default-edpm-compute-b-0

Issuing

Issuing certificate as Secret does not exist

openstack

cert-manager-certificates-key-manager

ovn-default-edpm-compute-b-0

Generated

Stored new private key in temporary Secret resource "ovn-default-edpm-compute-b-0-mqj8d"

openstack

cert-manager-certificaterequests-issuer-selfsigned

ovn-default-edpm-compute-b-0-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificates-request-manager

ovn-default-edpm-compute-b-0

Requested

Created new CertificateRequest resource "ovn-default-edpm-compute-b-0-1"

openstack

cert-manager-certificates-issuing

ovn-default-edpm-compute-b-0

Issuing

The certificate has been successfully issued

openstack

cert-manager-certificaterequests-issuer-ca

ovn-default-edpm-compute-b-0-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-venafi

ovn-default-edpm-compute-b-0-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-vault

ovn-default-edpm-compute-b-0-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificates-request-manager

neutron-metadata-default-edpm-compute-b-0

Requested

Created new CertificateRequest resource "neutron-metadata-default-edpm-compute-b-0-1"

openstack

cert-manager-certificates-key-manager

neutron-metadata-default-edpm-compute-b-0

Generated

Stored new private key in temporary Secret resource "neutron-metadata-default-edpm-compute-b-0-wqtcz"

openstack

cert-manager-certificaterequests-issuer-acme

neutron-metadata-default-edpm-compute-b-0-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-venafi

neutron-metadata-default-edpm-compute-b-0-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-vault

neutron-metadata-default-edpm-compute-b-0-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-selfsigned

neutron-metadata-default-edpm-compute-b-0-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-ca

neutron-metadata-default-edpm-compute-b-0-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificates-trigger

neutron-metadata-default-edpm-compute-b-0

Issuing

Issuing certificate as Secret does not exist

openstack

cert-manager-certificaterequests-approver

neutron-metadata-default-edpm-compute-b-0-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack

cert-manager-certificaterequests-issuer-ca

neutron-metadata-default-edpm-compute-b-0-1

CertificateIssued

Certificate fetched from issuer successfully

openstack

cert-manager-certificates-trigger

libvirt-default-edpm-compute-b-0

Issuing

Issuing certificate as Secret does not exist

openstack

cert-manager-certificates-issuing

neutron-metadata-default-edpm-compute-b-0

Issuing

The certificate has been successfully issued

openstack

cert-manager-certificaterequests-approver

libvirt-default-edpm-compute-b-0-1

cert-manager.io

Certificate request has been approved by cert-manager.io

openstack

cert-manager-certificates-key-manager

libvirt-default-edpm-compute-b-0

Generated

Stored new private key in temporary Secret resource "libvirt-default-edpm-compute-b-0-fbl72"

openstack

cert-manager-certificaterequests-issuer-ca

libvirt-default-edpm-compute-b-0-1

CertificateIssued

Certificate fetched from issuer successfully

openstack

cert-manager-certificaterequests-issuer-vault

libvirt-default-edpm-compute-b-0-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-acme

libvirt-default-edpm-compute-b-0-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-venafi

libvirt-default-edpm-compute-b-0-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificates-issuing

libvirt-default-edpm-compute-b-0

Issuing

The certificate has been successfully issued

openstack

cert-manager-certificates-request-manager

libvirt-default-edpm-compute-b-0

Requested

Created new CertificateRequest resource "libvirt-default-edpm-compute-b-0-1"

openstack

cert-manager-certificaterequests-issuer-ca

libvirt-default-edpm-compute-b-0-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

cert-manager-certificaterequests-issuer-selfsigned

libvirt-default-edpm-compute-b-0-1

WaitingForApproval

Not signing CertificateRequest until it is Approved

openstack

multus

install-certs-dataplane-step-2-edpm-a-5xzqr

AddedInterface

Add eth0 [10.128.1.26/23] from ovn-kubernetes

openstack

job-controller

install-certs-dataplane-step-2-edpm-a

SuccessfulCreate

Created pod: install-certs-dataplane-step-2-edpm-a-5xzqr

openstack

job-controller

install-certs-dataplane-step-2-edpm-b

SuccessfulCreate

Created pod: install-certs-dataplane-step-2-edpm-b-w5v2r

openstack

kubelet

install-certs-dataplane-step-2-edpm-a-5xzqr

Created

Created container: install-certs-dataplane-step-2-edpm-a

openstack

kubelet

install-certs-dataplane-step-2-edpm-a-5xzqr

Started

Started container install-certs-dataplane-step-2-edpm-a

openstack

kubelet

install-certs-dataplane-step-2-edpm-b-w5v2r

Pulling

Pulling image "quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:a8da62d734643927e28a259fa0b64b81a85cf108a29154c4ddaf9cf271c05610"

openstack

multus

install-certs-dataplane-step-2-edpm-a-5xzqr

AddedInterface

Add ctlplane [192.168.122.30/24] from openstack/ctlplane

openstack

kubelet

install-certs-dataplane-step-2-edpm-a-5xzqr

Pulling

Pulling image "quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:a8da62d734643927e28a259fa0b64b81a85cf108a29154c4ddaf9cf271c05610"

openstack

kubelet

install-certs-dataplane-step-2-edpm-a-5xzqr

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:a8da62d734643927e28a259fa0b64b81a85cf108a29154c4ddaf9cf271c05610" in 444ms (444ms including waiting). Image size: 487820744 bytes.

openstack

kubelet

install-certs-dataplane-step-2-edpm-b-w5v2r

Created

Created container: install-certs-dataplane-step-2-edpm-b

openstack

multus

install-certs-dataplane-step-2-edpm-b-w5v2r

AddedInterface

Add ctlplane [192.168.122.31/24] from openstack/ctlplane

openstack

kubelet

install-certs-dataplane-step-2-edpm-b-w5v2r

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:a8da62d734643927e28a259fa0b64b81a85cf108a29154c4ddaf9cf271c05610" in 451ms (451ms including waiting). Image size: 487820744 bytes.

openstack

multus

install-certs-dataplane-step-2-edpm-b-w5v2r

AddedInterface

Add eth0 [10.128.1.27/23] from ovn-kubernetes

openstack

kubelet

install-certs-dataplane-step-2-edpm-b-w5v2r

Started

Started container install-certs-dataplane-step-2-edpm-b

openstack

job-controller

install-certs-dataplane-step-2-edpm-a

Completed

Job completed

openstack

job-controller

ovn-dataplane-step-2-edpm-a

SuccessfulCreate

Created pod: ovn-dataplane-step-2-edpm-a-r2hdg

openstack

multus

ovn-dataplane-step-2-edpm-a-r2hdg

AddedInterface

Add eth0 [10.128.1.28/23] from ovn-kubernetes

openstack

job-controller

install-certs-dataplane-step-2-edpm-b

Completed

Job completed

openstack

kubelet

ovn-dataplane-step-2-edpm-a-r2hdg

Started

Started container ovn-dataplane-step-2-edpm-a

openstack

kubelet

ovn-dataplane-step-2-edpm-a-r2hdg

Created

Created container: ovn-dataplane-step-2-edpm-a

openstack

kubelet

ovn-dataplane-step-2-edpm-a-r2hdg

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:a8da62d734643927e28a259fa0b64b81a85cf108a29154c4ddaf9cf271c05610" in 448ms (448ms including waiting). Image size: 487820744 bytes.

openstack

kubelet

ovn-dataplane-step-2-edpm-a-r2hdg

Pulling

Pulling image "quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:a8da62d734643927e28a259fa0b64b81a85cf108a29154c4ddaf9cf271c05610"

openstack

multus

ovn-dataplane-step-2-edpm-a-r2hdg

AddedInterface

Add ctlplane [192.168.122.30/24] from openstack/ctlplane

openstack

job-controller

ovn-dataplane-step-2-edpm-b

SuccessfulCreate

Created pod: ovn-dataplane-step-2-edpm-b-tm4wb

openstack

kubelet

ovn-dataplane-step-2-edpm-b-tm4wb

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:a8da62d734643927e28a259fa0b64b81a85cf108a29154c4ddaf9cf271c05610" in 442ms (442ms including waiting). Image size: 487820744 bytes.

openstack

kubelet

ovn-dataplane-step-2-edpm-b-tm4wb

Pulling

Pulling image "quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:a8da62d734643927e28a259fa0b64b81a85cf108a29154c4ddaf9cf271c05610"

openstack

multus

ovn-dataplane-step-2-edpm-b-tm4wb

AddedInterface

Add ctlplane [192.168.122.31/24] from openstack/ctlplane

openstack

multus

ovn-dataplane-step-2-edpm-b-tm4wb

AddedInterface

Add eth0 [10.128.1.29/23] from ovn-kubernetes

openstack

kubelet

ovn-dataplane-step-2-edpm-b-tm4wb

Started

Started container ovn-dataplane-step-2-edpm-b

openstack

kubelet

ovn-dataplane-step-2-edpm-b-tm4wb

Created

Created container: ovn-dataplane-step-2-edpm-b

openshift-machine-config-operator

machine-config-operator

machine-config-operator

ConfigMapUpdated

Updated ConfigMap/kube-rbac-proxy -n openshift-machine-config-operator: cause by changes in data.config-file.yaml

openstack

job-controller

ovn-dataplane-step-2-edpm-a

Completed

Job completed

openstack

job-controller

neutron-metadata-dataplane-step-2-edpm-a

SuccessfulCreate

Created pod: neutron-metadata-dataplane-step-2-edpm-a-jnw8j

openstack

multus

neutron-metadata-dataplane-step-2-edpm-a-jnw8j

AddedInterface

Add eth0 [10.128.1.30/23] from ovn-kubernetes

openstack

multus

neutron-metadata-dataplane-step-2-edpm-a-jnw8j

AddedInterface

Add ctlplane [192.168.122.30/24] from openstack/ctlplane

openstack

kubelet

neutron-metadata-dataplane-step-2-edpm-a-jnw8j

Pulling

Pulling image "quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:a8da62d734643927e28a259fa0b64b81a85cf108a29154c4ddaf9cf271c05610"

openstack

job-controller

ovn-dataplane-step-2-edpm-b

Completed

Job completed

openstack

job-controller

neutron-metadata-dataplane-step-2-edpm-b

SuccessfulCreate

Created pod: neutron-metadata-dataplane-step-2-edpm-b-cm55j

openstack

kubelet

neutron-metadata-dataplane-step-2-edpm-b-cm55j

Pulling

Pulling image "quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:a8da62d734643927e28a259fa0b64b81a85cf108a29154c4ddaf9cf271c05610"

openstack

kubelet

neutron-metadata-dataplane-step-2-edpm-a-jnw8j

Started

Started container neutron-metadata-dataplane-step-2-edpm-a

openstack

kubelet

neutron-metadata-dataplane-step-2-edpm-b-cm55j

Started

Started container neutron-metadata-dataplane-step-2-edpm-b

openstack

kubelet

neutron-metadata-dataplane-step-2-edpm-b-cm55j

Created

Created container: neutron-metadata-dataplane-step-2-edpm-b

openstack

kubelet

neutron-metadata-dataplane-step-2-edpm-b-cm55j

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:a8da62d734643927e28a259fa0b64b81a85cf108a29154c4ddaf9cf271c05610" in 460ms (460ms including waiting). Image size: 487820744 bytes.

openstack

multus

neutron-metadata-dataplane-step-2-edpm-b-cm55j

AddedInterface

Add ctlplane [192.168.122.31/24] from openstack/ctlplane

openstack

multus

neutron-metadata-dataplane-step-2-edpm-b-cm55j

AddedInterface

Add eth0 [10.128.1.31/23] from ovn-kubernetes

openstack

kubelet

neutron-metadata-dataplane-step-2-edpm-a-jnw8j

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:a8da62d734643927e28a259fa0b64b81a85cf108a29154c4ddaf9cf271c05610" in 722ms (722ms including waiting). Image size: 487820744 bytes.

openstack

kubelet

neutron-metadata-dataplane-step-2-edpm-a-jnw8j

Created

Created container: neutron-metadata-dataplane-step-2-edpm-a

openstack

job-controller

neutron-metadata-dataplane-step-2-edpm-b

SuccessfulCreate

Created pod: neutron-metadata-dataplane-step-2-edpm-b-4vq4r

openstack

job-controller

neutron-metadata-dataplane-step-2-edpm-a

SuccessfulCreate

Created pod: neutron-metadata-dataplane-step-2-edpm-a-kwdqb

openstack

multus

neutron-metadata-dataplane-step-2-edpm-a-kwdqb

AddedInterface

Add eth0 [10.128.1.33/23] from ovn-kubernetes

openstack

multus

neutron-metadata-dataplane-step-2-edpm-b-4vq4r

AddedInterface

Add eth0 [10.128.1.32/23] from ovn-kubernetes

openstack

multus

neutron-metadata-dataplane-step-2-edpm-b-4vq4r

AddedInterface

Add ctlplane [192.168.122.30/24] from openstack/ctlplane

openstack

kubelet

neutron-metadata-dataplane-step-2-edpm-b-4vq4r

Pulling

Pulling image "quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:a8da62d734643927e28a259fa0b64b81a85cf108a29154c4ddaf9cf271c05610"

openstack

multus

neutron-metadata-dataplane-step-2-edpm-a-kwdqb

AddedInterface

Add ctlplane [192.168.122.31/24] from openstack/ctlplane

openstack

kubelet

neutron-metadata-dataplane-step-2-edpm-a-kwdqb

Pulling

Pulling image "quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:a8da62d734643927e28a259fa0b64b81a85cf108a29154c4ddaf9cf271c05610"

openstack

kubelet

neutron-metadata-dataplane-step-2-edpm-b-4vq4r

Created

Created container: neutron-metadata-dataplane-step-2-edpm-b

openstack

kubelet

neutron-metadata-dataplane-step-2-edpm-b-4vq4r

Started

Started container neutron-metadata-dataplane-step-2-edpm-b

openstack

kubelet

neutron-metadata-dataplane-step-2-edpm-a-kwdqb

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:a8da62d734643927e28a259fa0b64b81a85cf108a29154c4ddaf9cf271c05610" in 660ms (660ms including waiting). Image size: 487820744 bytes.

openstack

kubelet

neutron-metadata-dataplane-step-2-edpm-a-kwdqb

Created

Created container: neutron-metadata-dataplane-step-2-edpm-a

openstack

kubelet

neutron-metadata-dataplane-step-2-edpm-a-kwdqb

Started

Started container neutron-metadata-dataplane-step-2-edpm-a

openstack

kubelet

neutron-metadata-dataplane-step-2-edpm-b-4vq4r

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:a8da62d734643927e28a259fa0b64b81a85cf108a29154c4ddaf9cf271c05610" in 1.096s (1.096s including waiting). Image size: 487820744 bytes.

openstack

job-controller

neutron-metadata-dataplane-step-2-edpm-b

SuccessfulCreate

Created pod: neutron-metadata-dataplane-step-2-edpm-b-xbtq9

openstack

multus

neutron-metadata-dataplane-step-2-edpm-a-48qbs

AddedInterface

Add ctlplane [192.168.122.30/24] from openstack/ctlplane

openstack

multus

neutron-metadata-dataplane-step-2-edpm-a-48qbs

AddedInterface

Add eth0 [10.128.1.34/23] from ovn-kubernetes

openstack

multus

neutron-metadata-dataplane-step-2-edpm-b-xbtq9

AddedInterface

Add eth0 [10.128.1.35/23] from ovn-kubernetes

openstack

kubelet

neutron-metadata-dataplane-step-2-edpm-a-48qbs

Pulling

Pulling image "quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:a8da62d734643927e28a259fa0b64b81a85cf108a29154c4ddaf9cf271c05610"

openstack

job-controller

neutron-metadata-dataplane-step-2-edpm-a

SuccessfulCreate

Created pod: neutron-metadata-dataplane-step-2-edpm-a-48qbs

openstack

kubelet

neutron-metadata-dataplane-step-2-edpm-b-xbtq9

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:a8da62d734643927e28a259fa0b64b81a85cf108a29154c4ddaf9cf271c05610" in 509ms (509ms including waiting). Image size: 487820744 bytes.

openstack

kubelet

neutron-metadata-dataplane-step-2-edpm-a-48qbs

Started

Started container neutron-metadata-dataplane-step-2-edpm-a

openstack

kubelet

neutron-metadata-dataplane-step-2-edpm-a-48qbs

Created

Created container: neutron-metadata-dataplane-step-2-edpm-a

openstack

kubelet

neutron-metadata-dataplane-step-2-edpm-a-48qbs

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:a8da62d734643927e28a259fa0b64b81a85cf108a29154c4ddaf9cf271c05610" in 652ms (652ms including waiting). Image size: 487820744 bytes.

openstack

kubelet

neutron-metadata-dataplane-step-2-edpm-b-xbtq9

Created

Created container: neutron-metadata-dataplane-step-2-edpm-b

openstack

kubelet

neutron-metadata-dataplane-step-2-edpm-b-xbtq9

Pulling

Pulling image "quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:a8da62d734643927e28a259fa0b64b81a85cf108a29154c4ddaf9cf271c05610"

openstack

kubelet

neutron-metadata-dataplane-step-2-edpm-b-xbtq9

Started

Started container neutron-metadata-dataplane-step-2-edpm-b

openstack

multus

neutron-metadata-dataplane-step-2-edpm-b-xbtq9

AddedInterface

Add ctlplane [192.168.122.31/24] from openstack/ctlplane

openshift-machine-config-operator

machine-config-operator

machine-config-operator

ConfigMapUpdated

Updated ConfigMap/kube-rbac-proxy -n openshift-machine-config-operator: cause by changes in data.config-file.yaml

openstack

multus

neutron-metadata-dataplane-step-2-edpm-b-qvqxd

AddedInterface

Add ctlplane [192.168.122.30/24] from openstack/ctlplane

openstack

multus

neutron-metadata-dataplane-step-2-edpm-b-qvqxd

AddedInterface

Add eth0 [10.128.1.36/23] from ovn-kubernetes

openstack

job-controller

neutron-metadata-dataplane-step-2-edpm-b

SuccessfulCreate

Created pod: neutron-metadata-dataplane-step-2-edpm-b-qvqxd

openstack

kubelet

neutron-metadata-dataplane-step-2-edpm-b-qvqxd

Started

Started container neutron-metadata-dataplane-step-2-edpm-b

openstack

kubelet

neutron-metadata-dataplane-step-2-edpm-b-qvqxd

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:a8da62d734643927e28a259fa0b64b81a85cf108a29154c4ddaf9cf271c05610" in 565ms (565ms including waiting). Image size: 487820744 bytes.

openstack

kubelet

neutron-metadata-dataplane-step-2-edpm-b-qvqxd

Created

Created container: neutron-metadata-dataplane-step-2-edpm-b

openstack

kubelet

neutron-metadata-dataplane-step-2-edpm-b-qvqxd

Pulling

Pulling image "quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:a8da62d734643927e28a259fa0b64b81a85cf108a29154c4ddaf9cf271c05610"

openstack

multus

neutron-metadata-dataplane-step-2-edpm-a-kf242

AddedInterface

Add eth0 [10.128.1.37/23] from ovn-kubernetes

openstack

multus

neutron-metadata-dataplane-step-2-edpm-a-kf242

AddedInterface

Add ctlplane [192.168.122.31/24] from openstack/ctlplane

openstack

job-controller

neutron-metadata-dataplane-step-2-edpm-a

SuccessfulCreate

Created pod: neutron-metadata-dataplane-step-2-edpm-a-kf242

openstack

kubelet

neutron-metadata-dataplane-step-2-edpm-a-kf242

Pulling

Pulling image "quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:a8da62d734643927e28a259fa0b64b81a85cf108a29154c4ddaf9cf271c05610"

openstack

kubelet

neutron-metadata-dataplane-step-2-edpm-a-kf242

Started

Started container neutron-metadata-dataplane-step-2-edpm-a

openstack

kubelet

neutron-metadata-dataplane-step-2-edpm-a-kf242

Created

Created container: neutron-metadata-dataplane-step-2-edpm-a

openstack

kubelet

neutron-metadata-dataplane-step-2-edpm-a-kf242

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:a8da62d734643927e28a259fa0b64b81a85cf108a29154c4ddaf9cf271c05610" in 388ms (388ms including waiting). Image size: 487820744 bytes.

openstack

multus

neutron-metadata-dataplane-step-2-edpm-b-fq5nx

AddedInterface

Add ctlplane [192.168.122.30/24] from openstack/ctlplane

openstack

multus

neutron-metadata-dataplane-step-2-edpm-b-fq5nx

AddedInterface

Add eth0 [10.128.1.38/23] from ovn-kubernetes

openstack

job-controller

neutron-metadata-dataplane-step-2-edpm-b

SuccessfulCreate

Created pod: neutron-metadata-dataplane-step-2-edpm-b-fq5nx

openstack

kubelet

neutron-metadata-dataplane-step-2-edpm-b-fq5nx

Created

Created container: neutron-metadata-dataplane-step-2-edpm-b

openstack

kubelet

neutron-metadata-dataplane-step-2-edpm-b-fq5nx

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:a8da62d734643927e28a259fa0b64b81a85cf108a29154c4ddaf9cf271c05610" in 491ms (491ms including waiting). Image size: 487820744 bytes.

openstack

kubelet

neutron-metadata-dataplane-step-2-edpm-b-fq5nx

Pulling

Pulling image "quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:a8da62d734643927e28a259fa0b64b81a85cf108a29154c4ddaf9cf271c05610"

openstack

kubelet

neutron-metadata-dataplane-step-2-edpm-b-fq5nx

Started

Started container neutron-metadata-dataplane-step-2-edpm-b

openstack

multus

neutron-metadata-dataplane-step-2-edpm-a-tk7z9

AddedInterface

Add ctlplane [192.168.122.31/24] from openstack/ctlplane

openstack

multus

neutron-metadata-dataplane-step-2-edpm-a-tk7z9

AddedInterface

Add eth0 [10.128.1.39/23] from ovn-kubernetes

openstack

job-controller

neutron-metadata-dataplane-step-2-edpm-a

SuccessfulCreate

Created pod: neutron-metadata-dataplane-step-2-edpm-a-tk7z9

openstack

kubelet

neutron-metadata-dataplane-step-2-edpm-a-tk7z9

Pulling

Pulling image "quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:a8da62d734643927e28a259fa0b64b81a85cf108a29154c4ddaf9cf271c05610"

openstack

kubelet

neutron-metadata-dataplane-step-2-edpm-a-tk7z9

Started

Started container neutron-metadata-dataplane-step-2-edpm-a

openstack

kubelet

neutron-metadata-dataplane-step-2-edpm-a-tk7z9

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:a8da62d734643927e28a259fa0b64b81a85cf108a29154c4ddaf9cf271c05610" in 585ms (585ms including waiting). Image size: 487820744 bytes.

openstack

kubelet

neutron-metadata-dataplane-step-2-edpm-a-tk7z9

Created

Created container: neutron-metadata-dataplane-step-2-edpm-a

openshift-machine-config-operator

machine-config-operator

machine-config-operator

ConfigMapUpdated

Updated ConfigMap/kube-rbac-proxy -n openshift-machine-config-operator: cause by changes in data.config-file.yaml

openstack

job-controller

neutron-metadata-dataplane-step-2-edpm-b

SuccessfulCreate

Created pod: neutron-metadata-dataplane-step-2-edpm-b-5g8zh

openstack

multus

neutron-metadata-dataplane-step-2-edpm-b-5g8zh

AddedInterface

Add eth0 [10.128.1.40/23] from ovn-kubernetes

openstack

multus

neutron-metadata-dataplane-step-2-edpm-b-5g8zh

AddedInterface

Add ctlplane [192.168.122.30/24] from openstack/ctlplane

openstack

kubelet

neutron-metadata-dataplane-step-2-edpm-b-5g8zh

Pulling

Pulling image "quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:a8da62d734643927e28a259fa0b64b81a85cf108a29154c4ddaf9cf271c05610"

openstack

kubelet

neutron-metadata-dataplane-step-2-edpm-b-5g8zh

Started

Started container neutron-metadata-dataplane-step-2-edpm-b

openstack

kubelet

neutron-metadata-dataplane-step-2-edpm-b-5g8zh

Created

Created container: neutron-metadata-dataplane-step-2-edpm-b

openstack

kubelet

neutron-metadata-dataplane-step-2-edpm-b-5g8zh

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:a8da62d734643927e28a259fa0b64b81a85cf108a29154c4ddaf9cf271c05610" in 650ms (650ms including waiting). Image size: 487820744 bytes.

openstack

multus

neutron-metadata-dataplane-step-2-edpm-a-8999k

AddedInterface

Add eth0 [10.128.1.41/23] from ovn-kubernetes

openstack

job-controller

neutron-metadata-dataplane-step-2-edpm-a

SuccessfulCreate

Created pod: neutron-metadata-dataplane-step-2-edpm-a-8999k

openstack

multus

neutron-metadata-dataplane-step-2-edpm-a-8999k

AddedInterface

Add ctlplane [192.168.122.31/24] from openstack/ctlplane

openstack

kubelet

neutron-metadata-dataplane-step-2-edpm-a-8999k

Pulling

Pulling image "quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:a8da62d734643927e28a259fa0b64b81a85cf108a29154c4ddaf9cf271c05610"

openstack

kubelet

neutron-metadata-dataplane-step-2-edpm-a-8999k

Started

Started container neutron-metadata-dataplane-step-2-edpm-a

openstack

kubelet

neutron-metadata-dataplane-step-2-edpm-a-8999k

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:a8da62d734643927e28a259fa0b64b81a85cf108a29154c4ddaf9cf271c05610" in 412ms (412ms including waiting). Image size: 487820744 bytes.

openstack

kubelet

neutron-metadata-dataplane-step-2-edpm-a-8999k

Created

Created container: neutron-metadata-dataplane-step-2-edpm-a

openshift-machine-config-operator

machine-config-operator

machine-config-operator

ConfigMapUpdated

Updated ConfigMap/kube-rbac-proxy -n openshift-machine-config-operator: cause by changes in data.config-file.yaml

openshift-machine-config-operator

machine-config-operator

machine-config-operator

ConfigMapUpdated

Updated ConfigMap/kube-rbac-proxy -n openshift-machine-config-operator: cause by changes in data.config-file.yaml

openstack

job-controller

neutron-metadata-dataplane-step-2-edpm-b

SuccessfulCreate

Created pod: neutron-metadata-dataplane-step-2-edpm-b-nzczt

openstack

multus

neutron-metadata-dataplane-step-2-edpm-b-nzczt

AddedInterface

Add ctlplane [192.168.122.30/24] from openstack/ctlplane

openstack

kubelet

neutron-metadata-dataplane-step-2-edpm-b-nzczt

Pulling

Pulling image "quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:a8da62d734643927e28a259fa0b64b81a85cf108a29154c4ddaf9cf271c05610"

openstack

multus

neutron-metadata-dataplane-step-2-edpm-b-nzczt

AddedInterface

Add eth0 [10.128.1.42/23] from ovn-kubernetes

openstack

kubelet

neutron-metadata-dataplane-step-2-edpm-b-nzczt

Started

Started container neutron-metadata-dataplane-step-2-edpm-b

openstack

kubelet

neutron-metadata-dataplane-step-2-edpm-b-nzczt

Created

Created container: neutron-metadata-dataplane-step-2-edpm-b

openstack

kubelet

neutron-metadata-dataplane-step-2-edpm-b-nzczt

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:a8da62d734643927e28a259fa0b64b81a85cf108a29154c4ddaf9cf271c05610" in 706ms (706ms including waiting). Image size: 487820744 bytes.

openstack

multus

neutron-metadata-dataplane-step-2-edpm-a-9lz8x

AddedInterface

Add ctlplane [192.168.122.31/24] from openstack/ctlplane

openstack

multus

neutron-metadata-dataplane-step-2-edpm-a-9lz8x

AddedInterface

Add eth0 [10.128.1.43/23] from ovn-kubernetes

openstack

kubelet

neutron-metadata-dataplane-step-2-edpm-a-9lz8x

Pulling

Pulling image "quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:a8da62d734643927e28a259fa0b64b81a85cf108a29154c4ddaf9cf271c05610"

openstack

job-controller

neutron-metadata-dataplane-step-2-edpm-a

SuccessfulCreate

Created pod: neutron-metadata-dataplane-step-2-edpm-a-9lz8x

openstack

kubelet

neutron-metadata-dataplane-step-2-edpm-a-9lz8x

Started

Started container neutron-metadata-dataplane-step-2-edpm-a

openstack

kubelet

neutron-metadata-dataplane-step-2-edpm-a-9lz8x

Pulled

Successfully pulled image "quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:a8da62d734643927e28a259fa0b64b81a85cf108a29154c4ddaf9cf271c05610" in 1.455s (1.455s including waiting). Image size: 487820744 bytes.

openstack

kubelet

neutron-metadata-dataplane-step-2-edpm-a-9lz8x

Created

Created container: neutron-metadata-dataplane-step-2-edpm-a

openstack

job-controller

neutron-metadata-dataplane-step-2-edpm-b

BackoffLimitExceeded

Job has reached the specified backoff limit

openstack

job-controller

neutron-metadata-dataplane-step-2-edpm-a

BackoffLimitExceeded

Job has reached the specified backoff limit

openshift-machine-config-operator

machine-config-operator

machine-config-operator

ConfigMapUpdated

Updated ConfigMap/kube-rbac-proxy -n openshift-machine-config-operator: cause by changes in data.config-file.yaml

openshift-machine-config-operator

machine-config-operator

machine-config-operator

ConfigMapUpdated

Updated ConfigMap/kube-rbac-proxy -n openshift-machine-config-operator: cause by changes in data.config-file.yaml

openshift-machine-config-operator

machine-config-operator

machine-config-operator

ConfigMapUpdated

Updated ConfigMap/kube-rbac-proxy -n openshift-machine-config-operator: cause by changes in data.config-file.yaml

openshift-machine-config-operator

machine-config-operator

machine-config-operator

ConfigMapUpdated

Updated ConfigMap/kube-rbac-proxy -n openshift-machine-config-operator: cause by changes in data.config-file.yaml

openshift-machine-config-operator

machine-config-operator

machine-config-operator

ConfigMapUpdated

Updated ConfigMap/kube-rbac-proxy -n openshift-machine-config-operator: cause by changes in data.config-file.yaml

openshift-machine-config-operator

machine-config-operator

machine-config-operator

ConfigMapUpdated

Updated ConfigMap/kube-rbac-proxy -n openshift-machine-config-operator: cause by changes in data.config-file.yaml

openshift-machine-config-operator

machine-config-operator

machine-config-operator

ConfigMapUpdated

Updated ConfigMap/kube-rbac-proxy -n openshift-machine-config-operator: cause by changes in data.config-file.yaml

openstack

cronjob-controller

keystone-cron

SuccessfulCreate

Created job keystone-cron-29566681

openstack

job-controller

keystone-cron-29566681

SuccessfulCreate

Created pod: keystone-cron-29566681-szrxn

openstack

multus

keystone-cron-29566681-szrxn

AddedInterface

Add eth0 [10.128.1.44/23] from ovn-kubernetes

openstack

kubelet

keystone-cron-29566681-szrxn

Started

Started container keystone-cron

openstack

kubelet

keystone-cron-29566681-szrxn

Created

Created container: keystone-cron

openstack

kubelet

keystone-cron-29566681-szrxn

Pulled

Container image "quay.io/podified-antelope-centos9/openstack-keystone@sha256:3a148c87899d4cbbb6bbe1203ad6c237fb295b5f42abda425dc0329305723414" already present on machine

openstack

cronjob-controller

keystone-cron

SawCompletedJob

Saw completed job: keystone-cron-29566681, condition: Complete

openstack

job-controller

keystone-cron-29566681

Completed

Job completed

openshift-machine-config-operator

machine-config-operator

machine-config-operator

ConfigMapUpdated

Updated ConfigMap/kube-rbac-proxy -n openshift-machine-config-operator: cause by changes in data.config-file.yaml

openshift-machine-config-operator

machine-config-operator

machine-config-operator

ConfigMapUpdated

Updated ConfigMap/kube-rbac-proxy -n openshift-machine-config-operator: cause by changes in data.config-file.yaml

openshift-machine-config-operator

machine-config-operator

machine-config-operator

ConfigMapUpdated

Updated ConfigMap/kube-rbac-proxy -n openshift-machine-config-operator: cause by changes in data.config-file.yaml

openshift-machine-config-operator

machine-config-operator

machine-config-operator

ConfigMapUpdated

Updated ConfigMap/kube-rbac-proxy -n openshift-machine-config-operator: cause by changes in data.config-file.yaml

openshift-machine-config-operator

machine-config-operator

machine-config-operator

ConfigMapUpdated

Updated ConfigMap/kube-rbac-proxy -n openshift-machine-config-operator: cause by changes in data.config-file.yaml

openshift-machine-config-operator

machine-config-operator

machine-config-operator

ConfigMapUpdated

Updated ConfigMap/kube-rbac-proxy -n openshift-machine-config-operator: cause by changes in data.config-file.yaml

openshift-machine-config-operator

machine-config-operator

machine-config-operator

ConfigMapUpdated

Updated ConfigMap/kube-rbac-proxy -n openshift-machine-config-operator: cause by changes in data.config-file.yaml

openshift-machine-config-operator

machine-config-operator

machine-config-operator

ConfigMapUpdated

Updated ConfigMap/kube-rbac-proxy -n openshift-machine-config-operator: cause by changes in data.config-file.yaml

openshift-kube-controller-manager

cluster-policy-controller-namespace-security-allocation-controller

kube-controller-manager-master-0

CreatedSCCRanges

created SCC ranges for openshift-must-gather-6k5lb namespace